Hubbry Logo
logo
Linear trend estimation
Community hub

Linear trend estimation

logo
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Contribute something to knowledge base
Hub AI

Linear trend estimation AI simulator

(@Linear trend estimation_simulator)

Linear trend estimation

Linear trend estimation is a statistical technique used to analyze data patterns. Data patterns, or trends, occur when the information gathered tends to increase or decrease over time or is influenced by changes in an external factor. Linear trend estimation essentially creates a straight line on a graph of data that models the general direction that the data is heading.

Given a set of data, there are a variety of functions that can be chosen to fit the data. The simplest function is a straight line with the dependent variable (typically the measured data) on the vertical axis and the independent variable (often time) on the horizontal axis.

The least-squares fit is a common method to fit a straight line through the data. This method minimizes the sum of the squared errors in the data series . Given a set of points in time and data values observed for those points in time, values of and are chosen to minimize the sum of squared errors

This formula first calculates the difference between the observed data and the estimate , the difference at each data point is squared, and then added together, giving the "sum of squares" measurement of error. The values of and derived from the data parameterize the simple linear estimator . The term "trend" refers to the slope in the least squares estimator.

To analyze a (time) series of data, it can be assumed that it may be represented as trend plus noise:

where and are unknown constants and the 's are randomly distributed errors. If one can reject the null hypothesis that the errors are non-stationary, then the non-stationary series is called trend-stationary. The least-squares method assumes the errors are independently distributed with a normal distribution. If this is not the case, hypothesis tests about the unknown parameters and may be inaccurate. It is simplest if the 's all have the same distribution, but if not (if some have higher variance, meaning that those data points are effectively less certain), then this can be taken into account during the least-squares fitting by weighting each point by the inverse of the variance of that point.

Commonly, where only a single time series exists to be analyzed, the variance of the 's is estimated by fitting a trend to obtain the estimated parameter values and thus allowing the predicted values

to be subtracted from the data (thus detrending the data), leaving the residuals as the detrended data, and estimating the variance of the 's from the residuals — this is often the only way of estimating the variance of the 's.

See all
User Avatar
No comments yet.