๐Business Forecasting Unit 2 โ Fundamentals of Time Series Analysis
Time series analysis is a crucial tool for understanding and predicting patterns in data collected over time. It involves examining trends, seasonality, and other components to uncover insights and make forecasts. This fundamental skill is essential for business forecasting and decision-making.
Key concepts include stationarity, autocorrelation, and decomposition of time series components. Various models like ARIMA and exponential smoothing are used to capture different patterns. Preprocessing techniques and accuracy evaluation methods ensure reliable forecasts for real-world applications in finance, economics, and more.
Study Guides for Unit 2 โ Fundamentals of Time Series Analysis
Time series data consists of observations collected sequentially over time at regular intervals (hourly, daily, monthly, yearly)
Univariate time series involves a single variable measured over time, while multivariate time series involves multiple variables
Autocorrelation measures the correlation between a time series and its lagged values
Positive autocorrelation indicates that high values tend to be followed by high values and low values by low values
Negative autocorrelation suggests that high values are likely to be followed by low values and vice versa
Stationarity refers to the property of a time series where its statistical properties (mean, variance, autocorrelation) remain constant over time
Trend represents the long-term increase or decrease in the data over time (population growth, economic growth)
Seasonality refers to the recurring patterns or cycles within a fixed period (sales during holiday seasons, temperature variations throughout the year)
White noise is a series of uncorrelated random variables with zero mean and constant variance
Components of Time Series
Trend component captures the long-term increase or decrease in the data over time
Can be linear, where the data increases or decreases at a constant rate
Can be non-linear, where the rate of change varies over time (exponential growth, logarithmic growth)
Seasonal component represents the recurring patterns or cycles within a fixed period
Additive seasonality assumes that the seasonal fluctuations are constant over time and independent of the trend
Multiplicative seasonality assumes that the seasonal fluctuations are proportional to the level of the series and change with the trend
Cyclical component captures the medium-term fluctuations that are not of fixed period
Often related to economic or business cycles (expansion, recession)
Typically longer than seasonal patterns and not as predictable
Irregular component represents the random fluctuations or noise in the data that cannot be explained by the other components
Caused by unexpected events or measurement errors (natural disasters, policy changes, data collection issues)
Stationarity and Its Importance
Stationarity is a crucial assumption for many time series models and forecasting techniques
A stationary time series has constant mean, variance, and autocorrelation over time
Mean stationarity: The mean of the series remains constant and does not depend on time
Variance stationarity: The variance of the series remains constant and does not change over time
Autocorrelation stationarity: The autocorrelation structure remains constant over time
Non-stationary time series can lead to spurious relationships and unreliable forecasts
Stationarity tests, such as the Augmented Dickey-Fuller (ADF) test and the Kwiatkowski-Phillips-Schmidt-Shin (KPSS) test, help determine if a series is stationary
Differencing is a common technique to transform a non-stationary series into a stationary one by taking the differences between consecutive observations
First-order differencing involves subtracting each observation from its previous observation
Higher-order differencing may be necessary for more complex non-stationary patterns
Time Series Patterns and Models
Autoregressive (AR) models express the current value of a series as a linear combination of its past values and an error term
AR(1) model: $y_t = c + \phi_1 y_{t-1} + \epsilon_t$, where $y_t$ is the current value, $c$ is a constant, $\phi_1$ is the autoregressive coefficient, and $\epsilon_t$ is the error term
Higher-order AR models (AR(p)) include more lagged values of the series
Moving Average (MA) models express the current value of a series as a linear combination of the current and past error terms
MA(1) model: $y_t = \mu + \epsilon_t + \theta_1 \epsilon_{t-1}$, where $\mu$ is the mean, $\theta_1$ is the moving average coefficient, and $\epsilon_t$ is the error term
Higher-order MA models (MA(q)) include more lagged error terms
Autoregressive Moving Average (ARMA) models combine both AR and MA components