Week 6 - Time Series Methods
Due date: Lab 6 - Sunday, Oct 26, 5pm ET
Prepare
📖 Read Time Series Analysis with R, Chapter 7: Structural Decomposition
📖 Read Modeling time series with tidy resampling
📖 Read Introducing Modeltime: Tidy Time Series Forecasting using Tidymodels
📖 Read Forecasting: Principles and Practice, Chapter 7: Time series regression models
📖 Read Forecasting: Principles and Practice, Chapter 9: ARIMA Models
Participate
Perform
Study
Short Answer Questions
Instructions: Answer the following questions in 2-3 sentences each.
- Define Time Series Data: What distinguishes time series data from typical cross-sectional or panel data, and why does this distinction matter for analysis?
- Explain the Limitation of OLS for Time Series: Why is a simple linear regression model (Ordinary Least Squares) generally not suitable for analyzing time series data?
- Describe the Components of Time Series Dynamics: What are the three primary components into which the dynamic evolution of time series variables is often broken down? Briefly explain each.
- Differentiate Deterministic and Stochastic Processes: How do deterministic and stochastic processes differ in the context of time series, and provide an example of a stochastic process.
- What is White Noise? Characterize a white noise series. What are its key statistical properties, particularly concerning correlation and mean?
- Explain the Concept of a Random Walk: Describe a random walk process and its primary characteristic regarding the impact of shocks.
- What is an AR(1) Process? How is the current value in an AR(1) process determined, and what does it imply about the relationship between observations over time?
- Distinguish between ACF and PACF: What is the primary difference in what the Autocorrelation Function (ACF) and Partial Autocorrelation Function (PACF) measure in a time series?
- Explain Stationarity and its Importance: What does it mean for a time series to be stationary, and why is this property crucial for applying many time series models?
- Describe the ARIMA Model Components: What do the ‘AR’, ‘I’, and ‘MA’ components of an ARIMA(p, d, q) model represent?
Short-Answer Answer Key
Define Time Series Data: Time series data is a dataset where each observation includes a time measurement, and these measurements are ordered. This ordered temporal measurement implies that error terms may be correlated and not independent, which violates fundamental assumptions of standard statistical methods.
Explain the Limitation of OLS for Time Series: Simple linear regression (OLS) assumes independent measurements. In time series, observations are often serially correlated, meaning error terms are not independent, which seriously compromises the fundamental OLS assumptions and leads to biased variance estimates.
Describe the Components of Time Series Dynamics: Time series dynamics are often broken into three parts: trend ($T_t$), which represents the long-term direction; seasonality ($S_t$), which accounts for regular, periodic fluctuations; and noise ($I_t$), which represents the irregular or random component after accounting for trend and seasonality.
Differentiate Deterministic and Stochastic Processes: Deterministic processes always produce the same output from a given starting point, meaning their future state is perfectly predictable. Stochastic processes, conversely, have indeterminacy and are described by statistical distributions, with examples including white noise or random walks.
What is White Noise? A white noise series consists of serially uncorrelated random variables with a zero mean and finite variance. This implies that current errors have no dependence on past errors, and their expected value is zero.
Explain the Concept of a Random Walk: A random walk is a stochastic process where the current value is the previous value plus a random shock. Its defining characteristic is that the effect of any past shock is permanent and does not dissipate, giving it an infinite memory.
What is an AR(1) Process? In an AR(1) process, the current value of the time series is a linear function of its immediately preceding value plus a random error term. This means that the current observation is directly influenced by the observation from the previous time step.
Distinguish between ACF and PACF: The Autocorrelation Function (ACF) measures the correlation between a time series and its lagged values, including indirect correlations passed through intermediate lags. The Partial Autocorrelation Function (PACF), however, eliminates these pass-through effects, focusing solely on the direct, independent relationship between an observation and a lagged observation.
Explain Stationarity and its Importance: A time series is strictly stationary if its statistical properties (like mean, variance, and autocorrelation) do not change over time. This property is crucial because many time series models and statistical tests assume stationarity to ensure reliable and consistent results.
Describe the ARIMA Model Components: In an ARIMA(p, d, q) model, ‘AR’ (AutoRegressive) denotes the number of lagged observations included in the model; ‘I’ (Integrated) refers to the number of times the raw observations are differenced to make the series stationary; and ‘MA’ (Moving Average) represents the number of lagged forecast errors in the model.
Essay Questions
Discuss the fundamental challenges that time series data present for traditional econometric methods, particularly concerning the assumptions of Ordinary Least Squares (OLS) regression. How do characteristics like trends, seasonality, and lag structures necessitate different analytical approaches?
Compare and contrast the concepts of strict stationarity and covariance (weak) stationarity in time series. Explain why stationarity is a critical assumption for many time series models, and describe how unit root tests are used in practice to assess this property.
Explain the concept of an Autoregressive (AR) process, including the meaning of an AR(p) model. Detail the conditions required for an AR process to be stationary, specifically referencing the roots of the characteristic equation and their relation to the unit circle.
Describe the core idea behind Exponential Smoothing (ETS) models. Explain how additive components for trend and seasonality can be incorporated into the basic exponential smoothing formula, and discuss what the ‘smoothing factor’ ($\alpha$) signifies in this context.
Outline a typical workflow for building and evaluating time series models using the timetk and modeltime packages within the Tidymodels framework. Include discussions on data manipulation, feature engineering, model specification, forecasting, and accuracy evaluation, highlighting key functions or concepts at each stage.
Back to course schedule ⏎