14.7 Nonstationarity I: Trends
If a series is nonstationary, conventional hypothesis tests, confidence intervals and forecasts can be strongly misleading. The assumption of stationarity is violated if a series exhibits trends or breaks and the resulting complications in an econometric analysis depend on the specific type of the nonstationarity. This section focuses on time series that exhibit trends.
A series is said to exhibit a trend if it has a persistent long-term movement. One distinguishes between deterministic and stochastic trends.
A trend is deterministic if it is a nonrandom function of time.
A trend is said to be stochastic if it is a random function of time.
The figures we have produced in Chapter 14.2 reveal that many economic time series show a trending behavior that is probably best modeled by stochastic trends. This is why the book focuses on the treatment of stochastic trends.
The Random Walk Model of a Trend
The simplest way to model a time series YtYt that has stochastic trend is the random walk Yt=Yt−1+ut, where the ut are i.i.d. errors with E(ut|Yt−1,Yt−2,…)=0. Note that E(Yt|Yt−1,Yt−2…)=E(Yt−1|Yt−1,Yt−2…)+E(ut|Yt−1,Yt−2…)=Yt−1so the best forecast for Yt is yesterday’s observation Yt−1. Hence the difference between Yt and Yt−1 is unpredictable. The path followed by Yt consists of random steps ut, hence it is called a random walk.
Assume that Y0, the starting value of the random walk is 0. Another way to write (14.6) is Y0=0Y1=0+u1Y2=0+u1+u2⋮Yt=t∑i=1ui. Therefore we have Var(Yt)=Var(u1+u2+⋯+ut)=tσ2u.Thus the variance of a random walk depends on t which violates the assumption presented in Key Concept 14.5: a random walk is nonstationary.
Obviously, (14.6) is a special case of an AR(1) model where β1=1. One can show that a time series that follows an AR(1) model is stationary if |β1|<1. In a general AR(p) model, stationarity is linked to the roots of the polynomial 1−β1z−β2z2−β3z3−⋯−βpzp. If all roots are greater than 1 in absolute value, the AR(p) series is stationary. If at least one root equals 1, the AR(p) is said to have a unit root and thus has a stochastic trend.
It is straightforward to simulate random walks in R using arima.sim(). The function matplot() is convenient for simple plots of the columns of a matrix.
# simulate and plot random walks starting at 0
set.seed(1)
RWs <- ts(replicate(n = 4,
arima.sim(model = list(order = c(0, 1 ,0)), n = 100)))
matplot(RWs,
type ="l",
col = c("steelblue", "darkgreen", "darkred", "orange"),
lty = 1,
lwd = 2,
main = "Four Random Walks",
xlab = "Time",
ylab = "Value")
Hide Source
Hide Plot
a random walk model with a drift which allows to model the tendency of a series to move upwards or downwards. If β0 is positive, the series drifts upwards and it follows a downward trend if β0 is negative.
# simulate and plot random walks with drift
set.seed(1)
RWsd <- ts(replicate(n = 4,
arima.sim(model = list(order = c(0, 1, 0)), n = 100)))
matplot(RWsd,
type="l",
col = c("steelblue", "darkgreen", "darkred", "orange"),
lty = 1,
lwd = 2,
main = "Four Random Walks with Drift",
xlab = "Time",
ylab = "Value")
Hide Source
Hide Plot
Problems Caused by Stochastic Trends
OLS estimation of the coefficients on regressors that have a stochastic trend is problematic because the distribution of the estimator and its t-statistic is non-normal, even asymptotically. This has various consequences:
Downward bias of autoregressive coefficients:
If Yt is a random walk, β1 can be consistently estimated by OLS but the estimator is biased toward zero. This bias is roughly E(ˆβ1)≈1−5.3/T which is substantial for sample sizes typically encountered in macroeconomics. This estimation bias causes forecasts of Yt to perform worse than a pure random walk model.
Non-normally distributed t-statistics:
The nonnormal distribution of the estimated coefficient of a stochastic regressor translates to a nonnormal distribution of its t-statistic so that normal critical values are invalid and therefore usual confidence intervals and hypothesis tests are invalid, too, and the true distribution of the t-statistic cannot be readily determined.
Spurious Regression:
When two stochastically trending time series are regressed onto each other, the estimated relationship may appear highly significant using conventional normal critical values although the series are unrelated. This is what econometricians call a spurious relationship.
As an example for spurious regression, consider again the green and the red random walks that we have simulated above. We know that there is no relationship between both series: they are generated independently of each other.
# plot spurious relationship
matplot(RWs[, c(2, 3)],
lty = 1,
lwd = 2,
type = "l",
col = c("darkgreen", "darkred"),
xlab = "Time",
ylab = "",
main = "A Spurious Relationship")
Hide Source
Hide Plot
# estimate spurious AR model
summary(dynlm(RWs[, 2] ~ L(RWs[, 3])))$coefficients
## Estimate Std. Error t value Pr(>|t|)
## (Intercept) -3.459488 0.3635104 -9.516889 1.354156e-15
## L(RWs[, 3]) 1.047195 0.1450874 7.217687 1.135828e-10
The result is obviously spurious: the coefficient on Greent−1 is estimated to be about 1 and the p-value of 1.14⋅10−10 of the corresponding t-test indicates that the coefficient is highly significant while its true value is in fact zero.
As an empirical example, consider the U.S. unemployment rate and the Japanese industrial production. Both series show an upward trending behavior from the mid-1960s through the early 1980s.
# plot U.S. unemployment rate & Japanese industrial production
plot(merge(as.zoo(USUnemp), as.zoo(JPIndProd)),
plot.type = "single",
col = c("darkred", "steelblue"),
lwd = 2,
xlab = "Date",
ylab = "",
main = "Spurious Regression: Macroeconomic Time series")
# add a legend
legend("topleft",
legend = c("USUnemp", "JPIndProd"),
col = c("darkred", "steelblue"),
lwd = c(2, 2))
Hide Source
Hide Plot
# estimate regression using data from 1962 to 1985
SR_Unemp1 <- dynlm(ts(USUnemp["1962::1985"]) ~ ts(JPIndProd["1962::1985"]))
coeftest(SR_Unemp1, vcov = sandwich)
##
## t test of coefficients:
##
## Estimate Std. Error t value Pr(>|t|)
## (Intercept) -2.37452 1.12041 -2.1193 0.0367 *
## ts(JPIndProd["1962::1985"]) 2.22057 0.29233 7.5961 2.227e-11 ***
## ---
## Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
A simple regression of the U.S. unemployment rate on Japanese industrial production using data from 1962 to 1985 yields
^U.S.URt=−2.37(1.12)+2.22(0.29)log(JapaneseIPt).
This appears to be a significant relationship: the t-statistic of the coefficient on log(JapaneseIPt) is bigger than 7.
# Estimate regression using data from 1986 to 2012
SR_Unemp2 <- dynlm(ts(USUnemp["1986::2012"]) ~ ts(JPIndProd["1986::2012"]))
coeftest(SR_Unemp2, vcov = sandwich)
##
## t test of coefficients:
##
## Estimate Std. Error t value Pr(>|t|)
## (Intercept) 41.7763 5.4066 7.7270 6.596e-12 ***
## ts(JPIndProd["1986::2012"]) -7.7771 1.1714 -6.6391 1.386e-09 ***
## ---
## Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
When estimating the same model, this time with data from 1986 to 2012, we obtain
^U.S.URt=41.78(5.41)−7.78(1.17)log(JapaneseIP)t
which surprisingly is quite different. (14.8) indicates a moderate positive relationship, in contrast to the large negative coefficient in (14.9). This phenomenon can be attributed to stochastic trends in the series: since there is no economic reasoning that relates both trends, both regressions may be spurious.
Testing for a Unit AR Root
A formal test for a stochastic trend has been proposed by Dickey & Fuller (1979) which thus is termed the Dickey-Fuller test. As discussed above, a time series that follows an AR(1) model with β1=1 has a stochastic trend. Thus, the testing problem is H0:β1=1 vs. H1:|β1|<1. The null hypothesis is that the AR(1) model has a unit root and the alternative hypothesis is that it is stationary. One often rewrites the AR(1) model by subtracting Yt−1 on both sides: Yt=β0+β1Yt−1+ut ⇔ ΔYt=β0+δYt−1+ui where δ=β1−1. The testing problem then becomes H0:δ=0 vs. H1:δ<0which is convenient since the corresponding test statistic is reported by many relevant R functions.11
The Dickey-Fuller test can also be applied in an AR(p) model. The Augmented Dickey-Fuller (ADF) test is summarized in Key Concept 14.8.
Key Concept 14.8
The ADF Test for a Unit Root
Consider the regression ΔYt=β0+δYt−1+γ1Δ1Yt−1+γ2ΔYt−2+⋯+γpΔYt−p+ut.
The ADF test for a unit autoregressive root tests the hypothesis H0:δ=0 (stochastic trend) against the one-sided alternative H1:δ<0 (stationarity) using the usual OLS t-statistic.
If it is assumed that Yt is stationary around a deterministic linear time trend, the model is augmented by the regressor t: ΔYt=β0+at+δYt−1+γ1Δ1Yt−1+γ2ΔYt−2+⋯+γpΔYt−p+ut,where again H0:δ=0 is tested against H1:δ<0.
The optimal lag length p can be estimated using information criteria. In (14.10), p=0 (no lags of ΔYt are used as regressors) corresponds to a simple AR(1).
Under the null, the t-statistic corresponding to H0:δ=0 does not have a normal distribution. The critical values can only be obtained from simulation and differ for regressions (14.10) and (14.11) since the distribution of the ADF test statistic is sensitive to the deterministic components included in the regression.Critical Values for the ADF Statistic
Key Concept 14.8 states that the critical values for the ADF test in the regressions (14.10) and (14.11) can only be determined using simulation. The idea of the simulation study is to simulate a large number of ADF test statistics and use them to estimate quantiles of their asymptotic distribution. This section shows how this can be done using R.
First, consider an AR(1) model with drift. The procedure is as follows:
- Simulate N random walks with n observations using the data generating process
Yt=β0+β1Yt−1+ut,
t=1,…,n where N and n are large numbers.
- For each random walk, estimate the regression
ΔYt=β0+β1Yt−1+ut
and compute ADF test statistic. Save all N test statistics.
Estimate quantiles of the distribution of the ADF test statistic using the N test statistics obtained from the simulation.
Loosely speaking, the precision of the estimated quantiles depends on two factors: n, the length of the underlying series and N, the number of test statistics used. Since we are interested in estimating quantiles of the asymptotic distribution (the Dickey-Fuller distribution) of the ADF test statistic so both using many observations and large number of simulated test statistics will increase the precision of the estimated quantiles. We choose n=N=1000 as the computational burden grows quickly with n and N.
# repetitions
N <- 1000
# observations
n <- 1000
# define drift and trend
drift <- 0.5
trend <- 1:n
# simulate N random walks with drift
RWD <- ts(replicate(n = N,
(n-1) * drift + arima.sim(model = list(order = c(0, 1, 0)),
n = n - 1)))
# compute ADF test statistics and store them in 'ADFD'
ADFD <- numeric(N)
for(i in 1:ncol(RWD)) {
ADFD[i] <- summary(
dynlm(diff(RWD[, i], 1) ~ L(RWD[, i], 1)))$coef[2, 3]
}
# simulate N random walks with drift + trend
RWDT <- ts(replicate(n = N,
trend + drift + arima.sim(model = list(order = c(0, 1, 0)),
n = n - 1)))
# compute ADF test statistics and store them in 'ADFDT'
ADFDT <- numeric(N)
for(i in 1:ncol(RWDT)) {
ADFDT[i] <- summary(
dynlm(diff(RWDT[, i], 1) ~ L(RWDT[, i], 1) + trend(RWDT[, i], scale = F))
)$coef[2, 3]
}
# estimate quantiles for ADF regression with a drift
round(quantile(ADFD, c(0.1, 0.05, 0.01)), 2)
## 10% 5% 1%
## -2.62 -2.83 -3.39
# estimate quantiles for ADF regression with drift and trend
round(quantile(ADFDT, c(0.1,0.05,0.01)),2)
## 10% 5% 1%
## -3.11 -3.43 -3.97
The estimated quantiles are close to the large-sample critical values of the ADF test statistic reported in Table 14.4 of the book.
Deterministic Regressors | 10% | 5% | 1% |
---|---|---|---|
Intercept only | -2.57 | -2.86 | -3.43 |
Intercept and time trend | -3.12 | -3.41 | -3.96 |
The results show that using standard normal critical values is erroneous: the 5% critical value of the standard normal distribution is −1.64. For the Dickey-Fuller distributions the estimated critical values are −2.87 (drift) and −3.43 (drift and linear time trend). This implies that a true null (the series has a stochastic trend) would be rejected far too often if inappropriate normal critical values were used.
We may use the simulated test statistics for a graphical comparison of the standard normal density and (estimates of) both Dickey-Fuller densities.
# plot standard normal density
curve(dnorm(x),
from = -6, to = 3,
ylim = c(0, 0.6),
lty = 2,
ylab = "Density",
xlab = "t-Statistic",
main = "Distributions of ADF Test Statistics",
col = "darkred",
lwd = 2)
# plot density estimates of both Dickey-Fuller distributions
lines(density(ADFD), lwd = 2, col = "darkgreen")
lines(density(ADFDT), lwd = 2, col = "blue")
# add a legend
legend("topleft",
c("N(0,1)", "Drift", "Drift+Trend"),
col = c("darkred", "darkgreen", "blue"),
lty = c(2, 1, 1),
lwd = 2)
Hide Source
Hide Plot
The deviations from the standard normal distribution are significant: both Dickey-Fuller distributions are skewed to the left and have a heavier left tail than the standard normal distribution.
Does U.S. GDP Have a Unit Root?
As an empirical example, we use the ADF test to assess whether there is a stochastic trend in U.S. GDP using the regression Δlog(GDPt)=β0+αt+β1log(GDPt−1)+β2Δlog(GDPt−1)+β3Δlog(GDPt−2)+ut.# generate log GDP series
LogGDP <- ts(log(GDP["1962::2012"]))
# estimate the model
coeftest(
dynlm(diff(LogGDP) ~ trend(LogGDP, scale = F) + L(LogGDP)
+ diff(L(LogGDP)) + diff(L(LogGDP), 2)))
##
## t test of coefficients:
##
## Estimate Std. Error t value Pr(>|t|)
## (Intercept) 0.27877045 0.11793233 2.3638 0.019066 *
## trend(LogGDP, scale = F) 0.00023818 0.00011090 2.1476 0.032970 *
## L(LogGDP) -0.03332452 0.01441436 -2.3119 0.021822 *
## diff(L(LogGDP)) 0.08317976 0.11295542 0.7364 0.462371
## diff(L(LogGDP), 2) 0.18763384 0.07055574 2.6594 0.008476 **
## ---
## Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
The estimation yields
Δlog(GDPt)=0.28(0.118)+0.0002(0.0001)t−0.033(0.014)log(GDPt−1)+0.083(0.113)Δlog(GDPt−1)+0.188(0.071)Δlog(GDPt−2)+ut,
so the ADF test statistic is t=−0.033/0.014=−2.35. The corresponding 5% critical value from Table 14.2 is −3.41 so we cannot reject the null hypothesis that log(GDP) has a stochastic trend in favor of the alternative that it is stationary around a deterministic linear time trend.
The ADF test can be done conveniently using ur.df() from the package urca.
# test for unit root in GDP using 'ur.df()' from the package 'urca'
summary(ur.df(LogGDP,
type = "trend",
lags = 2,
selectlags = "Fixed"))
##
## ###############################################
## # Augmented Dickey-Fuller Test Unit Root Test #
## ###############################################
##
## Test regression trend
##
##
## Call:
## lm(formula = z.diff ~ z.lag.1 + 1 + tt + z.diff.lag)
##
## Residuals:
## Min 1Q Median 3Q Max
## -0.025580 -0.004109 0.000321 0.004869 0.032781
##
## Coefficients:
## Estimate Std. Error t value Pr(>|t|)
## (Intercept) 0.2790086 0.1180427 2.364 0.019076 *
## z.lag.1 -0.0333245 0.0144144 -2.312 0.021822 *
## tt 0.0002382 0.0001109 2.148 0.032970 *
## z.diff.lag1 0.2708136 0.0697696 3.882 0.000142 ***
## z.diff.lag2 0.1876338 0.0705557 2.659 0.008476 **
## ---
## Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
##
## Residual standard error: 0.007704 on 196 degrees of freedom
## Multiple R-squared: 0.1783, Adjusted R-squared: 0.1616
## F-statistic: 10.63 on 4 and 196 DF, p-value: 8.076e-08
##
##
## Value of test-statistic is: -2.3119 11.2558 4.267
##
## Critical values for test statistics:
## 1pct 5pct 10pct
## tau3 -3.99 -3.43 -3.13
## phi2 6.22 4.75 4.07
## phi3 8.43 6.49 5.47
The first test statistic at the bottom of the output is the one we are interested in. The number of test statistics reported depends on the test regression. For type = “trend”, the second statistics corresponds to the test that there is no unit root and no time trend while the third one corresponds to a test of the hypothesis that there is a unit root, no time trend and no drift term.
References
Dickey, D. A., & Fuller, W. A. (1979). Distribution of the Estimators for Autoregressive Time Series with a Unit Root. Journal of the American Statistical Association, 74(366), pp. 427–431.
The t-statistic of the Dickey-Fuller test is computed using homoskedasticity-only standard errors since under the null hypothesis, the usual t-statistic is robust to conditional heteroskedasticity.↩