This book is in Open Review. We want your feedback to make the book better for you and other students. You may annotate some text by selecting it with the cursor and then click the on the pop-up menu. You can also see the annotations of others: click the in the upper right hand corner of the page

14.7 Nonstationarity I: Trends

If a series is nonstationary, conventional hypothesis tests, confidence intervals and forecasts can be strongly misleading. The assumption of stationarity is violated if a series exhibits trends or breaks and the resulting complications in an econometric analysis depend on the specific type of the nonstationarity. This section focuses on time series that exhibit trends.

A series is said to exhibit a trend if it has a persistent long-term movement. One distinguishes between deterministic and stochastic trends.

  • A trend is deterministic if it is a nonrandom function of time.

  • A trend is said to be stochastic if it is a random function of time.

The figures we have produced in Chapter 14.2 reveal that many economic time series show a trending behavior that is probably best modeled by stochastic trends. This is why the book focuses on the treatment of stochastic trends.

The Random Walk Model of a Trend

The simplest way to model a time series YtYt that has stochastic trend is the random walk Yt=Yt1+ut, where the ut are i.i.d. errors with E(ut|Yt1,Yt2,)=0. Note that E(Yt|Yt1,Yt2)=E(Yt1|Yt1,Yt2)+E(ut|Yt1,Yt2)=Yt1

so the best forecast for Yt is yesterday’s observation Yt1. Hence the difference between Yt and Yt1 is unpredictable. The path followed by Yt consists of random steps ut, hence it is called a random walk.

Assume that Y0, the starting value of the random walk is 0. Another way to write (14.6) is Y0=0Y1=0+u1Y2=0+u1+u2Yt=ti=1ui. Therefore we have Var(Yt)=Var(u1+u2++ut)=tσ2u.

Thus the variance of a random walk depends on t which violates the assumption presented in Key Concept 14.5: a random walk is nonstationary.

Obviously, (14.6) is a special case of an AR(1) model where β1=1. One can show that a time series that follows an AR(1) model is stationary if |β1|<1. In a general AR(p) model, stationarity is linked to the roots of the polynomial 1β1zβ2z2β3z3βpzp. If all roots are greater than 1 in absolute value, the AR(p) series is stationary. If at least one root equals 1, the AR(p) is said to have a unit root and thus has a stochastic trend.

It is straightforward to simulate random walks in R using arima.sim(). The function matplot() is convenient for simple plots of the columns of a matrix.

# simulate and plot random walks starting at 0
set.seed(1)

RWs <- ts(replicate(n = 4, 
            arima.sim(model = list(order = c(0, 1 ,0)), n = 100)))

matplot(RWs, 
        type ="l", 
        col = c("steelblue", "darkgreen", "darkred", "orange"), 
        lty = 1, 
        lwd = 2,
        main = "Four Random Walks",
        xlab = "Time",
        ylab = "Value")Hide Source
Hide Plot
Adding a constant to (14.6) yields Yt=β0+Yt1+ut,

a random walk model with a drift which allows to model the tendency of a series to move upwards or downwards. If β0 is positive, the series drifts upwards and it follows a downward trend if β0 is negative.

# simulate and plot random walks with drift
set.seed(1)

RWsd <- ts(replicate(n = 4, 
           arima.sim(model = list(order = c(0, 1, 0)), n = 100)))

matplot(RWsd, 
        type="l", 
        col = c("steelblue", "darkgreen", "darkred", "orange"), 
        lty = 1, 
        lwd = 2,
        main = "Four Random Walks with Drift",
        xlab = "Time",
        ylab = "Value")Hide Source
Hide Plot

Testing for a Unit AR Root

A formal test for a stochastic trend has been proposed by Dickey & Fuller (1979) which thus is termed the Dickey-Fuller test. As discussed above, a time series that follows an AR(1) model with β1=1 has a stochastic trend. Thus, the testing problem is H0:β1=1   vs.   H1:|β1|<1. The null hypothesis is that the AR(1) model has a unit root and the alternative hypothesis is that it is stationary. One often rewrites the AR(1) model by subtracting Yt1 on both sides: Yt=β0+β1Yt1+ut    ΔYt=β0+δYt1+ui where δ=β11. The testing problem then becomes H0:δ=0   vs.   H1:δ<0

which is convenient since the corresponding test statistic is reported by many relevant R functions.11

The Dickey-Fuller test can also be applied in an AR(p) model. The Augmented Dickey-Fuller (ADF) test is summarized in Key Concept 14.8.

Key Concept 14.8

The ADF Test for a Unit Root

Consider the regression ΔYt=β0+δYt1+γ1Δ1Yt1+γ2ΔYt2++γpΔYtp+ut.

The ADF test for a unit autoregressive root tests the hypothesis H0:δ=0 (stochastic trend) against the one-sided alternative H1:δ<0 (stationarity) using the usual OLS t-statistic.

If it is assumed that Yt is stationary around a deterministic linear time trend, the model is augmented by the regressor t: ΔYt=β0+at+δYt1+γ1Δ1Yt1+γ2ΔYt2++γpΔYtp+ut,

where again H0:δ=0 is tested against H1:δ<0.

The optimal lag length p can be estimated using information criteria. In (14.10), p=0 (no lags of ΔYt are used as regressors) corresponds to a simple AR(1).

Under the null, the t-statistic corresponding to H0:δ=0 does not have a normal distribution. The critical values can only be obtained from simulation and differ for regressions (14.10) and (14.11) since the distribution of the ADF test statistic is sensitive to the deterministic components included in the regression.

Critical Values for the ADF Statistic

Key Concept 14.8 states that the critical values for the ADF test in the regressions (14.10) and (14.11) can only be determined using simulation. The idea of the simulation study is to simulate a large number of ADF test statistics and use them to estimate quantiles of their asymptotic distribution. This section shows how this can be done using R.

First, consider an AR(1) model with drift. The procedure is as follows:

  • Simulate N random walks with n observations using the data generating process Yt=β0+β1Yt1+ut,

    t=1,,n where N and n are large numbers.

  • For each random walk, estimate the regression ΔYt=β0+β1Yt1+ut

    and compute ADF test statistic. Save all N test statistics.

  • Estimate quantiles of the distribution of the ADF test statistic using the N test statistics obtained from the simulation.

For the case with drift and linear time trend we replace the data generating process by Yt=β0+α(β11)t+β1Yt1+ut and estimate ΔYt=β0+αt+δ1Yt1+ut.

Loosely speaking, the precision of the estimated quantiles depends on two factors: n, the length of the underlying series and N, the number of test statistics used. Since we are interested in estimating quantiles of the asymptotic distribution (the Dickey-Fuller distribution) of the ADF test statistic so both using many observations and large number of simulated test statistics will increase the precision of the estimated quantiles. We choose n=N=1000 as the computational burden grows quickly with n and N.

# repetitions
N <- 1000

# observations
n <- 1000

# define drift and trend 
drift <- 0.5
trend <- 1:n

# simulate N random walks with drift 
RWD <- ts(replicate(n = N, 
            (n-1) * drift + arima.sim(model = list(order = c(0, 1, 0)),
                                      n = n - 1)))

# compute ADF test statistics and store them in 'ADFD'
ADFD <- numeric(N)

for(i in 1:ncol(RWD)) {
  ADFD[i] <- summary(
    dynlm(diff(RWD[, i], 1) ~ L(RWD[, i], 1)))$coef[2, 3]
}

# simulate N random walks with drift + trend
RWDT <- ts(replicate(n = N, 
                    trend + drift + arima.sim(model = list(order = c(0, 1, 0)), 
                                              n = n - 1)))

# compute ADF test statistics and store them in 'ADFDT'
ADFDT <- numeric(N)

for(i in 1:ncol(RWDT)) {
  ADFDT[i] <- summary(
    dynlm(diff(RWDT[, i], 1) ~ L(RWDT[, i], 1) + trend(RWDT[, i], scale = F))
  )$coef[2, 3]
}
# estimate quantiles for ADF regression with a drift
round(quantile(ADFD, c(0.1, 0.05, 0.01)), 2)
##   10%    5%    1% 
## -2.62 -2.83 -3.39
# estimate quantiles for ADF regression with drift and trend
round(quantile(ADFDT, c(0.1,0.05,0.01)),2)
##   10%    5%    1% 
## -3.11 -3.43 -3.97

The estimated quantiles are close to the large-sample critical values of the ADF test statistic reported in Table 14.4 of the book.

Table 14.2: Large Sample Critical Values of ADF Test
Deterministic Regressors 10% 5% 1%
Intercept only -2.57 -2.86 -3.43
Intercept and time trend -3.12 -3.41 -3.96

The results show that using standard normal critical values is erroneous: the 5% critical value of the standard normal distribution is 1.64. For the Dickey-Fuller distributions the estimated critical values are 2.87 (drift) and 3.43 (drift and linear time trend). This implies that a true null (the series has a stochastic trend) would be rejected far too often if inappropriate normal critical values were used.

We may use the simulated test statistics for a graphical comparison of the standard normal density and (estimates of) both Dickey-Fuller densities.

# plot standard normal density
curve(dnorm(x), 
      from = -6, to = 3, 
      ylim = c(0, 0.6), 
      lty = 2,
      ylab = "Density",
      xlab = "t-Statistic",
      main = "Distributions of ADF Test Statistics",
      col = "darkred", 
      lwd = 2)

# plot density estimates of both Dickey-Fuller distributions
lines(density(ADFD), lwd = 2, col = "darkgreen")
lines(density(ADFDT), lwd = 2, col = "blue")

# add a legend
legend("topleft", 
       c("N(0,1)", "Drift", "Drift+Trend"),
       col = c("darkred", "darkgreen", "blue"),
       lty = c(2, 1, 1),
       lwd = 2)Hide Source
Hide Plot

The deviations from the standard normal distribution are significant: both Dickey-Fuller distributions are skewed to the left and have a heavier left tail than the standard normal distribution.

Does U.S. GDP Have a Unit Root?

As an empirical example, we use the ADF test to assess whether there is a stochastic trend in U.S. GDP using the regression Δlog(GDPt)=β0+αt+β1log(GDPt1)+β2Δlog(GDPt1)+β3Δlog(GDPt2)+ut.
# generate log GDP series
LogGDP <- ts(log(GDP["1962::2012"]))

# estimate the model
coeftest(
  dynlm(diff(LogGDP) ~ trend(LogGDP, scale = F) + L(LogGDP) 
                     + diff(L(LogGDP)) + diff(L(LogGDP), 2)))
## 
## t test of coefficients:
## 
##                             Estimate  Std. Error t value Pr(>|t|)   
## (Intercept)               0.27877045  0.11793233  2.3638 0.019066 * 
## trend(LogGDP, scale = F)  0.00023818  0.00011090  2.1476 0.032970 * 
## L(LogGDP)                -0.03332452  0.01441436 -2.3119 0.021822 * 
## diff(L(LogGDP))           0.08317976  0.11295542  0.7364 0.462371   
## diff(L(LogGDP), 2)        0.18763384  0.07055574  2.6594 0.008476 **
## ---
## Signif. codes:  0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
The estimation yields Δlog(GDPt)=0.28(0.118)+0.0002(0.0001)t0.033(0.014)log(GDPt1)+0.083(0.113)Δlog(GDPt1)+0.188(0.071)Δlog(GDPt2)+ut,

so the ADF test statistic is t=0.033/0.014=2.35. The corresponding 5% critical value from Table 14.2 is 3.41 so we cannot reject the null hypothesis that log(GDP) has a stochastic trend in favor of the alternative that it is stationary around a deterministic linear time trend.

The ADF test can be done conveniently using ur.df() from the package urca.

# test for unit root in GDP using 'ur.df()' from the package 'urca'
summary(ur.df(LogGDP, 
              type = "trend", 
              lags = 2, 
              selectlags = "Fixed"))
## 
## ############################################### 
## # Augmented Dickey-Fuller Test Unit Root Test # 
## ############################################### 
## 
## Test regression trend 
## 
## 
## Call:
## lm(formula = z.diff ~ z.lag.1 + 1 + tt + z.diff.lag)
## 
## Residuals:
##       Min        1Q    Median        3Q       Max 
## -0.025580 -0.004109  0.000321  0.004869  0.032781 
## 
## Coefficients:
##               Estimate Std. Error t value Pr(>|t|)    
## (Intercept)  0.2790086  0.1180427   2.364 0.019076 *  
## z.lag.1     -0.0333245  0.0144144  -2.312 0.021822 *  
## tt           0.0002382  0.0001109   2.148 0.032970 *  
## z.diff.lag1  0.2708136  0.0697696   3.882 0.000142 ***
## z.diff.lag2  0.1876338  0.0705557   2.659 0.008476 ** 
## ---
## Signif. codes:  0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
## 
## Residual standard error: 0.007704 on 196 degrees of freedom
## Multiple R-squared:  0.1783, Adjusted R-squared:  0.1616 
## F-statistic: 10.63 on 4 and 196 DF,  p-value: 8.076e-08
## 
## 
## Value of test-statistic is: -2.3119 11.2558 4.267 
## 
## Critical values for test statistics: 
##       1pct  5pct 10pct
## tau3 -3.99 -3.43 -3.13
## phi2  6.22  4.75  4.07
## phi3  8.43  6.49  5.47

The first test statistic at the bottom of the output is the one we are interested in. The number of test statistics reported depends on the test regression. For type = “trend”, the second statistics corresponds to the test that there is no unit root and no time trend while the third one corresponds to a test of the hypothesis that there is a unit root, no time trend and no drift term.

References

Dickey, D. A., & Fuller, W. A. (1979). Distribution of the Estimators for Autoregressive Time Series with a Unit Root. Journal of the American Statistical Association, 74(366), pp. 427–431.


  1. The t-statistic of the Dickey-Fuller test is computed using homoskedasticity-only standard errors since under the null hypothesis, the usual t-statistic is robust to conditional heteroskedasticity.