Time Series 3

Paulo Fagandini

Nova SBE

Recap

Assumptions

SLR MLR TS
Linearity in \(\beta\) SLR1 MLR1 TS1
Random sampling SLR2 MLR2
Sample variation in \(X\) SLR3
No perfect collinearity MLR4 TS2
\(E[u|X]=0\) SLR4 MLR4 TS3
\(V[u|X]=\sigma^2\) SLR5 MLR5 TS4
\(Corr[u_su_t|X]=0\) for \(t\neq s\) TS5

Stationarity and Weakly Dependent Time Series

The problem

Large samples in time series (normal frequency time series) are hard to come by!

Stationary Process

Stationary process

The stochastic process \(\{x_t\}_{t\in\mathbb{N}}\) is stationray if for any monotonically increasing function \(g:\mathbb{N}\mapsto\mathbb{N}\), the joint distribution of \[(x_{g(1)},x_{g(2)},\dots,x_{g(m)})\] is equal to the distribution of \[(x_{g(1)+h},x_{g(2)+h},\dots,x_{g(m)}+h)\] for any \(h\in\mathbb{N}\).

Stationary Process

A stochastic process that is not stationary is said to be nonstationary.🥁

It is difficult to identify a stationary process because you do not have all the data! However it can be very simple to identify a non-stationary process, for example a time series with a trend.

For some results, a weaker form of stationary is enough…

Covariance Stationary Process

Important

A stochastic process \(\{x_t, t\in\mathbb{N}\}\) with a finite second moment \(E[x_t^2]<\infty\), is covariance stationary if:

  1. \(E[x_t]\) is constant
  2. \(V[x_t]\) is constant
  3. \(\forall t,h\in\mathbb{N}\) we have \(Cov(x_t,x_{t+h})=\phi(h)\), i.e. it depends only on \(h\), and not \(t\).

Covariance Stationary Process

If a stationary process has a finite second moment, then it is Covariance Stationary as well. Note that converse is not true, i.e. a Covariance Stationary Process needs not be Stationary.

Weakly dependent Time Series

Important

A stationary time series process \(\{x_t, t\in\mathbb{N}\}\) is said to be weakly dependent if \(x_t\) and \(x_{t+h}\) are “almost independent” as \(h\rightarrow\infty\).

We could say that intuitively a time series is weakly dependent if the correlation between \(x_t\) and \(x_{t+h}\) gets smaller and smaller as \(h\) increases.

Weakly dependent Time Series

Important

A stochastic process \(\{x_t, t\in\mathbb{N}\}\) is said to be asymtotically uncorrelated if

\[Corr(x_t,x_{t+h})\underset{h\rightarrow\infty}{\rightarrow}0\]

This is basically how we can characterize weakly dependnece.

Moving Averages

Moving average process

The moving average process is an example of a weakly dependent process:

\[x_t = e_t + \alpha_1 e_{t-1},\quad t\in\mathbb{N}\]

where \(\{e_t\}_{t\in\mathbb{N}}\) is an iid sequence with zero mean and variance \(\sigma_e^2\). In particular, this process is called a moving average proces of order one MA(1)

Autoregressive processes

Consider now the following process

Important

##Autoregressive procees of order one AR(1) \[y_t=\rho y_{t-1}+e_t,\quad t\in\mathbb{N}\] with \(e_t\) and iid sequence with \(E[e]=0\) and \(V[e]=\sigma_e^2\). Let \(e_t\) be independent of \(y_0\) and \(E[y_0]=0\). This is called an autoregressive process of order one AR(1)

We need a crucial assumption for the weak dependence of the AR(1) process, which is \(|\rho|<1\). In this case we say \(\{y_t\}\) is a stable AR(1) process.

Autoregressive processes

we can show easily that \[Corr(y_t,y_{t+h})=\frac{Cov(y_t,y_{t+h})}{\sigma_y^2}=\rho_1^h\]

Note that a consequence of this is that for any \(t\) \(Corr(y_t,y_{t+1})=\rho\)! and moreover, as \(|\rho|<1\), \(\rho^h\underset{h\rightarrow\infty}{\rightarrow}0\).

We have that the AR(1) process is weakly dependent.

Highly persistent Time Series in Regression Analysis

Random Walk

What happens if in the AR(1) model \(|\rho|\geq 1\)? We needed that for weakly dependence…

Sadly, many economic series are characterized with a \(\rho = 1\). Was it all for nothing then?

Let’s see, the model would be something like

\[ y_t = y_{t-1}+e_t\] with a well behaved error term. This process is known as a random walk.

Random Walk

Let’s characterize this process, what is \(E[y_t]\)?

\[y_t=y_{t-1}+e_t\quad \Rightarrow\quad E[y_t]=E[y_0],\ V[y_t]=\sigma_e^2 t\]

Note the consequences of this! \(y_0\) affects all the future values of \(y\), and moreover, the best predition for the future is the current value, \(E[y_{t+h}|y_t]=y_t\).

Also, the variance explodes! It can be shown further that \[Corr(y_t,y_{t+h})=\sqrt{\frac{t}{(t+h)}}\]

Random Walk

A random walk does not satisfy the requirement of an asymptotically uncorrelated sequence.

Indeed, a random walk is a special case of what is known as unit root process (so \(\rho=1\) in an AR(1) model).

Remember that a trending and a highly persisten series are not the same, however many times time series have both features, they are highly persistent and they have a trend. let’s see an exmaple of this.

Random walk with drift

A random walk with drift is a time series process characterized by the following expression:

\[ y_t = \alpha_0 + y_{t-1} + e_t,\quad t\in\mathbb{N}\]

Let \(\{e_t\}\) and \(y_0\) satisfy the properties for the normal random walk process model. Now if we iterate as we did previously:

\[y_t = \alpha_0 t + e_t + e_{t-1}+\dots+e_1+y_0\] and if \(y_0=0\), \(E[y_t]=\alpha_0 t\). That is, the expected value of \(y_t\) is growing with \(t\) (or decreasing if \(\alpha_0<1\))

References

You can find these contents and examples in chapter 11, sections 1 and 3 of the book.