Chapter 21 Homework 1: Properties of Stochastic Process: Problems and Tentative Solutions
Exercise 21.1 (Weakly stationary process) Consider real valued random variable Ai,Bi, i=1,⋯,k such that E(Ai)=E(Bi)=0 and Var(Ai)=Var(Bi)=σ2>0 for i=1,⋯,k. Moreover, assume they are mutually uncorrelated, that is, E(AiAl)=E(BiBl)=0, for i≠l, and E(AiBl)=0, for all i,l. Define the stochastic process X={Xt,t∈R} by Xt=∑ki=1(Aicos(ωit)+Bsin(ωit)), where ωi are real constants i=1,⋯,k. Show that X is weakly stationary.
Proof. Firstly, consider the expectation, for any t∈R, we have E(Xt)=E(k∑i=1(Aicos(ωit)+Bisin(ωit)))=k∑i=1cos(ωit)E(Ai)+sin(ωit)E(Bi)=0 As a function of t, E(Xt) is a constant.
Then, for the covariance between Xt and Xt+r of any t,t+r∈R, since E(Xt)=E(Xt+r)=0, we have Cov(Xt,Xt+r)=E(XtXt+r)=E[{k∑i=1(Aicos(ωit)+Bisin(ωit))}×{k∑i=1(Aicos(ωi(t+r))+Bisin(ωi(t+r))}]=k∑i=1E(A2i)cos(ωit)cos(ωi(t+r))+E(B2i)sin(ωit)sin(ωi(t+r))=k∑i=1σ2icos(ωir) which is a function of r only. Thus, the stochastic process Xt is weakly stationary.
Exercise 21.2 (Weakly but not strongly stationary process) Consider a discrete-time, real-valued stochastic process X={Xn:n≥1} defined by Xn=cos(nU), where U is uniformly distributed on (−π,π). Show that X is weakly stationary but not strongly stationary.
Proof. Firstly, consider the expectation of Xn for any n∈N+, we have E(Xn)=E(cos(nU))=∫π−πcos(nu)⋅12πdu=0 which is a constant as a function of n.
Then, for the covariance between Xn and Xn+m with any n∈N+ and m∈N, E(Xn)=E(Xn+m)=0, so Cov(Xn,Xn+m)=E(XtXtr)=∫π−πcos(nu)cos((n+m)u)⋅12πdu=14π[∫π−π(cos((2n+m)u)+cos(mu))du]=14π[sin((2n+m)u)2n+m|π−π+sin(mu)m|π−π]=0 which is a function of m. Therefore, the stochastic process is weakly stationary.
The process is not strongly stationary. For example, consider the event [X1≤−1+ϵ,X3≥1−ϵ] for some small ϵ>0. The event has zero probability because X1≤1−ϵ happens when U is near −π or π, but X3≥1−ϵ happens when U is near −2π3,0 or −2π3. Therefore, we can pick small enough ϵ such that those regions are not overlapped. However, the event [X2≤−1+ϵ,X4≥1−ϵ] have positive probability because we can pick U around π2 to make both conditions hold.
Exercise 21.3 (Find correlation function from spectral density) Consider a weakly stationary process X={Xt:t∈R} with zero mean and unit variance. Find the correlation function of X if the spectral density function f of X is given by:
- f(u)=0.5exp(−|u|), u∈R
- f(u)=ϕ(α2+u2)−1, u∈R
- f(u)=12σ(πα)−1exp(−u2/(4α)), u∈R
- f(u)=12σ(πα)−1exp(−u2/(4α)), u∈R
Proof. Since we have variance equal to 1, the correlation function is just covariance function. Now using the one-to-one correspondence between the spectral density and covariance function c(t)=∫∞−∞exp(itu)f(u)du we can get the correlation function for each case.
For part (a), we have c(t)=∫∞−∞0.5exp(−|u|+itu)du=EU1(exp(itu))=11+t2 where since U1∼Laplace(0,1) we have the last equation. Therefore, in this case, c(t)=11+t2.
For part(b), we have c(t)=∫∞−∞ϕexp(itu)α2+u2du=ϕπαEU2(exp(itu))=ϕπαexp(α|t|) where since U2∼Cauchy(0,α) we have the last equation. Therefore, c(t)=ϕπαexp(α|t|).
Finally, for part (c) c(t)=∫∞−∞σexp(itu−u2/(4α))2παdu=σ√παEU3(exp(itu))=σexp(−αt2)√πα where since U3∼N(0,2α) we have the last equation. Therefore, c(t)=σexp(−αt2)√πα.
Proof. For any stochastic process, the strong stationary implies weak stationary. Therefore, for Gaussian process, we only need to show the weak stationary implies strong stationary.
Consider a weakly stationary Gaussian process Xt with mean function m(⋅) and covariance function C(⋅,⋅). By definition, any finite dimensional distributions of it is F(X1,⋯,Xk)=Φ(m(X),C(X)) where Φ(m(X),C(X)) represent the distribution function of a multivariate normal distribution with mean vector m(X)=(m(X1),⋯,m(Xk))T and covariance matrix C(X), whose i,jth term is C(Xi,Xj) for 1≤i,j≤k. Similarly, the distribution for X1+t0,⋯,Xk+t0 is also multivariate normal with mean vector mt0(X)=(m(X1+t0),⋯,m(Xk+t0))T and covariance matrix Ct0(X), whose i,jth term is C(Xi+t0,Xj+t0). From the weakly stationary assumption, we have m(X)=mt0(X) and C(X)=Ct0(X), and since they are all multivariate normal distribution, we have F(X1,⋯,Xk)=F(X1+t0,⋯,Xk+t0) for any k∈N+ and t0∈T. Thus, the Gaussian process is also strongly stationary.
Exercise 21.5 (Necessary and sufficient condition for a Gaussian process to be Markovian) By definition, a continuous-time real-valued stochastic process X={Xt:t∈R} is called a Markov process if for all n, for all x1,⋯,xn, and all increasing sequences t1<⋯<tn of index points, Pr(Xtn≤x|Xt1=x1,⋯,Xtn−1=xn−1)=Pr(Xtn≤x|Xtn−1=xn−1) Let Z be a real valued Gaussian process. Show that Z is a Markov process if and only if E(Ztn|Zt1=x1,⋯,Ztn−1=xn−1)=E(Ztn|Ztn−1=xn−1)
Proof. (⟸) This is trivial since if Z is a Markov process, then by definition for all n and for all x,x1,⋯,xn−1 and any sequence of index points t1<⋯<tn, we have Pr(Ztn≤x|Zt1=x1,⋯,Ztn−1=xn−1)=Pr(Ztn≤x|Ztn−1=xn−1). Then of course E(Ztn|Zt1=x1,⋯,Ztn−1=xn−1)=E(Ztn|Ztn−1=xn−1). We have the necessarity.
(⟹) Since the f.d.d.s. of Guassian process is multivariate normal distribution, we can actually write down the explict form of the distribution of Ztn|Zt1,⋯,Ztn−1 and Ztn|Ztn−1 using the properties of multivariate normal distribution. Actually, let us denote the joint distribution of (Ztn,Ztn−1) as (ZtnZtn−1)∼N((μtnμtn−1),(Σ11Σ12Σ21Σ22)) Then, Ztn|Ztn−1=xn−1∼N(μtn+Σ12Σ−122(xn−1−μn−1),Σ11−Σ12Σ−122Σ21) for any xn−1.
Similarly, denote the joint distribution of (Ztn,Ztn−1) where Ztn−1=(Ztn−1,⋯,Zt1) as \begin{equation} \begin{pmatrix} Z_{t_n}\\\mathbf{Z}_{t_{n-1}}\end{pmatrix}\sim N(\begin{pmatrix} \mu_{t_n}\\ \boldsymbol{\mu}_{t_{n-1}}\end{pmatrix},\begin{pmatrix} \Lambda_{11} & \Lambda_{12}\\ \Lambda_{21} & \Lambda_{22}\end{pmatrix}) \tag{21.14} \end{equation} Then, Z_{t_n}|\mathbf{Z}_{t_{n-1}}=\mathbf{x}_{n-1}\sim N(\mu_{t_n}+\Lambda_{12}\Lambda_{22}^{-1}(\mathbf{x}_{n-1}-\boldsymbol{\mu}_{n-1}),\Lambda_{11}-\Lambda_{12}\Lambda_{22}^{-1}\Lambda_{21}) for any vector \mathbf{x}_{n-1}=(x_1,\cdots,x_{n-1}).
Now, from the assumption, we know \begin{equation} \mu_{t_n}+\Sigma_{12}\Sigma_{22}^{-1}(x_{n-1}-\mu_{n-1})=\mu_{t_n}+\Lambda_{12}\Lambda_{22}^{-1}(\mathbf{x}_{n-1}-\boldsymbol{\mu}_{n-1}) \tag{21.15} \end{equation} for any x_{n-1} and \mathbf{x}_{n-1}. Then specifically we can choose x_{n-1}=\mu_{n-1}+\Sigma_{21} and \mathbf{x}_{n-1}=\boldsymbol{\mu}_{n-1}+\Lambda_{21} and subsitituted in () we have \begin{equation} \Sigma_{12}\Sigma_{22}^{-1}\Sigma_{21}=\Lambda_{12}\Lambda_{22}^{-1}\Lambda_{21} \tag{21.16} \end{equation} from which we can see the two distributions in (21.13) and (21.14) are identical. Therefore, the condition for Markov process is satisfied and Z is indeed a Markov process. We have the sufficiency.
Exercise 21.6 (Condition for Markovian) Show that any stochastic process X=\{X_n,n=0,\cdots\} with independent increments is a Markov process.
Proof. Consider a stochastic process X=\{X(n,\omega):n\in\mathbb{N}_+,\omega\in\Omega\} that has independent increments. Then for any n\geq 1 and any s, we have \begin{equation} \begin{split} &P(X_n=s|X_0=x_0,X_1=x_1,\cdots,X_{n-1}=x_{n-1})\\ &=P(X_n-X_{n-1}=s-x_{n-1}|X_0=x_0,X_1-X_0=x_1-x_0,\\ &X_2-X_1=x_2-x_1,\cdots,X_{n-1}-X_{n-2}=x_{n-1}-x_{n-2})\\ &=P(X_n-X_{n-1}=s-x_{n-1})\quad (by\,independence)\\ &=P(X_n-X_{n-1}=s-x_{n-1}|X_{n-1}=x_{n-1})\\ &=P(X_n=s|X_{n-1}=x_{n-1}) \end{split} \tag{21.17} \end{equation} Therefore, by defintion of Markov process, we know that X is actually a Markov process.
Exercise 21.7 (View Brownian motion as a special Gaussian process) Let W=\{W_t:t\geq 0\} be a Brownian motion (see Definition 2.5). Show that a Brownian motion can be viewed as a Gaussian process with mean 0 and Cov(W_s,W_t)=\min\{s,t\}.
Proof. Consider any finite dimensional distribution of W, denoted as F_{W_{t_1},\cdots,W_{t_k}}(w_{t_1},\cdots,w_{t_k}), w.l.o.g. we can assume t_1<\cdots<t_k and for simplicity of notation, we can surpass t and just denote the f.d.d.s. as F_{W_1,\cdots,W_k}(w_1,\cdots,w_k). Firstly, from the independent increment property and W_t-W_s\sim N(0,t-s) for 0\leq s\leq t, since W_1,\cdots,W_k can be expressed as a linear combination of independent normally distributed random variable W_1,W_2-W_1,\cdots,W_k-W_{k-1}, we have the f.d.d.s. of W is normally distributed. The mean is E(W_t)=0 and variance Var(W_t)=t, while the covariance is, assume w.l.o.g. s<t \begin{equation} \begin{split} Cov(W_sW_t)&=E(W_sW_t)=E(W_s(W_t-W_s+W_s))\\ &=E(W_s^2)+E(W_s)E(W_t-W_s)=s \end{split} \tag{21.18} \end{equation} Thus, the Brownian motion is a Gaussian process with mean function 0 and covariance function Cov(W_s,W_t)=\min\{s,t\}.
Exercise 21.8 (Moments of Brownian motion) Show that for a Brownian motion E(|W_s-W_t|^{2n})=C_n|s-t|^n, where C_n=\frac{2n!}{2^nn!}.
Proof. Suppose Y\sim N(0,\nu), we compute E(|Y|^{2n}) as follows: \begin{equation} \begin{split} E(|y|^{2n})&=\frac{1}{\sqrt{2\pi\nu}}\int_{-\infty}^{\infty}|y|^{2n}\exp(-\frac{y^2}{2\nu})dy\\ &=\frac{2}{\sqrt{2\pi\nu}}\int_{0}^{\infty}y^{2n}\exp(-\frac{y^2}{2\nu})dy\quad (y=\sqrt{2\nu z})\\ &=\frac{2}{\sqrt{2\pi\nu}}\int_{0}^{\infty}(2\nu z)^{n}\exp(-\frac{y^2}{2\nu})\frac{\nu dz}{\sqrt{2\nu z}}\\ &=\frac{\nu^n2^n}{\pi}\int_{0}^{\infty}z^{n-\frac{1}{2}}e^{-z}dz\\ &=\frac{\nu^n2^n}{\pi}\Gamma(n+\frac{1}{2})\\ &=\frac{(2n)!}{2^nn!}\nu^n=C_n\nu^n \end{split} \tag{21.19} \end{equation}
Now since W_s-W_t follows N(0,|s-t|), using the above result we immediately have E(|W_s-W_t|^{2n})=C_n|s-t|^n where C_n=\frac{(2n)!}{2^nn!}.