16.4 Estimation of SI Model
Consider the SI model (16.1) - (16.5). The asset specific parameters to be estimated are αi, βi and σ2ϵ,i, (i=1,…,N), and the market parameters to be estimated are μM and σ2M. These parameters can be estimated using the plug-in principle, linear regression, and maximum likelihood. All methods give essentially the same estimators for the SI model parameters.
16.4.1 Plug-in principle estimates
Let {(rit,rMt)}Tt=1 denote a sample of size T of observed returns on asset i and the market return which are assumed to be generated from the SI model. (16.1) - (16.12). Recall, the plug-in principle says to estimate population model parameters using appropriate sample statistics. For the market parameters, the plug-in principle estimates are the same as the CER model estimates ˆμM=1TT∑t=1rMt,ˆσ2M=1T−1T∑t=1(rMt−ˆμM)2. From (16.13) - (16.6) we see that αi and βi are functions of population parameters αi=μi−βiμM,βi=cov(Rit,RMt)var(RMt)=σiMσ2M. The corresponding plug-in principle estimates are then: ˆαi=ˆμi−ˆβiˆμM,ˆβi=ˆσiMˆσ2M, where ˆμi=1TT∑t=1rit,ˆσiM=1T−1T∑t=1(rit−ˆμi)(rMt−ˆμM).
Given the plug-in principle estimates ˆαi and ˆβi, the plug-in principle estimate of ϵit is ˆεit=rit−ˆαi−ˆβirMt,t=1,…,T. Using (16.23), the plug-in principle estimate of σ2ϵ,i is the sample variance of {ˆϵit}Tt=1 (adjusted for the number of degrees of freedom): ˆσ2ε,i=1T−2T∑t=1ˆε2t=1T−2T∑t=1(rit−ˆαi−ˆβirMt)2.
Plug-in principle estimates of R2 based on (16.15) can be computed using ˆR2=ˆβ2iˆσ2Mˆσ2i=1−ˆσ2ε,iˆσ2i.
Consider computing the plug-in principle estimates for αi, βi and σ2ϵ,i from the example data using the formulas (16.21), (16.22) and (16.24), respectively. First, extract the sample statistics ˆμi, ˆσiM, ˆμM, and ˆσ2M:
colnames(siRetS)[1:4]
assetNames = colMeans(siRetS)
muhat = diag(covmatHat)
sig2hat = covmatHat[assetNames, "SP500"] covAssetsSp500 =
Next, estimate ˆβi using
covAssetsSp500/sig2hat["SP500"]
betaHat = betaHat
## BA JWN MSFT SBUX
## 0.978 1.485 1.303 1.057
Here, we see that ˆβBA and ˆβSBUX are very close to one and that ˆβJWN and ˆβMSFT are slightly bigger than one. Using the estimates of ˆβi and the sample statistics ˆμi and ˆμM the estimates for ˆαi are
muhat[assetNames] - betaHat*muhat["SP500"]
alphaHat = alphaHat
## BA JWN MSFT SBUX
## 0.00516 0.01231 0.00544 0.01785
All of the estimates of ˆαi are close to zero. The estimates of σ2ϵ,i can be computed using:
rep(0, length(assetNames))
sig2eHat =names(sig2eHat) = assetNames
for (aName in assetNames) {
siRetS[, aName] - alphaHat[aName] - betaHat[aName]*siRetS[, "SP500"]
eHat = crossprod(eHat)/(length(eHat) - 2)
sig2eHat[aName] =
} sig2eHat
## BA JWN MSFT SBUX
## 0.00581 0.00994 0.00646 0.00941
Lastly, the estimates of R2 can be computed using
1 - sig2eHat/sig2hat[assetNames]
R2 = R2
## BA JWN MSFT SBUX
## 0.270 0.334 0.373 0.210
◼
16.4.2 Least squares estimates
The SI model representation (16.1) shows that returns are a linear function of the market return and an asset specific error term Rit=αi+βiRMt+ϵit, here αi is the intercept and βi is the slope. Least squares regression is a method for estimating αi and βi by finding the “best fitting” line to the scatterplot of returns where Rit is on the vertical axis and RMt is on the horizontal axis.
{[}Insert Figure here{]} To be completed…
To see how the method of least squares determines the “best fitting” line, consider the scatterplot of the sample returns on Boeing and the S&P 500 index illustrated in Figure xxx. In the figure, the black line is a fitted line with initial guess ˆαBA=0 and ˆβBA=0.5. The differences between the observed returns (blue dots) and the values on the fitted line are the estimated errors ˆϵBA,t=rBA,t−ˆαBA−ˆβBArMt=rBA,t−0−0.5×RMt. Some estimated errors are big and some are small. The overall fit of the line can be measured using a statistic based on all t=1,…,T of the estimated errors. A natural choice is the sum of the errors ∑Tt=1ˆϵt. However, this choice can be misleading due to the canceling out of large positive and negative errors. To avoid this problem, it is better to measure the overall fit using ∑Tt=1^|ϵt| or ∑Tt=1ˆϵ2t. Then the best fitting line can be determined by finding the intercept and slope values that minimize ∑Tt=1^|ϵt| or ∑Tt=1ˆϵ2t.
The method of least squares regression defines the “best fitting” line by finding the intercept and slope values that minimize the sum of squared errors SSE(ˆαi,ˆβi)=T∑t=1ˆϵ2it=T∑t=1(rit−ˆαi−ˆβirMt)2. Because SSE(ˆα,ˆβ) is a continuous and differential function of ˆαi and ˆβi, the minimizing values of ˆαi and ˆβi can be determined using simple calculus. The first order conditions for a minimum are: 0=∂SSE(ˆαi,ˆβi)∂ˆαi=−2T∑t=1(rit−ˆαi−ˆβirMt)=−2T∑t=1ˆεit,0=∂SSE(ˆαi,ˆβi)∂ˆβi=−2T∑t=1(rit−ˆαi−ˆβirMt)rMt=−2T∑t=1ˆεitrMt. These are two linear equations in two unknowns which can be re-expressed as ˆαiT+ˆβiT∑t=1rMt=T∑t=1rit,ˆαiT∑t=1rMt+ˆβiT∑t=1r2Mt=T∑t=1ritrMt. Using matrix algebra, we can write these equations as: (T∑Tt=1rMt∑Tt=1rMt∑Tt=1r2Mt)(ˆαiˆβi)=(∑Tt=1rit∑Tt=1ritrMt), which is of the form Ax=b with A=(T∑Tt=1rMt∑Tt=1rMt∑Tt=1r2Mt),x=(ˆαiˆβi),b=(∑Tt=1rit∑Tt=1ritrMt). Hence, we can determine ˆαi and ˆβi by solving x=A−1b. Now,98 A−1=1det(A)(∑Tt=1r2Mt−∑Tt=1rMt−∑Tt=1rMtT),det(A)=TT∑t=1r2Mt−(T∑t=1rMt)2=TT∑t=1(rMt−ˆμM)2,ˆμM=1TT∑t=1rMt. Consequently, (ˆαiˆβi)=1T∑Tt=1(rMt−ˆμM)2(∑Tt=1r2Mt−∑Tt=1rMt−∑Tt=1rMtT)(∑Tt=1rit∑Tt=1ritrMt) and so ˆαi=∑Tt=1r2Mt∑Tt=1rit−∑Tt=1r2Mt∑Tt=1ritrMtT∑Tt=1(rMt−ˆμM)2,ˆβi=T∑Tt=1ritrMt−∑Tt=1rMt∑Tt=1ritT∑Tt=1(rMt−ˆμM)2. After a little bit of algebra (see end-of-chapter exercises) it can be shown that ˆαi=ˆμi−ˆβiˆμM,ˆβi=ˆσiMˆσ2M, which are plug-in estimates for ˆαi and ˆβi determined earlier. Hence, the least squares estimates of ˆαi and ˆβi are identical to the plug-in estimates.
The solution for the least squares estimates in (16.30) has an elegant representation using matrix algebra. To see this, define the T×1 vectors ri=(ri1,…,riT)′, rM=(rM1,…,rMT)′ and 1=(1,…,1)′. Then we can re-write (16.29) as (1′11′rM1′rMr′MrM)(ˆαiˆβi)=(1′rir′Mri) or X′Xˆγi=X′ri where X=(1rM)is a T×2 matrix and ˆγ=(ˆαi,ˆβi)′. Provided X′X is invertible, solving (16.33) for ˆγi gives the least squares estimates in matrix form: ˆγi=(X′X)−1X′ri. The matrix form solution (16.34) is especially convenient for computation in R.
The least squares estimates of ϵt, σ2ϵ,i and R2 are the same as the plug-in estimators (16.23), (16.24) and (16.25), respectively. In the context of least squares estimation, the estimate ˆσϵ,i=√ˆσ2ϵ,i is called the standard error of the regression and measures the typical magnitude of ˆϵt (difference between observed return and fitted regression line).
16.4.3 Simple linear regression in R
- don’t do regression examples until statistical theory is discussed
- computing least squares estimates using matrix algebra formulas
- computing least squares estimates using
lm()
- See discussion from my regression chapter in MFTSR
- describe structure of
lm()
function, extractor and method functions
- Do analysis of example data To be completed…
16.4.4 Maximum likelihood estimates
The SI model parameters can also be estimated using the method of maximum likelihood, which was introduced in chapter (GARCH estimation chapter). To construct the likelihood function, we use property () of the SI model that conditional on RMt=rMt the distribution of Rit is normal with mean αi+βirMt and variance σ2ϵ,i. The pdf of Rit|RMt=rmt is then f(rit|rmt,θi)=(2πσ2ε,i)−1/2exp(−12σ2ε,i(rit−αi+βirMt)2),t=1,…,T, where θi=(αi,βi,σ2ϵ,i)′. Given a sample {(rit,rMt)}Tt=1={ri,rM} of observed returns on asset i and the market return, which are assumed to be generated from the SI model, the joint density of asset returns given the market returns is f(ri|rm)=T∏t=1(2πσ2ε,i)−1/2exp(−12σ2ε,i(rit−αi+βirMt)2)=(2πσ2ε,i)−T/2exp(−12σ2ε,iT∑t=1(rit−αi+βirMt)2)=(2πσ2ε,i)−T/2exp(−12σ2ε,iSSE(αi,βi)). where SSE(αi,βi) is the sum of squared residuals (16.26) used to determine the least squares estimates. The log-likelihood function for θi is then lnL(θi|ri,rM)=−T2ln(2π)−T2ln(σ2ε,i)−12σ2ε,iSSE(αi,βi). From (16.35), it can be seen that the values of αi and βi that maximize the log-likelihood are the values that minimize SSE(αi,βi). Hence, the ML estimates of αi and βi are the least squares estimates.
To find the ML estimate for σ2ϵ,i, plug the ML estimates of αi and βi into (16.35) giving lnL(ˆαi,ˆβi,σ2ϵ,i|ri,rM)=−T2ln(2π)−T2ln(σ2ε,i)−12σ2ε,iSSE(ˆαi,ˆβi). Maximization with respect to σ2ϵ,i gives the first order condition ∂lnL(ˆαi,ˆβi,σ2ϵ,i|ri,rM)∂σ2ϵ,i=−T2ˆσ2ϵ,i+12(ˆσ2ϵ,i)2SSE(ˆαi,ˆβi)=0. Solving for ˆσ2ϵ,i gives the ML estimate for σ2ϵ,i: ˆσ2ϵ,i=SSE(ˆαi,ˆβi)T=1TT∑t=1ˆϵ2t, which is plug-in principle estimate (16.24) not adjusted for degrees-of-freedom.
The matrix A is invertible provided det(A)≠0. This requires the sample variance of RMt to be non-zero. ↩︎