16.4 Estimation of SI Model

Consider the SI model (16.1) - (16.5). The asset specific parameters to be estimated are αi, βi and σ2ϵ,i, (i=1,,N), and the market parameters to be estimated are μM and σ2M. These parameters can be estimated using the plug-in principle, linear regression, and maximum likelihood. All methods give essentially the same estimators for the SI model parameters.

16.4.1 Plug-in principle estimates

Let {(rit,rMt)}Tt=1 denote a sample of size T of observed returns on asset i and the market return which are assumed to be generated from the SI model. (16.1) - (16.12). Recall, the plug-in principle says to estimate population model parameters using appropriate sample statistics. For the market parameters, the plug-in principle estimates are the same as the CER model estimates ˆμM=1TTt=1rMt,ˆσ2M=1T1Tt=1(rMtˆμM)2. From (16.13) - (16.6) we see that αi and βi are functions of population parameters αi=μiβiμM,βi=cov(Rit,RMt)var(RMt)=σiMσ2M. The corresponding plug-in principle estimates are then: ˆαi=ˆμiˆβiˆμM,ˆβi=ˆσiMˆσ2M, where ˆμi=1TTt=1rit,ˆσiM=1T1Tt=1(ritˆμi)(rMtˆμM).

Given the plug-in principle estimates ˆαi and ˆβi, the plug-in principle estimate of ϵit is ˆεit=ritˆαiˆβirMt,t=1,,T. Using (16.23), the plug-in principle estimate of σ2ϵ,i is the sample variance of {ˆϵit}Tt=1 (adjusted for the number of degrees of freedom): ˆσ2ε,i=1T2Tt=1ˆε2t=1T2Tt=1(ritˆαiˆβirMt)2.

Plug-in principle estimates of R2 based on (16.15) can be computed using ˆR2=ˆβ2iˆσ2Mˆσ2i=1ˆσ2ε,iˆσ2i.

Example 3.2 (Computing plug-in principle estimators for SI model parameters)

Consider computing the plug-in principle estimates for αi, βi and σ2ϵ,i from the example data using the formulas (16.21), (16.22) and (16.24), respectively. First, extract the sample statistics ˆμi, ˆσiM, ˆμM, and ˆσ2M:

assetNames = colnames(siRetS)[1:4] 
muhat = colMeans(siRetS) 
sig2hat = diag(covmatHat)
covAssetsSp500 = covmatHat[assetNames, "SP500"]

Next, estimate ˆβi using

betaHat = covAssetsSp500/sig2hat["SP500"]
betaHat
##    BA   JWN  MSFT  SBUX 
## 0.978 1.485 1.303 1.057

Here, we see that ˆβBA and ˆβSBUX are very close to one and that ˆβJWN and ˆβMSFT are slightly bigger than one. Using the estimates of ˆβi and the sample statistics ˆμi and ˆμM the estimates for ˆαi are

alphaHat = muhat[assetNames] - betaHat*muhat["SP500"] 
alphaHat
##      BA     JWN    MSFT    SBUX 
## 0.00516 0.01231 0.00544 0.01785

All of the estimates of ˆαi are close to zero. The estimates of σ2ϵ,i can be computed using:

sig2eHat = rep(0, length(assetNames)) 
names(sig2eHat) = assetNames 
for (aName in assetNames) {   
     eHat = siRetS[, aName] - alphaHat[aName] - betaHat[aName]*siRetS[, "SP500"]   
     sig2eHat[aName] = crossprod(eHat)/(length(eHat) - 2) 
} 
sig2eHat
##      BA     JWN    MSFT    SBUX 
## 0.00581 0.00994 0.00646 0.00941

Lastly, the estimates of R2 can be computed using

R2 = 1 - sig2eHat/sig2hat[assetNames] 
R2
##    BA   JWN  MSFT  SBUX 
## 0.270 0.334 0.373 0.210

16.4.2 Least squares estimates

The SI model representation (16.1) shows that returns are a linear function of the market return and an asset specific error term Rit=αi+βiRMt+ϵit, here αi is the intercept and βi is the slope. Least squares regression is a method for estimating αi and βi by finding the “best fitting” line to the scatterplot of returns where Rit is on the vertical axis and RMt is on the horizontal axis.

{[}Insert Figure here{]} To be completed…

To see how the method of least squares determines the “best fitting” line, consider the scatterplot of the sample returns on Boeing and the S&P 500 index illustrated in Figure xxx. In the figure, the black line is a fitted line with initial guess ˆαBA=0 and ˆβBA=0.5. The differences between the observed returns (blue dots) and the values on the fitted line are the estimated errors ˆϵBA,t=rBA,tˆαBAˆβBArMt=rBA,t00.5×RMt. Some estimated errors are big and some are small. The overall fit of the line can be measured using a statistic based on all t=1,,T of the estimated errors. A natural choice is the sum of the errors Tt=1ˆϵt. However, this choice can be misleading due to the canceling out of large positive and negative errors. To avoid this problem, it is better to measure the overall fit using Tt=1^|ϵt| or Tt=1ˆϵ2t. Then the best fitting line can be determined by finding the intercept and slope values that minimize Tt=1^|ϵt| or Tt=1ˆϵ2t.

The method of least squares regression defines the “best fitting” line by finding the intercept and slope values that minimize the sum of squared errors SSE(ˆαi,ˆβi)=Tt=1ˆϵ2it=Tt=1(ritˆαiˆβirMt)2. Because SSE(ˆα,ˆβ) is a continuous and differential function of ˆαi and ˆβi, the minimizing values of ˆαi and ˆβi can be determined using simple calculus. The first order conditions for a minimum are: 0=SSE(ˆαi,ˆβi)ˆαi=2Tt=1(ritˆαiˆβirMt)=2Tt=1ˆεit,0=SSE(ˆαi,ˆβi)ˆβi=2Tt=1(ritˆαiˆβirMt)rMt=2Tt=1ˆεitrMt. These are two linear equations in two unknowns which can be re-expressed as ˆαiT+ˆβiTt=1rMt=Tt=1rit,ˆαiTt=1rMt+ˆβiTt=1r2Mt=Tt=1ritrMt. Using matrix algebra, we can write these equations as: (TTt=1rMtTt=1rMtTt=1r2Mt)(ˆαiˆβi)=(Tt=1ritTt=1ritrMt), which is of the form Ax=b with A=(TTt=1rMtTt=1rMtTt=1r2Mt),x=(ˆαiˆβi),b=(Tt=1ritTt=1ritrMt). Hence, we can determine ˆαi and ˆβi by solving x=A1b. Now,98 A1=1det(A)(Tt=1r2MtTt=1rMtTt=1rMtT),det(A)=TTt=1r2Mt(Tt=1rMt)2=TTt=1(rMtˆμM)2,ˆμM=1TTt=1rMt. Consequently, (ˆαiˆβi)=1TTt=1(rMtˆμM)2(Tt=1r2MtTt=1rMtTt=1rMtT)(Tt=1ritTt=1ritrMt) and so ˆαi=Tt=1r2MtTt=1ritTt=1r2MtTt=1ritrMtTTt=1(rMtˆμM)2,ˆβi=TTt=1ritrMtTt=1rMtTt=1ritTTt=1(rMtˆμM)2. After a little bit of algebra (see end-of-chapter exercises) it can be shown that ˆαi=ˆμiˆβiˆμM,ˆβi=ˆσiMˆσ2M, which are plug-in estimates for ˆαi and ˆβi determined earlier. Hence, the least squares estimates of ˆαi and ˆβi are identical to the plug-in estimates.

The solution for the least squares estimates in (16.30) has an elegant representation using matrix algebra. To see this, define the T×1 vectors ri=(ri1,,riT), rM=(rM1,,rMT) and 1=(1,,1). Then we can re-write (16.29) as (111rM1rMrMrM)(ˆαiˆβi)=(1rirMri) or XXˆγi=Xri where X=(1rM)is a T×2 matrix and ˆγ=(ˆαi,ˆβi). Provided XX is invertible, solving (16.33) for ˆγi gives the least squares estimates in matrix form: ˆγi=(XX)1Xri. The matrix form solution (16.34) is especially convenient for computation in R.

The least squares estimates of ϵt, σ2ϵ,i and R2 are the same as the plug-in estimators (16.23), (16.24) and (16.25), respectively. In the context of least squares estimation, the estimate ˆσϵ,i=ˆσ2ϵ,i is called the standard error of the regression and measures the typical magnitude of ˆϵt (difference between observed return and fitted regression line).

16.4.3 Simple linear regression in R

  • don’t do regression examples until statistical theory is discussed
  • computing least squares estimates using matrix algebra formulas
  • computing least squares estimates using lm()
    • See discussion from my regression chapter in MFTSR
    • describe structure of lm() function, extractor and method functions
  • Do analysis of example data To be completed…

16.4.4 Maximum likelihood estimates

The SI model parameters can also be estimated using the method of maximum likelihood, which was introduced in chapter (GARCH estimation chapter). To construct the likelihood function, we use property () of the SI model that conditional on RMt=rMt the distribution of Rit is normal with mean αi+βirMt and variance σ2ϵ,i. The pdf of Rit|RMt=rmt is then f(rit|rmt,θi)=(2πσ2ε,i)1/2exp(12σ2ε,i(ritαi+βirMt)2),t=1,,T, where θi=(αi,βi,σ2ϵ,i). Given a sample {(rit,rMt)}Tt=1={ri,rM} of observed returns on asset i and the market return, which are assumed to be generated from the SI model, the joint density of asset returns given the market returns is f(ri|rm)=Tt=1(2πσ2ε,i)1/2exp(12σ2ε,i(ritαi+βirMt)2)=(2πσ2ε,i)T/2exp(12σ2ε,iTt=1(ritαi+βirMt)2)=(2πσ2ε,i)T/2exp(12σ2ε,iSSE(αi,βi)). where SSE(αi,βi) is the sum of squared residuals (16.26) used to determine the least squares estimates. The log-likelihood function for θi is then lnL(θi|ri,rM)=T2ln(2π)T2ln(σ2ε,i)12σ2ε,iSSE(αi,βi). From (16.35), it can be seen that the values of αi and βi that maximize the log-likelihood are the values that minimize SSE(αi,βi). Hence, the ML estimates of αi and βi are the least squares estimates.

To find the ML estimate for σ2ϵ,i, plug the ML estimates of αi and βi into (16.35) giving lnL(ˆαi,ˆβi,σ2ϵ,i|ri,rM)=T2ln(2π)T2ln(σ2ε,i)12σ2ε,iSSE(ˆαi,ˆβi). Maximization with respect to σ2ϵ,i gives the first order condition lnL(ˆαi,ˆβi,σ2ϵ,i|ri,rM)σ2ϵ,i=T2ˆσ2ϵ,i+12(ˆσ2ϵ,i)2SSE(ˆαi,ˆβi)=0. Solving for ˆσ2ϵ,i gives the ML estimate for σ2ϵ,i: ˆσ2ϵ,i=SSE(ˆαi,ˆβi)T=1TTt=1ˆϵ2t, which is plug-in principle estimate (16.24) not adjusted for degrees-of-freedom.


  1. The matrix A is invertible provided det(A)0. This requires the sample variance of RMt to be non-zero. ↩︎