You are on page 1of 8

Solutions to the Review Questions at the End of Chapter 4 1.

In the same way as we make assumptions about the true value of beta and not the estimated values, we make assumptions about the true unobservable disturbance terms rather than their estimated counterparts, the residuals.
t = yt y t We know the exact value of the residuals, since they are defined by u . So we do not need to make any assumptions about the residuals since we already know their value. We make assumptions about the unobservable error terms since it is always the true value of the population disturbances that we are really interested in, although we never actually know what these are.

2. We would like to see no pattern in the residual plot If there is a pattern in the residual plot, this is an indication that there is still some !action" or variability left in yt that has not been explained by our model. #his indicates that potentially it may be possible to form a better model, perhaps using additional or completely different explanatory variables, or by using lags of either the dependent or of one or more of the explanatory variables. $ecall that the two plots shown on pages 1%& and 1%', where the residuals followed a cyclical pattern, and when they followed an alternating pattern are used as indications that the residuals are positively and negatively autocorrelated respectively. (nother problem if there is a !pattern" in the residuals is that, if it does indicate the presence of autocorrelation, then this may suggest that our standard error estimates for the coefficients could be wrong and hence any inferences we make about the coefficients could be misleading. ). #he t*ratios for the coefficients in this model are given in the third row after the standard errors. #hey are calculated by dividing the individual coefficients by their standard errors.
t + ,.-). / ,.0,2 x2t * ,..'1 x)t y
R 2 =0.96,R 2 =0.89

1,.0)-2 1,.2'12 1,.&-)2 t*ratios 1.01.). *1.1& #he problem appears to be that the regression parameters are all individually insignificant 1i.e. not significantly different from 3ero2, although the value of $2 and its ad4usted version are both very high, so that the regression taken as a whole seems to indicate a good fit. #his looks like a classic example of what we term near multicollinearity. #his is where the individual regressors are very closely related, so that it becomes difficult to disentangle the effect of each individual variable upon the dependent variable. #he solution to near multicollinearity that is usually suggested is that since the problem is really one of insufficient information in the sample to determine each of the coefficients, then one should go out and get more data. In other words, we should switch to a higher fre5uency of data for analysis

1/8

Introductory Econometrics for Finance !ris "roo#s 2008

1e.g. weekly instead of monthly, monthly instead of 5uarterly etc.2. (n alternative is also to get more data by using a longer sample period 1i.e. one going further back in time2, or to combine the two independent variables in a ratio 1e.g. x2t 6 x)t 2. 7ther, more ad hoc methods for dealing with the possible existence of near multicollinearity were discussed in 8hapter 09 Ignore it9 if the model is otherwise ade5uate, i.e. statistically and in terms of each coefficient being of a plausible magnitude and having an appropriate sign. Sometimes, the existence of multicollinearity does not reduce the t*ratios on variables that would have been significant without the multicollinearity sufficiently to make them insignificant. It is worth stating that the presence of near multicollinearity does not affect the :;<= properties of the 7;S estimator > i.e. it will still be consistent, unbiased and efficient since the presence of near multicollinearity does not violate any of the 8;$? assumptions 1*0. @owever, in the presence of near multicollinearity, it will be hard to obtain small standard errors. #his will not matter if the aim of the model*building exercise is to produce forecasts from the estimated model, since the forecasts will be unaffected by the presence of near multicollinearity so long as this relationship between the explanatory variables continues to hold over the forecasted sample. Arop one of the collinear variables * so that the problem disappears. @owever, this may be unacceptable to the researcher if there were strong a priori theoretical reasons for including both variables in the model. (lso, if the removed variable was relevant in the data generating process for y, an omitted variable bias would result. #ransform the highly correlated variables into a ratio and include only the ratio and not the individual variables in the regression. (gain, this may be unacceptable if financial theory suggests that changes in the dependent variable should occur following changes in the individual explanatory variables, and not a ratio of them.

0. 1a2 #he assumption of homoscedasticity is that the variance of the errors is 2 constant and finite over time. #echnically, we write Var %u t $ = u . 1b2 #he coefficient estimates would still be the !correct" ones 1assuming that the other assumptions re5uired to demonstrate 7;S optimality are satisfied2, but the problem would be that the standard errors could be wrong. @ence if we were trying to test hypotheses about the true parameter values, we could end up drawing the wrong conclusions. In fact, for all of the variables except the constant, the standard errors would typically be too small, so that we would end up re4ecting the null hypothesis too many times. 1c2 #here are a number of ways to proceed in practice, including

2/8

Introductory Econometrics for Finance !ris "roo#s 2008

* <sing heteroscedasticity robust standard errors which correct for the problem by enlarging the standard errors relative to what they would have been for the situation where the error variance is positively related to one of the explanatory variables. * #ransforming the data into logs, which has the effect of reducing the effect of large errors relative to small ones. %. 1a2 #his is where there is a relationship between the ith and jth residuals. $ecall that one of the assumptions of the 8;$? was that such a relationship did not exist. We want our residuals to be random, and if there is evidence of autocorrelation in the residuals, then it implies that we could predict the sign of the next residual and get the right answer more than half the time on average 1b2 #he Aurbin Watson test is a test for first order autocorrelation. #he test is calculated as follows. Bou would run whatever regression you were interested in, and obtain the residuals. #hen calculate the statistic

DW = t =2

( u

t 1 ) u
2 t

u
t =2

Bou would then need to look up the two critical values from the Aurbin Watson tables, and these would depend on how many variables and how many observations and how many regressors 1excluding the constant this time2 you had in the model. #he re4ection 6 non*re4ection rule would be given by selecting the appropriate region from the following diagram9

1c2 We have -, observations, and the number of regressors excluding the constant term is ). #he appropriate lower and upper limits are 1.0. and 1.-' respectively, so the Aurbin Watson is lower than the lower limit. It is thus clear that we re4ect the null hypothesis of no autocorrelation. So it looks like the residuals are positively autocorrelated.

&/8

Introductory Econometrics for Finance !ris "roo#s 2008

1d2 y t = 1 + 2 x 2t + & x &t + ' x ' t + u t #he problem with a model entirely in first differences, is that once we calculate the long run solution, all the first difference terms drop out 1as in the long run we assume that the values of all variables have converged on their own long run values so that yt + yt*1 etc.2 #hus when we try to calculate the long run solution to this model, we cannot do it because there isnCt a long run solution to this model 1e2 y t = 1 + 2 x 2t + & x &t + ' x 't + ) x 2t 1 + 6 X &t 1 + ( X 't 1 + v t #he answer is yes, there is no reason why we cannot use Aurbin Watson in this case. Bou may have said no here because there are lagged values of the regressors 1the x variables2 variables in the regression. In fact this would be wrong since there are no lags of the A=D=EA=E# 1 y2 variable and hence AW can still be used. -.
y t = 1 + 2 x 2t + & x &t + ' y t 1 + ) x 2t 1 + 6 x &t 1 + ( x rt ' + u t

#he ma4or steps involved in calculating the long run solution are to * set the disturbance term e5ual to its expected value of 3ero * drop the time subscripts * remove all difference terms altogether since these will all be 3ero by the definition of the long run in this context. Following these steps, we obtain
0 = 1 + ' y + ) x 2 + 6 x& + ( x&

We now want to rearrange this to have all the terms in x2 together and so that y is the sub4ect of the formula9
' y = 1 ) x 2 6 x& ( x & ' y = 1 ) x 2 % 6 + ( $ x & % + ' $ y = 1 ) x2 6 x& ' ' '

#he last e5uation above is the long run solution. &. $amseyCs $=S=# test is a test of whether the functional form of the regression is appropriate. In other words, we test whether the relationship between the dependent variable and the independent variables really should be linear or whether a non*linear form would be more appropriate. #he test works by adding powers of the fitted values from the regression into a second '/8 Introductory Econometrics for Finance !ris "roo#s 2008

regression. If the appropriate model was a linear one, then the powers of the fitted values would not be significant in this second regression. If we fail $amseyCs $=S=# test, then the easiest !solution" is probably to transform all of the variables into logarithms. #his has the effect of turning a multiplicative model into an additive one. If this still fails, then we really have to admit that the relationship between the dependent variable and the independent variables was probably not linear after all so that we have to either estimate a non*linear model for the data 1which is beyond the scope of this course2 or we have to go back to the drawing board and run a different regression containing different variables. .. 1a2 It is important to note that we did not need to assume normality in order to derive the sample estimates of and or in calculating their standard errors. We needed the normality assumption at the later stage when we come to test hypotheses about the regression coefficients, either singly or 4ointly, so that the test statistics we calculate would indeed have the distribution 1 t or F2 that we said they would. 1b2 7ne solution would be to use a techni5ue for estimation and inference which did not re5uire normality. :ut these techni5ues are often highly complex and also their properties are not so well understood, so we do not know with such certainty how well the methods will perform in different circumstances. 7ne pragmatic approach to failing the normality test is to plot the estimated residuals of the model, and look for one or more very extreme outliers. #hese would be residuals that are much !bigger" 1either very big and positive, or very big and negative2 than the rest. It is, fortunately for us, often the case that one or two very extreme outliers will cause a violation of the normality assumption. #he reason that one or two extreme outliers can cause a violation of the normality assumption is that they would lead the 1absolute value of the2 skewness and 6 or kurtosis estimates to be very large. 7nce we spot a few extreme residuals, we should look at the dates when these outliers occurred. If we have a good theoretical reason for doing so, we can add in separate dummy variables for big outliers caused by, for example, wars, changes of government, stock market crashes, changes in market microstructure 1e.g. the !big bang" of 1'.-2. #he effect of the dummy variable is exactly the same as if we had removed the observation from the sample altogether and estimated the regression on the remainder. If we only remove observations in this way, then we make sure that we do not lose any useful pieces of information represented by sample points.

)/8

Introductory Econometrics for Finance !ris "roo#s 2008

'. 1a2 Darameter structural stability refers to whether the coefficient estimates for a regression e5uation are stable over time. If the regression is not structurally stable, it implies that the coefficient estimates would be different for some sub*samples of the data compared to others. #his is clearly not what we want to find since when we estimate a regression, we are implicitly assuming that the regression parameters are constant over the entire sample period under consideration. 1b2 1'.1?1*1''%?12 rt = ,.,21% + 1.0'1 rmt 1'.1?1*1'.&?1, rt = ,.,1-) + 1.),. rmt 1'.&?11*1''%?12 rt = ,.,)-, / 1.-1) rmt

RSS+,.1.' T+1., RSS+,.,&' T+.2 RSS+,.,.2 T+'.

1c2 If we define the coefficient estimates for the first and second halves of the sample as 1 and 1, and 2 and 2 respectively, then the null and alternative hypotheses are @, 9 1 + 2 and 1 + 2 and @1 9 1 2 or 1 2

1d2 #he test statistic is calculated as #est stat. + RSS % RSS1 + RSS 2 $ %T 2k $ 0.189 %0.0(9 + 0.082$ 180 ' * = * = 1).&0' RSS1 + RSS 2 k 0.0(9 + 0.082 2 #his follows an F distribution with 1k,T*2k2 degrees of freedom. F12,1&-2 + ).,% at the %G level. 8learly we re4ect the null hypothesis that the coefficients are e5ual in the two sub*periods. 1,. #he data we have are 1'.1?1*1''%?12 rt = ,.,21% + 1.0'1 Rmt 1'.1?1*1''0?12 rt = ,.,212 + 1.0&. Rmt 1'.2?1*1''%?12 rt = ,.,21& / 1.%2) Rmt

RSS+,.1.' T+1., RSS+,.10. T+1-. RSS+,.1.2 T+1-.

First, the forward predictive failure test * i.e. we are trying to see if the model for 1'.1?1*1''0?12 can predict 1''%?1*1''%?12. #he test statistic is given by 6/8 Introductory Econometrics for Finance !ris "roo#s 2008

RSS RSS1 T1 k 0.189 0.1'8 168 2 * = * = &.8&2 RSS1 T2 0.1'8 12

Where T1 is the number of observations in the first period 1i.e. the period that we actually estimate the model over2, and T2 is the number of observations we are trying to !predict". #he test statistic follows an F*distribution with 1T2, T1* k2 degrees of freedom. F112, 1--2 + 1..1 at the %G level. So we re4ect the null hypothesis that the model can predict the observations for 1''%. We would conclude that our model is no use for predicting this period, and from a practical point of view, we would have to consider whether this failure is a result of a*typical behaviour of the series out*of*sample 1i.e. during 1''%2, or whether it results from a genuine deficiency in the model. #he backward predictive failure test is a little more difficult to understand, although no more difficult to implement. #he test statistic is given by
RSS RSS1 T1 k 0.189 0.182 168 2 * = * = 0.)&2 RSS 1 T2 0.182 12

Eow we need to be a little careful in our interpretation of what exactly are the !first" and !second" sample periods. It would be possible to define T1 as always being the first sample period. :ut I think it easier to say that T1 is always the sample over which we estimate the model 1even though it now comes after the hold*out*sample2. #hus T2 is still the sample that we are trying to predict, even though it comes first. Bou can use either notation, but you need to be clear and consistent. If you wanted to choose the other way to the one I suggest, then you would need to change the subscript 1 everywhere in the formula above so that it was 2, and change every 2 so that it was a 1. =ither way, we conclude that there is little evidence against the null hypothesis. #hus our model is able to ade5uately back*cast the first 12 observations of the sample. 11. :y definition, variables having associated parameters that are not significantly different from 3ero are not, from a statistical perspective, helping to explain variations in the dependent variable about its mean value. 7ne could therefore argue that empirically, they serve no purpose in the fitted regression model. :ut leaving such variables in the model will use up valuable degrees of freedom, implying that the standard errors on all of the other parameters in the regression model, will be unnecessarily higher as a result. If the number of degrees of freedom is relatively small, then saving a couple by deleting two variables with insignificant parameters could be useful. 7n the other hand, if the number of degrees of freedom is already very large, the impact of these additional irrelevant variables on the others is likely to be inconse5uential. 12. (n outlier dummy variable will take the value one for one observation in the sample and 3ero for all others. #he 8how test involves splitting the sample into two parts. If we then try to run the regression on both the sub*parts but the model contains such an outlier dummy, then the observations on that (/8 Introductory Econometrics for Finance !ris "roo#s 2008

dummy will be 3ero everywhere for one of the regressions. For that sub* sample, the outlier dummy would show perfect multicollinearity with the intercept and therefore the model could not be estimated.

8/8

Introductory Econometrics for Finance !ris "roo#s 2008

You might also like