Chapter Three Metrics (I)
Chapter Three Metrics (I)
3.1 Introduction
good, Xi is consumer’s income, and β ' s are unknown parameters and ui is the
disturbance.
Equation (3.1) is a multiple regression with three explanatory variables. In general
for K-explanatory variable we can write the model as follows:
Y i =β 0 +β 1 X 1 i +β 2 X 2 i +β 3 X 3 i +.. . .. .. . .+ β k X ki +ui ------- (3.2)
X k =( i=1 , 2 ,3 , .. . .. .. , K )
Where i are explanatory variables, Yi is the dependent
3. Homoscedasticity: The variance of each ui is the same for all the x i values.
E( u 2 )=σ 2
i.e. i u (constant)
We can’t exclusively list all the assumptions but the above assumptions are some
of the basic assumptions that enable us to proceed our analysis.
and β 2 are also sometimes known as regression slopes of the regression. Note that,
β 2 for example measures the effect on E(Y ) of a unit change in X 2 when X 1 is
held constant.
Y^ = β^ 0 + β^ 1 X 1i + β^ 2 X 2i + ei ……………………………………. (3.6)
∂ [ ∑ e2i ]
=−2 ∑ (Y i − β^ 0 − β^ 1 X 1i − β^ 2 X 2 i )=0
∂ β^0 ………………………. (3.8)
∂ [ ∑ e2i ]
=−2 ∑ X 1 i ( Y i − β^ 0− β^ 1 X 1i− β^ 2 X 2i )=0
∂ β^
1 ……………………. (3.9)
∂ [ ∑ e2i ]
=−2 ∑ X 2 i ( Y i − β^ 0− β^ 1 X 1i − β^ 2 X 2i )=0
∂ β^
2 ………… ………..(3.10)
Summing from 1 to n, the multiple regression equation produces three Normal
Equations:
∑ Y =n β^ 0+ β^ 1 ΣX 1 i + β^ 2 ΣX 2 i …………………………………….(3.11)
∑ X 1 Y i = β^ 0 ΣX1 i + β^ 1 ΣX 21i + β^ 2 ΣX 2i X 1i …………………………(3.12)
∑ X 2i Y i = β^ 0 ΣX2 i + β^ 1 ΣX 1 i X 2i + β^ 2 ΣX 22i ………………………...(3.13)
^
From (3.11) we obtain β 0
β^ 0 =Ȳ − β^ 1 X̄ 1 − β^ 2 X̄ 2 ------------------------------------------------- (3.14)
Σx 2 . Σx 2 −Σ( x 1 x2 )2
1 2 …………………………..…………….. (3.21)
Σx2 y . Σx 2−Σx1 x 2 . Σx1 y
β^ 2 =
1
Σx 2 . Σx 2 −Σ( x 1 x 2 )2
1 2 ………………….……………………… (3.22)
^ ^
We can also express β 1 and β 2 in terms of covariance and variances of
Y , X 1 and X 2
=Σe i ( y i− β^ 1 x 1i − β^ 2 x 2i )
=Σe i y − β^ 1 Σx 1i e i− β^ 2 Σei x 2i
=Σe i y i since Σei x 1i=Σei x 2i =0
=Σy i ( y i − β^ 1 x 1 i− β^ 2 x 2i )
i. e Σe2i =Σy2 − β^ 1 Σx 1i y i − β^ 2 Σx 2i y i
ESS β^ 1 Σx 1i y i + β^ 2 Σx2 i y i
2
∴ R= =
TSS Σy2 ----------------------------------(3.27)
As in simple regression, R2 is also viewed as a measure of the prediction ability of
the model over the sample period, or as a measure of how well the estimated
regression fits the data. The value of R2 is also equal to the squared sample
there is a close association between the values of Y t and the values of predicted by
^
the model, Y t . In this case, the model is said to “fit” the data well. If R 2 is low,
there is no association between the values of Y t and the values predicted by the
^
model, Y t and the model does not fit the data well.
2
3.3.3 Adjusted Coefficient of Determination ( R̄ )
2
One difficulty with R is that it can be made large by adding more and more
variables, even if the variables added have no economic justification.
Algebraically, it is the fact that as the variables is added the sum of squared errors
2
(RSS) goes down or it can remain unchanged, but this is rare, and thus R goes up.
2
If the model contains n-1 variables then R =1. The manipulation of model just to
2
obtain a high R is not wise. An alternative measure of goodness of fit, called the
2 2
adjusted R and often symbolized as R̄ , is usually reported by regression
programs. It is computed as:
Σe2i / n−k
2
R̄ =1− 2
Σy / n−1
=1−( 1−R 2 ) ( n−k
n−1
)--------------------------------(3.28)
This measure does not always goes up when a variable is added because of the
degree of freedom term n-k is the numerator. As the number of variables k
2
increases, RSS goes down, but so does n-k. The effect on R̄ depends on the
2
amount by which R falls. While solving one problem, this corrected measure of
2
goodness of fit unfortunately introduces another one. It loses its interpretation; R̄
2
is no longer the percent of variation explained. This modified R̄ is sometimes
used and misused as a device for selecting the appropriate set of explanatory
variables.
So far we have discussed the regression models containing one or two explanatory
variables. Let us now generalize the model assuming that it contains k variables.
It will be of the form:
Y = β0 + β 1 X 1 + β 2 X 2 +. .. .. .+ β k X k +U
There are k parameters to be estimated. The system of normal equations consist of
k+1 equations, in which the unknowns are the parameters β 0 , β 1 , β2 . .. . .. . βk and the
known terms will be the sums of squares and the sums of products of all variables
in the structural equations.
Least square estimators of the unknown parameters are obtained by minimizing
the sum of the squared residuals.
Σe2i =Σ ( y i− β^ 0 − β^ 1 X 1 − β^ 2 X 2 −.. . .. .− β^ k X k )2
∂ Σe2i
=−2 Σ(Y i − β^ 0− β^ 1 X 1− β^ 2 X 2−. .. . ..− β^ k X k )( x i )=0
∂β ^
1
……………………………………………………..
∂ Σe2i
=−2 Σ(Y i − β^ 0− β^ 1 X 1− β^ 2 X 2−. .. . ..− β^ k X k )( x ki )=0
∂ β^ k
The general form of the above equations (except first ) may be written as:
∂ Σe2i
=−2 Σ(Y i − β^ 0− β^ 1 X 1i−−−−−− β^ k X ki )=0
∂ β^ j ; where ( j=1,2,....k )
The normal equations of the general linear regression model are
ΣY i =n β^ 0 + β^ 1 ΣX 1i + β^ 2 ΣX 2i +. .. .. . .. .. . .. .. . .. .. . .. .. . .. .. .+ β^ k ΣX ki
ΣY i X 1i = β^ 0 ΣX 1 i + β^ 1 ΣX +.. . .. .. . .. .. . .. .. . .. .. . .. .. .. . .. .+ β^ k ΣX 1 i X ki
1i 2
: : : : :
: : : : :
ΣY i X ki= β^ 0 ΣX ki + β^ 1 ΣX 1 X ki +∑ X 2 i X ki .. .. . .. .. . .. .. . .. .+ β^ k ΣX
i ki 2
Solving the above normal equations will result in algebraic complexity. But we
can solve this easily using matrix. Hence in the next section we will discuss the
matrix approach to linear regression model.
[] []
β^ 0 e1
^
β e2
1
^
β= and e= .
.
. .
^
β en
k
^ ^
Thus we can write: Y = X β+e and e=Y −X β
We have to minimize:
n
∑ e2i =e 21+ e22+ e23 +.. . .. .. . .+ e2n
i=1
[]
e1
e2
=[ e1 , e 2 .. . .. . en ] . = e'e
.
en
=∑ e2i =e ' e
⇒ β^ =CY …………………………………………….(3.33)
3. Minimum variance
Before showing all the OLS estimators are best (possess the minimum variance
property), it is important to derive their variance.
We know that, var( β )=Ε [ ( β−β ) ] =Ε [ ( β−β )( β−β )' ]
^ ^ 2 ^ ^
Ε [( β−
^ β )( β−β
^ )' ]=
[ ]
Ε ( β^ 1−β 1 )2 Ε [ ( β^ 1−β 1 )( β^ 2−β 2 ) ] .. .. . .. Ε [( β^ 1 −β 1 )( β^ k −β k ) ]
Ε [ ( β^ 2 −β 2 )( β^ 1 −β 1 ) ] Ε( β^ −β ) Ε [( β^ −β )( β^ −β ) ]
2
2 2 .. .. . .. 2 2 k k
: : :
: : :
Ε [( β^ k −β k )( β^ 1−β 1 ) ] Ε [ ( β^ k−β k )( β^ 2 −β 2 ) ] .. .. . .. . Ε ( β^ k−β k )2
[ ]
var ( β^ 1 ) cov ( β^ 1 , β^ 2 ) . . .. .. . cov ( β^ 1 , β^ k )
cov ( β^ 2 , β^ 1 ) var ( β^ 2 ) . . .. .. . cov ( β^ , β^ )
2 k
=: : :
: : :
cov ( β^ k , β^ 1 ) cov ( β^ k , β^ 2 ) . . .. .. . var ( β^ k )
The above matrix is a symmetric matrix containing variances along its main
diagonal and covariance of the estimators everywhere else. This matrix is,
therefore, called the Variance-covariance matrix of least squares estimators of the
regression slopes. Thus,
var ( β^ )=Ε [ ( β−β
^ ^
)( β−β )' ] ……………………………………………(3.35)
2
Note: (σ u being a scalar can be moved in front or behind of a matrix while
[ ]
n ΣX 1n . .. .. . . ΣX kn
2
1n
ΣX 1 n ΣX . .. .. . . ΣX 1 n X kn
: : :
: : :
2
−1 kn
ΣX kn ΣX 1n X kn . .. .. . . ΣX
Where, ( X ' X ) =
^
We can, therefore, obtain the variance of any estimator say β 1 by taking the ith term
−1 2
from the principal diagonal of ( X ' X ) and then multiplying it by σ u .
Where, the X’s are in their absolute form. When the x’s are in deviation form we
can write the multiple regression in matrix form as ;
^ x' x )−1 x ' y
β=(
^ ^
The above column matrix β doesn’t include the constant term β 0 .Under such
conditions the variances of slope parameters in deviation form can be written as:
var( β^ )=σ 2u (x ' x )−1 …………………………………………………….(2.38)
(the proof is the same as (3.37) above). In general we can illustrate the variance of
the parameters by taking two explanatory variables.
The multiple regressions when written in deviation form that has two explanatory
variables is;
y 1 = β^ 1 x 1 + β^ 2 x 2
^
( β−β)=¿ [( β^ 1−β1 ) ¿] ¿ ¿¿
In this model; ¿
^
( β−β ) ' =[( β^ 1 −β 1 )( β^ 2 −β 2 ) ]
)'=¿ [ ( β1 −β 1 ) ¿ ] ¿ ¿ ¿
^ ^ ^
∴ ( β−β )( β−β
¿
]= Εalignl [( β^ 1−β1 ) ( β^ 1−β 1)( β^ 2−β 2 ) ¿ ] ¿ ¿¿
2
Ε [( β−
^ β)( β−β)'
^
and ¿
=
[ var ( β^ 1 ) cov ( β^ 1 , β^ 2 )
cov ( β^ 1 , β^ 2 ) var( β^ 2 ) ]
In case of two explanatory variables, x in the deviation form shall be:
[ ]
−1
−1 Σx21 Σx1 x 2
∴ σ 2u ( x ' x) =σ 2u
Σx1 x 2 Σx22
σ 2u ( x ' −1
x) =
σ 2u
[ Σx 22
−Σx 1 x 2
−Σx 1 x 2
Σx 21 ]
Σx 21 Σx 1 x 2
| |
Or Σx 1 x 2 Σx 22
σ 2u Σx22
var ( β^ 1 )=
i.e., Σx 12 Σx 22−( Σx 1 Σx 2 )2 ……………………………………(3.39)
σ 2u Σx 12
var ( β^ 2 )=
and, Σx 21 Σx 22−( Σx 1 Σx 2 )2 ………………. …….…….(3.40)
(−) σ 2u Σx 1 x 2
cov ( β^ 1 , β^ 2 )=
Σx 21 Σx 22−( Σx 1 Σx 2 )2 …………………………………….
(3.41)
2
The only unknown part in variances and covariance of the estimators is σ u .
In the above model we have three parameters including the constant term and
σ^ 2= n−3
{ } Σe i2
^
Minimum variance of β
^
To show that all the β i ' s in the β vector are Best Estimators, we have also to
prove that the variances obtained in (3.37) are the smallest amongst all other
possible linear unbiased estimators. We follow the same procedure as followed in
case of single explanatory variable model where, we first assumed an alternative
linear unbiased estimator and then it was established that its variance is greater
than the estimator of the regression model.
^ ^
Assume that β is an alternative unbiased and linear estimator of β . Suppose that
^^
β= [ ( X ' X )−1 X ' +B ] Y
Where B is (k x n) matrix of known constants.
^^
∴ β= [( X ' X )−1 X '+B ] [ Xβ+U ]
^^
β=( X ' X )−1 X ' ( Xβ+U )+B( Xβ +U )
^
Ε( β^ )=Ε [( X ' X )−1 X ' ( Xβ+U )+B ( Xβ+U ) ]
=Ε [ ( X ' X )−1 X ' Xβ+( X ' X )−1 X ' U +BX β +BU ]
=β + BX β , [since E(U) = 0].……………………………….(3.44)
^ ^
Since our assumption regarding an alternative β is that it is to be an unbiased
^ ^
estimator of β , therefore, Ε( β ) should be equal to β ; in other words ( β XB) should
be a null matrix.
^
Thus we say, BX should be =0 if ( β )= [( X ' X ) X '+B ] Y is to be an unbiased
^ −1
=Ε [ {( X ' X )−1 X ' U +BU }{( X ' X )−1 X ' U +BU } ' ¿ ¿
( ∵ BX=0)
=Ε [ {( X ' X )−1 X ' U + BU }{U ' X ( X ' X )−1 +U ' B ' }]
=σ 2u [ ( X ' X )−1 X ' X ( X ' X )−1 +BX ( X ' X )−1 +( X ' X )−1 X ' B' +BB' ]
^ ^ ^ 2
Or, in other words, var( β ) is greater than var( β ) by an expression σ u BB ' and it
^
proves that β is the best estimator.
We know, y i =Y i−Ȳ
1
∴ Σy 2i =ΣY 2i − ( ΣY i )2
n
In matrix notation
1
Σy2i =Y ' Y − ( ΣY i )2
n ………………………………………………(3.48)
Equation (3.48) gives the total sum of squares variations in the model.
2 2
Explained sum of squares=Σy i −Σe i
1
=Y ' Y − ( Σy )2 −e ' e
n
1
= β^ ' X ' Y − ( ΣY i )2
n ……………………….(3.49)
Explained sum of squares
R2 =
Since Total sum of squares
1
β^ ' X ' Y − (ΣY i )2 ^
n β ' X ' Y −n Ȳ
∴ R 2= =
1 Y ' Y −n Ȳ 2
Y ' Y − ( ΣY i )2
n ……………………(3.50)
Dear Students! We hope that from the discussion made so far on multiple
regression model, in general, you may make the following summary of results.
^ ^ ^
Let Y = β0 + β 1 X 1 + β 2 X 2 +e i ………………………………… (3.51)
A. H 0 : β 1 =0
H 1 : β 1 ≠0
B. H 0 : β 2 =0
H 1 : β 2 ≠0
The null hypothesis (A) states that, holding X2 constant X1 has no (linear)
influence on Y. Similarly hypothesis (B) states that holding X1 constant, X2 has no
influence on the dependent variable Yi.To test these null hypothesis we will use
the following tests:
i- Standard error test: under this and the following testing methods we
^ ^
test only for β 1 .The test for β 2 will be done in the same way.
^ 1 ^
√ ∑ ∑
x12 i
σ^ 2 ∑ x 22i
x 22i −( ∑ x1 x 2)
2
; where
σ^ 2=
Σe 2i
n−3
In this section we extend this idea to joint test of the relevance of all the included
explanatory variables. Now consider the following:
Y = β0 +β 1 X 1 +β 2 X 2 +. .. .. . .. .+β k X k +U i
H 0 : β 1 =β 2=β 3 =. .. .. . .. .. . .=β k =0
H 1 : at least one of the β k is non-zero
that β 2 =0 , it was assumed tacitly that the testing was based on different sample
^
from the one used in testing the significance of β 3 under the null hypothesis that
β 3 =0 . But to test the joint hypothesis of the above, we shall be violating the
The test procedure for any set of hypothesis can be based on a comparison of the
sum of squared errors from the original, the unrestricted multiple regression
model to the sum of squared errors from a regression model in which the null
hypothesis is assumed to be true. When a null hypothesis is assumed to be true,
we in effect place conditions or constraints, on the values that the parameters can
take, and the sum of squared errors increases. The idea of the test is that if these
sum of squared errors are substantially different, then the assumption that the joint
null hypothesis is true has significantly reduced the ability of the model to fit the
data, and the data do not support the null hypothesis.
If the null hypothesis is true, we expect that the data are compliable with the
conditions placed on the parameters. Thus, there would be little change in the sum
of squared errors when the null hypothesis is assumed to be true.
Let the Restricted Residual Sum of Square (RRSS) be the sum of squared errors
in the model obtained by assuming that the null hypothesis is true and URSS be
the sum of the squared error of the original unrestricted model i.e. unrestricted
residual sum of square (URSS). It is always true that RRSS - URSS¿ 0.
^ ^ ^ ^ ^
Consider Y = β0 + β 1 X 1 + β 2 X 2 +. .. .. . .. .+ β k X k +e i .
This model is called unrestricted. The test of joint hypothesis is that:
H 0 : β 1 =β 2=β 3 =. .. .. . .. .. . .=β k =0
1
Gujurati, 3rd ed.pp
^ ^ ^ ^ ^
We know that: Y = β0 + β 1 X 1i + β 2 X 2i +.. .. .. . ..+ β k X ki
Y i =Y^ +e
e i=Y i−Y^ i
Σe2i =Σ (Y i−Y^ i )2
This sum of squared error is called unrestricted residual sum of square (URSS).
This is the case when the null hypothesis is not true. If the null hypothesis is
assumed to be true, i.e. when all the slope coefficients are zero.
Y = β^ 0 +e i
β^ 0 =
∑ Y i =Ȳ →
n (applying OLS)…………………………….(3.52)
e=Y − β^ 0 ^
but β 0 =Ȳ
e=Y − β^
Σe2i =Σ (Y i−Y^ i )2 =Σy2 =TSS
The sum of squared error when the null hypothesis is assumed to be true is called
Restricted Residual Sum of Square (RRSS) and this is equal to the total sum of
square (TSS).
RRSS−URSS / K−1
~ F (k −1, n−k )
The ratio: URSS /n−K ……………………… (3.53);
(has an F-ditribution with k-1 and n-k degrees of freedom for the numerator and denominator respectively)
RRSS=TSS
URSS=Σe2i =Σy2 − β^ 1 Σ yx 1 − β^ 2 Σ yx 2 +.. .. . .. .. . β^ k Σ yx k =RSS
(TSS−RSS)/ k−1
F=
RSS /n−k
ESS/ k −1
F=
RSS / n−k ………………………………………………. (3.54)
This implies the computed value of F can be calculated either as a ratio of ESS &
TSS or R2 & 1-R2. If the null hypothesis is not true, then the difference between
RRSS and URSS (TSS & RSS) becomes large, implying that the constraints
placed on the model by the null hypothesis have large effect on the ability of the
model to fit the data, and the value of F tends to be large. Thus, we reject the null
hypothesis if the F test static becomes too large. This value is compared with the
critical value of F which leaves the probability of α in the upper tail of the F-
distribution with k-1 and n-k degree of freedom.
If the computed value of F is greater than the critical value of F (k-1, n-k), then the
parameters of the model are jointly significant or the dependent variable Y is
linearly related to the independent variables included in the model.
Y=α + β 1 X 1 + β 2 X 2 + β 3 X 3 +U
Σx1=0
Σx2=0
Σx3=0
Σyi2=594
Σx1x2=240
Σx12=270
Σx22=630
Σx32=750
Σx3yi=319
Σx2yi=492
Σx2x3=-420
Σx1x3=-330
Σx3yi=-625
From the table, the means of the variables are computed and given below:
[] [ ]
β^ 1 x 11 x21 x 31
^ β^ , x= x 12
β=
x22 x 32
2
^β : : :
3 x1 n x 2n x3 n
Where, ; so that
[ ] [ ]
Σx21 Σx 1 x2 Σx1 x 3 Σx1 y
2
( x ' x )= Σx1 x 2 Σx 1 Σx2 x 3 and x ' y= Σx2 y
2 Σx 3 y
Σx1 x 3 Σx 2 x3 Σx3
[ ] [ ]
270 240 −330 319
( x' x )= 240 630 −420 and x ' y= 492
−330 −420 750 −625
Note: the calculations may be made easier by taking 30 as common factor from
all the elements of matrix (x’x). This will not affect the final results.
[ ]
270 240 −330
|x' x|= 240 630 −420 =4716000
−330 −420 750
[ ]
0.0085 −0.0012 0.0031
−1
( x' x ) = −0.0012 0.0027 0.0009
0.0031 0.0009 0.0032
And
α=Ȳ − β^ 1 X̄ 1− β^ 2 X̄ 2− β^ 3 X̄ 3
=52−(0 . 2063)( 42)−(0 .3309 )(62)−(−0 .5572 )(200)
=52−8 .6633−20 .5139+111. 4562=134 .2789
−1 2
(ii) The elements in the principal diagonal of ( x' x ) when multiplied σ u
give the variances of the regression parameters, i.e.,
2
Σe 17.11
var( β^ 1 )=σ2u (0.0085) ¿ } var( β^ 2 )=σ2u(0.0027) ¿ }¿ ¿ σ^ 2u= i = =2.851¿
n−k 6
var( β^ 1 )=0 .0243 , SE( β^ 1 )=0 .1560
var( β^ 2 )=0 . 0077 , SE ( β^ 2 )=0 .0877
var( β^ )=0 . 0093 , SE ( β^ )=0 .0962
3 3
1
β^ ' X ' Y − ( ΣY i )2 ^
2 n β 1 Σx1 y + β^ 2 Σx2 y + β^ 3 Σx3 y
R= =
1 2 Σy 2i 575 . 98
Y ' Y − ( ΣY i ) = =0 . 97
(iii) n 594
(iv) The estimated relation may be put in the following form:
Y^ =134 .28+0 .2063 X 1 +0 .3309 X 2−0 . 5572 X 3
SE( β^ i ) (0 .1560 ) (0 . 0877 ) (0 . 0962) R 2=0. 97
t∗ (1 .3221 ) (3 .7719 ) (5 .7949 )
[ ]
y 7 .59 3. 12 26 . 99
x 1 − 29. 16 30 . 80
x2 − − 133 . 00
The first raw and the first column of the above matrix shows ∑ y 2 and the first
raw and the second column shows ∑ y x 1i and so on.
Consider the following model
β β vi
Y 1 =AY 2 1 Y 2 2 e
a. Estimate β 1 and β 2
^ ^
b. Compute variance of β 1 and β 2
c. Compute coefficient of determination
d. Report the regression result.
Solution: It is difficult to estimate the above model as it is, to estimate the above
model easily let’s take the natural log of the above model;
ln Y 1 =ln A+β 1 lnY 2 +β 2 ln Y 3 +V i
And let: β 0 =ln A , Y =ln Y 1 , X 1 =lnY 2 and X 2 =lnY 3 the above model
becomes: Y = β0 + βX 1 + βX 2 +V i
The above matrix is based on the transformed model. Using values in the matrix
we can now estimate the parameters of the original model.
^β=¿ [ β^ 1 ¿ ] ¿ ¿ ¿
¿
∴ x ' x=
[ Σx 12
Σx 1 x 2
Σx 1 x 2
Σx 22 ] and x ' y=
[ ]
Σx 1 y
Σx 2 y
x ' x=
[2930 .16. 80 30 .80
133 .00 ]
and x ' y=
3. 12
26 . 99 [ ]
29.16 30.80
|x' x|=| | =2929.64
30.80 133.00
∴ ( x' x )−1=
1
=
[
133 . 00 −30 . 80
2929 . 64 −30 . 80 29 .16
≈
0 .0454 −0. 0105
−0. 0105 0.0099 ] [ ]
x)−1 x ' y=¿ [ β 1 ¿ ] ¿ ¿¿
^ ^
β=(x'
(a) ¿
2 α^ Σx2 y + β^ Σx 3 y
R =
(c). Σy2i
Example 3:
Since the x’s and y’s in the above formula are in deviation form we have to find
the corresponding deviation forms of the above given values.
We know that:
Σx1 x 2 =ΣX 1 X 2−n X̄ 1 X̄ 2
=4300−(25)(10)(16)
=300
Σx1 y=ΣX 1 Y −n X̄ 1 Ȳ
=8400−25(10 )(32)
=400
Σx2 y =ΣX 2 Y −n X̄ 2 Ȳ
=13500−25(16 )(32)
=700
Σx21 =ΣX 21 −n X̄ 21
=3200−25(10 )2
¿ 700
Σx22 =ΣX 22 −n X̄ 22
=7300−25(16 )2
¿ 900
Now we can compute the parameters.
Σx2 yΣx 22−Σx 2 yΣx1 x 2
β^ 1=
Σx21 Σx22 −( Σx 1 x 2 )2
=32−(0.278)(10)−(0.685)(16)
=18.26
σ^ 2 Σx 22
var ( β^ 1 )=
b. Σx 12 Σx 22−( Σx 1 x 2 )2
Σe 2i
2
^ =
⇒σ
n−k Where k is the number of parameter
2 Σe 2i
σ^ =
n−3
1809 . 3
=
25−3
=82. 24
(82 . 24 )(900)
⇒ var( β^ 1 )= =0 . 137
540 , 000
( 82. 24 )(700)
= =0 . 1067
540 , 000
α
=0 . 05 2 =0 . 025
The critical value of t from the t-table at 2 level of significance and
22 degree of freedom is 2.074.
t c=2. 074
t∗¿ 0 . 755
⇒ t∗¿ t c
The decision rule if t∗¿ t c is to reject the alternative hypothesis that says β is
different from zero and to accept the null hypothesis that says β is equal to zero.
^
The conclusion is β 1 is statistically insignificant or the sample we use to estimate
β^ 1 is drawn from the population of Y & X in which there is no relationship
1
(1−0. 24 )(24 )
=1−
22
=0. 178
e. Let’s set first the joint hypothesis as
H 0 : β 1 =β 2=0
F*(2,22) = 3.47
Fc(2,22)=3.44
⇒ F*>Fc, the decision rule is to reject H 0 and accept H1. We can say that
the model is significant i.e. the dependent variable is, at least, linearly
related to one of the explanatory variables.