Topic 3 Multiple Regression Analysis Estimation
Topic 3 Multiple Regression Analysis Estimation
Analysis: Estimation
2
Multiple Regression Model
• Multiple Regression Model
The equation that describes how the dependent variable y is related to the independent variables x1,
x2, . . . xp and an error term is:
Classical Linear Regression Model (CLRM)
A-1: The regression model is linear in the parameters; it may or may not be linear in the
variables Y and the Xs.
A-2: The regressors are assumed to be fixed or nonstochastic in the sense that their values
are fixed in repeated sampling. Fixed X values or X values independent of the error term.
Hence, this means we require zero covariance between and each X variables.
A-3: Given the values of the X variables, the expected, or mean, value of the error term is
zero. That is,
A-4: The variance of each ui, given the values of X, is constant, or homoscedastic (homo
means equal and scedastic means variance). That is,
Classical Linear Regression Model (CLRM)
A-5: There is no correlation between two error terms. That is, there is no autocorrelation.
Symbolically,
A-6: There are no perfect linear relationships among the X variables. This is the
assumption of no multicollinearity. For example, relationships like are ruled out.
A-7: The regression model is correctly specified. Alternatively, there is no specification
bias or specification error in the model used in empirical analysis. It is implicitly
assumed that the number of observations, n, is greater than the number of parameters
estimated.
Although it is not a part of the CLRM, it is assumed that the error term follows the normal
distribution with zero mean and (constant) variance . Symbolically,
A-8:
Classical Linear Regression Model (CLRM)
On the basis of Assumptions A-1 to A-7, it can be shown that the method of ordinary least
squares (OLS), the method most popularly used in practice, provides estimators of the
parameters of the population regression function (PRF) that have several desirable
statistical properties, such as:
1. The estimators are linear, that is, they are linear functions of the dependent variable Y.
Linear estimators are easy to understand and deal with compared to nonlinear estimators.
2. The estimators are unbiased, that is, in repeated applications of the method, on average,
the estimators are equal to their true values.
3. In the class of linear unbiased estimators, OLS estimators have minimum variance. As a
result, the true parameter values can be estimated with least possible uncertainty; an
unbiased estimator with the least variance is called an efficient estimator.
In short, under the assumed conditions, OLS estimators are BLUE: best linear unbiased
estimators. This is the essence of the well-known Gauss–Markov theorem, which
provides a theoretical justification for the method of least squares.
Multiple Regression Analysis: Estimation
6
Per student spending is likely to be correlated with average family income at a given high school because of
school financing.
Omitting average family income in regression would lead to biased estimate of the effect of spending on
average test scores.
In a simple regression model, effect of per student spending would partly include the effect of family
income on test scores.
Multiple Regression Analysis: Estimation
8
Model assumes a constant elasticity relationship between CEO salary and the sales of his or her firm.
Model assumes a quadratic relationship between CEO salary and his or her tenure with the firm.
Interpretation
Holding ACT fixed, another point on high school grade point average is associated with another .453 points college
grade point average
Or: If we compare two students with the same ACT, but the hsGPA of student A is one point higher, we predict
student A to have a colGPA that is .453 higher than that of student B
Holding high school grade point average fixed, another 10 points on ACT are associated with less than one point on
college GPA
11
Multiple Regression Equation
• Multiple Regression Equation
The equation that describes how the mean value of y is related to x1, x2,
. . . xk is:
E(y) = 0 + 1x1 + 2x2 + . . . + kxk
12 Estimated Multiple Regression Equation
• Estimated Multiple Regression Equation
A simple random sample is used to compute sample statistics b0, b1, b2, . . . , bk
that are used as the point estimators of the parameters b0, b1, b2, . . . , bk.
13
Estimation Process
Multiple Regression Model
Sample Data:
E(y) = 0 + 1x1 + 2x2 +. .+ kxk + e x1 x2 . . . xk y
Multiple Regression Equation . . . .
E(y) = 0 + 1x1 + 2x2 +. . .+ kxk . . . .
Unknown parameters are
b0 , b1 , b2 , . . . , bk
Estimated Multiple
Regression Equation
b0, b1, b2, . . . , bk = b0 + b1x1 + b2x2 + . . . + bkxk
provide estimates of Sample statistics are
b 0 , b 1 , b2 , . . . , b k b0, b1, b2, . . . , bk
14 Least Squares Method
Least Squares Criterion
min
where
y = annual salary ($1000s)
x1 = years of experience
x2 = score on programmer aptitude test
18
Solving for the Estimates of 0, 1, 2
Least Squares
Input Data Output
x1 x2 y Computer b0 =
Package b1 =
4 78 24 for Solving
7 100 43 b2 =
Multiple
. . .
Regression R2 =
. . .
3 89 30 Problems etc.
19
Solving for the Estimates of 0, 1, 2
• Regression Equation Output
b1 = 1.404
b2 = 0.251
where:
SST = total sum of squares
SSR = sum of squares due to regression
SSE = sum of squares due to error
25
Multiple Coefficient of Determination
• ANOVA Output
Analysis of Variance
SOURCE DF SS MS F P
Regression 2 500.3285 250.164 42.76 0.000
Residual Error 17 99.45697 5.850
Total 19 599.7855
26
Multiple Coefficient of Determination
R2 = SSR/SST
R2 = 500.3285/599.7855 = .83418
27 Adjusted Multiple Coefficient of Determination
• Adding independent variables, even ones that are not statistically significant,
causes the prediction errors to become smaller, thus reducing the sum of
squares due to error, SSE.
• Because SSR = SST – SSE, when SSE becomes smaller, SSR becomes larger,
causing R2 = SSR/SST to increase.
• The adjusted multiple coefficient of determination compensates for the number
of independent variables in the model.
28 Adjusted Multiple Coefficient of Determination
2 2 𝑛 −1
𝑅𝑎 =1−(1 − 𝑅 )
𝑛 −𝑘 −1
2 20 −1
𝑅𝑎 =1− ( 1−.834179 ) =.814671
20 −2 −1
29
Assumptions About the Error Term
• The error is a random variable with mean of zero.
• The variance of , denoted by 2, is the same for all values of the
independent variables.
• The values of are independent.
• The error is a normally distributed random variable reflecting the deviation
between the y value and the expected value of y given by 0 + 1x1 + 2x2 + . .
+ kxk.
Components
30 of OLS Variances:
The R-squared of this regression will be the higher when xj can be better explained by the other
independent variables.
The sampling variance of the slope estimator for xj will be higher when xj can be better
explained by the other independent variables.
Under perfect multicollinearity, the variance of the slope estimator will approach infinity.