Download 1 THE MULTIPLE LINEAR REGRESSION MODEL Notation: p

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Data assimilation wikipedia , lookup

Regression toward the mean wikipedia , lookup

Time series wikipedia , lookup

Choice modelling wikipedia , lookup

Regression analysis wikipedia , lookup

Coefficient of determination wikipedia , lookup

Linear regression wikipedia , lookup

Transcript
1
THE MULTIPLE LINEAR REGRESSION MODEL
Notation:
p predictors x1, x2, … , xp
k-1 non-constant terms u1, u2, … , uk-1
Each uj is a function of x1, x2, … , xp: uj = uj(x1, x2, … , xp)
For convenience, we often set u0 = 1 (constant function/term)
The Basic Multiple Linear Regression Model: Two assumptions:
1. E(Y|x) = η0 + η1u1 + … + ηk-1uk-1 (Linear Mean Function)
2. Var(Y|x) = σ2 (Or: Var(Y|u) = σ2)
(Constant Variance)
Assumption (1) in vector notation:
# u0 &
% (
% u1 (
.
u = % (=
% . (
% (
% . (
$uk"1'
# 1 &
% (
% u1 (
% . (,
% . (
% (
% . (
$uk"1'
$ "0 '
& )
& "1 )
.
η=& )
& . )
& )
& . )
%"k#1(
Then ηT = [η0 η1 … ηk-1] and
!
ηT!u = η0 + η1u1 +!… + ηk-1uk-1,
so (1) becomes:
(1') E(Y| x) = ηTu
If we have data with ith observation xi1, xi2, … , xip, yi, recall
" x i1 %
$ '
$x i2 '
$ .'
xi = $ ' = [xi1, xi2, … , xip]T
.
$ '
$ .'
$#x ip '&
Define similarly
!
uij = uj (xi1, xi2, … , xip) = the value of the jth term for the ith observation, and
2
# ui0 &
%
(
% ui1 (
. (
ui = %
% . (
%
(
% . (
$ui,k"1'
So in particular, the model says
!
E(Y|xi) = ηTui
Estimation of Parameters: Analogously to the case of simple linear regression, consider
functions of the form
y = h0 + h1u1 + … + hk-1uk-1 = hTu.
(The graph of such an equation is called a "hyperplane.")
The least squares estimate of η is the vector
$ "ˆ 0 '
& )
& "ˆ )
"ˆ = & 1 )
& M )
&"ˆ )
% k #1 (
that minimizes the "objective function"
n
T
RSS(h) = # (y i " h u i )
2
i=1
n
Recall: In simple linear regression, the solution had SXX =
# (x
i
" x ) 2 in the
i=1
denominator. So the formula won't work if all xi's = x . In this case, there is not a unique
solution to the least squares problem. (Draw a picture in the case n = 2!)
In multiple regression: There is a unique solution "ˆ provided:
i) k < n
(the number of terms is less than the number of observations)
ii) no uj is (as a function) a linear combination of the other uj's
When (ii) is violated, we say there is (strict) multicollinearity.
If there is a unique solution, it is called the ordinary least squares (OLS) estimate of the
(vector of) coefficients.