Lecture 3 Multivariate Multiple Regression
Lecture 3 Multivariate Multiple Regression
➢ The 𝑌(𝑗) vectors are column vectors that contain the values of the jth response
variable for each of the n trials or observations.
➢ Similarly, each 𝜀(𝑗) vector contains the random error terms obtained for each
of the n trials when considering the jth response variable.
➢ Each 𝛽(𝑗) vector is comprised of the unknown regression coefficients for the
regression model obtained for the jth response variable.
➢ Z is the design matrix.
2
We can generalize a multiple linear regression model for each response using the
notation above as
𝑌(𝑗) = 𝑍𝛽(𝑗) + 𝜀(𝑗) , 𝑗 = 1,2, … … … . , 𝑚 (3)
Combining each single response model, the following matrix model for multivariate
linear regression can be constructed.
For simplicity, we will write the multivariate linear regression model as,
𝑌 = 𝑍 𝛽 + 𝜀
(𝑛 × 𝑚) (𝑛 × (𝑟 + 1)) (𝑟 + 1) × 𝑚) (𝑛 × 𝑚)
Parameter Estimation
➢ Given the outcomes 𝑌 and the values of the predictor variables Z with full
column rank, we determine the least squares estimates 𝛽̂(𝑗) exclusively from
the observations 𝑌(𝑗) on the jth response.
➢ In conformity with the single-response solution, we take
𝛽̂0𝑗
4
and
1
4
1 1 1 1 1 25
𝑍 ′ 𝑦(1) = [ ] 3 =[ ]
0 1 2 3 4 70
8
[9 ]
−1
−1
1 1 1 1 1 5
𝑍 ′ 𝑦(2) = [ ] 2 =[ ]
0 1 2 3 4 20
3
[2]
0.6 −0.2 25 1
𝛽̂(1) = (𝑍 ′ 𝑍)−1 𝑍 ′ 𝑦(1) = [ ][ ] = [ ]
−0.2 0.1 70 2
0.6 −0.2 5 −1
𝛽̂(2) = (𝑍 ′ 𝑍)−1 𝑍 ′ 𝑦(2) = [ ][ ] = [ ]
−0.2 0.1 20 1
Hence,
1 −1
𝛽̂ = [𝛽̂(1) ⋮ 𝛽̂(2) ] = [ ] = (𝑍 ′ 𝑍)−1 𝑍 ′ [𝑦(1) ⋮ 𝑦(2) ]
2 1
The fitted values are generated from 𝑦̂1 = 1 + 2𝑧1 and 𝑦̂2 = −1 + 𝑧2
Collectively,
1 0 1 −1
1 1 3 0
1 −1
𝑌̂ = 𝑧𝛽̂ = 1 2 [ ]= 5 1
2 1
1 3 7 2
[1 4] [9 3 ]
and
5
1 −1 1 −1 0 0
4 −1 3 0 1 −1
𝜀̂ = 𝑌 − 𝑌̂ = 3 2 − 5 1 = −2 1
8 3 7 2 1 1
[9 2 ] [9 3 ] [ 0 −1]