more regression algebra

More Regression Algebra James H. Steiger Department of Psychology - PowerPoint PPT Presentation

More Regression Algebra James H. Steiger Department of Psychology and Human Development Vanderbilt University James H. Steiger (Vanderbilt University) 1 / 22 More Regression Algebra Introduction 1 Random Multiple Linear Regression: The


  1. More Regression Algebra James H. Steiger Department of Psychology and Human Development Vanderbilt University James H. Steiger (Vanderbilt University) 1 / 22

  2. More Regression Algebra Introduction 1 Random Multiple Linear Regression: The Model 2 Random Multiple Linear Regression: Solution for β 3 Orthogonality Properties 4 Least Squares β Weights Imply Orthogonality Orthogonality Implies A Least Squares β Error Covariance Structure 5 Coefficient of Determination 6 Additivity of Covariances 7 Applications 8 Regression Component Analysis James H. Steiger (Vanderbilt University) 2 / 22

  3. Introduction Introduction A number of important multivariate methods build on the algebra of multivariate linear regression, because they are least squares multiple regression systems, i.e., systems where one or more criteria are predicted as linear combinations of one or more predictors, with optimal prediction defined by a least squares criterion. James H. Steiger (Vanderbilt University) 3 / 22

  4. Introduction Introduction A number of important multivariate methods build on the algebra of multivariate linear regression, because they are least squares multiple regression systems, i.e., systems where one or more criteria are predicted as linear combinations of one or more predictors, with optimal prediction defined by a least squares criterion. In this module, we discuss some key results in multiple regression and multivariate regression that have significant implications in the context of other multivariate methods. James H. Steiger (Vanderbilt University) 3 / 22

  5. Introduction Introduction A number of important multivariate methods build on the algebra of multivariate linear regression, because they are least squares multiple regression systems, i.e., systems where one or more criteria are predicted as linear combinations of one or more predictors, with optimal prediction defined by a least squares criterion. In this module, we discuss some key results in multiple regression and multivariate regression that have significant implications in the context of other multivariate methods. We illustrate the algebra with a couple of theoretical derivations. James H. Steiger (Vanderbilt University) 3 / 22

  6. Random Multiple Linear Regression: The Model The Model Unlike the fixed score multiple regression model frequently employed, this one assumes that both predictor and criterion variables are random. James H. Steiger (Vanderbilt University) 4 / 22

  7. Random Multiple Linear Regression: The Model The Model Unlike the fixed score multiple regression model frequently employed, this one assumes that both predictor and criterion variables are random. Suppose we have a random variable y that we wish to predict from a set of random variables that are in the random vector x . James H. Steiger (Vanderbilt University) 4 / 22

  8. Random Multiple Linear Regression: The Model The Model Unlike the fixed score multiple regression model frequently employed, this one assumes that both predictor and criterion variables are random. Suppose we have a random variable y that we wish to predict from a set of random variables that are in the random vector x . To simplify matters, assume all variables are in deviation score form, i.e., have means of zero. James H. Steiger (Vanderbilt University) 4 / 22

  9. Random Multiple Linear Regression: The Model The Model Unlike the fixed score multiple regression model frequently employed, this one assumes that both predictor and criterion variables are random. Suppose we have a random variable y that we wish to predict from a set of random variables that are in the random vector x . To simplify matters, assume all variables are in deviation score form, i.e., have means of zero. The prediction system is linear, so we may write y = β ′ x + e (1) James H. Steiger (Vanderbilt University) 4 / 22

  10. Random Multiple Linear Regression: Solution for β Solution for β Weights We choose β to minimize the expected squared error, i.e., to minimize E( e 2 ). James H. Steiger (Vanderbilt University) 5 / 22

  11. Random Multiple Linear Regression: Solution for β Solution for β Weights We choose β to minimize the expected squared error, i.e., to minimize E( e 2 ). It is easy to see (C.P.) that E( e 2 ) = σ 2 y − 2 σ yx β + β ′ Σ xx β (2) James H. Steiger (Vanderbilt University) 5 / 22

  12. Random Multiple Linear Regression: Solution for β Solution for β Weights We choose β to minimize the expected squared error, i.e., to minimize E( e 2 ). It is easy to see (C.P.) that E( e 2 ) = σ 2 y − 2 σ yx β + β ′ Σ xx β (2) Minimizing this involves taking the partial derivative of E( e 2 ) with respect to β , setting the resulting equation to zero, and solving for β . The well-known result is that β = Σ − 1 xx σ xy (3) James H. Steiger (Vanderbilt University) 5 / 22

  13. Random Multiple Linear Regression: Solution for β Multiple Linear Regression: Solution for β Weights The preceding result assumed a single criterion variable y . James H. Steiger (Vanderbilt University) 6 / 22

  14. Random Multiple Linear Regression: Solution for β Multiple Linear Regression: Solution for β Weights The preceding result assumed a single criterion variable y . In least squares multivariate linear regression, we have 2 or more criteria, so the model becomes y = β ′ x + e (4) James H. Steiger (Vanderbilt University) 6 / 22

  15. Random Multiple Linear Regression: Solution for β Multiple Linear Regression: Solution for β Weights The preceding result assumed a single criterion variable y . In least squares multivariate linear regression, we have 2 or more criteria, so the model becomes y = β ′ x + e (4) In this case, we wish to select β to minimize the overall average squared error, i.e., to minimize Tr E( ee ′ ). It turns out that the solution is essentially the same as before, i.e., β = Σ − 1 (5) xx Σ xy James H. Steiger (Vanderbilt University) 6 / 22

  16. Orthogonality Properties Least Squares β Weights Imply Orthogonality Orthogonality Properties Least Squares β Weights Imply Orthogonality Suppose we have linear regression system where β = Σ − 1 xx Σ xy . There are a number of immediate consequences. James H. Steiger (Vanderbilt University) 7 / 22

  17. Orthogonality Properties Least Squares β Weights Imply Orthogonality Orthogonality Properties Least Squares β Weights Imply Orthogonality Suppose we have linear regression system where β = Σ − 1 xx Σ xy . There are a number of immediate consequences. One consequence is that x and e are orthogonal, because their covariance matrix is a null matrix. Cov( x , e ) = E( xe ′ ) E( x ( y − β ′ x ) ′ ) = = E( xy ′ ) − E( xx ′ β ) Σ xy − Σ xx Σ − 1 = xx Σ xy = Σ xy − I Σ xy = 0 James H. Steiger (Vanderbilt University) 7 / 22

  18. Orthogonality Properties Least Squares β Weights Imply Orthogonality Orthogonality Properties Least Squares β Weights Imply Orthogonality Suppose we have linear regression system where β = Σ − 1 xx Σ xy . There are a number of immediate consequences. One consequence is that x and e are orthogonal, because their covariance matrix is a null matrix. Cov( x , e ) = E( xe ′ ) E( x ( y − β ′ x ) ′ ) = = E( xy ′ ) − E( xx ′ β ) Σ xy − Σ xx Σ − 1 = xx Σ xy = Σ xy − I Σ xy = 0 Of course, if x and e are orthogonal, ˆ y and e must also be orthogonal. James H. Steiger (Vanderbilt University) 7 / 22

  19. Orthogonality Properties Orthogonality Implies A Least Squares β Orthogonality Implies a Least Squares β We have seen that a least squares β implies orthogonality. James H. Steiger (Vanderbilt University) 8 / 22

  20. Orthogonality Properties Orthogonality Implies A Least Squares β Orthogonality Implies a Least Squares β We have seen that a least squares β implies orthogonality. It turns out that, in a linear system of the form y = β ′ x + e , orthogonality of x and e implies that the β must be the least squares β . (You can prove this as a homework assignment.) James H. Steiger (Vanderbilt University) 8 / 22

  21. Error Covariance Structure Error Covariance Structure As a straightforward consequence of the formula for a least squares β , the covariance matrix of the errors in least squares regression is Σ yy − β ′ Σ xx β = Σ ee Σ yy − Σ yx Σ − 1 = xx Σ xy James H. Steiger (Vanderbilt University) 9 / 22

  22. Error Covariance Structure Error Covariance Structure As a straightforward consequence of the formula for a least squares β , the covariance matrix of the errors in least squares regression is Σ yy − β ′ Σ xx β = Σ ee Σ yy − Σ yx Σ − 1 = xx Σ xy In this case, Σ ee is the partial covariance matrix of the variables in y , with those in x partialled out. James H. Steiger (Vanderbilt University) 9 / 22

  23. Coefficient of Determination Coefficient of Determination The coefficient of determination R 2 pop is the square of the correlation between the predicted scores and the criterion scores. James H. Steiger (Vanderbilt University) 10 / 22

  24. Coefficient of Determination Coefficient of Determination The coefficient of determination R 2 pop is the square of the correlation between the predicted scores and the criterion scores. As a generalization of something we showed in Psychology 310, it is easy to prove (C.P.) that Cov( y j , ˆ y j ) = Var(ˆ y j ), and we shall use that fact below. James H. Steiger (Vanderbilt University) 10 / 22

Recommend


More recommend