You are working with the text-only light edition of "H.Lohninger: Teach/Me Data Analysis, Springer-Verlag, Berlin-New York-Tokyo, 1999. ISBN 3-540-14743-8". Click here for further information.

Multiple Linear Regression
Introduction

Multiple linear regression (MLR) is similar to simple regression, the only difference being the use of more than one input variable. In order to calculate the relationship between n input variables xi and the target variable y we could use the linear equation

y = a0 + a1x1 + a2x2 + ... + anxn + e

or

The parameter e defines the error, or the residual,  with a mean value of zero. This equation defines a hyperplane in n-dimensional space. The parameters of this plane have to be adjusted so that the plane optimally fits the data. In order to obtain the best fit, the parameters a0 to an are adjusted such that the sum of the squared errors is minimized. The assumptions are the same as for simple regression. The estimated parameters can again be discussed by using the ANOVA table.
 
 

Last Update: 2006-Jšn-17