You are working with the text-only light edition of "H.Lohninger: Teach/Me Data Analysis, Springer-Verlag, Berlin-New York-Tokyo, 1999. ISBN 3-540-14743-8". Click here for further information. |
See also: PCA, matrix algebra, matrix inversion, eigenvectors - advanced stuff |
Eigenvectors and eigenvalues are defined only for square matrices Z and obey the following equation:
Ze = el
with e (a vector) being the eigenvector, and l (a scalar) the eigenvalue.
This formal definition may look somewhat abstract, and its benefits are of course not evident from the equation above. However, eigenanalysis is important in various technical and scientific fields. In the context of data analysis, eigenvectors are used to obtain a more stable set of descriptors for viewing the data. In most cases eigenanalysis is based on some form of a scatter matrix Z=X^{T}X, with X being a matrix of n rows (observations) and p columns (variables).
A square matrix of dimension p by p can have at most p eigenvectors. This can be denoted in matrix notation
ZE = Ediag(l_{1}, ... l_{p})
The eigenvectors are orthogonal to each other, and the product E^{T}E is the identity matrix I (E is an orthonormal matrix). The matrix Z can be expressed by its eigenvectors
Z = Ediag(l_{1}, ... l_{p})E^{T}
Note that the inverse matrix of Z can be written simply as
Z = Ediag(1/l_{1}, ... 1/l_{p})E^{T}
Last Update: 2004-Jul-03