Abstract
Principal component analysis (PCA) and minor component analysis (MCA) are a powerful methodology for a wide variety of applications such as pattern recognition and signal processing. In this paper, we first propose a differential equation for the generalized eigenvalue problem. We prove that the stable points of this differential equation are the eigenvectors corresponding to the largest eigenvalue. Based on this generalized differential equation, a class of PCA and MCA learning algorithms can be obtained. We demonstrate that many existing PCA and MCA learning algorithms are special cases of this class, and this class includes some new and simpler MCA learning algorithms. Our results show that all the learning algorithms of this class have the same order of convergence speed, and they are robust to implementation error. 1045-9227/00$10.00
Original language | English |
---|---|
Pages (from-to) | 200-204 |
Number of pages | 5 |
Journal | IEEE Transactions on Neural Networks |
Volume | 11 |
Issue number | 1 |
DOIs | |
Publication status | Published - Jan 2000 |
Scopus Subject Areas
- Software
- Computer Science Applications
- Computer Networks and Communications
- Artificial Intelligence