Abstract
A new efficient algorithm is presented for joint diagonalization of several matrices. The algorithm is based on the Frobenius-norm formulation of the joint diagonalization problem, and addresses diagonalization with a general, non-orthogonal transformation. The iterative scheme of the algorithm is based on a multiplicative update which ensures the invertibility of the diagonalizer. The algorithm's efficiency stems from the special approximation of the cost function resulting in a sparse, block-diagonal Hessian to be used in the computation of the quasi-Newton update step. Extensive numerical simulations illustrate the performance of the algorithm and provide a comparison to other leading diagonalization methods. The results of such comparison demonstrate that the proposed algorithm is a viable alternative to existing state-of-the-art joint diagonalization algorithms. The practical use of our algorithm is shown for blind source separation problems.
Original language | English |
---|---|
Pages (from-to) | 777-800 |
Number of pages | 24 |
Journal | Journal of Machine Learning Research |
Volume | 5 |
Publication status | Published - 2004 Jul 1 |
Bibliographical note
Publisher Copyright:© 2004 Andreas Ziehe, Pavel Laskov, Guido Nolte, and Klaus-Robert Müller.
Keywords
- Blind source separation
- Common principle component analysis
- Independent component analysis
- Joint diagonalization
- Levenberg-marquardt algorithm
- Newton method
- Nonlinear least squares
ASJC Scopus subject areas
- Software
- Control and Systems Engineering
- Statistics and Probability
- Artificial Intelligence