Abstract
Principal components analysis is an important and well-studied subject in statistics and signal processing. The literature has an abundance of algorithms for solving this problem, where most of these algorithms could be grouped into one of the following three approaches: adaptation based on Hebbian updates and deflation, optimization of a second-order statistical criterion (like reconstruction error or output variance), and fixed point update rules with deflation. In this paper, we take a completely different approach that avoids deflation and the optimization of a cost function using gradients. The proposed method updates the eigenvector and eigenvalue matrices simultaneously with every new sample such that the estimates approximately track their true values as would be calculated from the current sample estimate of the data covariance matrix. The performance of this algorithm is compared with that of traditional methods like Sanger's rule and APEX, as well as a structurally similar matrix perturbation-based method.
Highlights
Principal components analysis (PCA) is a well-known statistical technique that has been widely applied to solve important signal processing problems like feature extraction, signal estimation, detection, and speech separation [1, 2, 3, 4]
Most of the analytical methods require extensive matrix operations and they are unsuited for real-time applications. In many applications such as direction of arrival (DOA) tracking, adaptive subspace estimation, and so forth, signal statistics change over time rendering the block methods virtually unacceptable
We recently explored a simultaneous principal component extraction algorithm called SIPEX [14] which reduced the gradient search only to the space of orthonormal matrices by using Givens rotations
Summary
Principal components analysis (PCA) is a well-known statistical technique that has been widely applied to solve important signal processing problems like feature extraction, signal estimation, detection, and speech separation [1, 2, 3, 4]. The recently proposed fixed-point PCA algorithm [13] showed fast convergence with little or no change in complexity compared with gradient methods This method and most of the existing methods in literature rely on using the standard deflation technique, which brings in sequential convergence of principal components that potentially reduces the overall speed of convergence. We will present an algorithm that undertakes a similar perturbation approach, but in contrast, the covariance matrix will be decomposed into its eigenvectors and eigenvalues at all times, which will reduce the perturbation step to be employed on the diagonal eigenvalue matrix This further restriction of structure, as expected, alleviates the difficulties encountered in the operation of the previous first-order perturbation algorithm, resulting in a fast converging and accurate subspace tracking algorithm. We conclude the paper with remarks and observations about the algorithm
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.