Abstract
Linear discriminant analysis (LDA) as a well-known supervised dimensionality reduction method has been widely applied in many fields. However, the lack of sparsity in the LDA solution makes interpretation of the results challenging. In this paper, we propose a new model for sparse uncorrelated LDA (ULDA). Our model is based on the characterization of all solutions of the generalized ULDA. We incorporate sparsity into the ULDA transformation by seeking the solution with minimum ℓ <sub xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">1</sub> -norm from all minimum dimension solutions of the generalized ULDA. The problem is then formulated as an ℓ <sub xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">1</sub> -minimization problem with orthogonality constraint. To solve this problem, we devise two algorithms: 1) by applying the linearized alternating direction method of multipliers and 2) by applying the accelerated linearized Bregman method. Simulation studies and high-dimensional real data examples demonstrate that our algorithms not only compute extremely sparse solutions but also perform well in classification.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
More From: IEEE Transactions on Neural Networks and Learning Systems
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.