Abstract

The fused least absolute shrinkage and selection operator (LASSO) simultaneously pursuing the joint sparsity of coefficients and their successive differences has attracted significant attention for analytics purposes. Although it is extensively used, especially when the number of features exceeds the sample size, tuning the regularization parameters, which depends on noise level σ, is a challenging task since σ is difficult to estimate accurately. To tackle this problem, in this paper, we propose and study square root fused LASSO, which combines the square root loss function and joint penalty functions. In theory, we show that the proposed method can achieve the same error rate as that of fused LASSO by proving its estimation and prediction error bounds. In addition, the error rate of square root fused LASSO is lower than those of LASSO and square root LASSO via simultaneous feature selection and clustering. The choices of the regularization parameters are also shown to be free of σ. In terms of computation, this work develops a novel algorithm based on the alternating direction method of multipliers algorithm with theoretical guarantee of its convergence. Experiments on simulation and real-world datasets demonstrate the superiority of square root fused LASSO over fused LASSO and other state-of-the-art feature selection methods.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.