Weighted L1 and L0 Regularization Using Proximal Operator Splitting Methods
Berkessa, Zewude A.; Waldmann, Patrik (2024-11-18)
Berkessa, Zewude A.
Waldmann, Patrik
OpenReview.net
18.11.2024
Berkessa, Zewude A., & Waldmann, P. (2024). Weighted L1 and L0 regularization using proximal operator splitting methods. Transactions on machine learning research, 11, 1-27. https://openreview.net/forum?id=9m2k96cDMK
https://creativecommons.org/licenses/by/4.0/
LIcensed under Creative Commons Attribution 4.0 International.
https://creativecommons.org/licenses/by/4.0/
LIcensed under Creative Commons Attribution 4.0 International.
https://creativecommons.org/licenses/by/4.0/
Julkaisun pysyvä osoite on
https://urn.fi/URN:NBN:fi:oulu-202412057078
https://urn.fi/URN:NBN:fi:oulu-202412057078
Tiivistelmä
Abstract
This paper develops a joint weighted L1- and L0-norm (WL1L0) regularization method by leveraging proximal operators and translation mapping techniques to mitigate the bias introduced by the L1-norm in applications to high-dimensional data. A weighting parameter is incorporated to control the influence of both regularizers. Our broadly applicable model is nonconvex and nonsmooth, but we show convergence for the alternating direction method of multipliers (ADMM) and the strictly contractive Peaceman–Rachford splitting method (SCPRSM). Moreover, we evaluate the effectiveness of our model on both simulated and real high-dimensional genomic datasets by comparing with adaptive versions of the least absolute shrinkage and selection operator (LASSO), elastic net (EN), smoothly clipped absolute deviation (SCAD) and minimax concave penalty (MCP). The results show that WL1L0 outperforms the LASSO, EN, SCAD and MCP by consistently achieving the lowest mean squared error (MSE) across all datasets, indicating its superior ability to handling large high-dimensional data. Furthermore, the WL1L0-SCPRSM also achieves the sparsest solution.
This paper develops a joint weighted L1- and L0-norm (WL1L0) regularization method by leveraging proximal operators and translation mapping techniques to mitigate the bias introduced by the L1-norm in applications to high-dimensional data. A weighting parameter is incorporated to control the influence of both regularizers. Our broadly applicable model is nonconvex and nonsmooth, but we show convergence for the alternating direction method of multipliers (ADMM) and the strictly contractive Peaceman–Rachford splitting method (SCPRSM). Moreover, we evaluate the effectiveness of our model on both simulated and real high-dimensional genomic datasets by comparing with adaptive versions of the least absolute shrinkage and selection operator (LASSO), elastic net (EN), smoothly clipped absolute deviation (SCAD) and minimax concave penalty (MCP). The results show that WL1L0 outperforms the LASSO, EN, SCAD and MCP by consistently achieving the lowest mean squared error (MSE) across all datasets, indicating its superior ability to handling large high-dimensional data. Furthermore, the WL1L0-SCPRSM also achieves the sparsest solution.
Kokoelmat
- Avoin saatavuus [38865]