Unified Lf-Norm Robust Fitting for Linear Models
Loading...

Date
2025
Journal Title
Journal ISSN
Volume Title
Publisher
Institute of Electrical and Electronics Engineers Inc.
IEEE
IEEE
Open Access Color
Green Open Access
No
OpenAIRE Downloads
OpenAIRE Views
Publicly Funded
No
Abstract
In statistical learning, accurately estimating model parameters is crucial for reliable predictions. Managing residuals, the differences between observed and predicted values, is a key challenge. In regression, the residual penalty choice strongly affects model performance. The L<inf>2</inf>-norm penalty aligns with the least-squares approach, while the L<inf>1</inf>-norm provides robust fitting by minimizing the influence of outliers. To generalize models, the weights can be regularized using either the L<inf>2</inf>-norm or L<inf>1</inf>-norm, corresponding to Ridge and LASSO regularization, respectively. Many methods have been developed to penalize residuals and model weights, resulting in diverse cost functions optimized by specific numerical solvers. In this study, we propose the smooth L<inf>f</inf>-norm, a quasi-norm, as a unified framework for penalizing both residuals and model weights in linear models. Our efficient and robust numerical minimization scheme ensures fast and accurate fitting by minimizing our novel cost function. © 2025 Elsevier B.V., All rights reserved.
In statistical learning, accurately estimating model parameters is crucial for reliable predictions. Managing residuals, the differences between observed and predicted values, is a key challenge. In regression, the residual penalty choice strongly affects model performance. The L-2-norm penalty aligns with the least-squares approach, while the L-1-norm provides robust fitting by minimizing the influence of outliers. To generalize models, the weights can be regularized using either the L-2-norm or L-1-norm, corresponding to Ridge and LASSO regularization, respectively. Many methods have been developed to penalize residuals and model weights, resulting in diverse cost functions optimized by specific numerical solvers. In this study, we propose the smooth L-f-norm, a quasi-norm, as a unified framework for penalizing both residuals and model weights in linear models. Our efficient and robust numerical minimization scheme ensures fast and accurate fitting by minimizing our novel cost function.
In statistical learning, accurately estimating model parameters is crucial for reliable predictions. Managing residuals, the differences between observed and predicted values, is a key challenge. In regression, the residual penalty choice strongly affects model performance. The L-2-norm penalty aligns with the least-squares approach, while the L-1-norm provides robust fitting by minimizing the influence of outliers. To generalize models, the weights can be regularized using either the L-2-norm or L-1-norm, corresponding to Ridge and LASSO regularization, respectively. Many methods have been developed to penalize residuals and model weights, resulting in diverse cost functions optimized by specific numerical solvers. In this study, we propose the smooth L-f-norm, a quasi-norm, as a unified framework for penalizing both residuals and model weights in linear models. Our efficient and robust numerical minimization scheme ensures fast and accurate fitting by minimizing our novel cost function.
Description
Isik University
Keywords
Linear Regression, Robust Fitting, Regularization, Sparsity, L-1-Norm, L-F-Norm
Fields of Science
Citation
WoS Q
N/A
Scopus Q
N/A

OpenCitations Citation Count
N/A
Source
-- 33rd IEEE Conference on Signal Processing and Communications Applications, SIU 2025 -- Istanbul; Isik University Sile Campus -- 211450
33rd Conference on Signal Processing and Communications Applications-SIU-Annual -- Jun 25-28, 2025 -- Istanbul, Turkiye
33rd Conference on Signal Processing and Communications Applications-SIU-Annual -- Jun 25-28, 2025 -- Istanbul, Turkiye
Volume
Issue
Start Page
1
End Page
4
PlumX Metrics
Citations
Scopus : 0
Page Views
2
checked on Feb 23, 2026
Google Scholar™


