TY - JOUR
T1 - Embedded feature selection for robust probability learning machines
AU - Carrasco, Miguel
AU - Ivorra, Benjamin
AU - López, Julio
AU - Ramos, Angel M.
N1 - Publisher Copyright:
© 2024
PY - 2025/3
Y1 - 2025/3
N2 - Methods: Feature selection is essential for building effective machine learning models in binary classification. Eliminating unnecessary features can reduce the risk of overfitting and improve classification performance. Moreover, the data we handle typically contains a stochastic component, making it important to develop robust models that are insensitive to data perturbations. Although there are numerous methods and tools for feature selection, relatively few studies address embedded feature selection within robust classification models using penalization techniques. Objective: In this work, we introduce robust classifiers with integrated feature selection capabilities, utilizing probability machines based on different penalization techniques, such as the ℓ1-norm or the elastic-net, combined with a novel Direct Feature Elimination process to improve model resilience and efficiency. Findings: Numerical experiments on standard datasets demonstrate the effectiveness and robustness of the proposed models in classification tasks even when using a reduced number of features. These experiments were evaluated using original performance indicators, highlighting the models’ ability to maintain high performance with fewer features. Novelty: The study discusses the trade-offs involved in combining different penalties to select the most relevant features while minimizing empirical risk. In particular, the integration of elastic-net and ℓ1-norm penalties within a unified framework, combined with the original Direct Feature Elimination approach, presents a novel method for improving both model accuracy and robustness.
AB - Methods: Feature selection is essential for building effective machine learning models in binary classification. Eliminating unnecessary features can reduce the risk of overfitting and improve classification performance. Moreover, the data we handle typically contains a stochastic component, making it important to develop robust models that are insensitive to data perturbations. Although there are numerous methods and tools for feature selection, relatively few studies address embedded feature selection within robust classification models using penalization techniques. Objective: In this work, we introduce robust classifiers with integrated feature selection capabilities, utilizing probability machines based on different penalization techniques, such as the ℓ1-norm or the elastic-net, combined with a novel Direct Feature Elimination process to improve model resilience and efficiency. Findings: Numerical experiments on standard datasets demonstrate the effectiveness and robustness of the proposed models in classification tasks even when using a reduced number of features. These experiments were evaluated using original performance indicators, highlighting the models’ ability to maintain high performance with fewer features. Novelty: The study discusses the trade-offs involved in combining different penalties to select the most relevant features while minimizing empirical risk. In particular, the integration of elastic-net and ℓ1-norm penalties within a unified framework, combined with the original Direct Feature Elimination approach, presents a novel method for improving both model accuracy and robustness.
KW - Cobb–Douglas
KW - Feature selection
KW - Minimax Probability Machine
KW - Minimum Error Minimax Probability Machine
KW - Second-order cone programming
KW - Support vector machines
UR - http://www.scopus.com/inward/record.url?scp=85208936986&partnerID=8YFLogxK
U2 - 10.1016/j.patcog.2024.111157
DO - 10.1016/j.patcog.2024.111157
M3 - Article
AN - SCOPUS:85208936986
SN - 0031-3203
VL - 159
JO - Pattern Recognition
JF - Pattern Recognition
M1 - 111157
ER -