Double regularization methods for robust feature selection and SVM classification via DC programming
Abstract
In this work, two novel formulations for embedded feature selection are presented. A second-order cone programming approach for Support Vector Machines is extended by adding a second regularizer to encourage feature elimination. The one- and the zero-norm penalties are used in combination with the Tikhonov regularization under a robust setting designed to correctly classify instances, up to a predefined error rate, even for the worst data distribution. The use of the zero norm leads to a nonconvex formulation, which is solved by using Difference of Convex (DC) functions, extending DC programming to second-order cones. Experiments on high-dimensional microarray datasets were performed, and the best performance was obtained with our approaches compared with well-known feature selection methods for Support Vector Machines. WOS Core Collection ISI
Más información
| Título de la Revista: | INFORMATION SCIENCES | 
| Volumen: | 429 | 
| Editorial: | Elsevier Science Inc. | 
| Fecha de publicación: | 2018 | 
| Página de inicio: | 377 | 
| Página final: | 389 | 
| Idioma: | Ingles | 
| URL: | https://www.sciencedirect.com/science/article/pii/S0020025517310976?via%3Dihub | 
| DOI: | 
 doi.org/10.1016/j.ins.2017.11.035  | 
| Notas: | WOS Core Collection ISI |