We treat the feature selection problem in the support vector machine (SVM) framework by adopting an optimization model based on use of the ℓ pseudo-norm. The objective is to control the number of non-zero components of the normal vector to the separating hyperplane, while maintaining satisfactory classification accuracy. In our model the polyhedral norm ‖. ‖ [k], intermediate between ‖. ‖ 1 and ‖. ‖ ∞, plays a significant role, allowing us to come out with a DC (difference of convex) optimization problem that is tackled by means of DCA algorithm. The results of several numerical experiments on benchmark classification datasets are reported.
Feature selection in SVM via polyhedral k-norm
Gaudioso M.
;Gorgone E.;
2020-01-01
Abstract
We treat the feature selection problem in the support vector machine (SVM) framework by adopting an optimization model based on use of the ℓ pseudo-norm. The objective is to control the number of non-zero components of the normal vector to the separating hyperplane, while maintaining satisfactory classification accuracy. In our model the polyhedral norm ‖. ‖ [k], intermediate between ‖. ‖ 1 and ‖. ‖ ∞, plays a significant role, allowing us to come out with a DC (difference of convex) optimization problem that is tackled by means of DCA algorithm. The results of several numerical experiments on benchmark classification datasets are reported.File | Dimensione | Formato | |
---|---|---|---|
GGHU.pdf
Solo gestori archivio
Tipologia:
versione post-print (AAM)
Dimensione
337.47 kB
Formato
Adobe PDF
|
337.47 kB | Adobe PDF | Visualizza/Apri Richiedi una copia |
Gaudioso2020_Article_FeatureSelectionInSVMViaPolyhe.pdf
Solo gestori archivio
Descrizione: articolo principale
Tipologia:
versione editoriale (VoR)
Dimensione
347.57 kB
Formato
Adobe PDF
|
347.57 kB | Adobe PDF | Visualizza/Apri Richiedi una copia |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.