Mutual information has been successfully adopted in filter feature-selection methods to assess both the relevancy of a subset of features in predicting the target variable and the redundancy with respect to other variables. However, existing algorithms are mostly heuristic and do not offer any guarantee on the proposed solution. In this paper, we provide novel theoretical results showing that conditional mutual information naturally arises when bounding the ideal regression/classification errors achieved by different subsets of features. Leveraging on these insights, we propose a novel stopping condition for backward and forward greedy methods which ensures that the ideal prediction error using the selected feature subset remains bounded by a user-specified threshold. We provide numerical simulations to support our theoretical claims and compare to common heuristic methods.

Beraha, M., Metelli, A., Papini, M., Tirinzoni, A., Restelli, M. (2019). Feature Selection via Mutual Information: New Theoretical Insights. In Proceedings of the International Joint Conference on Neural Networks (pp.1-9). Institute of Electrical and Electronics Engineers Inc. [10.1109/IJCNN.2019.8852410].

Feature Selection via Mutual Information: New Theoretical Insights

Beraha M.
;
2019

Abstract

Mutual information has been successfully adopted in filter feature-selection methods to assess both the relevancy of a subset of features in predicting the target variable and the redundancy with respect to other variables. However, existing algorithms are mostly heuristic and do not offer any guarantee on the proposed solution. In this paper, we provide novel theoretical results showing that conditional mutual information naturally arises when bounding the ideal regression/classification errors achieved by different subsets of features. Leveraging on these insights, we propose a novel stopping condition for backward and forward greedy methods which ensures that the ideal prediction error using the selected feature subset remains bounded by a user-specified threshold. We provide numerical simulations to support our theoretical claims and compare to common heuristic methods.
paper
classification; feature selection; machine learning; mutual information; regression; supervised learning;
English
2019 International Joint Conference on Neural Networks, IJCNN 2019
2019
Proceedings of the International Joint Conference on Neural Networks
9781728119854
2019
2019-July
1
9
8852410
reserved
Beraha, M., Metelli, A., Papini, M., Tirinzoni, A., Restelli, M. (2019). Feature Selection via Mutual Information: New Theoretical Insights. In Proceedings of the International Joint Conference on Neural Networks (pp.1-9). Institute of Electrical and Electronics Engineers Inc. [10.1109/IJCNN.2019.8852410].
File in questo prodotto:
File Dimensione Formato  
Beraha-2019-IJCNN-VoR.pdf

Solo gestori archivio

Tipologia di allegato: Publisher’s Version (Version of Record, VoR)
Licenza: Tutti i diritti riservati
Dimensione 546.12 kB
Formato Adobe PDF
546.12 kB Adobe PDF   Visualizza/Apri   Richiedi una copia

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/10281/545384
Citazioni
  • Scopus 57
  • ???jsp.display-item.citation.isi??? 31
Social impact