A scoring scheme for online feature selection: Simulating model performance without retraining
Article Type
Research Article
Publication Title
IEEE Transactions on Neural Networks and Learning Systems
Abstract
Increasing the number of features increases the complexity of a model even if the additional feature does not improve its decision-making capacity. Irrelevant features may also cause overfitting and reduce interpretability of the concerned model. It is, therefore, important that the features are optimally selected before a model is built. In the case of online learning, new instances are periodically discovered, and the respective model is tactically retrained as required. Similarly, there are many real-life situations where hundreds of new features are discovered periodically, and the existing model needs to be retrained or tested for its performance improvement. Supervised selection of feature subset usually requires creation of multiple suboptimal models, thus incurring time-intensive computations. Unsupervised selections, although faster, largely rely on some subjective definition of feature relevance. In this paper, we introduce a score that accurately determines the importance of the features. The proposed score is appropriate for online feature selection scenarios for its low time complexity and ability to interpret performance improvement of the current model after the addition of a new feature, without invoking a retraining.
First Page
405
Last Page
414
DOI
10.1109/TNNLS.2016.2514270
Publication Date
2-1-2017
Recommended Citation
Sengupta, Debarka; Bandyopadhyay, Sanghamitra; and Sinha, Debajyoti, "A scoring scheme for online feature selection: Simulating model performance without retraining" (2017). Journal Articles. 2710.
https://digitalcommons.isical.ac.in/journal-articles/2710