Efficient Wrapper Feature Selection using Autoencoder and Model Based Elimination

Published in IEEE Letters of the Computer Society (LOCS), 2020

Recommended citation: Sharan Ramjee, Aly El Gamal. “Efficient Wrapper Feature Selection using Autoencoder and Model Based Elimination”. Submitted to IEEE Letters of the Computer Society (LOCS), May. 2020

[ArXiv] [PDF] [Code] [Cite]

Abstract

We propose a computationally efficient wrapper feature selection method - called Autoencoder and Model Based Elimination of features using Relevance and Redundancy scores (AMBER) - that uses a single ranker model along with autoencoders to perform greedy backward elimination of features, without requiring model retraining. The ranker model is used to prioritize the removal of features that are not critical to the classification task, while the autoencoders are used to prioritize the elimination of correlated features. We demonstrate the superior feature selection ability of AMBER on four well known datasets corresponding to different domain applications via comparing the accuracies with other computationally efficient state of the art feature selection techniques, and note how a surprisingly small number of features can lead to very high accuracies on some datasets.