@inproceedings{e201dac73d69415689d029596ae90859,
title = "Feature Importance Ranking for Deep Learning",
abstract = "Feature importance ranking has become a powerful tool for explainable AI. However, its nature of combinatorial optimization poses a great challenge for deep learning. In this paper, we propose a novel dual-net architecture consisting of operator and selector for discovery of an optimal feature subset of a fixed size and ranking the importance of those features in the optimal subset simultaneously. During learning, the operator is trained for a supervised learning task via optimal feature subset candidates generated by the selector that learns predicting the learning performance of the operator working on different optimal subset candidates. We develop an alternate learning algorithm that trains two nets jointly and incorporates a stochastic local search procedure into learning to address the combinatorial optimization challenge. In deployment, the selector generates an optimal feature subset and ranks feature importance, while the operator makes predictions based on the optimal subset for test data. A thorough evaluation on synthetic, benchmark and real data sets suggests that our approach outperforms several state-of-the-art feature importance ranking and supervised feature selection methods.",
author = "Maksymilian Wojtas and Ke Chen",
year = "2020",
month = sep,
day = "25",
language = "English",
series = "Advances in Neural Information Processing Systems",
publisher = "Morgan Kaufmann Publishers",
booktitle = "Advances in Neural Information Processing Systems 33",
note = "34th Conference on Neural Information Processing Systems, NeurIPS 2020 ; Conference date: 06-12-2020 Through 12-12-2020",
}