1. A Framework to Learn with Interpretation
- Author
-
Parekh, Jayneel, Mozharovskyi, Pavlo, d'Alché-Buc, Florence, Signal, Statistique et Apprentissage (S2A), Laboratoire Traitement et Communication de l'Information (LTCI), Institut Mines-Télécom [Paris] (IMT)-Télécom Paris-Institut Mines-Télécom [Paris] (IMT)-Télécom Paris, Département Images, Données, Signal (IDS), Télécom ParisTech, Institut Polytechnique de Paris (IP Paris), paper funded by DSAIDIS, paper funded by DSAIDIS chair, ANR-20-CE23-0028,LIMPID,Exploitation de machines interprétables pour l'amélioration des performances et la prise de décision(2020), Parekh, Jayneel, and Exploitation de machines interprétables pour l'amélioration des performances et la prise de décision - - LIMPID2020 - ANR-20-CE23-0028 - AAPG2020 - VALID
- Subjects
FOS: Computer and information sciences ,Computer Science - Machine Learning ,Statistics - Machine Learning ,Machine Learning (stat.ML) ,[INFO]Computer Science [cs] ,[INFO] Computer Science [cs] ,Machine Learning (cs.LG) - Abstract
International audience; To tackle interpretability in deep learning, we present a novel framework to jointly learn a predictive model and its associated interpretation model. The interpreter provides both local and global interpretability about the predictive model in terms of human-understandable high level attribute functions, with minimal loss of accuracy. This is achieved by a dedicated architecture and well chosen regularization penalties. We seek for a small-size dictionary of high level attribute functions that take as inputs the outputs of selected hidden layers and whose outputs feed a linear classifier. We impose strong conciseness on the activation of attributes with an entropy-based criterion while enforcing fidelity to both inputs and outputs of the predictive model. A detailed pipeline to visualize the learnt features is also developed. Moreover, besides generating interpretable models by design, our approach can be specialized to provide post-hoc interpretations for a pre-trained neural network. We validate our approach against several state-of-the-art methods on multiple datasets and show its efficacy on both kinds of tasks.
- Published
- 2021