z-logo
open-access-imgOpen Access
Optimizing for Interpretability in Deep Neural Networks with Tree Regularization
Author(s) -
WuMike,
ParbhooSonali,
C HughesMichael,
RothVolker,
Doshi-VelezFinale
Publication year - 2021
Publication title -
journal of artificial intelligence research/the journal of artificial intelligence research
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.79
H-Index - 123
eISSN - 1943-5037
pISSN - 1076-9757
DOI - 10.1613/jair.1.12558
Subject(s) - interpretability , computer science , artificial intelligence , machine learning , deep learning , regularization (linguistics) , tree (set theory) , decision tree , deep neural networks , artificial neural network , benchmark (surveying) , black box , leverage (statistics) , mathematics , mathematical analysis , geodesy , geography
Deep models have advanced prediction in many domains, but their lack of interpretability  remains a key barrier to the adoption in many real world applications. There exists a large  body of work aiming to help humans understand these black box functions to varying levels  of granularity – for example, through distillation, gradients, or adversarial examples. These  methods however, all tackle interpretability as a separate process after training. In this  work, we take a different approach and explicitly regularize deep models so that they are  well-approximated by processes that humans can step through in little time. Specifically,  we train several families of deep neural networks to resemble compact, axis-aligned decision  trees without significant compromises in accuracy. The resulting axis-aligned decision  functions uniquely make tree regularized models easy for humans to interpret. Moreover,  for situations in which a single, global tree is a poor estimator, we introduce a regional tree regularizer that encourages the deep model to resemble a compact, axis-aligned decision  tree in predefined, human-interpretable contexts. Using intuitive toy examples, benchmark  image datasets, and medical tasks for patients in critical care and with HIV, we demonstrate  that this new family of tree regularizers yield models that are easier for humans to simulate  than L1 or L2 penalties without sacrificing predictive power. 

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here