| | |
| | |
Stat |
Members: 3645 Articles: 2'501'711 Articles rated: 2609
20 April 2024 |
|
| | | |
|
Article overview
| |
|
Regularizing Black-box Models for Improved Interpretability (HILL 2019 Version) | Gregory Plumb
; Maruan Al-Shedivat
; Eric Xing
; Ameet Talwalkar
; | Date: |
31 May 2019 | Abstract: | Most of the work on interpretable machine learning has focused on designing
either inherently interpretable models, which typically trade-off accuracy for
interpretability, or post-hoc explanation systems, which lack guarantees about
their explanation quality. We propose an alternative to these approaches by
directly regularizing a black-box model for interpretability at training time.
Our approach explicitly connects three key aspects of interpretable machine
learning: (i) the model’s innate explainability, (ii) the explanation system
used at test time, and (iii) the metrics that measure explanation quality. Our
regularization results in substantial improvement in terms of the explanation
fidelity and stability metrics across a range of datasets and black-box
explanation systems while slightly improving accuracy. Further, if the
resulting model is still not sufficiently interpretable, the weight of the
regularization term can be adjusted to achieve the desired trade-off between
accuracy and interpretability. Finally, we justify theoretically that the
benefits of explanation-based regularization generalize to unseen points. | Source: | arXiv, 1906.1431 | Services: | Forum | Review | PDF | Favorites |
|
|
No review found.
Did you like this article?
Note: answers to reviews or questions about the article must be posted in the forum section.
Authors are not allowed to review their own article. They can use the forum section.
browser Mozilla/5.0 AppleWebKit/537.36 (KHTML, like Gecko; compatible; ClaudeBot/1.0; +claudebot@anthropic.com)
|
| |
|
|
|
| News, job offers and information for researchers and scientists:
| |