System disruptions
We are currently experiencing disruptions on the search portals due to high traffic. We are working to resolve the issue, you may temporarily encounter an error message.
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Well-Calibrated Rule Extractors
Jönköping University, School of Engineering, JTH, Department of Computing, Jönköping AI Lab (JAIL).ORCID iD: 0000-0003-0412-6199
Jönköping University, School of Engineering, JTH, Department of Computing, Jönköping AI Lab (JAIL).ORCID iD: 0000-0003-0274-9026
Jönköping University, School of Engineering, JTH, Department of Computing.ORCID iD: 0000-0003-2128-7090
2022 (English)In: Proceedings of the Eleventh Symposium on Conformal and Probabilistic Prediction with Applications: Volume 179: Conformal and Probabilistic Prediction with Applications, 24-26 August 2022, Brighton, UK / [ed] U. Johansson, H. Boström, K. A. Nguyen, Z. Luo & L. Carlsson, ML Research Press , 2022, Vol. 179, p. 72-91Conference paper, Published paper (Refereed)
Abstract [en]

While explainability is widely considered necessary for trustworthy predictive models, most explanation modules give only a limited understanding of the reasoning behind the predictions. In pedagogical rule extraction, an opaque model is approximated with a transparent model induced using original training instances, but with the predictions from the opaque model as targets. The result is an interpretable model revealing the exact reasoning used for every possible prediction. The pedagogical approach can be applied to any opaque model and use any learning algorithm producing transparent models as the actual rule extractor. Unfortunately, even if the extracted model is induced to mimic the opaque, test set fidelity may still be poor, thus clearly limiting the value of using the extracted model for explanations and analyses. In this paper, it is suggested to alleviate this problem by extracting probabilistic predictors with well-calibrated fitness estimates. For the calibration, Venn-Abers with its unique validity guarantees, is employed. Using a setup where decision trees are extracted from MLP neural networks, the suggested approach is first demonstrated in detail on one real-world data set. After that, a large-scale empirical evaluation using 25 publicly available benchmark data sets is presented. The results show that the method indeed extracts interpretable models with well-calibrated fitness estimates, i.e., the extracted model can be used for explaining the opaque. Specifically, in the setup used, every leaf in a decision tree contains a label and a well-calibrated probability interval for the fidelity. Consequently, a user could, in addition to obtaining explanations of individual predictions, find the parts of feature space where the decision tree is a good approximation of the MLP and not. In fact, using the sizes of the probability intervals, the models also provide an indication of how certain individual fitness estimates are.

Place, publisher, year, edition, pages
ML Research Press , 2022. Vol. 179, p. 72-91
Series
Proceedings of Machine Learning Research, E-ISSN 2640-3498 ; 179
Keywords [en]
Rule extraction, Fidelity, Interpretability, Explainability, Calibration, VennAbers predictors
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:hj:diva-58683Scopus ID: 2-s2.0-85164728534OAI: oai:DiVA.org:hj-58683DiVA, id: diva2:1705491
Conference
11th Symposium on Conformal and Probabilistic Prediction with Applications, 24-26 August 2022, Brighton, UK
Available from: 2022-10-24 Created: 2022-10-24 Last updated: 2023-08-17Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

ScopusFull-text

Authority records

Johansson, UlfLöfström, TuweStåhl, Niclas

Search in DiVA

By author/editor
Johansson, UlfLöfström, TuweStåhl, Niclas
By organisation
Jönköping AI Lab (JAIL)JTH, Department of Computing
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar

urn-nbn

Altmetric score

urn-nbn
Total: 387 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf