SlideShare a Scribd company logo
Wednesday, 27th March
A Post-Hoc Categorisation
of Predictive Models
John Mitros
University College Dublin
ioannis.mitros@insight-centre.org
Outline
• Introduc)on
• Overview of interpretability/explainability
• Post-hoc approaches for interpretability
• Common themes, connec)ng ideas, general picture
• Not an exhaus)ve survey of all the literature body
• Open challenges and possible future direc)ons
• Examples and use cases
• Recent approaches
2
Explainable AI
3
Image from https://xaitutorial2019.github.io/
Explainable vs. Interpretable
• Explainable ML:
• Post-hoc analysis of black box models
• Interpretable ML:
• Intrinsically interpretable a.k.a transparent
4Rudin, C. & Ertekin, Ş. Math. Prog. Comp. (2018) 10: 659. hNps://doi.org/10.1007/s12532-018-0143-8
Interpretability
• It is inherently a mul/faceted no/on whose meaning changes according to
the different applicability scenarios
• Interpretability needs to answer what the model has learned and why it
came to that conclusion
• Defini/on of interpretability:
• “interpretability is the degree to which a human can understand the cause of a
decision” (Miller 2017)
5
Interpretability
• Defini&on of interpretability:
• “interpretability is the degree to which a machine can explain the cause of a decision
into coherent logical arguments”
• inherently it involves a bijec&ve process from input to output and vice versa, where
the intermediate steps are transparent to the end user
!" " # → % &ℎ() " % → #
• logical fallacies should be avoided
6
Scope of Interpretability
7
Lipton, Z. C. 2016. The Mythos of Model Interpretability. ICML Workshop on Human Interpretability in Machine Learning
(WHI 2016), New York, NY
Scope of Interpretability
8
Hierarchical Structure of Interpretability
9
Examples of Post-hoc Explana2ons
10
Chen, C.; Li, O.; Barne1, A.; Su, J.; and Rudin, C. 2018. This looks like that: Deep learning for interpretable image
recogniHon. ICML
Group A
What has the model
learned?
(holisHc or modular level)
Model
Specific
Examples of Post-hoc Explana2ons
11
Rudin, C., and Ertekin, S ̧. 2018. Learning customized and op>mized lists of rules with mathema>cal programming.
Mathema'cal Programming Computa'on 10(4):659–702
Group A
What has the model
learned?
(holis>c or modular level)
Model
Specific
Examples of Post-hoc Explana2ons
12
Montavon, G.; Lapuschkin, S.; Binder, A.; Samek, W.; and Mu ̈ller, K.-R. 2017. Explaining nonlinear classificaIon de- cisions
with deep Taylor decomposiIon. Pa#ern Recogni- .on 65:211–222.
Group A
What has the model
learned?
(holisIc or modular level)
Model
Specific
Examples of Post-hoc Explana2ons
13
Ribeiro, M. T.; Singh, S.; and Guestrin, C. 2016. ”Why Should I Trust You?”: Explaining the PredicJons of Any Classifier.
ACM KDD
Group A
What has the model
learned?
(holisJc or modular level)
Model
AgnosJc
Examples of Post-hoc Explana2ons
14Ribeiro, M. T.; Singh, S.; and Guestrin, C. 2018. Anchors: High-Precision Model-Agnostic Explanations. AAAI Press 32:1527–
Group A
What has the model
learned?
(holisOc or modular level)
Model
AgnosOc
Examples of Post-hoc Explana2ons
15
Henelius, A.; Puolama ̈ki, K.; and Ukkonen, A. 2017. Interpre?ng Classifiers through ADribute Interac?ons in Datasets.
ICML
Group A
What has the model
learned?
(holis?c or modular level)
Model
Agnos?c
General Concepts & Methods
• Rule Sets
• Sensi+vity Analysis
• Induc+ve Logic/Programming
• Recently:
• Counterfactuals
• Adversarial approaches
• Game theory
16
Interes'ng Direc'ons
17
Interes'ng Direc'ons
18
Open Challenges
• No formal agreed upon defini1on
• The no1on of interpretability seems to be an ill-defined term?
• Having agreed upon defini1on avoids reinven1ng the wheel
• Easier to built upon and contribute to prior work
• Rigorous, agreed upon evalua1on metrics
• Clear dis1nc1on of human vs. machine based evalua1on metrics
• Provide a clear picture of what is working and what needs improvement
19
Open Challenges
• Stochas(c nature of the models, different random seeds lead to different
outcomes for the same models
• P Henderson, R Islam, P Bachman, J Pineau, Deep Reinforcement Learning That
MaAers, AAAI 20018
• Models are built on assump(ons à = f( )
• When do they break and how?
20
Open Challenges
• Humans are great storytellers/story makers
• Memory championship à Method of loci
• Often humans create stories from small indications which rely upon in order to
build explanations
• These explanations might not have any relation with the underlying actual model
• How to avoid specific cognitive biases?
• Framing effect
• Focusing effect
• Illusory correlation 21
Open Challenges
22
Open Challenges
23
Open Challenges
• Saliency maps can be misleading (Olah et al., 2018)
• Models are uncalibrated
• Need for more transparent approaches
• Bringing another to interpret the exisCng 24
References
1. Bodenhofer, Ulrich and Bauer, Peter. Towards an Axiomatic Treatment of
Interpretability, Fuzzy Systems, 2000.
2. Olah, Chris and Satyanarayan, Arvind. The Building Blocks of Interpretability, Distill.pub,
2018.
3. Zadrozny, Bianca and Elkan, Charles. Obtaining calibrated probability estimates from
decision trees and naive bayesian classifiers. In ICML, pp. 609–616, 2001.
4. Zadrozny, Bianca and Elkan, Charles. Transforming classifier scores into accurate
multiclass probability estimates. In KDD, pp. 694–699, 2002.
5. Naeini, Mahdi Pakdaman, Cooper, Gregory F, and Hauskrecht, Milos. Obtaining well
calibrated probabilities using bayesian binning. In AAAI, pp. 2901, 2015.
6. Platt, John et al. Probabilistic outputs for support vector machines and comparisons to
regularized likelihood methods. Advances in large margin classifiers, 10(3): 61–74, 1999.
7. Guo, Chuan and Pleiss, Geoff and Sun, Yu and Weinberger, Kilian Q. On Calibration of
Modern Neural Networks. In ICML 2017.
25
THANKS!
Ques-ons?
26
Preprint: https://arxiv.org/abs/1904.02495
ioannis.mitros@insight-centre.org

More Related Content

PDF
Discrete Structure
Syed Shah
 
PPTX
Dms introduction Sharmila Chidaravalli
Sharmila Chidaravalli
 
PDF
Rsqrd AI: Recent Advances in Explainable Machine Learning Research
Sanjana Chowdhury
 
PDF
Interactive Analysis of Word Vector Embeddings
gleicher
 
PDF
SMIU Lecture #1 & 2 Introduction to Discrete Structure and Truth Table.pdf
MuhammadUmerIhtisham
 
PDF
IUI 2016 Presentation Slide
University of Central Florida
 
PPTX
What is Gate exam
Amit Kumar , Jaipur Engineers
 
DOC
Cds exam syllabus 2012
Gaurav Singh Rana
 
Discrete Structure
Syed Shah
 
Dms introduction Sharmila Chidaravalli
Sharmila Chidaravalli
 
Rsqrd AI: Recent Advances in Explainable Machine Learning Research
Sanjana Chowdhury
 
Interactive Analysis of Word Vector Embeddings
gleicher
 
SMIU Lecture #1 & 2 Introduction to Discrete Structure and Truth Table.pdf
MuhammadUmerIhtisham
 
IUI 2016 Presentation Slide
University of Central Florida
 
Cds exam syllabus 2012
Gaurav Singh Rana
 

What's hot (20)

PDF
Semantics-based Graph Approach to Complex Question-Answering
Jinho Choi
 
PDF
A Review of Distributional models of word meaning (Lenci, 2018)
Shuntaro Yada
 
PDF
Introduction to gre quant
Crack The Quant
 
PPTX
Marking breakdown
Queensbury 6th Form
 
DOCX
Cara apakah a raven matriks uji
Agel Hatmaja
 
PPTX
Relationship of knowledge to learn in programming methodology and evaluation ...
Technological Ecosystems for Enhancing Multiculturality
 
DOCX
Computational methods couurseout line
Temesgen Geta
 
PPTX
OMSU vs. EU comparative curricula study
metamath
 
PPTX
Flowchart and concept maps
Shruti Nimbkar
 
DOCX
M phil
peeroz
 
PPTX
GRE Presentation
Western Academy
 
PDF
Montpellier
Ben Bolker
 
PPTX
Applied logic: A mastery learning approach
john6938
 
PPT
Numerical Analysis And Linear Algebra
Ghulam Murtaza
 
PPT
Pattern for gate 2014
sakshij91
 
POT
先進網際服務系統1008
youji0811
 
PDF
Lec 6 learning
Eyob Seyfu
 
PDF
Sample prac exam2013
hccit
 
PDF
computational science & engineering seminar, 16 oct 2013
Ben Bolker
 
Semantics-based Graph Approach to Complex Question-Answering
Jinho Choi
 
A Review of Distributional models of word meaning (Lenci, 2018)
Shuntaro Yada
 
Introduction to gre quant
Crack The Quant
 
Marking breakdown
Queensbury 6th Form
 
Cara apakah a raven matriks uji
Agel Hatmaja
 
Relationship of knowledge to learn in programming methodology and evaluation ...
Technological Ecosystems for Enhancing Multiculturality
 
Computational methods couurseout line
Temesgen Geta
 
OMSU vs. EU comparative curricula study
metamath
 
Flowchart and concept maps
Shruti Nimbkar
 
M phil
peeroz
 
GRE Presentation
Western Academy
 
Montpellier
Ben Bolker
 
Applied logic: A mastery learning approach
john6938
 
Numerical Analysis And Linear Algebra
Ghulam Murtaza
 
Pattern for gate 2014
sakshij91
 
先進網際服務系統1008
youji0811
 
Lec 6 learning
Eyob Seyfu
 
Sample prac exam2013
hccit
 
computational science & engineering seminar, 16 oct 2013
Ben Bolker
 
Ad

Similar to A Categorisation of Post-hoc Explanations for Predictive Models (20)

PPTX
Hima_Lakkaraju_XAI_ShortCourse.pptx
PhanThDuy
 
PDF
Can Machine Learning Models be Trusted? Explaining Decisions of ML Models
Darek Smyk
 
PDF
GDG Community Day 2023 - Interpretable ML in production
SARADINDU SENGUPTA
 
PDF
Reds interpretability report
Raouf KESKES
 
PPTX
Responsible AI in Industry: Practical Challenges and Lessons Learned
Krishnaram Kenthapadi
 
PPTX
Hima_CHIL.pptx machine learning motivation
AnimeshRoy876945
 
PPTX
Explainable AI in Industry (KDD 2019 Tutorial)
Krishnaram Kenthapadi
 
PDF
Human in the loop: Bayesian Rules Enabling Explainable AI
Pramit Choudhary
 
PPTX
Explainable AI in Industry (FAT* 2020 Tutorial)
Krishnaram Kenthapadi
 
PDF
ML Interpretability Inside Out
Mara Graziani
 
PDF
DALL-E 2 - OpenAI imagery automation first developed by Vishal Coodye in 2021...
MITAILibrary
 
PDF
C3 w5
Ajay Taneja
 
PPTX
Introduction to Interpretable Machine Learning
Nguyen Giang
 
PDF
Explainable_artificial_intelligence_A_survey.pdf
fayazahmed944049
 
PDF
Model Evaluation in the land of Deep Learning
Pramit Choudhary
 
PPTX
algorithmic-decisions, fairness, machine learning, provenance, transparency
Paolo Missier
 
PDF
graziani_bias.pdf
Mara Graziani
 
PPTX
Machine Learning Interpretability - Mateusz Dymczyk - H2O AI World London 2018
Sri Ambati
 
PDF
Machine Learning Interpretability / Explainability
Raouf KESKES
 
PDF
Interpretable Machine Learning
inovex GmbH
 
Hima_Lakkaraju_XAI_ShortCourse.pptx
PhanThDuy
 
Can Machine Learning Models be Trusted? Explaining Decisions of ML Models
Darek Smyk
 
GDG Community Day 2023 - Interpretable ML in production
SARADINDU SENGUPTA
 
Reds interpretability report
Raouf KESKES
 
Responsible AI in Industry: Practical Challenges and Lessons Learned
Krishnaram Kenthapadi
 
Hima_CHIL.pptx machine learning motivation
AnimeshRoy876945
 
Explainable AI in Industry (KDD 2019 Tutorial)
Krishnaram Kenthapadi
 
Human in the loop: Bayesian Rules Enabling Explainable AI
Pramit Choudhary
 
Explainable AI in Industry (FAT* 2020 Tutorial)
Krishnaram Kenthapadi
 
ML Interpretability Inside Out
Mara Graziani
 
DALL-E 2 - OpenAI imagery automation first developed by Vishal Coodye in 2021...
MITAILibrary
 
Introduction to Interpretable Machine Learning
Nguyen Giang
 
Explainable_artificial_intelligence_A_survey.pdf
fayazahmed944049
 
Model Evaluation in the land of Deep Learning
Pramit Choudhary
 
algorithmic-decisions, fairness, machine learning, provenance, transparency
Paolo Missier
 
graziani_bias.pdf
Mara Graziani
 
Machine Learning Interpretability - Mateusz Dymczyk - H2O AI World London 2018
Sri Ambati
 
Machine Learning Interpretability / Explainability
Raouf KESKES
 
Interpretable Machine Learning
inovex GmbH
 
Ad

Recently uploaded (20)

PPTX
short term project on AI Driven Data Analytics
JMJCollegeComputerde
 
PPT
Real Life Application of Set theory, Relations and Functions
manavparmar205
 
PDF
blockchain123456789012345678901234567890
tanvikhunt1003
 
PPTX
1intro to AI.pptx AI components & composition
ssuserb993e5
 
PPTX
Presentation (1) (1).pptx k8hhfftuiiigff
karthikjagath2005
 
PPT
Grade 5 PPT_Science_Q2_W6_Methods of reproduction.ppt
AaronBaluyut
 
PDF
Research about a FoodFolio app for personalized dietary tracking and health o...
AustinLiamAndres
 
PDF
A Systems Thinking Approach to Algorithmic Fairness.pdf
Epistamai
 
PPTX
Future_of_AI_Presentation for everyone.pptx
boranamanju07
 
PDF
717629748-Databricks-Certified-Data-Engineer-Professional-Dumps-by-Ball-21-03...
pedelli41
 
PDF
Chad Readey - An Independent Thinker
Chad Readey
 
PPTX
Blue and Dark Blue Modern Technology Presentation.pptx
ap177979
 
PDF
202501214233242351219 QASS Session 2.pdf
lauramejiamillan
 
PPTX
Employee Salary Presentation.l based on data science collection of data
barridevakumari2004
 
PPTX
Introduction-to-Python-Programming-Language (1).pptx
dhyeysapariya
 
PPTX
Introduction to Data Analytics and Data Science
KavithaCIT
 
PPTX
Probability systematic sampling methods.pptx
PrakashRajput19
 
PDF
Blue Futuristic Cyber Security Presentation.pdf
tanvikhunt1003
 
PPTX
IP_Journal_Articles_2025IP_Journal_Articles_2025
mishell212144
 
PPTX
Web dev -ppt that helps us understand web technology
shubhragoyal12
 
short term project on AI Driven Data Analytics
JMJCollegeComputerde
 
Real Life Application of Set theory, Relations and Functions
manavparmar205
 
blockchain123456789012345678901234567890
tanvikhunt1003
 
1intro to AI.pptx AI components & composition
ssuserb993e5
 
Presentation (1) (1).pptx k8hhfftuiiigff
karthikjagath2005
 
Grade 5 PPT_Science_Q2_W6_Methods of reproduction.ppt
AaronBaluyut
 
Research about a FoodFolio app for personalized dietary tracking and health o...
AustinLiamAndres
 
A Systems Thinking Approach to Algorithmic Fairness.pdf
Epistamai
 
Future_of_AI_Presentation for everyone.pptx
boranamanju07
 
717629748-Databricks-Certified-Data-Engineer-Professional-Dumps-by-Ball-21-03...
pedelli41
 
Chad Readey - An Independent Thinker
Chad Readey
 
Blue and Dark Blue Modern Technology Presentation.pptx
ap177979
 
202501214233242351219 QASS Session 2.pdf
lauramejiamillan
 
Employee Salary Presentation.l based on data science collection of data
barridevakumari2004
 
Introduction-to-Python-Programming-Language (1).pptx
dhyeysapariya
 
Introduction to Data Analytics and Data Science
KavithaCIT
 
Probability systematic sampling methods.pptx
PrakashRajput19
 
Blue Futuristic Cyber Security Presentation.pdf
tanvikhunt1003
 
IP_Journal_Articles_2025IP_Journal_Articles_2025
mishell212144
 
Web dev -ppt that helps us understand web technology
shubhragoyal12
 

A Categorisation of Post-hoc Explanations for Predictive Models

  • 1. Wednesday, 27th March A Post-Hoc Categorisation of Predictive Models John Mitros University College Dublin [email protected]
  • 2. Outline • Introduc)on • Overview of interpretability/explainability • Post-hoc approaches for interpretability • Common themes, connec)ng ideas, general picture • Not an exhaus)ve survey of all the literature body • Open challenges and possible future direc)ons • Examples and use cases • Recent approaches 2
  • 3. Explainable AI 3 Image from https://xaitutorial2019.github.io/
  • 4. Explainable vs. Interpretable • Explainable ML: • Post-hoc analysis of black box models • Interpretable ML: • Intrinsically interpretable a.k.a transparent 4Rudin, C. & Ertekin, Ş. Math. Prog. Comp. (2018) 10: 659. hNps://doi.org/10.1007/s12532-018-0143-8
  • 5. Interpretability • It is inherently a mul/faceted no/on whose meaning changes according to the different applicability scenarios • Interpretability needs to answer what the model has learned and why it came to that conclusion • Defini/on of interpretability: • “interpretability is the degree to which a human can understand the cause of a decision” (Miller 2017) 5
  • 6. Interpretability • Defini&on of interpretability: • “interpretability is the degree to which a machine can explain the cause of a decision into coherent logical arguments” • inherently it involves a bijec&ve process from input to output and vice versa, where the intermediate steps are transparent to the end user !" " # → % &ℎ() " % → # • logical fallacies should be avoided 6
  • 7. Scope of Interpretability 7 Lipton, Z. C. 2016. The Mythos of Model Interpretability. ICML Workshop on Human Interpretability in Machine Learning (WHI 2016), New York, NY
  • 9. Hierarchical Structure of Interpretability 9
  • 10. Examples of Post-hoc Explana2ons 10 Chen, C.; Li, O.; Barne1, A.; Su, J.; and Rudin, C. 2018. This looks like that: Deep learning for interpretable image recogniHon. ICML Group A What has the model learned? (holisHc or modular level) Model Specific
  • 11. Examples of Post-hoc Explana2ons 11 Rudin, C., and Ertekin, S ̧. 2018. Learning customized and op>mized lists of rules with mathema>cal programming. Mathema'cal Programming Computa'on 10(4):659–702 Group A What has the model learned? (holis>c or modular level) Model Specific
  • 12. Examples of Post-hoc Explana2ons 12 Montavon, G.; Lapuschkin, S.; Binder, A.; Samek, W.; and Mu ̈ller, K.-R. 2017. Explaining nonlinear classificaIon de- cisions with deep Taylor decomposiIon. Pa#ern Recogni- .on 65:211–222. Group A What has the model learned? (holisIc or modular level) Model Specific
  • 13. Examples of Post-hoc Explana2ons 13 Ribeiro, M. T.; Singh, S.; and Guestrin, C. 2016. ”Why Should I Trust You?”: Explaining the PredicJons of Any Classifier. ACM KDD Group A What has the model learned? (holisJc or modular level) Model AgnosJc
  • 14. Examples of Post-hoc Explana2ons 14Ribeiro, M. T.; Singh, S.; and Guestrin, C. 2018. Anchors: High-Precision Model-Agnostic Explanations. AAAI Press 32:1527– Group A What has the model learned? (holisOc or modular level) Model AgnosOc
  • 15. Examples of Post-hoc Explana2ons 15 Henelius, A.; Puolama ̈ki, K.; and Ukkonen, A. 2017. Interpre?ng Classifiers through ADribute Interac?ons in Datasets. ICML Group A What has the model learned? (holis?c or modular level) Model Agnos?c
  • 16. General Concepts & Methods • Rule Sets • Sensi+vity Analysis • Induc+ve Logic/Programming • Recently: • Counterfactuals • Adversarial approaches • Game theory 16
  • 19. Open Challenges • No formal agreed upon defini1on • The no1on of interpretability seems to be an ill-defined term? • Having agreed upon defini1on avoids reinven1ng the wheel • Easier to built upon and contribute to prior work • Rigorous, agreed upon evalua1on metrics • Clear dis1nc1on of human vs. machine based evalua1on metrics • Provide a clear picture of what is working and what needs improvement 19
  • 20. Open Challenges • Stochas(c nature of the models, different random seeds lead to different outcomes for the same models • P Henderson, R Islam, P Bachman, J Pineau, Deep Reinforcement Learning That MaAers, AAAI 20018 • Models are built on assump(ons à = f( ) • When do they break and how? 20
  • 21. Open Challenges • Humans are great storytellers/story makers • Memory championship à Method of loci • Often humans create stories from small indications which rely upon in order to build explanations • These explanations might not have any relation with the underlying actual model • How to avoid specific cognitive biases? • Framing effect • Focusing effect • Illusory correlation 21
  • 24. Open Challenges • Saliency maps can be misleading (Olah et al., 2018) • Models are uncalibrated • Need for more transparent approaches • Bringing another to interpret the exisCng 24
  • 25. References 1. Bodenhofer, Ulrich and Bauer, Peter. Towards an Axiomatic Treatment of Interpretability, Fuzzy Systems, 2000. 2. Olah, Chris and Satyanarayan, Arvind. The Building Blocks of Interpretability, Distill.pub, 2018. 3. Zadrozny, Bianca and Elkan, Charles. Obtaining calibrated probability estimates from decision trees and naive bayesian classifiers. In ICML, pp. 609–616, 2001. 4. Zadrozny, Bianca and Elkan, Charles. Transforming classifier scores into accurate multiclass probability estimates. In KDD, pp. 694–699, 2002. 5. Naeini, Mahdi Pakdaman, Cooper, Gregory F, and Hauskrecht, Milos. Obtaining well calibrated probabilities using bayesian binning. In AAAI, pp. 2901, 2015. 6. Platt, John et al. Probabilistic outputs for support vector machines and comparisons to regularized likelihood methods. Advances in large margin classifiers, 10(3): 61–74, 1999. 7. Guo, Chuan and Pleiss, Geoff and Sun, Yu and Weinberger, Kilian Q. On Calibration of Modern Neural Networks. In ICML 2017. 25