SlideShare a Scribd company logo
A Unified Approach
to Interpreting
Model Predictions
Rama Irsheidat
Scott M. Lundberg. et al.
Introduction SHAP
Additive Feature
Attribution Methods
Simple Properties
Uniquely Determine
Additive Feature
Attributions
02
01
04
03
TABLE OF CONTENTS
Computational and
User Study
Experiments
05
Conclusion
06
Why do we care so
much about
explainability in ML ?
Introduction
Example
Introduction
Example
Introduction
Understanding why a model makes a certain prediction can be as crucial as
the prediction’s accuracy in many applications
Introduction
The reason this problem exists is with in Large Datasets complex models tend
to be very accurate but hard to interpret
Introduction
Focus on explaining individual predictions one at a time.
Introduction
We are replacing the input to a summation that you would get in a linear
model with something that represents the importance of that feature in the
complicated model.
Interprets individual
model predictions based
on locally approximating
the model around a given
prediction.
Interprets the predictions
of deep networks.
Recursive prediction
explanation method for
deep learning.
Shapley regression
values
Quantitative
Input Influence
Shapley sampling
values
LIME
DeepLIFT
Layer-Wise
Relevance
Propagation
Feature importance for linear
models. This method requires
retraining the model on all
features. It assigns an
importance value to each
feature that represents the
effect on the model prediction
of including that feature.
Explaining any model by Applying sampling
approximations to Equation in Shapley reg. values ,
and Approximating the effect of removing a variable
from the model by integrating over samples from the
training dataset.
Proposes a sampling
approximation to Shapley
values that is nearly
identical to Shapley
sampling values.
Interprets individual
model predictions based
on locally approximating
the model around a given
prediction. Interprets the predictions
of deep networks.
Recursive prediction
explanation method for
deep learning.
Shapley regression
values
Quantitative
Input Influence
Shapley sampling
values
LIME
DeepLIFT
Layer-Wise
Relevance
Propagation
Feature importances for linear
models in the presence of
multicollinearity. This method
requires retraining the model
on all features. It assigns an
importance value to each
feature that represents the
effect on the model prediction
of including that feature.
Explain any model by Applying sampling
approximations to Equation in Shapley reg. values ,
and Approximating the effect of removing a variable
from the model by integrating over samples from the
training dataset.
Proposes a sampling
approximation to Shapley
values that is nearly
identical to Shapley
sampling values.
Additive Feature Attribution Methods
Additive Feature Attribution Methods
Interprets individual
model predictions based
on locally approximating
the model around a given
prediction. Interprets the predictions
of deep networks.
Recursive prediction
explanation method for
deep learning.
Shapley regression
values
Quantitative
Input Influence
Shapley sampling
values
LIME
DeepLIFT
Layer-Wise
Relevance
Propagation
Feature importances for linear
models in the presence of
multicollinearity. This method
requires retraining the model
on all features. It assigns an
importance value to each
feature that represents the
effect on the model prediction
of including that feature.
Explain any model by Applying sampling
approximations to Equation in Shapley reg. values ,
and Approximating the effect of removing a variable
from the model by integrating over samples from the
training dataset.
Proposes a sampling
approximation to Shapley
values that is nearly
identical to Shapley
sampling values.
Have some better
theoretical grounding
but slower
computation
Additive Feature Attribution Methods
Interprets individual
model predictions based
on locally approximating
the model around a given
prediction. Interprets the predictions
of deep networks.
Recursive prediction
explanation method for
deep learning.
Shapley regression
values
Quantitative
Input Influence
Shapley sampling
values
LIME
DeepLIFT
Layer-Wise
Relevance
Propagation
Feature importances for linear
models in the presence of
multicollinearity. This method
requires retraining the model
on all features. It assigns an
importance value to each
feature that represents the
effect on the model prediction
of including that feature.
Explain any model by Applying sampling
approximations to Equation in Shapley reg. values ,
and Approximating the effect of removing a variable
from the model by integrating over samples from the
training dataset.
Proposes a sampling
approximation to Shapley
values that is nearly
identical to Shapley
sampling values.
Have faster
estimation but less
guarantees
Interprets individual
model predictions based
on locally approximating
the model around a given
prediction.
Interprets the predictions
of deep networks.
Recursive prediction
explanation method for
deep learning.
Shapley regression
values
Quantitative
Input Influence
Shapley sampling
values
LIME
DeepLIFT
Layer-Wise
Relevance
Propagation
Feature importances for linear
models in the presence of
multicollinearity. This method
requires retraining the model
on all features. It assigns an
importance value to each
feature that represents the
effect on the model prediction
of including that feature.
Explain any model by Applying sampling
approximations to Equation in Shapley reg. values ,
and Approximating the effect of removing a variable
from the model by integrating over samples from the
training dataset.
Proposes a sampling
approximation to Shapley
values that is nearly
identical to Shapley
sampling values.
SHAP
How should we define importance of each
feature (φi (f , x))
Base rate of loan rejection or how often do people get their loans denied on
average?
Why am I 55 percent ?
We have to explain is this
35 percent difference
here .
So how can we do this?
We should just take the expected value of the output of our model (Base
rate), then we can just introduce a term into that conditional expectation.
Fact that John's 20, his
risk jumps up by 15
percent.
We should just take the expected value of the output of our model (Base
rate), then we can just introduce a term into that conditional expectation.
A very risky profession
and that jumps the risk
up to 70 percent.
We should just take the expected value of the output of our model (Base
rate), then we can just introduce a term into that conditional expectation.
We should just take the expected value of the output of our model (Base
rate), then we can just introduce a term into that conditional expectation.
He made a ton of money
in the stock market last
year. So his capital gains
pushes him down to 55
percent.
We have to explain is this
35 percent difference
here . So how can we do
this?
We've basically divided up how
we got from here to here by
conditioning one at a time on
all the features until we've
conditioned on all of them.
A Unified Approach to Interpreting Model Predictions (SHAP)
Example:
We can't just pick a particular order and think that we've solved it so what
do we do here?
Meaningless
Simple Properties Uniquely Determine Additive Feature Attributions
Means the output of the
explanation model matches
the original model for the
prediction being explained.
Requires features missing
in the original input to have
no impact.
If you change the original model such that a
feature has a larger impact in every possible
ordering, then that input's attribution
(importance) should not decrease.
Shapley Properties
Local accuracy Missingness
Consistency
SHAP values arise from averaging the φi values
across all possible orderings.
Very painful to compute.
Find an approximate solution
1. Model-Agnostic Approximations
1.1 Shapley sampling values
2.1 Kernel SHAP (Linear LIME + Shapley values)
Linear LIME (uses a linear explanation model) fit a linear model locally to the original
model that we're trying to explain.
Shapley values are the only possible solution that satisfies Properties 1-3 – local
accuracy, missingness and consistency.
This means we can now
estimate the Shapley
values using linear
regression.
2. Model-Specific Approximations
2.1 Linear SHAP
For linear models, SHAP values can be approximated directly from the model’s weight
coefficients.
2.2 Low-Order SHAP
3.2 Max SHAP
Calculating the probability that each input will increase the maximum value over every
other input.
2. Model-Specific Approximations
2.4 Deep SHAP (DeepLIFT + Shapley values)
DeepLIFT
Recursive prediction explanation method for deep learning that satisfies local accuracy
and missingness, we know that Shapley values represent the attribution values that
satisfy consistency.
Adapting DeepLIFT to become a compositional approximation of SHAP values, leading to
Deep SHAP.
Computational and User Study
Experiments
1. Computational Efficiency
Comparing Shapley sampling, SHAP, and LIME on both dense and sparse decision tree
models illustrates both the improved sample efficiency of Kernel SHAP and that values
from LIME can differ significantly from SHAP values that satisfy local accuracy and
consistency.
2. Consistency with Human Intuition
(A) Attributions of sickness score (B) Attributions of profit among three men
Participants were asked to assign importance for the output (the sickness score or
money won) among the inputs (i.e., symptoms or players). We found a much stronger
agreement between human explanations and SHAP than with other methods.
3. Explaining Class Differences
Explaining the output of a convolutional network trained on the MNIST digit dataset.
(A) Red areas increase the probability of that class, and blue areas decrease the
probability . Masked removes pixels in order to go from 8 to 3.
(B) The change in log odds when masking over 20 random images supports the use of
better estimates of SHAP values.
Conclusion
• The growing tension between the accuracy and interpretability of model
predictions has motivated the development of methods that help users
interpret predictions.
• The SHAP framework identifies the class of additive feature importance
methods (which includes six previous methods) and shows there is a
unique solution in this class that adheres to desirable properties.
• We presented several different estimation methods for SHAP values, along
with proofs and experiments showing that these values are desirable.
THANKS!

More Related Content

PPTX
Explainable AI in Healthcare
vonaurum
 
PPTX
Shap
Giovanni Bruner
 
PDF
Michael Lewrick , Patrick Link, Larry Leifer - The Design Thinking Toolbox_ A...
MridulGupta21235
 
PDF
Inferring the effect of an event using CausalImpact by Kay H. Brodersen
Big Data Spain
 
PDF
AI and ML Series - Introduction to Generative AI and LLMs - Session 1
DianaGray10
 
PPTX
Explainable AI in Industry (AAAI 2020 Tutorial)
Krishnaram Kenthapadi
 
PPTX
Adverse drug reaction
Viraj Shinde
 
PPTX
SOAP notes
Abdurahmansalah
 
Explainable AI in Healthcare
vonaurum
 
Michael Lewrick , Patrick Link, Larry Leifer - The Design Thinking Toolbox_ A...
MridulGupta21235
 
Inferring the effect of an event using CausalImpact by Kay H. Brodersen
Big Data Spain
 
AI and ML Series - Introduction to Generative AI and LLMs - Session 1
DianaGray10
 
Explainable AI in Industry (AAAI 2020 Tutorial)
Krishnaram Kenthapadi
 
Adverse drug reaction
Viraj Shinde
 
SOAP notes
Abdurahmansalah
 

What's hot (20)

PPTX
Using SHAP to Understand Black Box Models
Jonathan Bechtel
 
PDF
Unified Approach to Interpret Machine Learning Model: SHAP + LIME
Databricks
 
PPTX
eScience SHAP talk
Scott Lundberg
 
PPTX
Explainable AI in Industry (FAT* 2020 Tutorial)
Krishnaram Kenthapadi
 
PPTX
Explainable Machine Learning (Explainable ML)
Hayim Makabee
 
PDF
Interpretable machine learning : Methods for understanding complex models
Manojit Nandi
 
PPTX
Scott Lundberg, Microsoft Research - Explainable Machine Learning with Shaple...
Sri Ambati
 
PPTX
An Introduction to XAI! Towards Trusting Your ML Models!
Mansour Saffar
 
PDF
Achieving Algorithmic Transparency with Shapley Additive Explanations (H2O Lo...
Sri Ambati
 
PPTX
Explainable AI in Industry (KDD 2019 Tutorial)
Krishnaram Kenthapadi
 
PPTX
Interpretable machine learning
Sri Ambati
 
PDF
Missing data handling
QuantUniversity
 
PDF
Linear models for data science
Brad Klingenberg
 
PPTX
[BEDROCK] Claude Prompt Engineering Techniques.pptx
ssuserdd71c7
 
PPTX
Explainable AI
Arithmer Inc.
 
PDF
Data Science - Part III - EDA & Model Selection
Derek Kane
 
PPTX
Prediction of House Sales Price
Anirvan Ghosh
 
PPTX
Responsible AI in Industry: Practical Challenges and Lessons Learned
Krishnaram Kenthapadi
 
PDF
Applying Statistical Modeling and Machine Learning to Perform Time-Series For...
PyData
 
PPTX
Explainable AI in Industry (WWW 2020 Tutorial)
Krishnaram Kenthapadi
 
Using SHAP to Understand Black Box Models
Jonathan Bechtel
 
Unified Approach to Interpret Machine Learning Model: SHAP + LIME
Databricks
 
eScience SHAP talk
Scott Lundberg
 
Explainable AI in Industry (FAT* 2020 Tutorial)
Krishnaram Kenthapadi
 
Explainable Machine Learning (Explainable ML)
Hayim Makabee
 
Interpretable machine learning : Methods for understanding complex models
Manojit Nandi
 
Scott Lundberg, Microsoft Research - Explainable Machine Learning with Shaple...
Sri Ambati
 
An Introduction to XAI! Towards Trusting Your ML Models!
Mansour Saffar
 
Achieving Algorithmic Transparency with Shapley Additive Explanations (H2O Lo...
Sri Ambati
 
Explainable AI in Industry (KDD 2019 Tutorial)
Krishnaram Kenthapadi
 
Interpretable machine learning
Sri Ambati
 
Missing data handling
QuantUniversity
 
Linear models for data science
Brad Klingenberg
 
[BEDROCK] Claude Prompt Engineering Techniques.pptx
ssuserdd71c7
 
Explainable AI
Arithmer Inc.
 
Data Science - Part III - EDA & Model Selection
Derek Kane
 
Prediction of House Sales Price
Anirvan Ghosh
 
Responsible AI in Industry: Practical Challenges and Lessons Learned
Krishnaram Kenthapadi
 
Applying Statistical Modeling and Machine Learning to Perform Time-Series For...
PyData
 
Explainable AI in Industry (WWW 2020 Tutorial)
Krishnaram Kenthapadi
 
Ad

Similar to A Unified Approach to Interpreting Model Predictions (SHAP) (20)

PPTX
Understanding Black Box Models with Shapley Values
Jonathan Bechtel
 
PDF
SHAP (SHapley Additive exPlanations)-A Complete guide to SHAP Values
Universidad Jorge Tadeo Lozano
 
PPTX
Interpretable ML
Mayur Sand
 
PDF
Explainable AI (XAI) - A Perspective
Saurabh Kaushik
 
PDF
Machine learning Mind Map
Ashish Patel
 
PDF
Intepretable Machine Learning
Ankit Tewari
 
PDF
CounterFactual Explanations.pdf
Bong-Ho Lee
 
PDF
Machine Learning Model Validation (Aijun Zhang 2024).pdf
Aijun Zhang
 
PDF
WIA 2019 - Steering Model Selection with Visual Diagnostics
Women in Analytics Conference
 
PDF
Steering Model Selection with Visual Diagnostics: Women in Analytics 2019
Rebecca Bilbro
 
PDF
GENETIC ALGORITHM FOR FUNCTION APPROXIMATION: AN EXPERIMENTAL INVESTIGATION
ijaia
 
PPTX
Regularization_BY_MOHAMED_ESSAM.pptx
Mohamed Essam
 
PDF
Steering Model Selection with Visual Diagnostics
Melissa Moody
 
PPTX
MachineLlearning introduction
The IOT Academy
 
PPTX
Introduction to ml
SuyashSingh70
 
PDF
Machine Learning.pdf
BeyaNasr1
 
PDF
Top 100+ Google Data Science Interview Questions.pdf
Datacademy.ai
 
PPTX
Overfitting & Underfitting
SOUMIT KAR
 
PDF
Sample_Subjective_Questions_Answers (1).pdf
AaryanArora10
 
PPT
Bengkel smartPLS 2011
Adi Ali
 
Understanding Black Box Models with Shapley Values
Jonathan Bechtel
 
SHAP (SHapley Additive exPlanations)-A Complete guide to SHAP Values
Universidad Jorge Tadeo Lozano
 
Interpretable ML
Mayur Sand
 
Explainable AI (XAI) - A Perspective
Saurabh Kaushik
 
Machine learning Mind Map
Ashish Patel
 
Intepretable Machine Learning
Ankit Tewari
 
CounterFactual Explanations.pdf
Bong-Ho Lee
 
Machine Learning Model Validation (Aijun Zhang 2024).pdf
Aijun Zhang
 
WIA 2019 - Steering Model Selection with Visual Diagnostics
Women in Analytics Conference
 
Steering Model Selection with Visual Diagnostics: Women in Analytics 2019
Rebecca Bilbro
 
GENETIC ALGORITHM FOR FUNCTION APPROXIMATION: AN EXPERIMENTAL INVESTIGATION
ijaia
 
Regularization_BY_MOHAMED_ESSAM.pptx
Mohamed Essam
 
Steering Model Selection with Visual Diagnostics
Melissa Moody
 
MachineLlearning introduction
The IOT Academy
 
Introduction to ml
SuyashSingh70
 
Machine Learning.pdf
BeyaNasr1
 
Top 100+ Google Data Science Interview Questions.pdf
Datacademy.ai
 
Overfitting & Underfitting
SOUMIT KAR
 
Sample_Subjective_Questions_Answers (1).pdf
AaryanArora10
 
Bengkel smartPLS 2011
Adi Ali
 
Ad

Recently uploaded (20)

PPTX
Multiscale Segmentation of Survey Respondents: Seeing the Trees and the Fores...
Sione Palu
 
PDF
SUMMER INTERNSHIP REPORT[1] (AutoRecovered) (6) (1).pdf
pandeydiksha814
 
PDF
717629748-Databricks-Certified-Data-Engineer-Professional-Dumps-by-Ball-21-03...
pedelli41
 
PDF
Key_Statistical_Techniques_in_Analytics_by_CA_Suvidha_Chaplot.pdf
CA Suvidha Chaplot
 
PDF
Practical Measurement Systems Analysis (Gage R&R) for design
Rob Schubert
 
PPTX
IP_Journal_Articles_2025IP_Journal_Articles_2025
mishell212144
 
PPT
Real Life Application of Set theory, Relations and Functions
manavparmar205
 
PDF
oop_java (1) of ice or cse or eee ic.pdf
sabiquntoufiqlabonno
 
PPTX
short term internship project on Data visualization
JMJCollegeComputerde
 
PDF
Company Presentation pada Perusahaan ADB.pdf
didikfahmi
 
PDF
blockchain123456789012345678901234567890
tanvikhunt1003
 
PPTX
Introduction to computer chapter one 2017.pptx
mensunmarley
 
PPTX
Introduction-to-Python-Programming-Language (1).pptx
dhyeysapariya
 
PPTX
lecture 13 mind test academy it skills.pptx
ggesjmrasoolpark
 
PPTX
Data Security Breach: Immediate Action Plan
varmabhuvan266
 
PDF
Chad Readey - An Independent Thinker
Chad Readey
 
PDF
The_Future_of_Data_Analytics_by_CA_Suvidha_Chaplot_UPDATED.pdf
CA Suvidha Chaplot
 
PPTX
Introduction to Data Analytics and Data Science
KavithaCIT
 
PPTX
Complete_STATA_Introduction_Beginner.pptx
mbayekebe
 
PPTX
short term project on AI Driven Data Analytics
JMJCollegeComputerde
 
Multiscale Segmentation of Survey Respondents: Seeing the Trees and the Fores...
Sione Palu
 
SUMMER INTERNSHIP REPORT[1] (AutoRecovered) (6) (1).pdf
pandeydiksha814
 
717629748-Databricks-Certified-Data-Engineer-Professional-Dumps-by-Ball-21-03...
pedelli41
 
Key_Statistical_Techniques_in_Analytics_by_CA_Suvidha_Chaplot.pdf
CA Suvidha Chaplot
 
Practical Measurement Systems Analysis (Gage R&R) for design
Rob Schubert
 
IP_Journal_Articles_2025IP_Journal_Articles_2025
mishell212144
 
Real Life Application of Set theory, Relations and Functions
manavparmar205
 
oop_java (1) of ice or cse or eee ic.pdf
sabiquntoufiqlabonno
 
short term internship project on Data visualization
JMJCollegeComputerde
 
Company Presentation pada Perusahaan ADB.pdf
didikfahmi
 
blockchain123456789012345678901234567890
tanvikhunt1003
 
Introduction to computer chapter one 2017.pptx
mensunmarley
 
Introduction-to-Python-Programming-Language (1).pptx
dhyeysapariya
 
lecture 13 mind test academy it skills.pptx
ggesjmrasoolpark
 
Data Security Breach: Immediate Action Plan
varmabhuvan266
 
Chad Readey - An Independent Thinker
Chad Readey
 
The_Future_of_Data_Analytics_by_CA_Suvidha_Chaplot_UPDATED.pdf
CA Suvidha Chaplot
 
Introduction to Data Analytics and Data Science
KavithaCIT
 
Complete_STATA_Introduction_Beginner.pptx
mbayekebe
 
short term project on AI Driven Data Analytics
JMJCollegeComputerde
 

A Unified Approach to Interpreting Model Predictions (SHAP)

  • 1. A Unified Approach to Interpreting Model Predictions Rama Irsheidat Scott M. Lundberg. et al.
  • 2. Introduction SHAP Additive Feature Attribution Methods Simple Properties Uniquely Determine Additive Feature Attributions 02 01 04 03 TABLE OF CONTENTS Computational and User Study Experiments 05 Conclusion 06
  • 3. Why do we care so much about explainability in ML ?
  • 6. Introduction Understanding why a model makes a certain prediction can be as crucial as the prediction’s accuracy in many applications
  • 7. Introduction The reason this problem exists is with in Large Datasets complex models tend to be very accurate but hard to interpret
  • 8. Introduction Focus on explaining individual predictions one at a time.
  • 9. Introduction We are replacing the input to a summation that you would get in a linear model with something that represents the importance of that feature in the complicated model.
  • 10. Interprets individual model predictions based on locally approximating the model around a given prediction. Interprets the predictions of deep networks. Recursive prediction explanation method for deep learning. Shapley regression values Quantitative Input Influence Shapley sampling values LIME DeepLIFT Layer-Wise Relevance Propagation Feature importance for linear models. This method requires retraining the model on all features. It assigns an importance value to each feature that represents the effect on the model prediction of including that feature. Explaining any model by Applying sampling approximations to Equation in Shapley reg. values , and Approximating the effect of removing a variable from the model by integrating over samples from the training dataset. Proposes a sampling approximation to Shapley values that is nearly identical to Shapley sampling values.
  • 11. Interprets individual model predictions based on locally approximating the model around a given prediction. Interprets the predictions of deep networks. Recursive prediction explanation method for deep learning. Shapley regression values Quantitative Input Influence Shapley sampling values LIME DeepLIFT Layer-Wise Relevance Propagation Feature importances for linear models in the presence of multicollinearity. This method requires retraining the model on all features. It assigns an importance value to each feature that represents the effect on the model prediction of including that feature. Explain any model by Applying sampling approximations to Equation in Shapley reg. values , and Approximating the effect of removing a variable from the model by integrating over samples from the training dataset. Proposes a sampling approximation to Shapley values that is nearly identical to Shapley sampling values. Additive Feature Attribution Methods
  • 12. Additive Feature Attribution Methods Interprets individual model predictions based on locally approximating the model around a given prediction. Interprets the predictions of deep networks. Recursive prediction explanation method for deep learning. Shapley regression values Quantitative Input Influence Shapley sampling values LIME DeepLIFT Layer-Wise Relevance Propagation Feature importances for linear models in the presence of multicollinearity. This method requires retraining the model on all features. It assigns an importance value to each feature that represents the effect on the model prediction of including that feature. Explain any model by Applying sampling approximations to Equation in Shapley reg. values , and Approximating the effect of removing a variable from the model by integrating over samples from the training dataset. Proposes a sampling approximation to Shapley values that is nearly identical to Shapley sampling values. Have some better theoretical grounding but slower computation
  • 13. Additive Feature Attribution Methods Interprets individual model predictions based on locally approximating the model around a given prediction. Interprets the predictions of deep networks. Recursive prediction explanation method for deep learning. Shapley regression values Quantitative Input Influence Shapley sampling values LIME DeepLIFT Layer-Wise Relevance Propagation Feature importances for linear models in the presence of multicollinearity. This method requires retraining the model on all features. It assigns an importance value to each feature that represents the effect on the model prediction of including that feature. Explain any model by Applying sampling approximations to Equation in Shapley reg. values , and Approximating the effect of removing a variable from the model by integrating over samples from the training dataset. Proposes a sampling approximation to Shapley values that is nearly identical to Shapley sampling values. Have faster estimation but less guarantees
  • 14. Interprets individual model predictions based on locally approximating the model around a given prediction. Interprets the predictions of deep networks. Recursive prediction explanation method for deep learning. Shapley regression values Quantitative Input Influence Shapley sampling values LIME DeepLIFT Layer-Wise Relevance Propagation Feature importances for linear models in the presence of multicollinearity. This method requires retraining the model on all features. It assigns an importance value to each feature that represents the effect on the model prediction of including that feature. Explain any model by Applying sampling approximations to Equation in Shapley reg. values , and Approximating the effect of removing a variable from the model by integrating over samples from the training dataset. Proposes a sampling approximation to Shapley values that is nearly identical to Shapley sampling values.
  • 15. SHAP
  • 16. How should we define importance of each feature (φi (f , x))
  • 17. Base rate of loan rejection or how often do people get their loans denied on average?
  • 18. Why am I 55 percent ?
  • 19. We have to explain is this 35 percent difference here . So how can we do this?
  • 20. We should just take the expected value of the output of our model (Base rate), then we can just introduce a term into that conditional expectation. Fact that John's 20, his risk jumps up by 15 percent.
  • 21. We should just take the expected value of the output of our model (Base rate), then we can just introduce a term into that conditional expectation. A very risky profession and that jumps the risk up to 70 percent.
  • 22. We should just take the expected value of the output of our model (Base rate), then we can just introduce a term into that conditional expectation.
  • 23. We should just take the expected value of the output of our model (Base rate), then we can just introduce a term into that conditional expectation. He made a ton of money in the stock market last year. So his capital gains pushes him down to 55 percent.
  • 24. We have to explain is this 35 percent difference here . So how can we do this? We've basically divided up how we got from here to here by conditioning one at a time on all the features until we've conditioned on all of them.
  • 27. We can't just pick a particular order and think that we've solved it so what do we do here? Meaningless
  • 28. Simple Properties Uniquely Determine Additive Feature Attributions
  • 29. Means the output of the explanation model matches the original model for the prediction being explained. Requires features missing in the original input to have no impact. If you change the original model such that a feature has a larger impact in every possible ordering, then that input's attribution (importance) should not decrease. Shapley Properties Local accuracy Missingness Consistency
  • 30. SHAP values arise from averaging the φi values across all possible orderings. Very painful to compute.
  • 32. 1. Model-Agnostic Approximations 1.1 Shapley sampling values 2.1 Kernel SHAP (Linear LIME + Shapley values) Linear LIME (uses a linear explanation model) fit a linear model locally to the original model that we're trying to explain. Shapley values are the only possible solution that satisfies Properties 1-3 – local accuracy, missingness and consistency. This means we can now estimate the Shapley values using linear regression.
  • 33. 2. Model-Specific Approximations 2.1 Linear SHAP For linear models, SHAP values can be approximated directly from the model’s weight coefficients. 2.2 Low-Order SHAP 3.2 Max SHAP Calculating the probability that each input will increase the maximum value over every other input.
  • 34. 2. Model-Specific Approximations 2.4 Deep SHAP (DeepLIFT + Shapley values) DeepLIFT Recursive prediction explanation method for deep learning that satisfies local accuracy and missingness, we know that Shapley values represent the attribution values that satisfy consistency. Adapting DeepLIFT to become a compositional approximation of SHAP values, leading to Deep SHAP.
  • 35. Computational and User Study Experiments
  • 36. 1. Computational Efficiency Comparing Shapley sampling, SHAP, and LIME on both dense and sparse decision tree models illustrates both the improved sample efficiency of Kernel SHAP and that values from LIME can differ significantly from SHAP values that satisfy local accuracy and consistency.
  • 37. 2. Consistency with Human Intuition (A) Attributions of sickness score (B) Attributions of profit among three men Participants were asked to assign importance for the output (the sickness score or money won) among the inputs (i.e., symptoms or players). We found a much stronger agreement between human explanations and SHAP than with other methods.
  • 38. 3. Explaining Class Differences Explaining the output of a convolutional network trained on the MNIST digit dataset. (A) Red areas increase the probability of that class, and blue areas decrease the probability . Masked removes pixels in order to go from 8 to 3. (B) The change in log odds when masking over 20 random images supports the use of better estimates of SHAP values.
  • 39. Conclusion • The growing tension between the accuracy and interpretability of model predictions has motivated the development of methods that help users interpret predictions. • The SHAP framework identifies the class of additive feature importance methods (which includes six previous methods) and shows there is a unique solution in this class that adheres to desirable properties. • We presented several different estimation methods for SHAP values, along with proofs and experiments showing that these values are desirable.