SlideShare a Scribd company logo
Explaining Black-Box
Machine Learning Predictions
Sameer Singh
University of California, Irvine
Machine Learning is Everywhere…
Classification: Wolf or a Husky?
Machine
Learning
Model
Wolf!
Classification: Wolf or a Husky?
Machine
Learning
Model
Husky!
Only 1 mistake!
Classification: Wolf or a Husky?
More Complex: Question Answering
Is there a moustache in the picture?
> Yes
What is the moustache made of?
> Banana
Essentially black-boxes!
How can we trust the
predictions are correct?
How do we know they are
not breaking regulations?
How do we avoid
“stupid mistakes”?
Trust
How can we understand and
predict the behavior?
Predict
How do we improve it to
prevent potential mistakes?
Improve
Only 1 mistake!
Classification: Wolf or a Husky?
We’ve built a
snow detector…
Explaining Black-Box Machine Learning Predictions - Sameer Singh, Assistant Professor of Computer Science, UC Irvine
Visual Question Answering
What is the moustache made of?
> Banana
What are the eyes made of?
> Bananas
What?
> Banana What is?
> Banana
Text Classification
Why did this
happen?
From: Keith Richards
Subject: Christianity is the answer
NTTP-Posting-Host: x.x.com
I think Christianity is the one true religion.
If you’d like to know more, send me a note
Applying for a Loan
Machine
Learning
I would like to apply for a loan.
vvvvvvv
vvvvvvv
vvvvvvv
vvvvvvv
vvvvvvv
Here is my information.
vvvvvvv
vvvvvvv
vvvvvvv
vvvvvvv
vvvvvvv
Sorry, your request has been denied
Why? What were the reasons?
Currently
Cannot explain.. [0.25,-4.5,3.5,-10.4,…]
How did we get here?
Big Data and Deep Learning
Simple Data
X1
X2
Linear Classifiers
X1
X2
You can interpret it…
- Both have a positive effect
- X1 > X2
10X1 + X2 - 5 > 0if:
otherwise
Decision trees
X1
X2
X1 > 0.5
X2 > 0.5
You can interpret it…
- X2 is irrelevant if X1<0.5
- Otherwise X2 is enough
Looking at the structure
How can we trust the
predictions are correct?
Trust
How can we understand and
predict the behavior?
Predict
How do we improve it to
prevent potential mistakes?
Improve
Test whether the structure
agrees with our intuitions.
Structure tells us exactly what
will happen on any data.
Structure tells you where the
error is, thus how to fix it.
Arrival of Big Data
Big Data: Applications of ML
Big Data: More Complexity
X1
X2
http://www.mckinsey.com/industries/high-tech/our-insights/an-executives-guide-to-machine-learning
Big Data: More Dimensions
Savings
Income
Profession
Loan Amount
Age
Marital
Status
Past defaults
Credit scores
Recent defaults
This easily goes to hundreds
- Images: thousands
- Text: tens of thousands
- Video: millions
- … and so on
X1
X2
Complex Surfaces
Savings
Income
Profession
Loan Amount
Age
Married
Past defaults
Credit scores
Recent defaults
…
Lots of dimensions
+
Black-boxes!
Accuracy vs Interpretability
Interpretability
Accuracy
10X1 + X2 - 5 > 0
X1 > 0.5
X2 > 0.5
millions of weights,
complex features
Real-world use case
Research on
“interpretable models”
Deep Learning
Interpretability
Accuracy
Real-world use case
Research on
“interpretable models”
Focus on accuracy!
Human-level
Looking at the structure
How can we trust the
predictions are correct?
Trust
How can we understand and
predict the behavior?
Predict
How do we improve it to
prevent potential mistakes?
Improve
Test whether the structure
agrees with our intuitions.
Structure tells us exactly what
will happen on any data.
Structure tells you where the
error is, thus how to fix it.
Explaining Predictions
The LIME Algorithm
Being Model-Agnostic…
No assumptions about the internal structure…
X1 > 0.5
X2 > 0.5f(x)
Explain any existing, or future, model
Data Decision
LIME: Explain Any Classifier!
Interpretability
Accuracy
Real-world use case Make
everything
interpretable!
What an explanation looks like
Why did this
happen?
From: Keith Richards
Subject: Christianity is the answer
NTTP-Posting-Host: x.x.com
I think Christianity is the one true religion.
If you’d like to know more, send me a note
Being Model-Agnostic…
“Global” explanation is too complicated
Being Model-Agnostic…
“Global” explanation is too complicated
Being Model-Agnostic…
“Global” explanation is too complicated
Explanation is an interpretable model,
that is locally accurate
Google’s Object Detector
P( ) = 0.21P( ) = 0.24P( ) = 0.32
Only 1 mistake!
Classification: Wolf or a Husky?
Neural Network Explanations
We’ve built a great snow detector…
Understanding Behavior
We’ve built a great snow detector…
Comparing Classifiers
Classifier 1
Classifier 2
Explanations?
Look at Examples?
Deploy and Check?
“I have a gut feeling..”
Accuracy?
Change the model
Different data
Different parameters
Different “features”
…
Comparing Classifiers
Original Image “Bad” Classifier “Good” Classifier
Explanation for a bad classifier
From: Keith Richards
Subject: Christianity is the answer
NTTP-Posting-Host: x.x.com
I think Christianity is the one true religion.
If you’d like to know more, send me a note
After looking at the explanation,
we shouldn’t trust the model!
“Good” Explanation
It seems to be picking
up on more reasonable
things.. good!
Recent Work
Counter-examples and Counter-factuals
Understanding via Predicting
Users “understand” a model if they can
predict its behavior on unseen instances
Precision is much more
important than Coverage!
Precision
How accurate are the users guesses?
If the users guess wrong, they don’t understand
Coverage
How often do the users make confident guesses?
It’s okay not to be able to guess!
It’s much better not to guess than to guess
confidently, but be completely wrong!
Linear Explanations
This movie is not bad. This movie is not very good.
LIMELIME
D
D
…
D
This director is always bad.
This movie is not nice.
This stuff is rather honest.
This star is not bad.
Problem 1: Where is the explanation good?
Explanation is wrong in this region
This explanation is a better approximation
than the other one.
Problem 2: What is the coverage?
Explanation doesn’t apply here
→ Users will make mistakes!
Anchors: Precise Counter-factuals
Anchor: ”not bad” →
This movie is not bad.
This audio is not bad.
This novel is not bad.
This footage is not bad.
D(.|A)
Positive
This movie is not very good.
Anchor: ”not good” →
This poster is not ever good.
This picture is not rarely good.
This actor is not incredibly good.
D(.|A)
Negative
anchor
anchor
LIMELIME
D
D
An anchor is a sufficient condition
Clear (and adaptive) coverage Probabilistic guarantee avoids human mistakes
Salary Prediction
IF Education < High School
Then Predict Salary < 50K
Salary
71%
29% >$50K
<$50K
Visual QA
Encoder/Decoder LSTMs
Encoder/Decoder LSTMs
Encoder/Decoder LSTMs
What’s a Good Explanation?
We want to understand the models
Compact description
Lines, Decision Trees,
Simple Rules, etc.
When we read them,
we imagine instances
where they apply, and
where they don’t
Directly show useful examples?
What examples describe the behavior?
Closest Counter-example:
How can we change this example
to change the prediction?
Adversarial Examples
Goodfellow et al, "Explaining and Harnessing Adversarial Examples", ICLR 2015.
adversary predicted as "2"original MNIST digit "3"
+ .02 x =
adversarial noise
"inputs formed by applying small but
intentionally worst-case perturbations
to examples from the dataset, such
that the perturbed input results in the
model outputting an incorrect answer
with high confidence"
Adversarial Examples: Pros
Advantages:
◦ Applicable to any gradient -based classifier
◦ Useful to evaluate the robustness of the model against adversaries
◦ Small perturbations often lead to imperceivable adversarial examples
Adversarial Examples: Cons
Disadvantages:
◦ Examples are unnatural
◦ may not look anything you would naturally see in the "wild"
◦ Distance is not always meaningful
◦ E.g. color change or translation/rotation of an image
◦ Cannot be used for structured domains like text, code, etc.:
◦ E.g. replacing/removing words results in sentences that are not grammatical
◦ Do not provide insights into why the sample is an adversary
◦ How is the model working?
◦ How to fix the model?
Example: MNIST Digits
Example: Church vs Tower
Machine Translation
Debug Google Translate, remotely!
Explanations are important!
How can we trust the
predictions are correct?
Trust
How can we understand and
predict the behavior?
Predict
How do we improve it to
prevent potential mistakes?
Improve
Model Agnostic
Explanations
Thanks! sameer@uci.edu
sameersingh.org
Model Agnostic
Explanations
Work with Marco T. Ribeiro, Carlos Guestrin, Dheeru Dua, and Zhengli Zhao

More Related Content

PDF
ViT (Vision Transformer) Review [CDM]
Dongmin Choi
 
PDF
Machine Learning Explanations: LIME framework
Deep Learning Italia
 
PDF
Yurii Pashchenko: Unlocking the potential of Segment Anything Model (UA)
Lviv Startup Club
 
PDF
The current state of generative AI
Benjaminlapid1
 
PPTX
Module 31
UllasSS1
 
PDF
generative-ai-fundamentals and Large language models
AdventureWorld5
 
PPTX
eScience SHAP talk
Scott Lundberg
 
PPTX
Inheritance ppt
Nivegeetha
 
ViT (Vision Transformer) Review [CDM]
Dongmin Choi
 
Machine Learning Explanations: LIME framework
Deep Learning Italia
 
Yurii Pashchenko: Unlocking the potential of Segment Anything Model (UA)
Lviv Startup Club
 
The current state of generative AI
Benjaminlapid1
 
Module 31
UllasSS1
 
generative-ai-fundamentals and Large language models
AdventureWorld5
 
eScience SHAP talk
Scott Lundberg
 
Inheritance ppt
Nivegeetha
 

What's hot (20)

PDF
Interpretable Machine Learning Using LIME Framework - Kasia Kulma (PhD), Data...
Sri Ambati
 
PPTX
Difference between Artificial Intelligence, Machine Learning, Deep Learning a...
Sanjay Srivastava
 
PPTX
An Introduction to XAI! Towards Trusting Your ML Models!
Mansour Saffar
 
PPTX
Using SHAP to Understand Black Box Models
Jonathan Bechtel
 
PPTX
What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...
Simplilearn
 
PDF
Explainable AI - making ML and DL models more interpretable
Aditya Bhattacharya
 
PPT
Machine Learning presentation.
butest
 
PPTX
Intro to deep learning
David Voyles
 
PPT
2.17Mb ppt
butest
 
PPTX
Deep neural networks
Si Haem
 
PDF
Explainability and bias in AI
Bill Liu
 
PPTX
Explainable Machine Learning (Explainable ML)
Hayim Makabee
 
PDF
What is Machine Learning | Introduction to Machine Learning | Machine Learnin...
Simplilearn
 
PPTX
Transfer learning-presentation
Bushra Jbawi
 
PDF
Explainable AI (XAI) - A Perspective
Saurabh Kaushik
 
PDF
Explainable AI
Equifax Ltd
 
PDF
Using Large Language Models in 10 Lines of Code
Gautier Marti
 
PPTX
Deep Learning With Neural Networks
Aniket Maurya
 
PPT
Genetic algorithm
garima931
 
PDF
Ai black box
Marcel Blattner, PhD
 
Interpretable Machine Learning Using LIME Framework - Kasia Kulma (PhD), Data...
Sri Ambati
 
Difference between Artificial Intelligence, Machine Learning, Deep Learning a...
Sanjay Srivastava
 
An Introduction to XAI! Towards Trusting Your ML Models!
Mansour Saffar
 
Using SHAP to Understand Black Box Models
Jonathan Bechtel
 
What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...
Simplilearn
 
Explainable AI - making ML and DL models more interpretable
Aditya Bhattacharya
 
Machine Learning presentation.
butest
 
Intro to deep learning
David Voyles
 
2.17Mb ppt
butest
 
Deep neural networks
Si Haem
 
Explainability and bias in AI
Bill Liu
 
Explainable Machine Learning (Explainable ML)
Hayim Makabee
 
What is Machine Learning | Introduction to Machine Learning | Machine Learnin...
Simplilearn
 
Transfer learning-presentation
Bushra Jbawi
 
Explainable AI (XAI) - A Perspective
Saurabh Kaushik
 
Explainable AI
Equifax Ltd
 
Using Large Language Models in 10 Lines of Code
Gautier Marti
 
Deep Learning With Neural Networks
Aniket Maurya
 
Genetic algorithm
garima931
 
Ai black box
Marcel Blattner, PhD
 
Ad

Similar to Explaining Black-Box Machine Learning Predictions - Sameer Singh, Assistant Professor of Computer Science, UC Irvine (20)

PPTX
The Art of Speaking Data.
David Wellman
 
PPTX
Intro to modelling-supervised learning
Justin Sebok
 
PDF
AI in the Real World: Challenges, and Risks and how to handle them?
Srinath Perera
 
PDF
Deep Learning Class #0 - You Can Do It
Holberton School
 
PDF
DL Classe 0 - You can do it
Gregory Renard
 
PPTX
What If? Demystifying AI Decisions with Counterfactuals
davidmartens007
 
PPTX
Large Components in the Rearview Mirror
Michelle Brush
 
PPTX
Explainable AI in Industry (FAT* 2020 Tutorial)
Krishnaram Kenthapadi
 
PDF
Size Of Writing Paper. Writing Paper Sizes Chart. 2019-01-16
Kimberly Gomez
 
PDF
Qualitative Legal Prediction - Prof. Daniel Katz
smahboobani
 
PPTX
Taking Machine Learning from Batch to Real-Time (big data eXposed 2015)
Elad Rosenheim
 
PPTX
HarambeeNet: Data by the people, for the people
Michael Bernstein
 
PDF
Defeating online fraud and abuse – Continuous Intelligence in action
Thoughtworks
 
PPT
Machine Learning ICS 273A
butest
 
PPT
Machine Learning ICS 273A
butest
 
PDF
Knowing Ranking Factors won't be enough!
Mark Orr
 
PDF
IxDA Taiwan 6th slide
Stanley Chang
 
PPTX
Security Snake Oil Cycle 2019
Dave Cole
 
PPTX
Software estimation is crap
Ian Garrison
 
The Art of Speaking Data.
David Wellman
 
Intro to modelling-supervised learning
Justin Sebok
 
AI in the Real World: Challenges, and Risks and how to handle them?
Srinath Perera
 
Deep Learning Class #0 - You Can Do It
Holberton School
 
DL Classe 0 - You can do it
Gregory Renard
 
What If? Demystifying AI Decisions with Counterfactuals
davidmartens007
 
Large Components in the Rearview Mirror
Michelle Brush
 
Explainable AI in Industry (FAT* 2020 Tutorial)
Krishnaram Kenthapadi
 
Size Of Writing Paper. Writing Paper Sizes Chart. 2019-01-16
Kimberly Gomez
 
Qualitative Legal Prediction - Prof. Daniel Katz
smahboobani
 
Taking Machine Learning from Batch to Real-Time (big data eXposed 2015)
Elad Rosenheim
 
HarambeeNet: Data by the people, for the people
Michael Bernstein
 
Defeating online fraud and abuse – Continuous Intelligence in action
Thoughtworks
 
Machine Learning ICS 273A
butest
 
Machine Learning ICS 273A
butest
 
Knowing Ranking Factors won't be enough!
Mark Orr
 
IxDA Taiwan 6th slide
Stanley Chang
 
Security Snake Oil Cycle 2019
Dave Cole
 
Software estimation is crap
Ian Garrison
 
Ad

More from Sri Ambati (20)

PDF
H2O Label Genie Starter Track - Support Presentation
Sri Ambati
 
PDF
H2O.ai Agents : From Theory to Practice - Support Presentation
Sri Ambati
 
PDF
H2O Generative AI Starter Track - Support Presentation Slides.pdf
Sri Ambati
 
PDF
H2O Gen AI Ecosystem Overview - Level 1 - Slide Deck
Sri Ambati
 
PDF
An In-depth Exploration of Enterprise h2oGPTe Slide Deck
Sri Ambati
 
PDF
Intro to Enterprise h2oGPTe Presentation Slides
Sri Ambati
 
PDF
Enterprise h2o GPTe Learning Path Slide Deck
Sri Ambati
 
PDF
H2O Wave Course Starter - Presentation Slides
Sri Ambati
 
PDF
Large Language Models (LLMs) - Level 3 Slides
Sri Ambati
 
PDF
Data Science and Machine Learning Platforms (2024) Slides
Sri Ambati
 
PDF
Data Prep for H2O Driverless AI - Slides
Sri Ambati
 
PDF
H2O Cloud AI Developer Services - Slides (2024)
Sri Ambati
 
PDF
LLM Learning Path Level 2 - Presentation Slides
Sri Ambati
 
PDF
LLM Learning Path Level 1 - Presentation Slides
Sri Ambati
 
PDF
Hydrogen Torch - Starter Course - Presentation Slides
Sri Ambati
 
PDF
Presentation Resources - H2O Gen AI Ecosystem Overview - Level 2
Sri Ambati
 
PDF
H2O Driverless AI Starter Course - Slides and Assignments
Sri Ambati
 
PPTX
GenAISummit 2024 May 28 Sri Ambati Keynote: AGI Belongs to The Community in O...
Sri Ambati
 
PDF
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
Sri Ambati
 
PPTX
Generative AI Masterclass - Model Risk Management.pptx
Sri Ambati
 
H2O Label Genie Starter Track - Support Presentation
Sri Ambati
 
H2O.ai Agents : From Theory to Practice - Support Presentation
Sri Ambati
 
H2O Generative AI Starter Track - Support Presentation Slides.pdf
Sri Ambati
 
H2O Gen AI Ecosystem Overview - Level 1 - Slide Deck
Sri Ambati
 
An In-depth Exploration of Enterprise h2oGPTe Slide Deck
Sri Ambati
 
Intro to Enterprise h2oGPTe Presentation Slides
Sri Ambati
 
Enterprise h2o GPTe Learning Path Slide Deck
Sri Ambati
 
H2O Wave Course Starter - Presentation Slides
Sri Ambati
 
Large Language Models (LLMs) - Level 3 Slides
Sri Ambati
 
Data Science and Machine Learning Platforms (2024) Slides
Sri Ambati
 
Data Prep for H2O Driverless AI - Slides
Sri Ambati
 
H2O Cloud AI Developer Services - Slides (2024)
Sri Ambati
 
LLM Learning Path Level 2 - Presentation Slides
Sri Ambati
 
LLM Learning Path Level 1 - Presentation Slides
Sri Ambati
 
Hydrogen Torch - Starter Course - Presentation Slides
Sri Ambati
 
Presentation Resources - H2O Gen AI Ecosystem Overview - Level 2
Sri Ambati
 
H2O Driverless AI Starter Course - Slides and Assignments
Sri Ambati
 
GenAISummit 2024 May 28 Sri Ambati Keynote: AGI Belongs to The Community in O...
Sri Ambati
 
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
Sri Ambati
 
Generative AI Masterclass - Model Risk Management.pptx
Sri Ambati
 

Recently uploaded (20)

PDF
Accelerating Oracle Database 23ai Troubleshooting with Oracle AHF Fleet Insig...
Sandesh Rao
 
PPTX
The-Ethical-Hackers-Imperative-Safeguarding-the-Digital-Frontier.pptx
sujalchauhan1305
 
PDF
Oracle AI Vector Search- Getting Started and what's new in 2025- AIOUG Yatra ...
Sandesh Rao
 
PDF
How Open Source Changed My Career by abdelrahman ismail
a0m0rajab1
 
PDF
Using Anchore and DefectDojo to Stand Up Your DevSecOps Function
Anchore
 
PDF
MASTERDECK GRAPHSUMMIT SYDNEY (Public).pdf
Neo4j
 
PDF
Security features in Dell, HP, and Lenovo PC systems: A research-based compar...
Principled Technologies
 
PPTX
Introduction to Flutter by Ayush Desai.pptx
ayushdesai204
 
PDF
Doc9.....................................
SofiaCollazos
 
PDF
OFFOFFBOX™ – A New Era for African Film | Startup Presentation
ambaicciwalkerbrian
 
PDF
Make GenAI investments go further with the Dell AI Factory
Principled Technologies
 
PPTX
cloud computing vai.pptx for the project
vaibhavdobariyal79
 
PDF
Brief History of Internet - Early Days of Internet
sutharharshit158
 
PDF
The Future of Mobile Is Context-Aware—Are You Ready?
iProgrammer Solutions Private Limited
 
PPTX
OA presentation.pptx OA presentation.pptx
pateldhruv002338
 
PDF
Trying to figure out MCP by actually building an app from scratch with open s...
Julien SIMON
 
PDF
AI-Cloud-Business-Management-Platforms-The-Key-to-Efficiency-Growth.pdf
Artjoker Software Development Company
 
PPTX
AI and Robotics for Human Well-being.pptx
JAYMIN SUTHAR
 
PPTX
IT Runs Better with ThousandEyes AI-driven Assurance
ThousandEyes
 
PDF
Research-Fundamentals-and-Topic-Development.pdf
ayesha butalia
 
Accelerating Oracle Database 23ai Troubleshooting with Oracle AHF Fleet Insig...
Sandesh Rao
 
The-Ethical-Hackers-Imperative-Safeguarding-the-Digital-Frontier.pptx
sujalchauhan1305
 
Oracle AI Vector Search- Getting Started and what's new in 2025- AIOUG Yatra ...
Sandesh Rao
 
How Open Source Changed My Career by abdelrahman ismail
a0m0rajab1
 
Using Anchore and DefectDojo to Stand Up Your DevSecOps Function
Anchore
 
MASTERDECK GRAPHSUMMIT SYDNEY (Public).pdf
Neo4j
 
Security features in Dell, HP, and Lenovo PC systems: A research-based compar...
Principled Technologies
 
Introduction to Flutter by Ayush Desai.pptx
ayushdesai204
 
Doc9.....................................
SofiaCollazos
 
OFFOFFBOX™ – A New Era for African Film | Startup Presentation
ambaicciwalkerbrian
 
Make GenAI investments go further with the Dell AI Factory
Principled Technologies
 
cloud computing vai.pptx for the project
vaibhavdobariyal79
 
Brief History of Internet - Early Days of Internet
sutharharshit158
 
The Future of Mobile Is Context-Aware—Are You Ready?
iProgrammer Solutions Private Limited
 
OA presentation.pptx OA presentation.pptx
pateldhruv002338
 
Trying to figure out MCP by actually building an app from scratch with open s...
Julien SIMON
 
AI-Cloud-Business-Management-Platforms-The-Key-to-Efficiency-Growth.pdf
Artjoker Software Development Company
 
AI and Robotics for Human Well-being.pptx
JAYMIN SUTHAR
 
IT Runs Better with ThousandEyes AI-driven Assurance
ThousandEyes
 
Research-Fundamentals-and-Topic-Development.pdf
ayesha butalia
 

Explaining Black-Box Machine Learning Predictions - Sameer Singh, Assistant Professor of Computer Science, UC Irvine

Editor's Notes

  • #16: X2 = mX1 + c X1=0.5, m/2 + c = 0, m = -2c X2 = 10, c=10, m=-20, Y = 20X1 -X2 + 10
  • #20: Applications: Biology, Linguistics, Journalism, Econometrics, Humanities, Medicine, … Data types: Time series, Text, Images, Audio, Video, Graphs, User Histories, …
  • #23: Instead of tens.. Images: hundreds Text: thousands Video: millions … and so on