SlideShare a Scribd company logo
4
Most read
10
Most read
12
Most read
Machine Learning Interpretability
Marcel Spitzer Munich, 20.11.2018
2
Marcel Spitzer
Big Data Scientist @ inovex
● Applied Mathematics, Data Science
● SW Engineering, Data Products
● Big Data, Hadoop, Spark
mspitzer@inovex.de
@mspitzer243
3
4
Interpretation
is the process of
giving explanations
to humans.
~ Kim B., Google Brain, Interpretable
Machine Learning (ICML 2017)
https://people.csail.mit.edu/beenkim/papers/BeenK_FinaleDV_ICML2017_tutorial.pdf
5
“Interpretability is the degree to which an observer
can understand the cause of a decision.”
~ Miller T., 2017, Explanation in AI: Insights from the Social Sciences
➢ humans create decision systems
➢ humans are affected by decisions
➢ humans demand for explanations
https://arxiv.org/pdf/1706.07269.pdf
6
NIPS 2016 workshop on Interpretable Machine
Learning for Complex Systems
ICML 2016 Workshop on Human
Interpretability in Machine Learning
NIPS 2017 Workshop on Interpreting,
Explaining and Visualizing Deep Learning
NIPS 2017 symposium and workshop:
interpretable and Bayesian machine learning
ICML 2017 Workshop on Human
Interpretability in Machine Learning
ICML 2018 Workshop on Human
Interpretability in Machine Learning
https://people.csail.mit.edu/beenkim/papers/BeenK_FinaleDV_ICML2017_tutorial.pdf
7https://arxiv.org/pdf/1606.03490.pdf
The additional need for interpretability
8
safety system should provide sound decisions
curiosity understand something unexpected
debugging behaviour should be predictable
optimality optimize for true objectives
Why do we need interpretability?
https://people.csail.mit.edu/beenkim/papers/BeenK_FinaleDV_ICML2017_tutorial.pdf
9
When we may not need interpretability
https://christophm.github.io/interpretable-ml-book/interpretability-importance.html
low risk no significant consequences
awareness problem is well-studied
vulnerability prevent people from gaming the system
10
Use models that are intrinsically interpretable and
known to be easy for humans to understand.
Train a black box model and apply post-hoc
interpretability techniques to provide explanations.
1
2
11
Post-hoc interpretability techniques
Global Local
Model-specific
Model Internals,
Intrinsic Feature Importance
Rule Sets (Tree Structure)
Model-agnostic
Partial Dependence Plots,
Feature Importance (perm-based),
Global Surrogate Models
Individual Conditional Expectation,
Local Surrogate Models
12
Post-hoc interpretability techniques
Global Local
Model-specific
Model Internals,
Intrinsic Feature Importance
Rule Sets (Tree Structure)
Model-agnostic
Partial Dependence Plots,
Feature Importance (perm-based),
Global Surrogate Models
Individual Conditional Expectation,
Local Surrogate Models
➢ shows dependence of the response
on a feature per instance
➢ single curve results from varying a
certain feature for a given instance
➢ inconsistent pattern indicates
multicollinearity
13
Individual Conditional Expectation (ICE)
https://christophm.github.io/interpretable-ml-book/pdp.html
14
Partial Dependence Plots (PDP)
➢ PDP curve is the result of averaging
ICE curves
➢ very intuitive, easy to understand
➢ assumption of independence is a
strong drawback
https://christophm.github.io/interpretable-ml-book/pdp.html
➢ averages degradation measured by a certain loss function after repeatedly
permuting single features
➢ feature is important if the error significantly increases after a shuffle
15
Feature Shuffling
https://amunategui.github.io/variable-importance-shuffler/
➢ highly compressed, global insight
➢ tied to some loss function
➢ practically infeasible in high
dimensional domains (e.g.
image/speech recognition, NLP)
16
Feature Shuffling
https://christophm.github.io/interpretable-ml-book/feature-importance.html https://scikit-plot.readthedocs.io/en/stable/estimators.html#scikitplot.estimators.plot_feature_importances
17
Global Surrogate Models
https://www.oreilly.com/ideas/ideas-on-interpreting-machine-learning
➢ feeds original model with small
variations of instance to be explained
➢ sampled instances are weighted by
proximity to the instance of interest
➢ interpretable models are fit locally on
observed outcome
18
Local Surrogate Models: LIME
https://christophm.github.io/interpretable-ml-book/lime.html
19
Local Surrogate Models: LIME
https://www.oreilly.com/learning/introduction-to-local-interpretable-model-agnostic-explanations-lime
20
Recommendations for interpretability techniques
➢ Who is the recipient?
○ Lay-Men → rather intuitive, example-based local explanations
○ Analysts → global surrogates, perm-based feature importance
○ Authorities → intrinsically interpretable models
➢ What are the explanations used for?
○ Debug/Improve → PDP & ICE curves
○ Decision support → rule-based explanations
○ Auditing/Legal → intrinsically interpretable models
➢ Molnar C., 2018, Interpretable Machine Learning - A Guide for Making
Black Box Models Explainable
➢ Gill N., Hall P., 2018, An Introduction to Machine Learning Interpretability
➢ Zhao Q., Hastie T., 2017, Causal Interpretations of Black-Box Models
➢ Kim B., Doshi-Velez F., 2017, Interpretable Machine Learning: The fuss,
the concrete and the questions
➢ Ribeiro, M.T., Singh, S. and Guestrin, C., 2016, August. Why should i trust
you? Explaining the predictions of any classifier
21
Resources
Vielen Dank
Marcel Spitzer
Big Data Scientist
mspitzer@inovex.de
inovex GmbH
Schanzenstraße 6-20
Kupferhütte 1.13
51063 Köln

More Related Content

PDF
Interpretable Machine Learning Using LIME Framework - Kasia Kulma (PhD), Data...
PDF
DC02. Interpretation of predictions
PDF
Explainable AI - making ML and DL models more interpretable
PPTX
Explainable AI in Healthcare
PPTX
Introduction to Interpretable Machine Learning
PDF
Explainable AI (XAI) - A Perspective
PPTX
Explainable AI in Industry (KDD 2019 Tutorial)
PDF
Machine Learning Interpretability / Explainability
Interpretable Machine Learning Using LIME Framework - Kasia Kulma (PhD), Data...
DC02. Interpretation of predictions
Explainable AI - making ML and DL models more interpretable
Explainable AI in Healthcare
Introduction to Interpretable Machine Learning
Explainable AI (XAI) - A Perspective
Explainable AI in Industry (KDD 2019 Tutorial)
Machine Learning Interpretability / Explainability

What's hot (20)

PDF
Machine Learning Explanations: LIME framework
PPTX
PDF
Interpretable machine learning : Methods for understanding complex models
PDF
Explainability and bias in AI
PDF
Explainable AI (XAI)
PPTX
Explainable Machine Learning (Explainable ML)
PDF
Unified Approach to Interpret Machine Learning Model: SHAP + LIME
PPTX
Explainable AI
PPTX
Explaining Black-Box Machine Learning Predictions - Sameer Singh, Assistant P...
PDF
Intepretability / Explainable AI for Deep Neural Networks
PPTX
Interpretable machine learning
PPTX
An Introduction to XAI! Towards Trusting Your ML Models!
PDF
Introduction to XGBoost
PPTX
Explainability for Natural Language Processing
PDF
Visualizing Data Using t-SNE
PPTX
Overfitting & Underfitting
PPTX
Explainable AI
PPTX
Decision Tree Algorithm With Example | Decision Tree In Machine Learning | Da...
PDF
Linear regression
PPTX
Scott Lundberg, Microsoft Research - Explainable Machine Learning with Shaple...
Machine Learning Explanations: LIME framework
Interpretable machine learning : Methods for understanding complex models
Explainability and bias in AI
Explainable AI (XAI)
Explainable Machine Learning (Explainable ML)
Unified Approach to Interpret Machine Learning Model: SHAP + LIME
Explainable AI
Explaining Black-Box Machine Learning Predictions - Sameer Singh, Assistant P...
Intepretability / Explainable AI for Deep Neural Networks
Interpretable machine learning
An Introduction to XAI! Towards Trusting Your ML Models!
Introduction to XGBoost
Explainability for Natural Language Processing
Visualizing Data Using t-SNE
Overfitting & Underfitting
Explainable AI
Decision Tree Algorithm With Example | Decision Tree In Machine Learning | Da...
Linear regression
Scott Lundberg, Microsoft Research - Explainable Machine Learning with Shaple...
Ad

Similar to Machine Learning Interpretability (20)

PDF
Interpretable Machine Learning
PDF
Notes on data-intensive processing with Hadoop Mapreduce
PDF
Hadoop.mapreduce
PPTX
Brownfield Domain Driven Design
PDF
How Will AI Change the Role of the Data Scientist?
PDF
Lecture 05: Recurrent Neural Networks / Deep Learning by Pankaj Gupta
PDF
The Hitchhiker's Guide to Machine Learning with Python & Apache Spark
DOCX
CSc investigatory project
PDF
Recommender Trends 2014
PDF
Deprecating the state machine: building conversational AI with the Rasa stack...
PDF
Deprecating the state machine: building conversational AI with the Rasa stack
PPTX
Practical Tips for Interpreting Machine Learning Models - Patrick Hall, H2O.ai
PDF
Keynote at-icpc-2020
PDF
10 Things Every PHP Developer Should Know About Machine Learning
PDF
AI & The Virtuous Cycle of Compute
PDF
Our Concurrent Past; Our Distributed Future
PDF
Codemotion Berlin 2015 recap
PDF
Big Media: Multimedia goes Big Data
PPTX
4 D Computing: Life comes at us polydimensionally
PPTX
Synergy of Human and Artificial Intelligence in Software Engineering
Interpretable Machine Learning
Notes on data-intensive processing with Hadoop Mapreduce
Hadoop.mapreduce
Brownfield Domain Driven Design
How Will AI Change the Role of the Data Scientist?
Lecture 05: Recurrent Neural Networks / Deep Learning by Pankaj Gupta
The Hitchhiker's Guide to Machine Learning with Python & Apache Spark
CSc investigatory project
Recommender Trends 2014
Deprecating the state machine: building conversational AI with the Rasa stack...
Deprecating the state machine: building conversational AI with the Rasa stack
Practical Tips for Interpreting Machine Learning Models - Patrick Hall, H2O.ai
Keynote at-icpc-2020
10 Things Every PHP Developer Should Know About Machine Learning
AI & The Virtuous Cycle of Compute
Our Concurrent Past; Our Distributed Future
Codemotion Berlin 2015 recap
Big Media: Multimedia goes Big Data
4 D Computing: Life comes at us polydimensionally
Synergy of Human and Artificial Intelligence in Software Engineering
Ad

More from inovex GmbH (20)

PDF
lldb – Debugger auf Abwegen
PDF
Are you sure about that?! Uncertainty Quantification in AI
PDF
Why natural language is next step in the AI evolution
PDF
WWDC 2019 Recap
PDF
Network Policies
PDF
Jenkins X – CI/CD in wolkigen Umgebungen
PDF
AI auf Edge-Geraeten
PDF
Prometheus on Kubernetes
PDF
Deep Learning for Recommender Systems
PDF
Azure IoT Edge
PDF
Representation Learning von Zeitreihen
PDF
Talk to me – Chatbots und digitale Assistenten
PDF
Künstlich intelligent?
PDF
Dev + Ops = Go
PDF
Das Android Open Source Project
PDF
Performance evaluation of GANs in a semisupervised OCR use case
PDF
People & Products – Lessons learned from the daily IT madness
PDF
Infrastructure as (real) Code – Manage your K8s resources with Pulumi
PDF
Remote First – Der Arbeitsplatz in der Cloud
PDF
Data Science und Machine Learning im Kubernetes-Ökosystem
lldb – Debugger auf Abwegen
Are you sure about that?! Uncertainty Quantification in AI
Why natural language is next step in the AI evolution
WWDC 2019 Recap
Network Policies
Jenkins X – CI/CD in wolkigen Umgebungen
AI auf Edge-Geraeten
Prometheus on Kubernetes
Deep Learning for Recommender Systems
Azure IoT Edge
Representation Learning von Zeitreihen
Talk to me – Chatbots und digitale Assistenten
Künstlich intelligent?
Dev + Ops = Go
Das Android Open Source Project
Performance evaluation of GANs in a semisupervised OCR use case
People & Products – Lessons learned from the daily IT madness
Infrastructure as (real) Code – Manage your K8s resources with Pulumi
Remote First – Der Arbeitsplatz in der Cloud
Data Science und Machine Learning im Kubernetes-Ökosystem

Recently uploaded (20)

PDF
How Creative Agencies Leverage Project Management Software.pdf
PPTX
Essential Infomation Tech presentation.pptx
PPTX
ai tools demonstartion for schools and inter college
PDF
T3DD25 TYPO3 Content Blocks - Deep Dive by André Kraus
PPTX
Agentic AI Use Case- Contract Lifecycle Management (CLM).pptx
PDF
Upgrade and Innovation Strategies for SAP ERP Customers
PDF
2025 Textile ERP Trends: SAP, Odoo & Oracle
PDF
Internet Downloader Manager (IDM) Crack 6.42 Build 42 Updates Latest 2025
PDF
PTS Company Brochure 2025 (1).pdf.......
PDF
SAP S4 Hana Brochure 3 (PTS SYSTEMS AND SOLUTIONS)
PPTX
Oracle E-Business Suite: A Comprehensive Guide for Modern Enterprises
PPTX
Transform Your Business with a Software ERP System
PDF
How to Migrate SBCGlobal Email to Yahoo Easily
PPTX
Reimagine Home Health with the Power of Agentic AI​
PPTX
history of c programming in notes for students .pptx
PDF
Softaken Excel to vCard Converter Software.pdf
PDF
EN-Survey-Report-SAP-LeanIX-EA-Insights-2025.pdf
PDF
Design an Analysis of Algorithms I-SECS-1021-03
PDF
Flood Susceptibility Mapping Using Image-Based 2D-CNN Deep Learnin. Overview ...
PDF
Adobe Illustrator 28.6 Crack My Vision of Vector Design
How Creative Agencies Leverage Project Management Software.pdf
Essential Infomation Tech presentation.pptx
ai tools demonstartion for schools and inter college
T3DD25 TYPO3 Content Blocks - Deep Dive by André Kraus
Agentic AI Use Case- Contract Lifecycle Management (CLM).pptx
Upgrade and Innovation Strategies for SAP ERP Customers
2025 Textile ERP Trends: SAP, Odoo & Oracle
Internet Downloader Manager (IDM) Crack 6.42 Build 42 Updates Latest 2025
PTS Company Brochure 2025 (1).pdf.......
SAP S4 Hana Brochure 3 (PTS SYSTEMS AND SOLUTIONS)
Oracle E-Business Suite: A Comprehensive Guide for Modern Enterprises
Transform Your Business with a Software ERP System
How to Migrate SBCGlobal Email to Yahoo Easily
Reimagine Home Health with the Power of Agentic AI​
history of c programming in notes for students .pptx
Softaken Excel to vCard Converter Software.pdf
EN-Survey-Report-SAP-LeanIX-EA-Insights-2025.pdf
Design an Analysis of Algorithms I-SECS-1021-03
Flood Susceptibility Mapping Using Image-Based 2D-CNN Deep Learnin. Overview ...
Adobe Illustrator 28.6 Crack My Vision of Vector Design

Machine Learning Interpretability

  • 1. Machine Learning Interpretability Marcel Spitzer Munich, 20.11.2018
  • 2. 2 Marcel Spitzer Big Data Scientist @ inovex ● Applied Mathematics, Data Science ● SW Engineering, Data Products ● Big Data, Hadoop, Spark mspitzer@inovex.de @mspitzer243
  • 3. 3
  • 4. 4 Interpretation is the process of giving explanations to humans. ~ Kim B., Google Brain, Interpretable Machine Learning (ICML 2017) https://people.csail.mit.edu/beenkim/papers/BeenK_FinaleDV_ICML2017_tutorial.pdf
  • 5. 5 “Interpretability is the degree to which an observer can understand the cause of a decision.” ~ Miller T., 2017, Explanation in AI: Insights from the Social Sciences ➢ humans create decision systems ➢ humans are affected by decisions ➢ humans demand for explanations https://arxiv.org/pdf/1706.07269.pdf
  • 6. 6 NIPS 2016 workshop on Interpretable Machine Learning for Complex Systems ICML 2016 Workshop on Human Interpretability in Machine Learning NIPS 2017 Workshop on Interpreting, Explaining and Visualizing Deep Learning NIPS 2017 symposium and workshop: interpretable and Bayesian machine learning ICML 2017 Workshop on Human Interpretability in Machine Learning ICML 2018 Workshop on Human Interpretability in Machine Learning https://people.csail.mit.edu/beenkim/papers/BeenK_FinaleDV_ICML2017_tutorial.pdf
  • 8. 8 safety system should provide sound decisions curiosity understand something unexpected debugging behaviour should be predictable optimality optimize for true objectives Why do we need interpretability? https://people.csail.mit.edu/beenkim/papers/BeenK_FinaleDV_ICML2017_tutorial.pdf
  • 9. 9 When we may not need interpretability https://christophm.github.io/interpretable-ml-book/interpretability-importance.html low risk no significant consequences awareness problem is well-studied vulnerability prevent people from gaming the system
  • 10. 10 Use models that are intrinsically interpretable and known to be easy for humans to understand. Train a black box model and apply post-hoc interpretability techniques to provide explanations. 1 2
  • 11. 11 Post-hoc interpretability techniques Global Local Model-specific Model Internals, Intrinsic Feature Importance Rule Sets (Tree Structure) Model-agnostic Partial Dependence Plots, Feature Importance (perm-based), Global Surrogate Models Individual Conditional Expectation, Local Surrogate Models
  • 12. 12 Post-hoc interpretability techniques Global Local Model-specific Model Internals, Intrinsic Feature Importance Rule Sets (Tree Structure) Model-agnostic Partial Dependence Plots, Feature Importance (perm-based), Global Surrogate Models Individual Conditional Expectation, Local Surrogate Models
  • 13. ➢ shows dependence of the response on a feature per instance ➢ single curve results from varying a certain feature for a given instance ➢ inconsistent pattern indicates multicollinearity 13 Individual Conditional Expectation (ICE) https://christophm.github.io/interpretable-ml-book/pdp.html
  • 14. 14 Partial Dependence Plots (PDP) ➢ PDP curve is the result of averaging ICE curves ➢ very intuitive, easy to understand ➢ assumption of independence is a strong drawback https://christophm.github.io/interpretable-ml-book/pdp.html
  • 15. ➢ averages degradation measured by a certain loss function after repeatedly permuting single features ➢ feature is important if the error significantly increases after a shuffle 15 Feature Shuffling https://amunategui.github.io/variable-importance-shuffler/
  • 16. ➢ highly compressed, global insight ➢ tied to some loss function ➢ practically infeasible in high dimensional domains (e.g. image/speech recognition, NLP) 16 Feature Shuffling https://christophm.github.io/interpretable-ml-book/feature-importance.html https://scikit-plot.readthedocs.io/en/stable/estimators.html#scikitplot.estimators.plot_feature_importances
  • 18. ➢ feeds original model with small variations of instance to be explained ➢ sampled instances are weighted by proximity to the instance of interest ➢ interpretable models are fit locally on observed outcome 18 Local Surrogate Models: LIME https://christophm.github.io/interpretable-ml-book/lime.html
  • 19. 19 Local Surrogate Models: LIME https://www.oreilly.com/learning/introduction-to-local-interpretable-model-agnostic-explanations-lime
  • 20. 20 Recommendations for interpretability techniques ➢ Who is the recipient? ○ Lay-Men → rather intuitive, example-based local explanations ○ Analysts → global surrogates, perm-based feature importance ○ Authorities → intrinsically interpretable models ➢ What are the explanations used for? ○ Debug/Improve → PDP & ICE curves ○ Decision support → rule-based explanations ○ Auditing/Legal → intrinsically interpretable models
  • 21. ➢ Molnar C., 2018, Interpretable Machine Learning - A Guide for Making Black Box Models Explainable ➢ Gill N., Hall P., 2018, An Introduction to Machine Learning Interpretability ➢ Zhao Q., Hastie T., 2017, Causal Interpretations of Black-Box Models ➢ Kim B., Doshi-Velez F., 2017, Interpretable Machine Learning: The fuss, the concrete and the questions ➢ Ribeiro, M.T., Singh, S. and Guestrin, C., 2016, August. Why should i trust you? Explaining the predictions of any classifier 21 Resources
  • 22. Vielen Dank Marcel Spitzer Big Data Scientist mspitzer@inovex.de inovex GmbH Schanzenstraße 6-20 Kupferhütte 1.13 51063 Köln