SlideShare une entreprise Scribd logo
1  sur  19
Machine Learning
Interpretability
The good, the bad, and the ugly
Acknowledgments
Makers: Navdeep Gill, Mark Chan, Doug Deloy, Megan Kurka, Michal Kurka, Wen Phan, Sri Satish
Ambati, Lingyao Meng, Mathias Müller
Advisors: Leland Wilkinson, Trevor Hastie, Rob Tibshirani
Community: O’Reilly Strata and AI (accepted talks and tutorials), FAT/ML Conference (accepted
tutorial), ASA Symposium on Data Science and Statistics (invited talk), Joint Statistical Meetings
(accepted talk)
What is Machine Learning Interpretability?
“The ability to explain or to present in understandable terms to a human.”
-- Finale Doshi-Velez and Been Kim. “Towards a rigorous science of interpretable
machine learning.” arXiv preprint. 2017. https://arxiv.org/pdf/1702.08608.pdf
FAT*: https://www.fatml.org/resources/principles-for-accountable-algorithms
XAI: https://www.darpa.mil/program/explainable-artificial-intelligence
Why Should You Care About Machine Learning Interpretability?
“The now-contemplated field of data science amounts to a superset of the fields of statistics and
machine learning, which adds some technology for “scaling up” to “big data.” This chosen superset
is motivated by commercial rather than intellectual developments. Choosing in this way is likely
to miss out on the really important intellectual event of the next 50 years.”
-- David Donoho. “50 years of Data Science.” Tukey Centennial Workshop, 2015. http://bit.ly/2GQOh1J
Social Motivation: Interpretability plays a critical role in the increased convenience, automation, and organization
in our day-to-day lives promised by AI.
Commercial Motivation: Interpretability is required for regulated industry to adopt machine learning.
• Check and balance against accidental or intentional discrimination.
• “Right to explanation.”
• Hacking and adversarial attacks.
• Improved revenue, i.e. Equifax NeuroDecision: https://www.youtube.com/watch?v=9Z_GW9WDS2c
Why is Machine Learning Interpretability Difficult?
https://github.com/jphall663/GWU_data_mining/blob/master/05_neural_networks/notes/cnn-gwu.pdf
Machine learning algorithms intrinsically
consider high-degree interactions between
input features.
Disaggregating such functions into reason
codes based on single input features is
difficult.
Neural Network
Linear Models Machine Learning
For a given well-understood dataset there is usually one
best model.
For a given well-understood dataset there are usually many
good models. This is often referred to as “the multiplicity of
good models.”
-- Leo Breiman. “Statistical modeling: The two cultures (with comments
and a rejoinder by the author).” Statistical Science. 2001.
http://bit.ly/2pwz6m5
Why is Machine Learning Interpretability Difficult?
What is the Value Proposition of Machine Learning Interpretability?
Linear Models
Machine Learning
Exact explanations for
approximate models.
Approximate explanations
for exact models.
How Can Machine Learning Interpretability Be Practiced?
By seeing and understanding relationships and structures in training, test, and new data.
2-D Projections Correlation Graphs
https://github.com/jphall663/corr_graphhttps://www.cs.toronto.edu/~hinton/science.pdf
How Can Machine Learning Interpretability Be Practiced?
By training interpretable (“white-box”) models.
Decision Trees
• References:
• Breiman, L., J. Friedman, C. J. Stone, and R. A. Olshen. Classification and regression
trees. CRC press, 1984.
• The Elements of Statistical Learning (ESL):
https://web.stanford.edu/~hastie/ElemStatLearn/printings/ESLII_print12.pdf
• OSS:
• rpart: https://cran.r-project.org/web/packages/rpart/index.html
• scikit-learn (various functions): https://github.com/scikit-learn/scikit-learn
Monotonic Gradient Boosted Machines (GBMs)
• Reference: XGBoost Documentation:
http://xgboost.readthedocs.io/en/latest/tutorials/monotonic.html
• OSS: XGBoost: https://github.com/dmlc/xgboost
Logistic, elastic net, GAM, and quantile regression
• References:
• ESL
• Koenker, R. Quantile regression (No. 38). Cambridge University Press, 2005.
• OSS:
• gam: https://cran.r-project.org/web/packages/gam/index.html
• glmnet: https://cran.r-project.org/web/packages/glmnet/index.html
• h2o: https://github.com/h2oai/h2o-3
• quantreg: https://cran.r-project.org/web/packages/quantreg/index.html
• scikit-learn (various functions): https://github.com/scikit-learn/scikit-learn
Rule-based models
• Reference: An Introduction to Data Mining, Chapter 6:
https://www-users.cs.umn.edu/~kumar001/dmbook/ch6.pdf
• OSS:
• RuleFit:
http://statweb.stanford.edu/~jhf/R_RuleFit.html
• arules:
https://cran.r-project.org/web/packages/arules/index.html
• FP-Growth:
http://spark.apache.org/docs/2.2.0/mllib-frequent-pattern-mining.html
Supersparse Linear Integer Models (SLIMs)
• Reference: Supersparse Linear Integer Models for Optimized Medical
Scoring Systems:
https://link.springer.com/content/pdf/10.1007%2Fs10994-015-5528-6.pdf
How Can Machine Learning Interpretability Be Practiced?
With complimentary 2-D visualizations of trained models that enable understanding of learned high-degree
interactions.
Decision Tree Surrogate Models
Partial Dependence and Individual Conditional Expectation
https://github.com/jphall663/interpretable_machine_learning_with_python
How Can Machine Learning Interpretability Be Practiced?
By calculating approximate local feature importance values and ranking them to create reason codes for every
prediction.
Treeinterpreter:
https://github.com/TeamHG-Memex/eli5
LIME:
https://github.com/marcotcr/lime
shap:
https://github.com/slundberg/shap
How Can Machine Learning Interpretability Be Practiced?
By calculating approximate global feature importance to understand how each feature impacts the model in general.
Global feature importance indicates the impact of a feature on the model for the entire training data set.
How Can Machine Learning Interpretability Be Practiced?
By using sensitivity analysis to test machine learning model predictions for accuracy and stability. If you are using a
machine learning model, you should probably be conducting sensitivity analysis.
http://www.vias.org/tmdatanaleng/
Can Machine Learning Interpretability Be Tested?
Yes. By humans or …
Simulated data
You can use simulated data with known characteristics to test explanations. For instance, models trained on totally random data with
no relationship between a number of input variables and a prediction target should not give strong weight to any input variable nor
generate compelling local explanations or reason codes. Conversely, you can use simulated data with a known signal generating
function to test that explanations accurately represent that known function.
https://github.com/h2oai/mli-resources/tree/master/lime_shap_treeint_compare
Explanation stability under data perturbation
Trustworthy explanations likely should not change drastically for minor changes in input data. You can set and test thresholds for
allowable explanation value changes automatically by perturbing input data. (Explanations or reason code values can also be averaged
across a number of models to create more stable explanations.)
Explanation stability with increased prediction accuracy
If previously known, accurate explanations or reason codes from a simpler linear model are available, you can use them as a reference
for the accuracy of explanations from a related, but more complex and hopefully more accurate, model. You can perform tests to see
how accurate a model can become before its prediction’s reason codes veer away from known standards.
General Recommendations
• Consider deployment.
• A very direct path to interpretable machine learning today is to train a monotonic GBM with XGBoost and to use
Shapley explanations either in XGBoost or with the shap Python package.
(Or just buy H2O Driverless AI https://www.h2o.ai/driverless-ai/ ;)
• Use a combination of local and global explanatory techniques.
• Conduct sensitivity analysis and random data attacks on all machine learning models.
• Test your explanatory software.
• If possible, use model-specific explanatory techniques to generate reason codes.
• Open source explanation packages seem immature.
• Beware of uninterpretable features.
LIME Recommendations/Observations
• LIME can give an indication of its own trustworthiness using fit statistics.
• LIME can fail, particularly in the presence of extreme nonlinearity or high-degree interactions.
• LIME is difficult to deploy, but there are highly deployable variants, e.g. H2O’s K-LIME.
• Reason codes are offsets from a local intercept.
• Note that the intercept in LIME can account for the most important local phenomena.
• Generated LIME samples can contain large proportions of out-of-range data that can lead to
unrealistically high or low intercept values.
• Try LIME on discretized input features and on manually constructed interactions.
• Use cross-validation to construct standard deviations or even confidence intervals for reason code
values.
Treeinterpreter and Shapley Recommendations/Observations
• Treeinterpreter and Shapley explanations do not give an indication of their own trustworthiness. We can
only assume they are trustworthy …
• Treeinterpreter appears to fail when used with regularized (L1/L2) XGBoost models.
• Due to theoretical support and robust implementation, Shapely explanations may be suitable for
regulated applications.
• Reason codes are offsets from a global intercept.
Questions?
References and resources:
https://github.com/h2oai/mli-resources
Machine Learning Interpretability: The good, the bad, and the ugly

Contenu connexe

Tendances

What Is Machine Learning? | What Is Machine Learning And How Does It Work? | ...
What Is Machine Learning? | What Is Machine Learning And How Does It Work? | ...What Is Machine Learning? | What Is Machine Learning And How Does It Work? | ...
What Is Machine Learning? | What Is Machine Learning And How Does It Work? | ...Simplilearn
 
Explainable AI in Industry (WWW 2020 Tutorial)
Explainable AI in Industry (WWW 2020 Tutorial)Explainable AI in Industry (WWW 2020 Tutorial)
Explainable AI in Industry (WWW 2020 Tutorial)Krishnaram Kenthapadi
 
Explainable AI: Building trustworthy AI models?
Explainable AI: Building trustworthy AI models? Explainable AI: Building trustworthy AI models?
Explainable AI: Building trustworthy AI models? Raheel Ahmad
 
GANの概要とDCGANのアーキテクチャ/アルゴリズム
GANの概要とDCGANのアーキテクチャ/アルゴリズムGANの概要とDCGANのアーキテクチャ/アルゴリズム
GANの概要とDCGANのアーキテクチャ/アルゴリズムHirosaji
 
DC02. Interpretation of predictions
DC02. Interpretation of predictionsDC02. Interpretation of predictions
DC02. Interpretation of predictionsAnton Kulesh
 
Deep neural networks and tabular data
Deep neural networks and tabular dataDeep neural networks and tabular data
Deep neural networks and tabular dataJimmyLiang20
 
Deep Learning - The Past, Present and Future of Artificial Intelligence
Deep Learning - The Past, Present and Future of Artificial IntelligenceDeep Learning - The Past, Present and Future of Artificial Intelligence
Deep Learning - The Past, Present and Future of Artificial IntelligenceLukas Masuch
 
Interpretable machine learning
Interpretable machine learningInterpretable machine learning
Interpretable machine learningSri Ambati
 
Explainable AI - making ML and DL models more interpretable
Explainable AI - making ML and DL models more interpretableExplainable AI - making ML and DL models more interpretable
Explainable AI - making ML and DL models more interpretableAditya Bhattacharya
 
Explainable AI in Industry (AAAI 2020 Tutorial)
Explainable AI in Industry (AAAI 2020 Tutorial)Explainable AI in Industry (AAAI 2020 Tutorial)
Explainable AI in Industry (AAAI 2020 Tutorial)Krishnaram Kenthapadi
 
Introduction to machine learning
Introduction to machine learningIntroduction to machine learning
Introduction to machine learningKoundinya Desiraju
 
Algorithmic Impact Assessment: Fairness, Robustness and Explainability in Aut...
Algorithmic Impact Assessment: Fairness, Robustness and Explainability in Aut...Algorithmic Impact Assessment: Fairness, Robustness and Explainability in Aut...
Algorithmic Impact Assessment: Fairness, Robustness and Explainability in Aut...Adriano Soares Koshiyama
 
Scott Lundberg, Microsoft Research - Explainable Machine Learning with Shaple...
Scott Lundberg, Microsoft Research - Explainable Machine Learning with Shaple...Scott Lundberg, Microsoft Research - Explainable Machine Learning with Shaple...
Scott Lundberg, Microsoft Research - Explainable Machine Learning with Shaple...Sri Ambati
 
Unified Approach to Interpret Machine Learning Model: SHAP + LIME
Unified Approach to Interpret Machine Learning Model: SHAP + LIMEUnified Approach to Interpret Machine Learning Model: SHAP + LIME
Unified Approach to Interpret Machine Learning Model: SHAP + LIMEDatabricks
 
Explainability and bias in AI
Explainability and bias in AIExplainability and bias in AI
Explainability and bias in AIBill Liu
 
Introduction to Interpretable Machine Learning
Introduction to Interpretable Machine LearningIntroduction to Interpretable Machine Learning
Introduction to Interpretable Machine LearningNguyen Giang
 

Tendances (20)

What Is Machine Learning? | What Is Machine Learning And How Does It Work? | ...
What Is Machine Learning? | What Is Machine Learning And How Does It Work? | ...What Is Machine Learning? | What Is Machine Learning And How Does It Work? | ...
What Is Machine Learning? | What Is Machine Learning And How Does It Work? | ...
 
Explainable AI (XAI)
Explainable AI (XAI)Explainable AI (XAI)
Explainable AI (XAI)
 
Explainable AI
Explainable AIExplainable AI
Explainable AI
 
Explainable AI in Industry (WWW 2020 Tutorial)
Explainable AI in Industry (WWW 2020 Tutorial)Explainable AI in Industry (WWW 2020 Tutorial)
Explainable AI in Industry (WWW 2020 Tutorial)
 
Explainable AI: Building trustworthy AI models?
Explainable AI: Building trustworthy AI models? Explainable AI: Building trustworthy AI models?
Explainable AI: Building trustworthy AI models?
 
Explainable AI
Explainable AIExplainable AI
Explainable AI
 
GANの概要とDCGANのアーキテクチャ/アルゴリズム
GANの概要とDCGANのアーキテクチャ/アルゴリズムGANの概要とDCGANのアーキテクチャ/アルゴリズム
GANの概要とDCGANのアーキテクチャ/アルゴリズム
 
DC02. Interpretation of predictions
DC02. Interpretation of predictionsDC02. Interpretation of predictions
DC02. Interpretation of predictions
 
Deep neural networks and tabular data
Deep neural networks and tabular dataDeep neural networks and tabular data
Deep neural networks and tabular data
 
Deep Learning - The Past, Present and Future of Artificial Intelligence
Deep Learning - The Past, Present and Future of Artificial IntelligenceDeep Learning - The Past, Present and Future of Artificial Intelligence
Deep Learning - The Past, Present and Future of Artificial Intelligence
 
Interpretable machine learning
Interpretable machine learningInterpretable machine learning
Interpretable machine learning
 
Explainable AI - making ML and DL models more interpretable
Explainable AI - making ML and DL models more interpretableExplainable AI - making ML and DL models more interpretable
Explainable AI - making ML and DL models more interpretable
 
Explainable AI in Industry (AAAI 2020 Tutorial)
Explainable AI in Industry (AAAI 2020 Tutorial)Explainable AI in Industry (AAAI 2020 Tutorial)
Explainable AI in Industry (AAAI 2020 Tutorial)
 
Introduction to machine learning
Introduction to machine learningIntroduction to machine learning
Introduction to machine learning
 
Intro to LLMs
Intro to LLMsIntro to LLMs
Intro to LLMs
 
Algorithmic Impact Assessment: Fairness, Robustness and Explainability in Aut...
Algorithmic Impact Assessment: Fairness, Robustness and Explainability in Aut...Algorithmic Impact Assessment: Fairness, Robustness and Explainability in Aut...
Algorithmic Impact Assessment: Fairness, Robustness and Explainability in Aut...
 
Scott Lundberg, Microsoft Research - Explainable Machine Learning with Shaple...
Scott Lundberg, Microsoft Research - Explainable Machine Learning with Shaple...Scott Lundberg, Microsoft Research - Explainable Machine Learning with Shaple...
Scott Lundberg, Microsoft Research - Explainable Machine Learning with Shaple...
 
Unified Approach to Interpret Machine Learning Model: SHAP + LIME
Unified Approach to Interpret Machine Learning Model: SHAP + LIMEUnified Approach to Interpret Machine Learning Model: SHAP + LIME
Unified Approach to Interpret Machine Learning Model: SHAP + LIME
 
Explainability and bias in AI
Explainability and bias in AIExplainability and bias in AI
Explainability and bias in AI
 
Introduction to Interpretable Machine Learning
Introduction to Interpretable Machine LearningIntroduction to Interpretable Machine Learning
Introduction to Interpretable Machine Learning
 

Similaire à Machine Learning Interpretability: The good, the bad, and the ugly

Machine Learning Interpretability - Mateusz Dymczyk - H2O AI World London 2018
Machine Learning Interpretability - Mateusz Dymczyk - H2O AI World London 2018Machine Learning Interpretability - Mateusz Dymczyk - H2O AI World London 2018
Machine Learning Interpretability - Mateusz Dymczyk - H2O AI World London 2018Sri Ambati
 
Practical Tips for Interpreting Machine Learning Models - Patrick Hall, H2O.ai
Practical Tips for Interpreting Machine Learning Models - Patrick Hall, H2O.aiPractical Tips for Interpreting Machine Learning Models - Patrick Hall, H2O.ai
Practical Tips for Interpreting Machine Learning Models - Patrick Hall, H2O.aiSri Ambati
 
​​Explainability in AI and Recommender systems: let’s make it interactive!
​​Explainability in AI and Recommender systems: let’s make it interactive!​​Explainability in AI and Recommender systems: let’s make it interactive!
​​Explainability in AI and Recommender systems: let’s make it interactive!Eindhoven University of Technology / JADS
 
Synergy of Human and Artificial Intelligence in Software Engineering
Synergy of Human and Artificial Intelligence in Software EngineeringSynergy of Human and Artificial Intelligence in Software Engineering
Synergy of Human and Artificial Intelligence in Software EngineeringTao Xie
 
Model Evaluation in the land of Deep Learning
Model Evaluation in the land of Deep LearningModel Evaluation in the land of Deep Learning
Model Evaluation in the land of Deep LearningPramit Choudhary
 
Hacking Predictive Modeling - RoadSec 2018
Hacking Predictive Modeling - RoadSec 2018Hacking Predictive Modeling - RoadSec 2018
Hacking Predictive Modeling - RoadSec 2018HJ van Veen
 
The importance of model fairness and interpretability in AI systems
The importance of model fairness and interpretability in AI systemsThe importance of model fairness and interpretability in AI systems
The importance of model fairness and interpretability in AI systemsFrancesca Lazzeri, PhD
 
Train, explain, acclaim. Build a good model in three steps
Train, explain, acclaim.  Build a good model in three stepsTrain, explain, acclaim.  Build a good model in three steps
Train, explain, acclaim. Build a good model in three stepsPrzemek Biecek
 
Machine Learning basics
Machine Learning basicsMachine Learning basics
Machine Learning basicsNeeleEilers
 
How Does Generative AI Actually Work? (a quick semi-technical introduction to...
How Does Generative AI Actually Work? (a quick semi-technical introduction to...How Does Generative AI Actually Work? (a quick semi-technical introduction to...
How Does Generative AI Actually Work? (a quick semi-technical introduction to...ssuser4edc93
 
Human in the loop: Bayesian Rules Enabling Explainable AI
Human in the loop: Bayesian Rules Enabling Explainable AIHuman in the loop: Bayesian Rules Enabling Explainable AI
Human in the loop: Bayesian Rules Enabling Explainable AIPramit Choudhary
 
[DSC Europe 23] Dmitry Ustalov - Design and Evaluation of Large Language Models
[DSC Europe 23] Dmitry Ustalov - Design and Evaluation of Large Language Models[DSC Europe 23] Dmitry Ustalov - Design and Evaluation of Large Language Models
[DSC Europe 23] Dmitry Ustalov - Design and Evaluation of Large Language ModelsDataScienceConferenc1
 
Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!
Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!
Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!Sri Ambati
 
GDG Cloud Southlake #17: Meg Dickey-Kurdziolek: Explainable AI is for Everyone
GDG Cloud Southlake #17: Meg Dickey-Kurdziolek: Explainable AI is for EveryoneGDG Cloud Southlake #17: Meg Dickey-Kurdziolek: Explainable AI is for Everyone
GDG Cloud Southlake #17: Meg Dickey-Kurdziolek: Explainable AI is for EveryoneJames Anderson
 
SBQS 2013 Keynote: Cooperative Testing and Analysis
SBQS 2013 Keynote: Cooperative Testing and AnalysisSBQS 2013 Keynote: Cooperative Testing and Analysis
SBQS 2013 Keynote: Cooperative Testing and AnalysisTao Xie
 
ML crash course
ML crash courseML crash course
ML crash coursemikaelhuss
 
Recent Advances in Machine Learning: Bringing a New Level of Intelligence to ...
Recent Advances in Machine Learning: Bringing a New Level of Intelligence to ...Recent Advances in Machine Learning: Bringing a New Level of Intelligence to ...
Recent Advances in Machine Learning: Bringing a New Level of Intelligence to ...Brocade
 
2013 nas-ehs-data-integration-dc
2013 nas-ehs-data-integration-dc2013 nas-ehs-data-integration-dc
2013 nas-ehs-data-integration-dcc.titus.brown
 

Similaire à Machine Learning Interpretability: The good, the bad, and the ugly (20)

Machine Learning Interpretability - Mateusz Dymczyk - H2O AI World London 2018
Machine Learning Interpretability - Mateusz Dymczyk - H2O AI World London 2018Machine Learning Interpretability - Mateusz Dymczyk - H2O AI World London 2018
Machine Learning Interpretability - Mateusz Dymczyk - H2O AI World London 2018
 
Practical Tips for Interpreting Machine Learning Models - Patrick Hall, H2O.ai
Practical Tips for Interpreting Machine Learning Models - Patrick Hall, H2O.aiPractical Tips for Interpreting Machine Learning Models - Patrick Hall, H2O.ai
Practical Tips for Interpreting Machine Learning Models - Patrick Hall, H2O.ai
 
​​Explainability in AI and Recommender systems: let’s make it interactive!
​​Explainability in AI and Recommender systems: let’s make it interactive!​​Explainability in AI and Recommender systems: let’s make it interactive!
​​Explainability in AI and Recommender systems: let’s make it interactive!
 
Synergy of Human and Artificial Intelligence in Software Engineering
Synergy of Human and Artificial Intelligence in Software EngineeringSynergy of Human and Artificial Intelligence in Software Engineering
Synergy of Human and Artificial Intelligence in Software Engineering
 
Model Evaluation in the land of Deep Learning
Model Evaluation in the land of Deep LearningModel Evaluation in the land of Deep Learning
Model Evaluation in the land of Deep Learning
 
Hacking Predictive Modeling - RoadSec 2018
Hacking Predictive Modeling - RoadSec 2018Hacking Predictive Modeling - RoadSec 2018
Hacking Predictive Modeling - RoadSec 2018
 
The importance of model fairness and interpretability in AI systems
The importance of model fairness and interpretability in AI systemsThe importance of model fairness and interpretability in AI systems
The importance of model fairness and interpretability in AI systems
 
Train, explain, acclaim. Build a good model in three steps
Train, explain, acclaim.  Build a good model in three stepsTrain, explain, acclaim.  Build a good model in three steps
Train, explain, acclaim. Build a good model in three steps
 
Machine Learning basics
Machine Learning basicsMachine Learning basics
Machine Learning basics
 
How Does Generative AI Actually Work? (a quick semi-technical introduction to...
How Does Generative AI Actually Work? (a quick semi-technical introduction to...How Does Generative AI Actually Work? (a quick semi-technical introduction to...
How Does Generative AI Actually Work? (a quick semi-technical introduction to...
 
Human in the loop: Bayesian Rules Enabling Explainable AI
Human in the loop: Bayesian Rules Enabling Explainable AIHuman in the loop: Bayesian Rules Enabling Explainable AI
Human in the loop: Bayesian Rules Enabling Explainable AI
 
AI Presentation 1
AI Presentation 1AI Presentation 1
AI Presentation 1
 
ODSC APAC 2022 - Explainable AI
ODSC APAC 2022 - Explainable AIODSC APAC 2022 - Explainable AI
ODSC APAC 2022 - Explainable AI
 
[DSC Europe 23] Dmitry Ustalov - Design and Evaluation of Large Language Models
[DSC Europe 23] Dmitry Ustalov - Design and Evaluation of Large Language Models[DSC Europe 23] Dmitry Ustalov - Design and Evaluation of Large Language Models
[DSC Europe 23] Dmitry Ustalov - Design and Evaluation of Large Language Models
 
Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!
Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!
Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!
 
GDG Cloud Southlake #17: Meg Dickey-Kurdziolek: Explainable AI is for Everyone
GDG Cloud Southlake #17: Meg Dickey-Kurdziolek: Explainable AI is for EveryoneGDG Cloud Southlake #17: Meg Dickey-Kurdziolek: Explainable AI is for Everyone
GDG Cloud Southlake #17: Meg Dickey-Kurdziolek: Explainable AI is for Everyone
 
SBQS 2013 Keynote: Cooperative Testing and Analysis
SBQS 2013 Keynote: Cooperative Testing and AnalysisSBQS 2013 Keynote: Cooperative Testing and Analysis
SBQS 2013 Keynote: Cooperative Testing and Analysis
 
ML crash course
ML crash courseML crash course
ML crash course
 
Recent Advances in Machine Learning: Bringing a New Level of Intelligence to ...
Recent Advances in Machine Learning: Bringing a New Level of Intelligence to ...Recent Advances in Machine Learning: Bringing a New Level of Intelligence to ...
Recent Advances in Machine Learning: Bringing a New Level of Intelligence to ...
 
2013 nas-ehs-data-integration-dc
2013 nas-ehs-data-integration-dc2013 nas-ehs-data-integration-dc
2013 nas-ehs-data-integration-dc
 

Plus de Sri Ambati

H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo DayH2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo DaySri Ambati
 
Generative AI Masterclass - Model Risk Management.pptx
Generative AI Masterclass - Model Risk Management.pptxGenerative AI Masterclass - Model Risk Management.pptx
Generative AI Masterclass - Model Risk Management.pptxSri Ambati
 
AI and the Future of Software Development: A Sneak Peek
AI and the Future of Software Development: A Sneak Peek AI and the Future of Software Development: A Sneak Peek
AI and the Future of Software Development: A Sneak Peek Sri Ambati
 
LLMOps: Match report from the top of the 5th
LLMOps: Match report from the top of the 5thLLMOps: Match report from the top of the 5th
LLMOps: Match report from the top of the 5thSri Ambati
 
Building, Evaluating, and Optimizing your RAG App for Production
Building, Evaluating, and Optimizing your RAG App for ProductionBuilding, Evaluating, and Optimizing your RAG App for Production
Building, Evaluating, and Optimizing your RAG App for ProductionSri Ambati
 
Building LLM Solutions using Open Source and Closed Source Solutions in Coher...
Building LLM Solutions using Open Source and Closed Source Solutions in Coher...Building LLM Solutions using Open Source and Closed Source Solutions in Coher...
Building LLM Solutions using Open Source and Closed Source Solutions in Coher...Sri Ambati
 
Risk Management for LLMs
Risk Management for LLMsRisk Management for LLMs
Risk Management for LLMsSri Ambati
 
Open-Source AI: Community is the Way
Open-Source AI: Community is the WayOpen-Source AI: Community is the Way
Open-Source AI: Community is the WaySri Ambati
 
Building Custom GenAI Apps at H2O
Building Custom GenAI Apps at H2OBuilding Custom GenAI Apps at H2O
Building Custom GenAI Apps at H2OSri Ambati
 
Applied Gen AI for the Finance Vertical
Applied Gen AI for the Finance Vertical Applied Gen AI for the Finance Vertical
Applied Gen AI for the Finance Vertical Sri Ambati
 
Cutting Edge Tricks from LLM Papers
Cutting Edge Tricks from LLM PapersCutting Edge Tricks from LLM Papers
Cutting Edge Tricks from LLM PapersSri Ambati
 
Practitioner's Guide to LLMs: Exploring Use Cases and a Glimpse Beyond Curren...
Practitioner's Guide to LLMs: Exploring Use Cases and a Glimpse Beyond Curren...Practitioner's Guide to LLMs: Exploring Use Cases and a Glimpse Beyond Curren...
Practitioner's Guide to LLMs: Exploring Use Cases and a Glimpse Beyond Curren...Sri Ambati
 
Open Source h2oGPT with Retrieval Augmented Generation (RAG), Web Search, and...
Open Source h2oGPT with Retrieval Augmented Generation (RAG), Web Search, and...Open Source h2oGPT with Retrieval Augmented Generation (RAG), Web Search, and...
Open Source h2oGPT with Retrieval Augmented Generation (RAG), Web Search, and...Sri Ambati
 
KGM Mastering Classification and Regression with LLMs: Insights from Kaggle C...
KGM Mastering Classification and Regression with LLMs: Insights from Kaggle C...KGM Mastering Classification and Regression with LLMs: Insights from Kaggle C...
KGM Mastering Classification and Regression with LLMs: Insights from Kaggle C...Sri Ambati
 
LLM Interpretability
LLM Interpretability LLM Interpretability
LLM Interpretability Sri Ambati
 
Never Reply to an Email Again
Never Reply to an Email AgainNever Reply to an Email Again
Never Reply to an Email AgainSri Ambati
 
Introducción al Aprendizaje Automatico con H2O-3 (1)
Introducción al Aprendizaje Automatico con H2O-3 (1)Introducción al Aprendizaje Automatico con H2O-3 (1)
Introducción al Aprendizaje Automatico con H2O-3 (1)Sri Ambati
 
From Rapid Prototypes to an end-to-end Model Deployment: an AI Hedge Fund Use...
From Rapid Prototypes to an end-to-end Model Deployment: an AI Hedge Fund Use...From Rapid Prototypes to an end-to-end Model Deployment: an AI Hedge Fund Use...
From Rapid Prototypes to an end-to-end Model Deployment: an AI Hedge Fund Use...Sri Ambati
 
AI Foundations Course Module 1 - Shifting to the Next Step in Your AI Transfo...
AI Foundations Course Module 1 - Shifting to the Next Step in Your AI Transfo...AI Foundations Course Module 1 - Shifting to the Next Step in Your AI Transfo...
AI Foundations Course Module 1 - Shifting to the Next Step in Your AI Transfo...Sri Ambati
 
AI Foundations Course Module 1 - An AI Transformation Journey
AI Foundations Course Module 1 - An AI Transformation JourneyAI Foundations Course Module 1 - An AI Transformation Journey
AI Foundations Course Module 1 - An AI Transformation JourneySri Ambati
 

Plus de Sri Ambati (20)

H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo DayH2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
 
Generative AI Masterclass - Model Risk Management.pptx
Generative AI Masterclass - Model Risk Management.pptxGenerative AI Masterclass - Model Risk Management.pptx
Generative AI Masterclass - Model Risk Management.pptx
 
AI and the Future of Software Development: A Sneak Peek
AI and the Future of Software Development: A Sneak Peek AI and the Future of Software Development: A Sneak Peek
AI and the Future of Software Development: A Sneak Peek
 
LLMOps: Match report from the top of the 5th
LLMOps: Match report from the top of the 5thLLMOps: Match report from the top of the 5th
LLMOps: Match report from the top of the 5th
 
Building, Evaluating, and Optimizing your RAG App for Production
Building, Evaluating, and Optimizing your RAG App for ProductionBuilding, Evaluating, and Optimizing your RAG App for Production
Building, Evaluating, and Optimizing your RAG App for Production
 
Building LLM Solutions using Open Source and Closed Source Solutions in Coher...
Building LLM Solutions using Open Source and Closed Source Solutions in Coher...Building LLM Solutions using Open Source and Closed Source Solutions in Coher...
Building LLM Solutions using Open Source and Closed Source Solutions in Coher...
 
Risk Management for LLMs
Risk Management for LLMsRisk Management for LLMs
Risk Management for LLMs
 
Open-Source AI: Community is the Way
Open-Source AI: Community is the WayOpen-Source AI: Community is the Way
Open-Source AI: Community is the Way
 
Building Custom GenAI Apps at H2O
Building Custom GenAI Apps at H2OBuilding Custom GenAI Apps at H2O
Building Custom GenAI Apps at H2O
 
Applied Gen AI for the Finance Vertical
Applied Gen AI for the Finance Vertical Applied Gen AI for the Finance Vertical
Applied Gen AI for the Finance Vertical
 
Cutting Edge Tricks from LLM Papers
Cutting Edge Tricks from LLM PapersCutting Edge Tricks from LLM Papers
Cutting Edge Tricks from LLM Papers
 
Practitioner's Guide to LLMs: Exploring Use Cases and a Glimpse Beyond Curren...
Practitioner's Guide to LLMs: Exploring Use Cases and a Glimpse Beyond Curren...Practitioner's Guide to LLMs: Exploring Use Cases and a Glimpse Beyond Curren...
Practitioner's Guide to LLMs: Exploring Use Cases and a Glimpse Beyond Curren...
 
Open Source h2oGPT with Retrieval Augmented Generation (RAG), Web Search, and...
Open Source h2oGPT with Retrieval Augmented Generation (RAG), Web Search, and...Open Source h2oGPT with Retrieval Augmented Generation (RAG), Web Search, and...
Open Source h2oGPT with Retrieval Augmented Generation (RAG), Web Search, and...
 
KGM Mastering Classification and Regression with LLMs: Insights from Kaggle C...
KGM Mastering Classification and Regression with LLMs: Insights from Kaggle C...KGM Mastering Classification and Regression with LLMs: Insights from Kaggle C...
KGM Mastering Classification and Regression with LLMs: Insights from Kaggle C...
 
LLM Interpretability
LLM Interpretability LLM Interpretability
LLM Interpretability
 
Never Reply to an Email Again
Never Reply to an Email AgainNever Reply to an Email Again
Never Reply to an Email Again
 
Introducción al Aprendizaje Automatico con H2O-3 (1)
Introducción al Aprendizaje Automatico con H2O-3 (1)Introducción al Aprendizaje Automatico con H2O-3 (1)
Introducción al Aprendizaje Automatico con H2O-3 (1)
 
From Rapid Prototypes to an end-to-end Model Deployment: an AI Hedge Fund Use...
From Rapid Prototypes to an end-to-end Model Deployment: an AI Hedge Fund Use...From Rapid Prototypes to an end-to-end Model Deployment: an AI Hedge Fund Use...
From Rapid Prototypes to an end-to-end Model Deployment: an AI Hedge Fund Use...
 
AI Foundations Course Module 1 - Shifting to the Next Step in Your AI Transfo...
AI Foundations Course Module 1 - Shifting to the Next Step in Your AI Transfo...AI Foundations Course Module 1 - Shifting to the Next Step in Your AI Transfo...
AI Foundations Course Module 1 - Shifting to the Next Step in Your AI Transfo...
 
AI Foundations Course Module 1 - An AI Transformation Journey
AI Foundations Course Module 1 - An AI Transformation JourneyAI Foundations Course Module 1 - An AI Transformation Journey
AI Foundations Course Module 1 - An AI Transformation Journey
 

Dernier

Dev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebDev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebUiPathCommunity
 
Gen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfGen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfAddepto
 
The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsPixlogix Infotech
 
Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Enterprise Knowledge
 
Advanced Computer Architecture – An Introduction
Advanced Computer Architecture – An IntroductionAdvanced Computer Architecture – An Introduction
Advanced Computer Architecture – An IntroductionDilum Bandara
 
TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024Lonnie McRorey
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxLoriGlavin3
 
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024BookNet Canada
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .Alan Dix
 
SAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxSAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxNavinnSomaal
 
"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr BaganFwdays
 
Scanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsScanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsRizwan Syed
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubKalema Edgar
 
Story boards and shot lists for my a level piece
Story boards and shot lists for my a level pieceStory boards and shot lists for my a level piece
Story boards and shot lists for my a level piececharlottematthew16
 
Commit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyCommit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyAlfredo García Lavilla
 
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdfHyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdfPrecisely
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsMark Billinghurst
 
TrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data PrivacyTrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data PrivacyTrustArc
 
WordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your BrandWordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your Brandgvaughan
 

Dernier (20)

Dev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebDev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio Web
 
Gen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfGen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdf
 
The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and Cons
 
Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024
 
Advanced Computer Architecture – An Introduction
Advanced Computer Architecture – An IntroductionAdvanced Computer Architecture – An Introduction
Advanced Computer Architecture – An Introduction
 
TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
 
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .
 
SAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxSAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptx
 
"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan
 
DMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special EditionDMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special Edition
 
Scanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsScanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL Certs
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding Club
 
Story boards and shot lists for my a level piece
Story boards and shot lists for my a level pieceStory boards and shot lists for my a level piece
Story boards and shot lists for my a level piece
 
Commit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyCommit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easy
 
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdfHyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR Systems
 
TrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data PrivacyTrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data Privacy
 
WordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your BrandWordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your Brand
 

Machine Learning Interpretability: The good, the bad, and the ugly

  • 2. Acknowledgments Makers: Navdeep Gill, Mark Chan, Doug Deloy, Megan Kurka, Michal Kurka, Wen Phan, Sri Satish Ambati, Lingyao Meng, Mathias Müller Advisors: Leland Wilkinson, Trevor Hastie, Rob Tibshirani Community: O’Reilly Strata and AI (accepted talks and tutorials), FAT/ML Conference (accepted tutorial), ASA Symposium on Data Science and Statistics (invited talk), Joint Statistical Meetings (accepted talk)
  • 3. What is Machine Learning Interpretability? “The ability to explain or to present in understandable terms to a human.” -- Finale Doshi-Velez and Been Kim. “Towards a rigorous science of interpretable machine learning.” arXiv preprint. 2017. https://arxiv.org/pdf/1702.08608.pdf FAT*: https://www.fatml.org/resources/principles-for-accountable-algorithms XAI: https://www.darpa.mil/program/explainable-artificial-intelligence
  • 4. Why Should You Care About Machine Learning Interpretability? “The now-contemplated field of data science amounts to a superset of the fields of statistics and machine learning, which adds some technology for “scaling up” to “big data.” This chosen superset is motivated by commercial rather than intellectual developments. Choosing in this way is likely to miss out on the really important intellectual event of the next 50 years.” -- David Donoho. “50 years of Data Science.” Tukey Centennial Workshop, 2015. http://bit.ly/2GQOh1J Social Motivation: Interpretability plays a critical role in the increased convenience, automation, and organization in our day-to-day lives promised by AI. Commercial Motivation: Interpretability is required for regulated industry to adopt machine learning. • Check and balance against accidental or intentional discrimination. • “Right to explanation.” • Hacking and adversarial attacks. • Improved revenue, i.e. Equifax NeuroDecision: https://www.youtube.com/watch?v=9Z_GW9WDS2c
  • 5. Why is Machine Learning Interpretability Difficult? https://github.com/jphall663/GWU_data_mining/blob/master/05_neural_networks/notes/cnn-gwu.pdf Machine learning algorithms intrinsically consider high-degree interactions between input features. Disaggregating such functions into reason codes based on single input features is difficult. Neural Network
  • 6. Linear Models Machine Learning For a given well-understood dataset there is usually one best model. For a given well-understood dataset there are usually many good models. This is often referred to as “the multiplicity of good models.” -- Leo Breiman. “Statistical modeling: The two cultures (with comments and a rejoinder by the author).” Statistical Science. 2001. http://bit.ly/2pwz6m5 Why is Machine Learning Interpretability Difficult?
  • 7. What is the Value Proposition of Machine Learning Interpretability? Linear Models Machine Learning Exact explanations for approximate models. Approximate explanations for exact models.
  • 8. How Can Machine Learning Interpretability Be Practiced? By seeing and understanding relationships and structures in training, test, and new data. 2-D Projections Correlation Graphs https://github.com/jphall663/corr_graphhttps://www.cs.toronto.edu/~hinton/science.pdf
  • 9. How Can Machine Learning Interpretability Be Practiced? By training interpretable (“white-box”) models. Decision Trees • References: • Breiman, L., J. Friedman, C. J. Stone, and R. A. Olshen. Classification and regression trees. CRC press, 1984. • The Elements of Statistical Learning (ESL): https://web.stanford.edu/~hastie/ElemStatLearn/printings/ESLII_print12.pdf • OSS: • rpart: https://cran.r-project.org/web/packages/rpart/index.html • scikit-learn (various functions): https://github.com/scikit-learn/scikit-learn Monotonic Gradient Boosted Machines (GBMs) • Reference: XGBoost Documentation: http://xgboost.readthedocs.io/en/latest/tutorials/monotonic.html • OSS: XGBoost: https://github.com/dmlc/xgboost Logistic, elastic net, GAM, and quantile regression • References: • ESL • Koenker, R. Quantile regression (No. 38). Cambridge University Press, 2005. • OSS: • gam: https://cran.r-project.org/web/packages/gam/index.html • glmnet: https://cran.r-project.org/web/packages/glmnet/index.html • h2o: https://github.com/h2oai/h2o-3 • quantreg: https://cran.r-project.org/web/packages/quantreg/index.html • scikit-learn (various functions): https://github.com/scikit-learn/scikit-learn Rule-based models • Reference: An Introduction to Data Mining, Chapter 6: https://www-users.cs.umn.edu/~kumar001/dmbook/ch6.pdf • OSS: • RuleFit: http://statweb.stanford.edu/~jhf/R_RuleFit.html • arules: https://cran.r-project.org/web/packages/arules/index.html • FP-Growth: http://spark.apache.org/docs/2.2.0/mllib-frequent-pattern-mining.html Supersparse Linear Integer Models (SLIMs) • Reference: Supersparse Linear Integer Models for Optimized Medical Scoring Systems: https://link.springer.com/content/pdf/10.1007%2Fs10994-015-5528-6.pdf
  • 10. How Can Machine Learning Interpretability Be Practiced? With complimentary 2-D visualizations of trained models that enable understanding of learned high-degree interactions. Decision Tree Surrogate Models Partial Dependence and Individual Conditional Expectation https://github.com/jphall663/interpretable_machine_learning_with_python
  • 11. How Can Machine Learning Interpretability Be Practiced? By calculating approximate local feature importance values and ranking them to create reason codes for every prediction. Treeinterpreter: https://github.com/TeamHG-Memex/eli5 LIME: https://github.com/marcotcr/lime shap: https://github.com/slundberg/shap
  • 12. How Can Machine Learning Interpretability Be Practiced? By calculating approximate global feature importance to understand how each feature impacts the model in general. Global feature importance indicates the impact of a feature on the model for the entire training data set.
  • 13. How Can Machine Learning Interpretability Be Practiced? By using sensitivity analysis to test machine learning model predictions for accuracy and stability. If you are using a machine learning model, you should probably be conducting sensitivity analysis. http://www.vias.org/tmdatanaleng/
  • 14. Can Machine Learning Interpretability Be Tested? Yes. By humans or … Simulated data You can use simulated data with known characteristics to test explanations. For instance, models trained on totally random data with no relationship between a number of input variables and a prediction target should not give strong weight to any input variable nor generate compelling local explanations or reason codes. Conversely, you can use simulated data with a known signal generating function to test that explanations accurately represent that known function. https://github.com/h2oai/mli-resources/tree/master/lime_shap_treeint_compare Explanation stability under data perturbation Trustworthy explanations likely should not change drastically for minor changes in input data. You can set and test thresholds for allowable explanation value changes automatically by perturbing input data. (Explanations or reason code values can also be averaged across a number of models to create more stable explanations.) Explanation stability with increased prediction accuracy If previously known, accurate explanations or reason codes from a simpler linear model are available, you can use them as a reference for the accuracy of explanations from a related, but more complex and hopefully more accurate, model. You can perform tests to see how accurate a model can become before its prediction’s reason codes veer away from known standards.
  • 15. General Recommendations • Consider deployment. • A very direct path to interpretable machine learning today is to train a monotonic GBM with XGBoost and to use Shapley explanations either in XGBoost or with the shap Python package. (Or just buy H2O Driverless AI https://www.h2o.ai/driverless-ai/ ;) • Use a combination of local and global explanatory techniques. • Conduct sensitivity analysis and random data attacks on all machine learning models. • Test your explanatory software. • If possible, use model-specific explanatory techniques to generate reason codes. • Open source explanation packages seem immature. • Beware of uninterpretable features.
  • 16. LIME Recommendations/Observations • LIME can give an indication of its own trustworthiness using fit statistics. • LIME can fail, particularly in the presence of extreme nonlinearity or high-degree interactions. • LIME is difficult to deploy, but there are highly deployable variants, e.g. H2O’s K-LIME. • Reason codes are offsets from a local intercept. • Note that the intercept in LIME can account for the most important local phenomena. • Generated LIME samples can contain large proportions of out-of-range data that can lead to unrealistically high or low intercept values. • Try LIME on discretized input features and on manually constructed interactions. • Use cross-validation to construct standard deviations or even confidence intervals for reason code values.
  • 17. Treeinterpreter and Shapley Recommendations/Observations • Treeinterpreter and Shapley explanations do not give an indication of their own trustworthiness. We can only assume they are trustworthy … • Treeinterpreter appears to fail when used with regularized (L1/L2) XGBoost models. • Due to theoretical support and robust implementation, Shapely explanations may be suitable for regulated applications. • Reason codes are offsets from a global intercept.