📒
Machine & Deep Learning Compendium
  • The Machine & Deep Learning Compendium
    • Thanks Page
  • The Ops Compendium
  • Types Of Machine Learning
    • Overview
    • Model Families
    • Weakly Supervised
    • Semi Supervised
    • Active Learning
    • Online Learning
    • N-Shot Learning
    • Unlearning
  • Foundation Knowledge
    • Data Science
    • Data Science Tools
    • Management
    • Project & Program Management
    • Data Science Management
    • Calculus
    • Probability & Statistics
    • Probability
    • Hypothesis Testing
    • Feature Types
    • Multi Label Classification
    • Distribution
    • Distribution Transformation
    • Normalization & Scaling
    • Regularization
    • Information Theory
    • Game Theory
    • Multi CPU Processing
    • Benchmarking
  • Validation & Evaluation
    • Features
    • Evaluation Metrics
    • Datasets
    • Dataset Confidence
    • Hyper Parameter Optimization
    • Training Strategies
    • Calibration
    • Datasets Reliability & Correctness
    • Data & Model Tests
    • Fairness, Accountability, and Transparency
    • Interpretable & Explainable AI (XAI)
    • Federated Learning
  • Machine Learning
    • Algorithms 101
    • Meta Learning (AutoML)
    • Probabilistic, Regression
    • Data Mining
    • Process Mining
    • Label Algorithms
    • Clustering Algorithms
    • Anomaly Detection
    • Decision Trees
    • Active Learning Algorithms
    • Linear Separator Algorithms
    • Regression
    • Ensembles
    • Reinforcement Learning
    • Incremental Learning
    • Dimensionality Reduction Methods
    • Genetic Algorithms & Genetic Programming
    • Learning Classifier Systems
    • Recommender Systems
    • Timeseries
    • Fourier Transform
    • Digital Signal Processing (DSP)
    • Propensity Score Matching
    • Diffusion models
  • Classical Graph Models
    • Graph Theory
    • Social Network Analysis
  • Deep Learning
    • Deep Neural Nets Basics
    • Deep Neural Frameworks
    • Embedding
    • Deep Learning Models
    • Deep Network Optimization
    • Attention
    • Deep Neural Machine Vision
    • Deep Neural Tabular
    • Deep Neural Time Series
  • Audio
    • Basics
    • Terminology
    • Feature Engineering
    • Deep Neural Audio
    • Algorithms
  • Natural Language Processing
    • A Reality Check
    • NLP Tools
    • Foundation NLP
    • Name Matching
    • String Matching
    • TF-IDF
    • Language Detection Identification Generation (NLD, NLI, NLG)
    • Topics Modeling
    • Named Entity Recognition (NER)
    • SEARCH
    • Neural NLP
    • Tokenization
    • Decoding Algorithms For NLP
    • Multi Language
    • Augmentation
    • Knowledge Graphs
    • Annotation & Disagreement
    • Sentiment Analysis
    • Question Answering
    • Summarization
    • Chat Bots
    • Conversation
  • Generative AI
    • Methods
    • Gen AI Industry
    • Speech
    • Prompt
    • Fairness, Accountability, and Transparency In Prompts
    • Large Language Models (LLMs)
    • Vision
    • GPT
    • Mix N Match
    • Diffusion Models
    • GenAI Applications
    • Agents
    • RAG
    • Chat UI/UX
  • Experimental Design
    • Design Of Experiments
    • DOE Tools
    • A/B Testing
    • Multi Armed Bandits
    • Contextual Bandits
    • Factorial Design
  • Business Domains
    • Follow the regularized leader
    • Growth
    • Root Cause Effects (RCE/RCA)
    • Log Parsing / Templatization
    • Fraud Detection
    • Life Time Value (LTV)
    • Survival Analysis
    • Propaganda Detection
    • NYC TAXI
    • Drug Discovery
    • Intent Recognition
    • Churn Prediction
    • Electronic Network Frequency Analysis
    • Marketing
  • Product Management
    • Expanding Your Data Science Skills
    • Product Vision & Strategy
    • Product / Program Managers
    • Product Management Resources
    • Product Tools
    • User Experience Design (UX)
    • Business
    • Marketing
    • Ideation
  • MLOps (www.OpsCompendium.com)
  • DataOps (www.OpsCompendium.com)
  • Humor
Powered by GitBook
On this page
  • XAI
  • Lime
  • Anchor
  • Shap

Was this helpful?

  1. Validation & Evaluation

Interpretable & Explainable AI (XAI)

PreviousFairness, Accountability, and TransparencyNextFederated Learning

Last updated 1 year ago

Was this helpful?

XAI

  1. Interpretability and Explainability in Machine Learning / slides. Understanding, evaluating, rule based, prototype based, risk scores, generalized additive models, explaining black box, visualizing, feature importance, actionable explanations, casual models, human in the loop, connection with debugging.

  2. a tutorial by Hima Lakkaraju (tutorial , , )

  3. by Cinthia rudin

    1. A great on the topic by Shir Meir Lador

  4. From the above image: - a really good review for everything XAI - “a survey to help industry practitioners (but also data scientists more broadly) understand the field of explainable machine learning better and apply the right tools. Our latter sections build a narrative around a putative data scientist, and discuss how she might go about explaining her models by asking the right questions. From an organization viewpoint, after motivating the area broadly, we discuss the main developments, including the principles that allow us to study transparent models vs opaque models, as well as model-specific or model-agnostic post-hoc explainability approaches. We also briefly reflect on deep learning models, and conclude with a discussion about future research directions.”

  5. ,

  6. (great) transparent (simultability, decomposability, algorithmic transparency) post-hoc interpretability (text explanation, visual local, explanation by example,), evaluation, utility.

  7. “A growing number of techniques provide model interpretations, but can lead to wrong conclusions if applied incorrectly. We illustrate pitfalls of ML model interpretation such as bad model generalization, dependent features, feature interactions or unjustified causal interpretations. Our paper addresses ML practitioners by raising awareness of pitfalls and pointing out solutions for correct model interpretation, as well as ML researchers by discussing open issues for further research.” - mulner et al.

  8. *** This is very good, includes eli5, lime, shap, many others.

  9. Book:

  10. - White-box and black-box ML model explanation library. is an open source Python library aimed at machine learning model inspection and interpretation. The focus of the library is to provide high-quality implementations of black-box, white-box, local and global explanation methods for classification and regression models.

  11. youtube,

  12. are not always consistent and do not agree with each other, this article has a make-sense explanation and flow for using shap and its many plots.

  13. - Path attribution methods are a gradient-based way of explaining deep models. These methods require choosing a hyperparameter known as the baseline input. What does this hyperparameter mean, and how important is it? In this article, we investigate these questions using image classification networks as a case study. We discuss several different ways to choose a baseline input and the assumptions that are implicit in each baseline. Although we focus here on path attribution methods, our discussion of baselines is closely connected with the concept of missingness in the feature space - a concept that is critical to interpretability research.

  14. WHAT IF TOOL - GOOGLE, ,

  15. The Language Interpretability Tool (LIT) is an open-source platform for visualization and understanding of NLP models.

  16. - “trying to \textit{explain} black box models, rather than creating models that are \textit{interpretable} in the first place, is likely to perpetuate bad practices and can potentially cause catastrophic harm to society. There is a way forward -- it is to design models that are inherently interpretable. This manuscript clarifies the chasm between explaining black boxes and using inherently interpretable models, outlines several key reasons why explainable black boxes should be avoided in high-stakes decisions, identifies challenges to interpretable machine learning, and provides several example applications where interpretable models could potentially replace black box models in criminal justice, healthcare, and computer vision.”

  17. from

  18. () was designed to automatically perform feature selection on a dataset using randomized features, i.e., measuring valid features against their shadow/noisy counterparts.

  19. by Microsoft, .

  20. ,

  21. - explain transformers with 2 lines of code.

Lime

Anchor

Shap

  1. Theory:

  2. A series on Shap, Lime.

  3. Shapash

      1. Consistency - do different explainability methods give, on average, similar explanations?

      2. Stability - for similar instances, are the explanations similar?

      3. Compacity - do fewer features drive the model?

  4. Partial Shap

  5. Shap residuals

  6. SHAP advanced

    1. How to calculate Shap values per class based on this graph

NLP and IMAGE, - In the experiments in , we demonstrate that both machine learning experts and lay users greatly benefit from explanations similar to Figures 5 and 6 and are able to choose which models generalize better, improve models by changing them, and get crucial insights into the models' behavior.

- An anchor explanation is a rule that sufficiently “anchors” the prediction locally – such that changes to the rest of the feature values of the instance do not matter. In other words, for instances on which the anchor holds, the prediction is (almost) always the same.

How Shap values are calculated - .

Cooporative game theory & Shapely values, ,

Intro to shap and lime, ,

Part I:

Part II:

Part III:

Part VI:

Part V:

Part VI:

Part VII:

Part VIII:

Medium

**** In depth

(lime and shap)

- Yann Golhen

- francesco marini using 3 new metrics

? by Samuele Mazzanti - "Discover “ParShap”: an advanced method to detect which columns make your model underperform on new data" implemented in -stats.

- whatever you do start with this

- continue with this

*** how lime works behind the scenes
LIME to interpret models
github
our research paper
Anchor from the authors of Lime,
youtube
Medium
youtube
Calculating a Taxi fare using Shap
Shap explained
part 1
part 2
Explain Your Model with the SHAP Values
The SHAP with More Elegant Charts
How Is the Partial Dependent Plot Calculated?
An Explanation for eXplainable AI
Explain Any Models with the SHAP Values — Use the KernelExplainer
The SHAP Values with H2O Models
Explain Your Model with LIME
Explain Your Model with Microsoft’s InterpretML
Intro to lime and shap
SHAP
Github
Country happiness using shap
Stackoverflow example, predicting tags, pandas keras etc
Intro to shapely and shap
Fiddler on shap
shapash git -
a web app
.
making models understandable by everyone
using shapash for confidence on XAI.
Which Of Your Features Are Overfitting
pingouin
medium
Official shap tutorial on their plots, you can never read this too many times.
What are shap values on kaggle
Shap values on kaggle #2
A thorough post about the many ways of explaining a model, from regression, to bayes, to trees, forests, lime, beta, feature selection/elimination
Trusting models
Interpret using uncertainty
Shap in Python
A series of videos about XAI.
A curated document about XAI research resources.
course
Explainable Machine Learning: Understanding the Limits & Pushing the Boundaries
VIDEO
youtube
twitter
Stop Explaining Black Box Machine Learning Models for High Stakes Decisions and Use Interpretable Models Instead
talk
explainML tutorial
When not to trust explanations :)
Paper: Principles and practice of explainable models
Book: interpretable machine learning
christoph mulner
Interpretability overview,
Medium: the great debate
Paper: pitfalls to avoid when interpreting ML models
whitening a black box.
exploratory model analysis
Alibi-explain
Alibi
Hands on explainable ai
git
Explainable methods
The notebook!
Blog
More resources!
Visualizing the impact of feature attribution baseline
notebook
walkthrough
Language interpretability tool (LIT) -
Stop Explaining Black Box Machine Learning Models for High Stakes Decisions and Use Interpretable Models Instead
Using genetic algorithms
Google’s what-if tool
PAIR
Boruta
medium
InterpretML
git
Connecting Interpretability and Robustness in Decision Trees through Separation
git
Interpret Transformers
for cnns, 3 methods, activation maximization, saliency and class activation maps
Keras-vis