SlideShare une entreprise Scribd logo
1  sur  61
Télécharger pour lire hors ligne
The Importance of Model Fairness and
Interpretability in AI Systems
Spark Summit 2020
Machine Fairness
How to assess AI system’s fairness and mitigate any observed unfairness issues
www.aka.ms/Fairlearn-build2020
Resources
➢www.aka.ms/InterpretML-github
➢www.aka.ms/InterpretML-toolkit
➢www.aka.ms/InterpretML-doc
➢www.aka.ms/FairlearnAI
➢www.aka.ms/FairlearnAIDoc
➢www.aka.ms/ResponsibleMLDoc
➢www.aka.ms/AzureResponsibleML
➢www.aka.ms/AzureMLDoc
➢www.aka.ms/AzureMLservice
Agenda
WHY RESPONSIBLE AI?
INTERPRETML
FAIRLEARN TOOLKIT
AI will have a considerable impact on
business and society as a whole
Report
Why Responsible AI?
Microsoft’s AI Principles
Microsoft’s AI Principles
Transparency
AI systems should be explainable
AI systems should have algorithmic
accountability
Machine Learning Interpretability use cases
www.aka.ms/InterpretML-github
Interpretability Toolkit
www.aka.ms/InterpretML-github
Explain Models - www.aka.ms/InterpretML-toolkit
InterpretML Repo - www.aka.ms/InterpretML-github
Interpret
Tabular Data
Additional interpretability
techniques for tabular
extension
www.aka.ms/InterpretML-github github/InterpretML/Interpret-Community
Interpret-
Community
Interpret
Tabular Data
Additional interpretability
techniques for tabular
extension
github/InterpretML/Interpret github/InterpretML/Interpret-Community
Interpret-
Community
AzureML-
Interpret
wrapper
Azure Machine Learning SDK
InterpretML Repo - www.aka.ms/InterpretML-github
Interpret
Tabular Data
Additional interpretability
techniques for tabular
extension
github/InterpretML/Interpret github/InterpretML/Interpret-Community
Interpret-
Community
Interpret-
text-contrib
github/InterpretML/Interpret-text-contrib
Text Data
(Text Classification)
AzureML-
Interpret
wrapper
Azure Machine Learning SDK
InterpretML Repo - www.aka.ms/InterpretML-github
Types of
Models
supported
www.aka.ms/InterpretML-github
Interpretability Approaches
• Given an existing model/system, extract
explanations
• Typically provide approximate explanations
• Examples: LIME, SHAP, Permutation Feature
Importance, Sensitivity Analysis, Influence
Functions, etc.
Tools to explain the system
www.aka.ms/InterpretML-github
SHAP
(SHapley Additive
exPlanations)
• Not a new concept
• Concept based on game theory
• Mathematically sound
• Application in ML relatively new
www.aka.ms/InterpretML-github
SHAP
Modeling scenario:
House pricing
prediction
www.aka.ms/InterpretML-github
How much has each feature contributed
to the prediction compared to the average prediction?
House price
prediction:
€ 300,000
Average house
price prediction
for all apartments
is € 310,000
Delta here is
- €10,000
www.aka.ms/InterpretML-github
Parks contributed
+10k
Cats banned
contributed -50k
50𝑚2
contributed
+10k 2nd floor had net zero
contributions
www.aka.ms/InterpretML-github
SHAP
How do we calculate Shapley values?
• Take your feature of interest (e.g.,
cats-banned) and remove it from the
feature set
• Take the remaining features and
generate all possible coalitions
• Add and remove your feature of
interested to each of the coalitions
and calculate the difference it makes
www.aka.ms/InterpretML-github
SHAP Pros SHAP Cons
Based on a solid theory and distributes the
effects fairly
Computation time: 2k possible coalitions of the
feature values for k features
Contrastive explanations: Instead of comparing a
prediction to the average prediction of the entire
dataset, you could compare it to a subset or even
to a single data point.
Can be misinterpreted
Inclusion of unrealistic data instances when
features are correlated.
www.aka.ms/InterpretML-github
Interpretability Approaches
• Models inherently interpretable
• Typically provide exact explanations
• Examples: Linear Models, Decision Rule Lists,
Explainable Boosting Machines, Shallow
Decision Trees, etc.
Tools to explain the system
www.aka.ms/InterpretML-github
Linear Models
• (Generalized) Linear Models
• Current standard for interpretable
models
• Learns an additive relationship
between data and response:
• y = β0 + β1x1 + β2x2 + ... + βn xn
• βi terms are scalar – single number
that multiplies to each feature value
• Each feature contributes a “score”
that adds up to final output
Example Model :
House Price = 50,000 + 0.2*sq_ft +
10000*bedrooms +
7000*bathrooms
Using this model on a new house:
- 5 Bedrooms
- 3 Bathrooms
- 3000 Sq Feet
- House Price = 50000 + 0.2*3000
+ 10000*5 + 7000 * 3
- House Price = $121,600
www.aka.ms/InterpretML-github
Explainable Boosting Machine
• Generalized Additive Models (with Pairwise
Interactions)
• GAMs have existed since 1980s
• MSR invented new methods of training GAMs for
higher accuracy
• Think of it as a “more expressive” linear model
• Still additive!
• Linear Model: y = β0 + β1x1 + β2x2 + ... + βn xn
• Additive Model: y = f1(x1) + f2(x2) + ... + fn (xn)
• Additive Model with Pairwise Interactions
(EBM):
y = Ʃi fi (xi) + Ʃij fij (xi , xj )
• Full Complexity Model: y = f (x1, ..., xn)
Linear Model can only learn a
straight line!
www.aka.ms/InterpretML-github
FAIRNESS
More positive outcomes & avoiding
harmful outcomes of AI systems for
groups of people
• Allocation: extends or withholds opportunities, resources, or information.
• Quality of service: whether a system works as well for one person as it does
for another
• Stereotyping: reinforce existing societal stereotypes
• Denigration: actively derogatory or offensive
• Over or under representation: over-represent, under-represent, or even
erase particular groups of people
Crawford et al. 2017
Types of harm
What is
Fairlearn?
A new approach to measuring and mitigating
unfairness in systems that make predictions, serve
users, or make decisions about allocating resources,
opportunities, or information.
www.aka.ms/FairlearnAI
There are many ways that an AI system can behave unfairly.
Fairness in AI
Avoiding negative outcomes of AI systems for different groups of people
A model for screening loan or job application
might be much better at picking good candidates
among white men than among other groups.
A voice recognition system might
fail to work as well for women as it
does for men.
www.aka.ms/FairlearnAI
A toolkit that empowers developers of artificial intelligence systems to
assess their systems' fairness and mitigate any observed fairness issues.
Helps users identify and mitigate unfairness in their machine learning
models with a focus on group fairness.
Automatically
analyze a model’s
predictions
Provide the user with
insights into (un)fairness of
their model’s predictions
Support (algorithmic)
methods to mitigate
unfairness
www.aka.ms/FairlearnAI
• Allocation: extends or withholds opportunities, resources, or information.
• Quality of service: whether a system works as well for one person as it does
for another
• Stereotyping: reinforce existing societal stereotypes
• Denigration: actively derogatory or offensive
• Over or under representation: over-represent, under-represent, or even
erase particular groups of people
Crawford et al. 2017
Types of harm addressed by Fairlearn
www.aka.ms/FairlearnAI
Harm of “allocation”
Example Scenarios: Lending
The data set describes whether each individual repaid the loan or not.
• [Classification] Recommends whether a given individual should get a loan. The
trained model outputs: Yes/Maybe/No decision.
Example Scenarios: School Admissions
The data set describes what was the first-year GPA of each student.
• [Regression] For a given applicant, predicts their GPA at the end of the first year.
The trained model outputs a real-valued prediction that is used as a score to
screen applicants.
www.aka.ms/FairlearnAI
Harm of “quality of service”
Example Scenarios: News Recommendation
The training data indicates what article was presented to which user, whether the user
clicked, and how much time the user spent on the article.
• [Classification] Predict which article to show to each user to optimize click-
through rate. The trained model outputs: Yes/No decision.
Two kinds of (group) fairness: across users (quality of service), or across
publishers/topics of articles (quality of service, but also allocation).
www.aka.ms/FairlearnAI
Assessment
Evaluation
Visualizations
Fairness Criteria
Mitigation Algorithm
Mitigation
• Select a protected attribute
• Select a fairness metric (e.g., disparity in recall, precision, etc
and disparity in selection rate)
Fairlearn Toolkit
• Demographic Parity
• Equalized Odds
• Postprocessing Method
• Reduction Methodwww.aka.ms/FairlearnAI
Fairness assessment
through disparity metrics
Disparity in Performance
• How the model accuracy differs across different buckets of a sensitive feature
(e.g., how accuracy of the model differs for "female" vs. "male" vs. “unspecified"
data points)
Disparity in Selection Rate
• How the model predictions differ across different buckets of a sensitive feature
(e.g., how many "female" individuals have received prediction `approved` on their
loan application in contrast to "male" and “unspecified" data points?).
www.aka.ms/FairlearnAI
Assessment
Evaluation
Visualizations
Fairness Criteria
Mitigation Algorithm
Mitigation
• Select a protected attribute
• Select a fairness metric (e.g., disparity in recall, precision, etc
and disparity in selection rate)
Fairlearn Toolkit
• Demographic Parity
• Equalized Odds
• Postprocessing Method
• Reduction Methodwww.aka.ms/FairlearnAI
Demographic parity:
Applicants of each race (gender, ...) have the
same odds of getting approval on their loan applications
Loan approval decision is independent of protected attribute
Equalized odds:
Qualified applicants have the same odds of getting approval on their
loan applications regardless of race (gender, …)
Unqualified applicants have the same odds of getting approval on their
loan applications regardless of race (gender, …)
Fairness Criteria
www.aka.ms/FairlearnAI
Reductions
approach:
Wrapper
around
standard ML
algorithms
Input:
• any standard ML training algorithm (as a
black box)
• data set including sensitive feature
Output:
• a trained model that minimizes error
subject to fairness constraints
Advantages:
• doesn’t need to access the sensitive
feature at test time
• works for a wide range of disparity metrics
• allows extracting the full disparity-accuracy
frontier
Disadvantages:
• requires re-training: the black box is called
10-20 times
www.aka.ms/FairlearnAI
Post
processing:
Picking a fair
threshold rule
Input:
• an existing (already trained) scoring model
• data set including sensitive feature
Output:
• the most accurate among all fair threshold rules (a
separate threshold for each subpopulation)
Advantages:
• simplicity
• no need to re-train the model
Disadvantages:
• requires sensitive feature at test-time
• doesn’t allow trade-offs between disparity and
accuracy
www.aka.ms/FairlearnAI
Demo of
Interpretability for
Tabular Data in
Azure Databricks
Hardware Performance Dataset - The
Regression goal is to predict the
performance of certain combinations
of hardware parts.
www.aka.ms/InterpretML-github
www.aka.ms/DatabricksRuntimeML
Azure ML
ServeStore Prep and trainIngest
Power BIAzure Data Factory Azure Data Lake Storage
Azure Kubernetes Service
Model Serving
Cosmos DB, SQL DB
Operational Databases
Azure Synapse Analytics
Streaming data
Azure Databricks
Azure Machine
Learning
Apps
Batch data
Ad-hoc analysis
Azure Event Hubs
Reference architecture
Provisioning Azure Databricks WORKSPACE
© Microsoft Corporation
Fast, easy, and collaborative Apache Spark-based analytics platform
Azure Databricks for machine learning modeling
Tools InfrastructureFrameworks
Leverage powerful GPU-enabled VMs
pre-configured for deep neural
network training
Use HorovodEstimator via a native
runtime to enable build deep learning
models with a few lines of code
Full Python and Scala support for
transfer learning on images
Automatically store metadata in
Azure Database with geo-replication
for fault tolerance
Use built-in hyperparameter tuning
via Spark MLLib to quickly drive
model progress
Simultaneously collaborate within
notebooks environments to streamline
model development
Load images natively in Spark
DataFrames to automatically decode
them for manipulation at scale
Improve performance 10x-100x over
traditional Spark deployments with
an optimized environment
Seamlessly use TensorFlow, Microsoft
Cognitive Toolkit, Caffe2, Keras, and more
© Microsoft Corporation
www.aka.ms/DatabricksRuntimeML
Demo dataset
Breast Cancer Wisconsin
(Diagnostic) Data Set
Features are computed
from a digitized image of a
fine needle aspirate (FNA)
of a breast mass. They
describe characteristics of
the cell nuclei present in
the image. A few of the
images can be found
at [Web Link]
Attribute Information:
1) ID number
2) Diagnosis (M = malignant, B = benign)
3-32)
Ten real-valued features are computed for each cell
nucleus:
a) radius (mean of distances from center to points on the
perimeter)
b) texture (standard deviation of gray-scale values)
c) perimeter
d) area
e) smoothness (local variation in radius lengths)
f) compactness (perimeter^2 / area - 1.0)
g) concavity (severity of concave portions of the contour)
h) concave points (number of concave portions of the
contour)
i) symmetry
j) fractal dimension ("coastline approximation" - 1)
© Microsoft Corporation
Install azureml-interpret and azureml-contrib-interpret
packages.
© Microsoft Corporation
Train a sample model in a local Jupyter notebook.
© Microsoft Corporation
Call the explainer locally.
© Microsoft Corporation
Explain the entire model behavior (global explanation)
© Microsoft Corporation
© Microsoft Corporation
Explain an individual prediction (local explanation)
© Microsoft Corporation
© Microsoft Corporation
© Microsoft Corporation
Mehrnoosh Sameki
Senior Product Manager
Minsoo Thigpen
Product Manager
Ehi Nosakhare
Machine Learning Scientist
Article:
Machine Fairness
How to assess AI system’s fairness
and mitigate any observed
unfairness issues
www.aka.ms/Fairlearn-build2020
Resources
➢www.aka.ms/InterpretML-github
➢www.aka.ms/InterpretML-toolkit
➢www.aka.ms/InterpretML-doc
➢www.aka.ms/FairlearnAI
➢www.aka.ms/FairlearnAIDoc
➢www.aka.ms/ResponsibleMLDoc
➢www.aka.ms/AzureResponsibleML
➢www.aka.ms/AzureMLDoc
➢www.aka.ms/AzureMLservice
Thank You!

Contenu connexe

Tendances

Large-scale Parallel Collaborative Filtering and Clustering using MapReduce f...
Large-scale Parallel Collaborative Filtering and Clustering using MapReduce f...Large-scale Parallel Collaborative Filtering and Clustering using MapReduce f...
Large-scale Parallel Collaborative Filtering and Clustering using MapReduce f...Varad Meru
 
Building a Predictive Model
Building a Predictive ModelBuilding a Predictive Model
Building a Predictive ModelDKALab
 
Applications of Machine Learning
Applications of Machine LearningApplications of Machine Learning
Applications of Machine LearningHayim Makabee
 
End-to-End Machine Learning Project
End-to-End Machine Learning ProjectEnd-to-End Machine Learning Project
End-to-End Machine Learning ProjectEng Teong Cheah
 
Explainability and bias in AI
Explainability and bias in AIExplainability and bias in AI
Explainability and bias in AIBill Liu
 
Collaborative Filtering 2: Item-based CF
Collaborative Filtering 2: Item-based CFCollaborative Filtering 2: Item-based CF
Collaborative Filtering 2: Item-based CFYusuke Yamamoto
 
Recsys2016 Tutorial by Xavier and Deepak
Recsys2016 Tutorial by Xavier and DeepakRecsys2016 Tutorial by Xavier and Deepak
Recsys2016 Tutorial by Xavier and DeepakDeepak Agarwal
 
Matrix Factorization Techniques For Recommender Systems
Matrix Factorization Techniques For Recommender SystemsMatrix Factorization Techniques For Recommender Systems
Matrix Factorization Techniques For Recommender SystemsLei Guo
 
Matrix Factorization Technique for Recommender Systems
Matrix Factorization Technique for Recommender SystemsMatrix Factorization Technique for Recommender Systems
Matrix Factorization Technique for Recommender SystemsAladejubelo Oluwashina
 
A Research Platform for Coevolving Agents.doc
A Research Platform for Coevolving Agents.docA Research Platform for Coevolving Agents.doc
A Research Platform for Coevolving Agents.docbutest
 
Information Retrieval Models for Recommender Systems - PhD slides
Information Retrieval Models for Recommender Systems - PhD slidesInformation Retrieval Models for Recommender Systems - PhD slides
Information Retrieval Models for Recommender Systems - PhD slidesDaniel Valcarce
 
Recommendation systems
Recommendation systems  Recommendation systems
Recommendation systems Badr Hirchoua
 
REAL-TIME RECOMMENDATION SYSTEMS
REAL-TIME RECOMMENDATION SYSTEMS REAL-TIME RECOMMENDATION SYSTEMS
REAL-TIME RECOMMENDATION SYSTEMS BigDataCloud
 
Machine Learning for Recommender Systems MLSS 2015 Sydney
Machine Learning for Recommender Systems MLSS 2015 SydneyMachine Learning for Recommender Systems MLSS 2015 Sydney
Machine Learning for Recommender Systems MLSS 2015 SydneyAlexandros Karatzoglou
 
Movie Recommender System Using Artificial Intelligence
Movie Recommender System Using Artificial Intelligence Movie Recommender System Using Artificial Intelligence
Movie Recommender System Using Artificial Intelligence Shrutika Oswal
 
A Hybrid Recommendation system
A Hybrid Recommendation systemA Hybrid Recommendation system
A Hybrid Recommendation systemPranav Prakash
 
Item Based Collaborative Filtering Recommendation Algorithms
Item Based Collaborative Filtering Recommendation AlgorithmsItem Based Collaborative Filtering Recommendation Algorithms
Item Based Collaborative Filtering Recommendation Algorithmsnextlib
 

Tendances (19)

Large-scale Parallel Collaborative Filtering and Clustering using MapReduce f...
Large-scale Parallel Collaborative Filtering and Clustering using MapReduce f...Large-scale Parallel Collaborative Filtering and Clustering using MapReduce f...
Large-scale Parallel Collaborative Filtering and Clustering using MapReduce f...
 
Building a Predictive Model
Building a Predictive ModelBuilding a Predictive Model
Building a Predictive Model
 
Applications of Machine Learning
Applications of Machine LearningApplications of Machine Learning
Applications of Machine Learning
 
End-to-End Machine Learning Project
End-to-End Machine Learning ProjectEnd-to-End Machine Learning Project
End-to-End Machine Learning Project
 
Explainability and bias in AI
Explainability and bias in AIExplainability and bias in AI
Explainability and bias in AI
 
Collaborative Filtering 2: Item-based CF
Collaborative Filtering 2: Item-based CFCollaborative Filtering 2: Item-based CF
Collaborative Filtering 2: Item-based CF
 
Recsys2016 Tutorial by Xavier and Deepak
Recsys2016 Tutorial by Xavier and DeepakRecsys2016 Tutorial by Xavier and Deepak
Recsys2016 Tutorial by Xavier and Deepak
 
Matrix Factorization Techniques For Recommender Systems
Matrix Factorization Techniques For Recommender SystemsMatrix Factorization Techniques For Recommender Systems
Matrix Factorization Techniques For Recommender Systems
 
Matrix Factorization Technique for Recommender Systems
Matrix Factorization Technique for Recommender SystemsMatrix Factorization Technique for Recommender Systems
Matrix Factorization Technique for Recommender Systems
 
A Research Platform for Coevolving Agents.doc
A Research Platform for Coevolving Agents.docA Research Platform for Coevolving Agents.doc
A Research Platform for Coevolving Agents.doc
 
Btp 3rd Report
Btp 3rd ReportBtp 3rd Report
Btp 3rd Report
 
Information Retrieval Models for Recommender Systems - PhD slides
Information Retrieval Models for Recommender Systems - PhD slidesInformation Retrieval Models for Recommender Systems - PhD slides
Information Retrieval Models for Recommender Systems - PhD slides
 
Recommendation systems
Recommendation systems  Recommendation systems
Recommendation systems
 
REAL-TIME RECOMMENDATION SYSTEMS
REAL-TIME RECOMMENDATION SYSTEMS REAL-TIME RECOMMENDATION SYSTEMS
REAL-TIME RECOMMENDATION SYSTEMS
 
[0201699613]visual modeling with rational rose 2000 and uml
[0201699613]visual modeling with rational rose 2000 and uml[0201699613]visual modeling with rational rose 2000 and uml
[0201699613]visual modeling with rational rose 2000 and uml
 
Machine Learning for Recommender Systems MLSS 2015 Sydney
Machine Learning for Recommender Systems MLSS 2015 SydneyMachine Learning for Recommender Systems MLSS 2015 Sydney
Machine Learning for Recommender Systems MLSS 2015 Sydney
 
Movie Recommender System Using Artificial Intelligence
Movie Recommender System Using Artificial Intelligence Movie Recommender System Using Artificial Intelligence
Movie Recommender System Using Artificial Intelligence
 
A Hybrid Recommendation system
A Hybrid Recommendation systemA Hybrid Recommendation system
A Hybrid Recommendation system
 
Item Based Collaborative Filtering Recommendation Algorithms
Item Based Collaborative Filtering Recommendation AlgorithmsItem Based Collaborative Filtering Recommendation Algorithms
Item Based Collaborative Filtering Recommendation Algorithms
 

Similaire à Assessing and Mitigating Unfairness in AI Systems

The importance of model fairness and interpretability in AI systems
The importance of model fairness and interpretability in AI systemsThe importance of model fairness and interpretability in AI systems
The importance of model fairness and interpretability in AI systemsFrancesca Lazzeri, PhD
 
Human in the loop: Bayesian Rules Enabling Explainable AI
Human in the loop: Bayesian Rules Enabling Explainable AIHuman in the loop: Bayesian Rules Enabling Explainable AI
Human in the loop: Bayesian Rules Enabling Explainable AIPramit Choudhary
 
Towards Responsible AI - KC.pptx
Towards Responsible AI - KC.pptxTowards Responsible AI - KC.pptx
Towards Responsible AI - KC.pptxLuis775803
 
Machine Learning with Python- Methods for Machine Learning.pptx
Machine Learning with Python- Methods for Machine Learning.pptxMachine Learning with Python- Methods for Machine Learning.pptx
Machine Learning with Python- Methods for Machine Learning.pptxiaeronlineexm
 
Collaborative Filtering Recommendation System
Collaborative Filtering Recommendation SystemCollaborative Filtering Recommendation System
Collaborative Filtering Recommendation SystemMilind Gokhale
 
Towards Responsible AI - NY.pptx
Towards Responsible AI - NY.pptxTowards Responsible AI - NY.pptx
Towards Responsible AI - NY.pptxLuis775803
 
Practical Explainable AI: How to build trustworthy, transparent and unbiased ...
Practical Explainable AI: How to build trustworthy, transparent and unbiased ...Practical Explainable AI: How to build trustworthy, transparent and unbiased ...
Practical Explainable AI: How to build trustworthy, transparent and unbiased ...Raheel Ahmad
 
Introduction to machine learning
Introduction to machine learningIntroduction to machine learning
Introduction to machine learningSanghamitra Deb
 
Dad (Data Analysis And Design)
Dad (Data Analysis And Design)Dad (Data Analysis And Design)
Dad (Data Analysis And Design)Jill Lyons
 
Responsible AI in Industry (ICML 2021 Tutorial)
Responsible AI in Industry (ICML 2021 Tutorial)Responsible AI in Industry (ICML 2021 Tutorial)
Responsible AI in Industry (ICML 2021 Tutorial)Krishnaram Kenthapadi
 
Supervised learning techniques and applications
Supervised learning techniques and applicationsSupervised learning techniques and applications
Supervised learning techniques and applicationsBenjaminlapid1
 
Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!
Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!
Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!Sri Ambati
 
Responsible AI in Industry: Practical Challenges and Lessons Learned
Responsible AI in Industry: Practical Challenges and Lessons LearnedResponsible AI in Industry: Practical Challenges and Lessons Learned
Responsible AI in Industry: Practical Challenges and Lessons LearnedKrishnaram Kenthapadi
 
Understanding Products Driven by Machine Learning and AI: A Data Scientist's ...
Understanding Products Driven by Machine Learning and AI: A Data Scientist's ...Understanding Products Driven by Machine Learning and AI: A Data Scientist's ...
Understanding Products Driven by Machine Learning and AI: A Data Scientist's ...Amanda Casari
 
A Fairness-aware Machine Learning Interface for End-to-end Discrimination Dis...
A Fairness-aware Machine Learning Interface for End-to-end Discrimination Dis...A Fairness-aware Machine Learning Interface for End-to-end Discrimination Dis...
A Fairness-aware Machine Learning Interface for End-to-end Discrimination Dis...wajrcs
 
The Incredible Disappearing Data Scientist
The Incredible Disappearing Data ScientistThe Incredible Disappearing Data Scientist
The Incredible Disappearing Data ScientistRebecca Bilbro
 

Similaire à Assessing and Mitigating Unfairness in AI Systems (20)

The importance of model fairness and interpretability in AI systems
The importance of model fairness and interpretability in AI systemsThe importance of model fairness and interpretability in AI systems
The importance of model fairness and interpretability in AI systems
 
Human in the loop: Bayesian Rules Enabling Explainable AI
Human in the loop: Bayesian Rules Enabling Explainable AIHuman in the loop: Bayesian Rules Enabling Explainable AI
Human in the loop: Bayesian Rules Enabling Explainable AI
 
Towards Responsible AI - KC.pptx
Towards Responsible AI - KC.pptxTowards Responsible AI - KC.pptx
Towards Responsible AI - KC.pptx
 
Ai in finance
Ai in financeAi in finance
Ai in finance
 
Machine Learning with Python- Methods for Machine Learning.pptx
Machine Learning with Python- Methods for Machine Learning.pptxMachine Learning with Python- Methods for Machine Learning.pptx
Machine Learning with Python- Methods for Machine Learning.pptx
 
Introduction to Machine Learning
Introduction to Machine LearningIntroduction to Machine Learning
Introduction to Machine Learning
 
Collaborative Filtering Recommendation System
Collaborative Filtering Recommendation SystemCollaborative Filtering Recommendation System
Collaborative Filtering Recommendation System
 
Towards Responsible AI - NY.pptx
Towards Responsible AI - NY.pptxTowards Responsible AI - NY.pptx
Towards Responsible AI - NY.pptx
 
Machine learning
Machine learningMachine learning
Machine learning
 
Practical Explainable AI: How to build trustworthy, transparent and unbiased ...
Practical Explainable AI: How to build trustworthy, transparent and unbiased ...Practical Explainable AI: How to build trustworthy, transparent and unbiased ...
Practical Explainable AI: How to build trustworthy, transparent and unbiased ...
 
Classification
ClassificationClassification
Classification
 
Introduction to machine learning
Introduction to machine learningIntroduction to machine learning
Introduction to machine learning
 
Dad (Data Analysis And Design)
Dad (Data Analysis And Design)Dad (Data Analysis And Design)
Dad (Data Analysis And Design)
 
Responsible AI in Industry (ICML 2021 Tutorial)
Responsible AI in Industry (ICML 2021 Tutorial)Responsible AI in Industry (ICML 2021 Tutorial)
Responsible AI in Industry (ICML 2021 Tutorial)
 
Supervised learning techniques and applications
Supervised learning techniques and applicationsSupervised learning techniques and applications
Supervised learning techniques and applications
 
Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!
Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!
Get hands-on with Explainable AI at Machine Learning Interpretability(MLI) Gym!
 
Responsible AI in Industry: Practical Challenges and Lessons Learned
Responsible AI in Industry: Practical Challenges and Lessons LearnedResponsible AI in Industry: Practical Challenges and Lessons Learned
Responsible AI in Industry: Practical Challenges and Lessons Learned
 
Understanding Products Driven by Machine Learning and AI: A Data Scientist's ...
Understanding Products Driven by Machine Learning and AI: A Data Scientist's ...Understanding Products Driven by Machine Learning and AI: A Data Scientist's ...
Understanding Products Driven by Machine Learning and AI: A Data Scientist's ...
 
A Fairness-aware Machine Learning Interface for End-to-end Discrimination Dis...
A Fairness-aware Machine Learning Interface for End-to-end Discrimination Dis...A Fairness-aware Machine Learning Interface for End-to-end Discrimination Dis...
A Fairness-aware Machine Learning Interface for End-to-end Discrimination Dis...
 
The Incredible Disappearing Data Scientist
The Incredible Disappearing Data ScientistThe Incredible Disappearing Data Scientist
The Incredible Disappearing Data Scientist
 

Dernier

Kantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdf
Kantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdfKantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdf
Kantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdfSocial Samosa
 
20240419 - Measurecamp Amsterdam - SAM.pdf
20240419 - Measurecamp Amsterdam - SAM.pdf20240419 - Measurecamp Amsterdam - SAM.pdf
20240419 - Measurecamp Amsterdam - SAM.pdfHuman37
 
1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样
1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样
1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样vhwb25kk
 
2006_GasProcessing_HB (1).pdf HYDROCARBON PROCESSING
2006_GasProcessing_HB (1).pdf HYDROCARBON PROCESSING2006_GasProcessing_HB (1).pdf HYDROCARBON PROCESSING
2006_GasProcessing_HB (1).pdf HYDROCARBON PROCESSINGmarianagonzalez07
 
MK KOMUNIKASI DATA (TI)komdat komdat.docx
MK KOMUNIKASI DATA (TI)komdat komdat.docxMK KOMUNIKASI DATA (TI)komdat komdat.docx
MK KOMUNIKASI DATA (TI)komdat komdat.docxUnduhUnggah1
 
Top 5 Best Data Analytics Courses In Queens
Top 5 Best Data Analytics Courses In QueensTop 5 Best Data Analytics Courses In Queens
Top 5 Best Data Analytics Courses In Queensdataanalyticsqueen03
 
How we prevented account sharing with MFA
How we prevented account sharing with MFAHow we prevented account sharing with MFA
How we prevented account sharing with MFAAndrei Kaleshka
 
9654467111 Call Girls In Munirka Hotel And Home Service
9654467111 Call Girls In Munirka Hotel And Home Service9654467111 Call Girls In Munirka Hotel And Home Service
9654467111 Call Girls In Munirka Hotel And Home ServiceSapana Sha
 
办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一
办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一
办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一F sss
 
Easter Eggs From Star Wars and in cars 1 and 2
Easter Eggs From Star Wars and in cars 1 and 2Easter Eggs From Star Wars and in cars 1 and 2
Easter Eggs From Star Wars and in cars 1 and 217djon017
 
Predicting Salary Using Data Science: A Comprehensive Analysis.pdf
Predicting Salary Using Data Science: A Comprehensive Analysis.pdfPredicting Salary Using Data Science: A Comprehensive Analysis.pdf
Predicting Salary Using Data Science: A Comprehensive Analysis.pdfBoston Institute of Analytics
 
Call Us ➥97111√47426🤳Call Girls in Aerocity (Delhi NCR)
Call Us ➥97111√47426🤳Call Girls in Aerocity (Delhi NCR)Call Us ➥97111√47426🤳Call Girls in Aerocity (Delhi NCR)
Call Us ➥97111√47426🤳Call Girls in Aerocity (Delhi NCR)jennyeacort
 
Defining Constituents, Data Vizzes and Telling a Data Story
Defining Constituents, Data Vizzes and Telling a Data StoryDefining Constituents, Data Vizzes and Telling a Data Story
Defining Constituents, Data Vizzes and Telling a Data StoryJeremy Anderson
 
Customer Service Analytics - Make Sense of All Your Data.pptx
Customer Service Analytics - Make Sense of All Your Data.pptxCustomer Service Analytics - Make Sense of All Your Data.pptx
Customer Service Analytics - Make Sense of All Your Data.pptxEmmanuel Dauda
 
Call Girls In Dwarka 9654467111 Escorts Service
Call Girls In Dwarka 9654467111 Escorts ServiceCall Girls In Dwarka 9654467111 Escorts Service
Call Girls In Dwarka 9654467111 Escorts ServiceSapana Sha
 
Call Girls in Defence Colony Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Defence Colony Delhi 💯Call Us 🔝8264348440🔝Call Girls in Defence Colony Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Defence Colony Delhi 💯Call Us 🔝8264348440🔝soniya singh
 
DBA Basics: Getting Started with Performance Tuning.pdf
DBA Basics: Getting Started with Performance Tuning.pdfDBA Basics: Getting Started with Performance Tuning.pdf
DBA Basics: Getting Started with Performance Tuning.pdfJohn Sterrett
 
INTERNSHIP ON PURBASHA COMPOSITE TEX LTD
INTERNSHIP ON PURBASHA COMPOSITE TEX LTDINTERNSHIP ON PURBASHA COMPOSITE TEX LTD
INTERNSHIP ON PURBASHA COMPOSITE TEX LTDRafezzaman
 
Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...
Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...
Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...Sapana Sha
 
原版1:1定制南十字星大学毕业证(SCU毕业证)#文凭成绩单#真实留信学历认证永久存档
原版1:1定制南十字星大学毕业证(SCU毕业证)#文凭成绩单#真实留信学历认证永久存档原版1:1定制南十字星大学毕业证(SCU毕业证)#文凭成绩单#真实留信学历认证永久存档
原版1:1定制南十字星大学毕业证(SCU毕业证)#文凭成绩单#真实留信学历认证永久存档208367051
 

Dernier (20)

Kantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdf
Kantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdfKantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdf
Kantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdf
 
20240419 - Measurecamp Amsterdam - SAM.pdf
20240419 - Measurecamp Amsterdam - SAM.pdf20240419 - Measurecamp Amsterdam - SAM.pdf
20240419 - Measurecamp Amsterdam - SAM.pdf
 
1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样
1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样
1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样
 
2006_GasProcessing_HB (1).pdf HYDROCARBON PROCESSING
2006_GasProcessing_HB (1).pdf HYDROCARBON PROCESSING2006_GasProcessing_HB (1).pdf HYDROCARBON PROCESSING
2006_GasProcessing_HB (1).pdf HYDROCARBON PROCESSING
 
MK KOMUNIKASI DATA (TI)komdat komdat.docx
MK KOMUNIKASI DATA (TI)komdat komdat.docxMK KOMUNIKASI DATA (TI)komdat komdat.docx
MK KOMUNIKASI DATA (TI)komdat komdat.docx
 
Top 5 Best Data Analytics Courses In Queens
Top 5 Best Data Analytics Courses In QueensTop 5 Best Data Analytics Courses In Queens
Top 5 Best Data Analytics Courses In Queens
 
How we prevented account sharing with MFA
How we prevented account sharing with MFAHow we prevented account sharing with MFA
How we prevented account sharing with MFA
 
9654467111 Call Girls In Munirka Hotel And Home Service
9654467111 Call Girls In Munirka Hotel And Home Service9654467111 Call Girls In Munirka Hotel And Home Service
9654467111 Call Girls In Munirka Hotel And Home Service
 
办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一
办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一
办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一
 
Easter Eggs From Star Wars and in cars 1 and 2
Easter Eggs From Star Wars and in cars 1 and 2Easter Eggs From Star Wars and in cars 1 and 2
Easter Eggs From Star Wars and in cars 1 and 2
 
Predicting Salary Using Data Science: A Comprehensive Analysis.pdf
Predicting Salary Using Data Science: A Comprehensive Analysis.pdfPredicting Salary Using Data Science: A Comprehensive Analysis.pdf
Predicting Salary Using Data Science: A Comprehensive Analysis.pdf
 
Call Us ➥97111√47426🤳Call Girls in Aerocity (Delhi NCR)
Call Us ➥97111√47426🤳Call Girls in Aerocity (Delhi NCR)Call Us ➥97111√47426🤳Call Girls in Aerocity (Delhi NCR)
Call Us ➥97111√47426🤳Call Girls in Aerocity (Delhi NCR)
 
Defining Constituents, Data Vizzes and Telling a Data Story
Defining Constituents, Data Vizzes and Telling a Data StoryDefining Constituents, Data Vizzes and Telling a Data Story
Defining Constituents, Data Vizzes and Telling a Data Story
 
Customer Service Analytics - Make Sense of All Your Data.pptx
Customer Service Analytics - Make Sense of All Your Data.pptxCustomer Service Analytics - Make Sense of All Your Data.pptx
Customer Service Analytics - Make Sense of All Your Data.pptx
 
Call Girls In Dwarka 9654467111 Escorts Service
Call Girls In Dwarka 9654467111 Escorts ServiceCall Girls In Dwarka 9654467111 Escorts Service
Call Girls In Dwarka 9654467111 Escorts Service
 
Call Girls in Defence Colony Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Defence Colony Delhi 💯Call Us 🔝8264348440🔝Call Girls in Defence Colony Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Defence Colony Delhi 💯Call Us 🔝8264348440🔝
 
DBA Basics: Getting Started with Performance Tuning.pdf
DBA Basics: Getting Started with Performance Tuning.pdfDBA Basics: Getting Started with Performance Tuning.pdf
DBA Basics: Getting Started with Performance Tuning.pdf
 
INTERNSHIP ON PURBASHA COMPOSITE TEX LTD
INTERNSHIP ON PURBASHA COMPOSITE TEX LTDINTERNSHIP ON PURBASHA COMPOSITE TEX LTD
INTERNSHIP ON PURBASHA COMPOSITE TEX LTD
 
Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...
Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...
Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...
 
原版1:1定制南十字星大学毕业证(SCU毕业证)#文凭成绩单#真实留信学历认证永久存档
原版1:1定制南十字星大学毕业证(SCU毕业证)#文凭成绩单#真实留信学历认证永久存档原版1:1定制南十字星大学毕业证(SCU毕业证)#文凭成绩单#真实留信学历认证永久存档
原版1:1定制南十字星大学毕业证(SCU毕业证)#文凭成绩单#真实留信学历认证永久存档
 

Assessing and Mitigating Unfairness in AI Systems

  • 1.
  • 2. The Importance of Model Fairness and Interpretability in AI Systems Spark Summit 2020
  • 3. Machine Fairness How to assess AI system’s fairness and mitigate any observed unfairness issues www.aka.ms/Fairlearn-build2020
  • 6. AI will have a considerable impact on business and society as a whole
  • 10. Transparency AI systems should be explainable AI systems should have algorithmic accountability
  • 11. Machine Learning Interpretability use cases www.aka.ms/InterpretML-github
  • 13. Explain Models - www.aka.ms/InterpretML-toolkit
  • 14. InterpretML Repo - www.aka.ms/InterpretML-github Interpret Tabular Data Additional interpretability techniques for tabular extension www.aka.ms/InterpretML-github github/InterpretML/Interpret-Community Interpret- Community
  • 15. Interpret Tabular Data Additional interpretability techniques for tabular extension github/InterpretML/Interpret github/InterpretML/Interpret-Community Interpret- Community AzureML- Interpret wrapper Azure Machine Learning SDK InterpretML Repo - www.aka.ms/InterpretML-github
  • 16. Interpret Tabular Data Additional interpretability techniques for tabular extension github/InterpretML/Interpret github/InterpretML/Interpret-Community Interpret- Community Interpret- text-contrib github/InterpretML/Interpret-text-contrib Text Data (Text Classification) AzureML- Interpret wrapper Azure Machine Learning SDK InterpretML Repo - www.aka.ms/InterpretML-github
  • 18. Interpretability Approaches • Given an existing model/system, extract explanations • Typically provide approximate explanations • Examples: LIME, SHAP, Permutation Feature Importance, Sensitivity Analysis, Influence Functions, etc. Tools to explain the system www.aka.ms/InterpretML-github
  • 19. SHAP (SHapley Additive exPlanations) • Not a new concept • Concept based on game theory • Mathematically sound • Application in ML relatively new www.aka.ms/InterpretML-github
  • 21. How much has each feature contributed to the prediction compared to the average prediction? House price prediction: € 300,000 Average house price prediction for all apartments is € 310,000 Delta here is - €10,000 www.aka.ms/InterpretML-github
  • 22. Parks contributed +10k Cats banned contributed -50k 50𝑚2 contributed +10k 2nd floor had net zero contributions www.aka.ms/InterpretML-github
  • 23. SHAP How do we calculate Shapley values? • Take your feature of interest (e.g., cats-banned) and remove it from the feature set • Take the remaining features and generate all possible coalitions • Add and remove your feature of interested to each of the coalitions and calculate the difference it makes www.aka.ms/InterpretML-github
  • 24. SHAP Pros SHAP Cons Based on a solid theory and distributes the effects fairly Computation time: 2k possible coalitions of the feature values for k features Contrastive explanations: Instead of comparing a prediction to the average prediction of the entire dataset, you could compare it to a subset or even to a single data point. Can be misinterpreted Inclusion of unrealistic data instances when features are correlated. www.aka.ms/InterpretML-github
  • 25. Interpretability Approaches • Models inherently interpretable • Typically provide exact explanations • Examples: Linear Models, Decision Rule Lists, Explainable Boosting Machines, Shallow Decision Trees, etc. Tools to explain the system www.aka.ms/InterpretML-github
  • 26. Linear Models • (Generalized) Linear Models • Current standard for interpretable models • Learns an additive relationship between data and response: • y = β0 + β1x1 + β2x2 + ... + βn xn • βi terms are scalar – single number that multiplies to each feature value • Each feature contributes a “score” that adds up to final output Example Model : House Price = 50,000 + 0.2*sq_ft + 10000*bedrooms + 7000*bathrooms Using this model on a new house: - 5 Bedrooms - 3 Bathrooms - 3000 Sq Feet - House Price = 50000 + 0.2*3000 + 10000*5 + 7000 * 3 - House Price = $121,600 www.aka.ms/InterpretML-github
  • 27. Explainable Boosting Machine • Generalized Additive Models (with Pairwise Interactions) • GAMs have existed since 1980s • MSR invented new methods of training GAMs for higher accuracy • Think of it as a “more expressive” linear model • Still additive! • Linear Model: y = β0 + β1x1 + β2x2 + ... + βn xn • Additive Model: y = f1(x1) + f2(x2) + ... + fn (xn) • Additive Model with Pairwise Interactions (EBM): y = Ʃi fi (xi) + Ʃij fij (xi , xj ) • Full Complexity Model: y = f (x1, ..., xn) Linear Model can only learn a straight line! www.aka.ms/InterpretML-github
  • 28.
  • 29. FAIRNESS More positive outcomes & avoiding harmful outcomes of AI systems for groups of people
  • 30. • Allocation: extends or withholds opportunities, resources, or information. • Quality of service: whether a system works as well for one person as it does for another • Stereotyping: reinforce existing societal stereotypes • Denigration: actively derogatory or offensive • Over or under representation: over-represent, under-represent, or even erase particular groups of people Crawford et al. 2017 Types of harm
  • 31. What is Fairlearn? A new approach to measuring and mitigating unfairness in systems that make predictions, serve users, or make decisions about allocating resources, opportunities, or information. www.aka.ms/FairlearnAI
  • 32. There are many ways that an AI system can behave unfairly. Fairness in AI Avoiding negative outcomes of AI systems for different groups of people A model for screening loan or job application might be much better at picking good candidates among white men than among other groups. A voice recognition system might fail to work as well for women as it does for men. www.aka.ms/FairlearnAI
  • 33. A toolkit that empowers developers of artificial intelligence systems to assess their systems' fairness and mitigate any observed fairness issues. Helps users identify and mitigate unfairness in their machine learning models with a focus on group fairness. Automatically analyze a model’s predictions Provide the user with insights into (un)fairness of their model’s predictions Support (algorithmic) methods to mitigate unfairness www.aka.ms/FairlearnAI
  • 34. • Allocation: extends or withholds opportunities, resources, or information. • Quality of service: whether a system works as well for one person as it does for another • Stereotyping: reinforce existing societal stereotypes • Denigration: actively derogatory or offensive • Over or under representation: over-represent, under-represent, or even erase particular groups of people Crawford et al. 2017 Types of harm addressed by Fairlearn www.aka.ms/FairlearnAI
  • 35. Harm of “allocation” Example Scenarios: Lending The data set describes whether each individual repaid the loan or not. • [Classification] Recommends whether a given individual should get a loan. The trained model outputs: Yes/Maybe/No decision. Example Scenarios: School Admissions The data set describes what was the first-year GPA of each student. • [Regression] For a given applicant, predicts their GPA at the end of the first year. The trained model outputs a real-valued prediction that is used as a score to screen applicants. www.aka.ms/FairlearnAI
  • 36. Harm of “quality of service” Example Scenarios: News Recommendation The training data indicates what article was presented to which user, whether the user clicked, and how much time the user spent on the article. • [Classification] Predict which article to show to each user to optimize click- through rate. The trained model outputs: Yes/No decision. Two kinds of (group) fairness: across users (quality of service), or across publishers/topics of articles (quality of service, but also allocation). www.aka.ms/FairlearnAI
  • 37. Assessment Evaluation Visualizations Fairness Criteria Mitigation Algorithm Mitigation • Select a protected attribute • Select a fairness metric (e.g., disparity in recall, precision, etc and disparity in selection rate) Fairlearn Toolkit • Demographic Parity • Equalized Odds • Postprocessing Method • Reduction Methodwww.aka.ms/FairlearnAI
  • 38. Fairness assessment through disparity metrics Disparity in Performance • How the model accuracy differs across different buckets of a sensitive feature (e.g., how accuracy of the model differs for "female" vs. "male" vs. “unspecified" data points) Disparity in Selection Rate • How the model predictions differ across different buckets of a sensitive feature (e.g., how many "female" individuals have received prediction `approved` on their loan application in contrast to "male" and “unspecified" data points?). www.aka.ms/FairlearnAI
  • 39. Assessment Evaluation Visualizations Fairness Criteria Mitigation Algorithm Mitigation • Select a protected attribute • Select a fairness metric (e.g., disparity in recall, precision, etc and disparity in selection rate) Fairlearn Toolkit • Demographic Parity • Equalized Odds • Postprocessing Method • Reduction Methodwww.aka.ms/FairlearnAI
  • 40. Demographic parity: Applicants of each race (gender, ...) have the same odds of getting approval on their loan applications Loan approval decision is independent of protected attribute Equalized odds: Qualified applicants have the same odds of getting approval on their loan applications regardless of race (gender, …) Unqualified applicants have the same odds of getting approval on their loan applications regardless of race (gender, …) Fairness Criteria www.aka.ms/FairlearnAI
  • 41. Reductions approach: Wrapper around standard ML algorithms Input: • any standard ML training algorithm (as a black box) • data set including sensitive feature Output: • a trained model that minimizes error subject to fairness constraints Advantages: • doesn’t need to access the sensitive feature at test time • works for a wide range of disparity metrics • allows extracting the full disparity-accuracy frontier Disadvantages: • requires re-training: the black box is called 10-20 times www.aka.ms/FairlearnAI
  • 42. Post processing: Picking a fair threshold rule Input: • an existing (already trained) scoring model • data set including sensitive feature Output: • the most accurate among all fair threshold rules (a separate threshold for each subpopulation) Advantages: • simplicity • no need to re-train the model Disadvantages: • requires sensitive feature at test-time • doesn’t allow trade-offs between disparity and accuracy www.aka.ms/FairlearnAI
  • 43. Demo of Interpretability for Tabular Data in Azure Databricks Hardware Performance Dataset - The Regression goal is to predict the performance of certain combinations of hardware parts. www.aka.ms/InterpretML-github www.aka.ms/DatabricksRuntimeML
  • 44. Azure ML ServeStore Prep and trainIngest Power BIAzure Data Factory Azure Data Lake Storage Azure Kubernetes Service Model Serving Cosmos DB, SQL DB Operational Databases Azure Synapse Analytics Streaming data Azure Databricks Azure Machine Learning Apps Batch data Ad-hoc analysis Azure Event Hubs Reference architecture
  • 46. © Microsoft Corporation Fast, easy, and collaborative Apache Spark-based analytics platform Azure Databricks for machine learning modeling Tools InfrastructureFrameworks Leverage powerful GPU-enabled VMs pre-configured for deep neural network training Use HorovodEstimator via a native runtime to enable build deep learning models with a few lines of code Full Python and Scala support for transfer learning on images Automatically store metadata in Azure Database with geo-replication for fault tolerance Use built-in hyperparameter tuning via Spark MLLib to quickly drive model progress Simultaneously collaborate within notebooks environments to streamline model development Load images natively in Spark DataFrames to automatically decode them for manipulation at scale Improve performance 10x-100x over traditional Spark deployments with an optimized environment Seamlessly use TensorFlow, Microsoft Cognitive Toolkit, Caffe2, Keras, and more
  • 48. Demo dataset Breast Cancer Wisconsin (Diagnostic) Data Set Features are computed from a digitized image of a fine needle aspirate (FNA) of a breast mass. They describe characteristics of the cell nuclei present in the image. A few of the images can be found at [Web Link] Attribute Information: 1) ID number 2) Diagnosis (M = malignant, B = benign) 3-32) Ten real-valued features are computed for each cell nucleus: a) radius (mean of distances from center to points on the perimeter) b) texture (standard deviation of gray-scale values) c) perimeter d) area e) smoothness (local variation in radius lengths) f) compactness (perimeter^2 / area - 1.0) g) concavity (severity of concave portions of the contour) h) concave points (number of concave portions of the contour) i) symmetry j) fractal dimension ("coastline approximation" - 1)
  • 49. © Microsoft Corporation Install azureml-interpret and azureml-contrib-interpret packages.
  • 50. © Microsoft Corporation Train a sample model in a local Jupyter notebook.
  • 51. © Microsoft Corporation Call the explainer locally.
  • 52. © Microsoft Corporation Explain the entire model behavior (global explanation)
  • 54. © Microsoft Corporation Explain an individual prediction (local explanation)
  • 58. Mehrnoosh Sameki Senior Product Manager Minsoo Thigpen Product Manager Ehi Nosakhare Machine Learning Scientist
  • 59. Article: Machine Fairness How to assess AI system’s fairness and mitigate any observed unfairness issues www.aka.ms/Fairlearn-build2020