SlideShare une entreprise Scribd logo
1  sur  43
Algorithmic
Fairness:
A Brief
Introduction
Talk of AI and ethics is on the rise
2017
The Issues Perpetuating Discussion
What Really Caused it
to Explode?
Pro Publica and COMPAS
• Pro Publica
• Journalism
• Social Justice
• Audited COMPAS
• Northpoint
• COMPAS
• Recidivism Prediction
• “Cutting Edge” ML
• COMPAS Audit
• Was Calibrated
• Equal Accuracy
• Was Deployed
High Risk, No Offense
• Black: 45%
• White: 24%
Low Risk, Offense
• Black: 28%
• White: 48%
Problem: Disparate Impact
So why use these systems at all?
Probabilistic vs Causal Policies
Rain Dancing Problems (End Drought)
• Key Question: Does rain dancing cause rain?
• Solution: Causal Rule (If yes, then rain dance)
• Doesn’t require prediction of anything
Umbrella Problems (Stay Dry)
• Key Question: Will it rain today?
• Solution: Probabilistic cost/benefit (many variables)
• Requires prediction of rain, and knowledge of payouts
Takeaway: Some problems require prediction
for good results. ML predicts better than
other approaches.
Kleinberg, Jon. “Prediction Policy Problems.”
How do we get the best
of both?
First, we look at the difference?
Why are ML Policies Different from Causal Policies?
How Does
ML/AI Make
Decisions?
Risky
Not Risky
A Deeper Look: Probabilistic Decision Boundaries
Basics:
• (Un)Certainty of Cases
• Risk vs Class
• Red Errors Blue Errors
Takeaways:
• Estimated Functions aren’t
like known functions
• Different results for equally
deserving people/groups
• The uncertain space is most
interesting
• Upside
• Downside
Class
Boundary
Uncertain
Space
~0 < x < ~1
Creditworthy
Not CW
What Did Researchers Do in
Response to Their Dilemma?
They
Proposed a
New
Direction
Previous
Views/Strategies
New
Views/Strategies
Source of
Bias None
Human Biases
Garbage in, Garbage out
Protected
Features Unawareness Redundant Encodings
Fairness
Measures None
Blindness, Disparate
Impact (many)
Solution None Fair Classifiers (ADMs)
The Dream Machine: Best of Both Worlds
This Dream Led to New
Research
Hypothesis 1: Garbage
In, Garbage Out
Problem: Biased Data
• Structural Bias
• Cognitive Bias
• Historical Bias
• Stored in Protected Features
Solution:
Blind the model to Protected Features
Attempts to Blind Models
Problem: Discriminatory Decisions
Goal: Race-Blindness
Original Solution: Drop Protected Features
• Problem: Redundant Encodings or Proxies
Solution 2: Eliminate proxy information
• Problem: Poor performance
Solution 3: Eliminate Proxies’ influence with penalty
• Problem: Downstream effects
Race
IncomeCity
Income City
Race
IncomeCity
Problem 1) Drop Features
2) Eliminate Proxy Information
3) Penalize Model
What Went Wrong?
“disparate treatment doctrine does not
appear to do much to regulate
discriminatory data mining”
--Solon Barocas
“The race-aware predictor dominates
the other prediction functions”
--Jon Kleinberg
Questions:
Should we still race-blind models?
Are there other reasons to eliminate sensitive information?
Hypothesis 2:
Only Outcomes
Matter, Make
them “Fair”
• Problem: Error-Rate
Imbalances (COMPAS)
• Structural Bias
• Cognitive Bias
• Historical Bias
• Black-Box Algorithms
Repeat these Errors
• Solution:
• Move the Decision
Boundary to Fix Imbalances
Re-Thresholding Decision Boundaries
Make sure groups are in
line with predetermined
“fairness metrics”
Goal:
Move decision boundary until you
get the results you want
Popular Metrics:
• Model Blinding
• Demographic Parity
• Equal Opportunity
There are more than 20 different
ones
New Class
Boundary
Uncertain
Space
~0 < x < ~1
Creditworthy
Not CW
How Well Did This Work?
Quick Recap From Day 1
About ML
• ML is Function Estimation
• ML Decisions are Mostly Classification Problems
• These Classifications Are Always Uncertain
About Ethics
• Compas (Pro-Publica and Northpoint)
• Hypothesis 1: Biased Data, Erase Protected Information
• Didn’t have legal teeth or Practical Effectiveness
• Hypothesis 2: Error-Rate Imbalance, Move Decision Boundary
Researchers Original Solution: Fair Classifier
Max Profit
Maximize profits:
Obvious
Total Profit: 32,400
Decision Rules:
Different
True Positive Rate:
Unequal
Percent Approved:
Unequal
Blinding
Model
Group Unaware:
Class features and all
“proxy” information
removed
Total Profit: 25,600
Decision Rules: Same
True Positive Rate:
Unequal
Percent Approved:
Unequal
Demographic
Parity
Parity:
All groups have same
percentage approved
Total Profit: 30,800
Decision Rules:
Different
True Positive Rate:
Unequal
Percent Approved:
Equal
Equal
Opportunity
Opportunity:
Same percentage of
“credit-worthy” candidates
Total Profit: 30,400
Decision Rules:
Different
True Positive Rate:
Equal
Percent Approved:
Unequal
What Went Wrong? (Legally)
“smaller differences can constitute adverse
impact and greater differences may not,
depending on circumstances”
--Solon Barocas
What Went Wrong? (Conceptually)
“without precise definitions of beliefs about
the state of the world and…harms one wishes
to prevent, our results show that it is not
possible to make progress”
--Sorelle Friedler
Observed
Space
Construct
Space
Predicted
Space
GPA Success in High
School
College
Performance
Arrest Record Criminal Past Recidivism
Experience Job Knowledge Productivity
What Went Wrong? (Practically)
“since the optimal constrained algorithms differ
from the optimal unconstrained algorithm,
fairness has a cost”
--Sam Corbett-Davies
Do we need new laws to address AI Ethics?
Is it acceptable to use different rules for
different people?
What matters? The Rule, the Outcomes, Both…?
Can fairness exist outside of context?
What Happens in the
Broader Environment?
Feedback Loops (Exacerbating Feedback)
Model Trained on
Skewed Data
Poor Performance
on
Misrepresented
Class
Unobservability
Or
Over-observability
Under- and Over-
Representation
In Sample
Skewed Dataset
Example 1: Predictive Policing
• Targeting
• Arrest
• Data Skewed
• Poor Sample for Model
• Skewed Prediction
• Targeting
Over- Observation/Representation
Feedback Loops (Exacerbating Feedback)
Model Trained on
Skewed Data
Poor Performance
on
Misrepresented
Class
Unobservability
Or
Over-observability
Under- and Over-
Representation
In Sample
Skewed Dataset
Example 1: Predictive Policing
• Targeting
• Arrest
• Data Skewed
• Poor Sample for Model
• Skewed Prediction
• Targeting
Over- Observation/Representation
Example 2: Loans
• Group Traditionally Denied Loans
• Don’t Get Loans
• Can’t Pay Back
• Can’t Establish Credit
• Predicted as not “Credit-Worthy”
• Denied Loans
Under-Observation/Representation
A closer Look: Feedback and Inclusion/Exclusion
Survivorship Bias: Focus on those
that made it through selection
• Wald’s damaged planes
• Under/Over- Representation as
Qualitative and Quantitative
• Affects Interpretations, Observability,
and Solutions
A closer Look: Feedback and Inclusion/Exclusion
Survivorship Bias: Focus on those
that made it through selection
• Wald’s damaged planes
• Under/Over- Representation as
Qualitative and Quantitative
• Affects Interpretations, Observability,
and Solutions
A closer Look: Feedback and Inclusion/Exclusion
Questions:
• Can an organization be
responsible for responding to
realities it cannot observe?
• Is it acceptable to experiment
on candidates to observe their
outcomes?
Critical Thinking About Predictive Policing
Seeming Paradox:
Arrest data is a census, but feedback
loops result from (and exacerbate)
minority over-representation. How?
• Arrest as a proxy for crime
• Statistical vs Historical Bias
• Garbage in, Garbage out fails here
(Wrong Proxies)
Questions:
Would we say that the
sociological data is biased?
What could strengthen the proxy
relationship?
Arrest Data
Model
Arrest
Prediction
Arrest Data
Model
Arrest
Prediction
Crime
Prediction
Representative Not Representative
Dynamics, Gaming, and Context
College Admissions and ML
Features:
• GPA
• Extracurriculars
• ACT, SAT
Dynamics, Gaming, and Context
Goodhart’s Law (Gaming):
"When a measure becomes a target,
it ceases to be a good measure.“
Features:
• GPA:
• Grade Inflation
• Extracurriculars:
• Filling the Boxes
• ACT, SAT:
• Tutoring and
Teaching to Test
Dynamics, Gaming, and Context
Systemic Bias:
When a process supports or reproduces
specific outcomes.
Systemic
Bias
Features:
• GPA:
• Grade Inflation
• Quality of
Education
• Extracurriculars:
• Filling the Boxes
• Time and Money
• ACT, SAT:
• Tutoring and
Teaching to Test
• Less Focus on Test
Dynamics, Gaming, and Context
Testing Bias:
“Differential Validity of Test Scores for
Sub-Groups”
Systemic
Bias
Features:
• GPA:
• Grade Inflation
• Quality of Education
• Unsuitable Curriculum
• Extracurriculars:
• Filling the Boxes
• Time and Money
• Performance
interpretation
• ACT, SAT:
• Tutoring and Teaching
to Test
• Less Focus on Test
• Cultural Focus
Performance
Metrics
“the discussion in the technical community…is happening
without a moral framework…and you know, it’s kinda
amateur hour…, a lot of it is, lets design a loss function that
measures your utility and your fairness and whatever else
and just optimize the heck out of it”
-- Arvind Narayanan
Biggest Problem of All
A Few Starter Questions
• Is data the cause of bias in ML decisions?
• Do we need new laws (not disparate impact/treatment) to address
these systems?
• How much needs to be considered from outside an organization
when claiming that a system is fair?
• Is automated decision making ethically different than human decision
making?
• Can there be unfair models for reasons other than group differences?
• Should we study fairness with privacy, security, transparency…or not?

Contenu connexe

Tendances

Tendances (14)

Algorithmic Bias - What is it? Why should we care? What can we do about it?
Algorithmic Bias - What is it? Why should we care? What can we do about it? Algorithmic Bias - What is it? Why should we care? What can we do about it?
Algorithmic Bias - What is it? Why should we care? What can we do about it?
 
Fairness and Ethics in A
Fairness and Ethics in AFairness and Ethics in A
Fairness and Ethics in A
 
How can algorithms be biased?
How can algorithms be biased?How can algorithms be biased?
How can algorithms be biased?
 
Introduction to the ethics of machine learning
Introduction to the ethics of machine learningIntroduction to the ethics of machine learning
Introduction to the ethics of machine learning
 
Dual Approaches for Integrating Ethics into the Information Systems Curriculum
Dual Approaches for Integrating Ethics into the Information Systems CurriculumDual Approaches for Integrating Ethics into the Information Systems Curriculum
Dual Approaches for Integrating Ethics into the Information Systems Curriculum
 
A Tutorial to AI Ethics - Fairness, Bias & Perception
A Tutorial to AI Ethics - Fairness, Bias & Perception A Tutorial to AI Ethics - Fairness, Bias & Perception
A Tutorial to AI Ethics - Fairness, Bias & Perception
 
BigMLSchool: Trustworthy AI
BigMLSchool: Trustworthy AIBigMLSchool: Trustworthy AI
BigMLSchool: Trustworthy AI
 
Crowdsourcing: From Aggregation to Search Engine Evaluation
Crowdsourcing: From Aggregation to Search Engine EvaluationCrowdsourcing: From Aggregation to Search Engine Evaluation
Crowdsourcing: From Aggregation to Search Engine Evaluation
 
Agent-Based Modelling and Microsimulation: Ne’er the Twain Shall Meet?
Agent-Based Modelling and Microsimulation: Ne’er the Twain Shall Meet? Agent-Based Modelling and Microsimulation: Ne’er the Twain Shall Meet?
Agent-Based Modelling and Microsimulation: Ne’er the Twain Shall Meet?
 
Algorithmic Bias : What is it? Why should we care? What can we do about it?
Algorithmic Bias : What is it? Why should we care? What can we do about it?Algorithmic Bias : What is it? Why should we care? What can we do about it?
Algorithmic Bias : What is it? Why should we care? What can we do about it?
 
How do we Humans feel about AI?
How do we Humans feel about AI?How do we Humans feel about AI?
How do we Humans feel about AI?
 
Ethical Dilemmas in AI/ML-based systems
Ethical Dilemmas in AI/ML-based systemsEthical Dilemmas in AI/ML-based systems
Ethical Dilemmas in AI/ML-based systems
 
Designing more ethical and unbiased experiences
Designing more ethical and unbiased experiencesDesigning more ethical and unbiased experiences
Designing more ethical and unbiased experiences
 
Social Science Applications of Agent Based Modelling
Social Science Applications of Agent Based ModellingSocial Science Applications of Agent Based Modelling
Social Science Applications of Agent Based Modelling
 

Similaire à Algorithmic Fairness: A Brief Introduction

What is Data Science and How to Succeed in it
What is Data Science and How to Succeed in itWhat is Data Science and How to Succeed in it
What is Data Science and How to Succeed in it
Khosrow Hassibi
 
IDS Impact, Innovation and Learning Workshop March 2013: Day 2, Paper Session...
IDS Impact, Innovation and Learning Workshop March 2013: Day 2, Paper Session...IDS Impact, Innovation and Learning Workshop March 2013: Day 2, Paper Session...
IDS Impact, Innovation and Learning Workshop March 2013: Day 2, Paper Session...
Institute of Development Studies
 
Statistics in the age of data science, issues you can not ignore
Statistics in the age of data science, issues you can not ignoreStatistics in the age of data science, issues you can not ignore
Statistics in the age of data science, issues you can not ignore
Turi, Inc.
 
Data Science Consulting at ThoughtWorks -- NYC Open Data Meetup
Data Science Consulting at ThoughtWorks -- NYC Open Data MeetupData Science Consulting at ThoughtWorks -- NYC Open Data Meetup
Data Science Consulting at ThoughtWorks -- NYC Open Data Meetup
David Johnston
 
Baworld adapting to whats happening
Baworld adapting to whats happeningBaworld adapting to whats happening
Baworld adapting to whats happening
Dave Davis PMP, PgMP, PBA
 

Similaire à Algorithmic Fairness: A Brief Introduction (20)

Max Shron, Thinking with Data at the NYC Data Science Meetup
Max Shron, Thinking with Data at the NYC Data Science MeetupMax Shron, Thinking with Data at the NYC Data Science Meetup
Max Shron, Thinking with Data at the NYC Data Science Meetup
 
Engineering Ethics: Practicing Fairness
Engineering Ethics: Practicing FairnessEngineering Ethics: Practicing Fairness
Engineering Ethics: Practicing Fairness
 
What is Data Science and How to Succeed in it
What is Data Science and How to Succeed in itWhat is Data Science and How to Succeed in it
What is Data Science and How to Succeed in it
 
AI Models For Fun and Profit by Walmart Director of Artificial Intelligence
AI Models For Fun and Profit by Walmart Director of Artificial IntelligenceAI Models For Fun and Profit by Walmart Director of Artificial Intelligence
AI Models For Fun and Profit by Walmart Director of Artificial Intelligence
 
AI in the Real World: Challenges, and Risks and how to handle them?
AI in the Real World: Challenges, and Risks and how to handle them?AI in the Real World: Challenges, and Risks and how to handle them?
AI in the Real World: Challenges, and Risks and how to handle them?
 
DataEngConf SF16 - Three lessons learned from building a production machine l...
DataEngConf SF16 - Three lessons learned from building a production machine l...DataEngConf SF16 - Three lessons learned from building a production machine l...
DataEngConf SF16 - Three lessons learned from building a production machine l...
 
IDS Impact, Innovation and Learning Workshop March 2013: Day 2, Paper Session...
IDS Impact, Innovation and Learning Workshop March 2013: Day 2, Paper Session...IDS Impact, Innovation and Learning Workshop March 2013: Day 2, Paper Session...
IDS Impact, Innovation and Learning Workshop March 2013: Day 2, Paper Session...
 
Turning Information chaos into reliable data
Turning Information chaos into reliable dataTurning Information chaos into reliable data
Turning Information chaos into reliable data
 
Data Con LA 2022 - Real world consumer segmentation
Data Con LA 2022 - Real world consumer segmentationData Con LA 2022 - Real world consumer segmentation
Data Con LA 2022 - Real world consumer segmentation
 
Testing for cognitive bias in ai systems
Testing for cognitive bias in ai systemsTesting for cognitive bias in ai systems
Testing for cognitive bias in ai systems
 
Statistics in the age of data science, issues you can not ignore
Statistics in the age of data science, issues you can not ignoreStatistics in the age of data science, issues you can not ignore
Statistics in the age of data science, issues you can not ignore
 
Responsible AI in Industry: Practical Challenges and Lessons Learned
Responsible AI in Industry: Practical Challenges and Lessons LearnedResponsible AI in Industry: Practical Challenges and Lessons Learned
Responsible AI in Industry: Practical Challenges and Lessons Learned
 
Data Science Consulting at ThoughtWorks -- NYC Open Data Meetup
Data Science Consulting at ThoughtWorks -- NYC Open Data MeetupData Science Consulting at ThoughtWorks -- NYC Open Data Meetup
Data Science Consulting at ThoughtWorks -- NYC Open Data Meetup
 
CS194Lec0hbh6EDA.pptx
CS194Lec0hbh6EDA.pptxCS194Lec0hbh6EDA.pptx
CS194Lec0hbh6EDA.pptx
 
Baworld adapting to whats happening
Baworld adapting to whats happeningBaworld adapting to whats happening
Baworld adapting to whats happening
 
Data Science-final7
Data Science-final7Data Science-final7
Data Science-final7
 
Altron presentation on Emerging Technologies: Data Science and Artificial Int...
Altron presentation on Emerging Technologies: Data Science and Artificial Int...Altron presentation on Emerging Technologies: Data Science and Artificial Int...
Altron presentation on Emerging Technologies: Data Science and Artificial Int...
 
Module 1.3 data exploratory
Module 1.3  data exploratoryModule 1.3  data exploratory
Module 1.3 data exploratory
 
Entering the Data Analytics industry
Entering the Data Analytics industryEntering the Data Analytics industry
Entering the Data Analytics industry
 
The Human Side of Data By Colin Strong
The Human Side of Data By Colin StrongThe Human Side of Data By Colin Strong
The Human Side of Data By Colin Strong
 

Dernier

Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and Myths
Joaquim Jorge
 
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
?#DUbAI#??##{{(☎️+971_581248768%)**%*]'#abortion pills for sale in dubai@
 

Dernier (20)

Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and Myths
 
MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024
 
Manulife - Insurer Innovation Award 2024
Manulife - Insurer Innovation Award 2024Manulife - Insurer Innovation Award 2024
Manulife - Insurer Innovation Award 2024
 
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
 
Boost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfBoost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdf
 
presentation ICT roal in 21st century education
presentation ICT roal in 21st century educationpresentation ICT roal in 21st century education
presentation ICT roal in 21st century education
 
2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)
 
Boost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityBoost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivity
 
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt Robison
 
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost SavingRepurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
Deploy with confidence: VMware Cloud Foundation 5.1 on next gen Dell PowerEdg...
Deploy with confidence: VMware Cloud Foundation 5.1 on next gen Dell PowerEdg...Deploy with confidence: VMware Cloud Foundation 5.1 on next gen Dell PowerEdg...
Deploy with confidence: VMware Cloud Foundation 5.1 on next gen Dell PowerEdg...
 
GenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdfGenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdf
 
Real Time Object Detection Using Open CV
Real Time Object Detection Using Open CVReal Time Object Detection Using Open CV
Real Time Object Detection Using Open CV
 
HTML Injection Attacks: Impact and Mitigation Strategies
HTML Injection Attacks: Impact and Mitigation StrategiesHTML Injection Attacks: Impact and Mitigation Strategies
HTML Injection Attacks: Impact and Mitigation Strategies
 

Algorithmic Fairness: A Brief Introduction

  • 2. Talk of AI and ethics is on the rise
  • 4. What Really Caused it to Explode?
  • 5. Pro Publica and COMPAS • Pro Publica • Journalism • Social Justice • Audited COMPAS • Northpoint • COMPAS • Recidivism Prediction • “Cutting Edge” ML • COMPAS Audit • Was Calibrated • Equal Accuracy • Was Deployed High Risk, No Offense • Black: 45% • White: 24% Low Risk, Offense • Black: 28% • White: 48% Problem: Disparate Impact
  • 6. So why use these systems at all?
  • 7. Probabilistic vs Causal Policies Rain Dancing Problems (End Drought) • Key Question: Does rain dancing cause rain? • Solution: Causal Rule (If yes, then rain dance) • Doesn’t require prediction of anything Umbrella Problems (Stay Dry) • Key Question: Will it rain today? • Solution: Probabilistic cost/benefit (many variables) • Requires prediction of rain, and knowledge of payouts Takeaway: Some problems require prediction for good results. ML predicts better than other approaches. Kleinberg, Jon. “Prediction Policy Problems.”
  • 8. How do we get the best of both? First, we look at the difference?
  • 9. Why are ML Policies Different from Causal Policies?
  • 11. A Deeper Look: Probabilistic Decision Boundaries Basics: • (Un)Certainty of Cases • Risk vs Class • Red Errors Blue Errors Takeaways: • Estimated Functions aren’t like known functions • Different results for equally deserving people/groups • The uncertain space is most interesting • Upside • Downside Class Boundary Uncertain Space ~0 < x < ~1 Creditworthy Not CW
  • 12. What Did Researchers Do in Response to Their Dilemma?
  • 13. They Proposed a New Direction Previous Views/Strategies New Views/Strategies Source of Bias None Human Biases Garbage in, Garbage out Protected Features Unawareness Redundant Encodings Fairness Measures None Blindness, Disparate Impact (many) Solution None Fair Classifiers (ADMs)
  • 14. The Dream Machine: Best of Both Worlds
  • 15. This Dream Led to New Research
  • 16. Hypothesis 1: Garbage In, Garbage Out Problem: Biased Data • Structural Bias • Cognitive Bias • Historical Bias • Stored in Protected Features Solution: Blind the model to Protected Features
  • 17. Attempts to Blind Models Problem: Discriminatory Decisions Goal: Race-Blindness Original Solution: Drop Protected Features • Problem: Redundant Encodings or Proxies Solution 2: Eliminate proxy information • Problem: Poor performance Solution 3: Eliminate Proxies’ influence with penalty • Problem: Downstream effects Race IncomeCity Income City Race IncomeCity Problem 1) Drop Features 2) Eliminate Proxy Information 3) Penalize Model
  • 18. What Went Wrong? “disparate treatment doctrine does not appear to do much to regulate discriminatory data mining” --Solon Barocas “The race-aware predictor dominates the other prediction functions” --Jon Kleinberg Questions: Should we still race-blind models? Are there other reasons to eliminate sensitive information?
  • 19. Hypothesis 2: Only Outcomes Matter, Make them “Fair” • Problem: Error-Rate Imbalances (COMPAS) • Structural Bias • Cognitive Bias • Historical Bias • Black-Box Algorithms Repeat these Errors • Solution: • Move the Decision Boundary to Fix Imbalances
  • 20. Re-Thresholding Decision Boundaries Make sure groups are in line with predetermined “fairness metrics” Goal: Move decision boundary until you get the results you want Popular Metrics: • Model Blinding • Demographic Parity • Equal Opportunity There are more than 20 different ones New Class Boundary Uncertain Space ~0 < x < ~1 Creditworthy Not CW
  • 21. How Well Did This Work?
  • 22. Quick Recap From Day 1 About ML • ML is Function Estimation • ML Decisions are Mostly Classification Problems • These Classifications Are Always Uncertain About Ethics • Compas (Pro-Publica and Northpoint) • Hypothesis 1: Biased Data, Erase Protected Information • Didn’t have legal teeth or Practical Effectiveness • Hypothesis 2: Error-Rate Imbalance, Move Decision Boundary Researchers Original Solution: Fair Classifier
  • 23. Max Profit Maximize profits: Obvious Total Profit: 32,400 Decision Rules: Different True Positive Rate: Unequal Percent Approved: Unequal
  • 24. Blinding Model Group Unaware: Class features and all “proxy” information removed Total Profit: 25,600 Decision Rules: Same True Positive Rate: Unequal Percent Approved: Unequal
  • 25. Demographic Parity Parity: All groups have same percentage approved Total Profit: 30,800 Decision Rules: Different True Positive Rate: Unequal Percent Approved: Equal
  • 26. Equal Opportunity Opportunity: Same percentage of “credit-worthy” candidates Total Profit: 30,400 Decision Rules: Different True Positive Rate: Equal Percent Approved: Unequal
  • 27. What Went Wrong? (Legally) “smaller differences can constitute adverse impact and greater differences may not, depending on circumstances” --Solon Barocas
  • 28. What Went Wrong? (Conceptually) “without precise definitions of beliefs about the state of the world and…harms one wishes to prevent, our results show that it is not possible to make progress” --Sorelle Friedler Observed Space Construct Space Predicted Space GPA Success in High School College Performance Arrest Record Criminal Past Recidivism Experience Job Knowledge Productivity
  • 29. What Went Wrong? (Practically) “since the optimal constrained algorithms differ from the optimal unconstrained algorithm, fairness has a cost” --Sam Corbett-Davies
  • 30. Do we need new laws to address AI Ethics? Is it acceptable to use different rules for different people? What matters? The Rule, the Outcomes, Both…? Can fairness exist outside of context?
  • 31. What Happens in the Broader Environment?
  • 32. Feedback Loops (Exacerbating Feedback) Model Trained on Skewed Data Poor Performance on Misrepresented Class Unobservability Or Over-observability Under- and Over- Representation In Sample Skewed Dataset Example 1: Predictive Policing • Targeting • Arrest • Data Skewed • Poor Sample for Model • Skewed Prediction • Targeting Over- Observation/Representation
  • 33. Feedback Loops (Exacerbating Feedback) Model Trained on Skewed Data Poor Performance on Misrepresented Class Unobservability Or Over-observability Under- and Over- Representation In Sample Skewed Dataset Example 1: Predictive Policing • Targeting • Arrest • Data Skewed • Poor Sample for Model • Skewed Prediction • Targeting Over- Observation/Representation Example 2: Loans • Group Traditionally Denied Loans • Don’t Get Loans • Can’t Pay Back • Can’t Establish Credit • Predicted as not “Credit-Worthy” • Denied Loans Under-Observation/Representation
  • 34. A closer Look: Feedback and Inclusion/Exclusion Survivorship Bias: Focus on those that made it through selection • Wald’s damaged planes • Under/Over- Representation as Qualitative and Quantitative • Affects Interpretations, Observability, and Solutions
  • 35. A closer Look: Feedback and Inclusion/Exclusion Survivorship Bias: Focus on those that made it through selection • Wald’s damaged planes • Under/Over- Representation as Qualitative and Quantitative • Affects Interpretations, Observability, and Solutions
  • 36. A closer Look: Feedback and Inclusion/Exclusion Questions: • Can an organization be responsible for responding to realities it cannot observe? • Is it acceptable to experiment on candidates to observe their outcomes?
  • 37. Critical Thinking About Predictive Policing Seeming Paradox: Arrest data is a census, but feedback loops result from (and exacerbate) minority over-representation. How? • Arrest as a proxy for crime • Statistical vs Historical Bias • Garbage in, Garbage out fails here (Wrong Proxies) Questions: Would we say that the sociological data is biased? What could strengthen the proxy relationship? Arrest Data Model Arrest Prediction Arrest Data Model Arrest Prediction Crime Prediction Representative Not Representative
  • 38. Dynamics, Gaming, and Context College Admissions and ML Features: • GPA • Extracurriculars • ACT, SAT
  • 39. Dynamics, Gaming, and Context Goodhart’s Law (Gaming): "When a measure becomes a target, it ceases to be a good measure.“ Features: • GPA: • Grade Inflation • Extracurriculars: • Filling the Boxes • ACT, SAT: • Tutoring and Teaching to Test
  • 40. Dynamics, Gaming, and Context Systemic Bias: When a process supports or reproduces specific outcomes. Systemic Bias Features: • GPA: • Grade Inflation • Quality of Education • Extracurriculars: • Filling the Boxes • Time and Money • ACT, SAT: • Tutoring and Teaching to Test • Less Focus on Test
  • 41. Dynamics, Gaming, and Context Testing Bias: “Differential Validity of Test Scores for Sub-Groups” Systemic Bias Features: • GPA: • Grade Inflation • Quality of Education • Unsuitable Curriculum • Extracurriculars: • Filling the Boxes • Time and Money • Performance interpretation • ACT, SAT: • Tutoring and Teaching to Test • Less Focus on Test • Cultural Focus Performance Metrics
  • 42. “the discussion in the technical community…is happening without a moral framework…and you know, it’s kinda amateur hour…, a lot of it is, lets design a loss function that measures your utility and your fairness and whatever else and just optimize the heck out of it” -- Arvind Narayanan Biggest Problem of All
  • 43. A Few Starter Questions • Is data the cause of bias in ML decisions? • Do we need new laws (not disparate impact/treatment) to address these systems? • How much needs to be considered from outside an organization when claiming that a system is fair? • Is automated decision making ethically different than human decision making? • Can there be unfair models for reasons other than group differences? • Should we study fairness with privacy, security, transparency…or not?