SlideShare a Scribd company logo
Fairness in AI
“Facilis descensus Averno”
Henk Griffioen - 2019-05-07
Data Scientists have to put it the work
to not end up in ML hell
2
“The gates of hell are open night and day;
Smooth the descent, and easy is the way:
But to return, and view the cheerful skies,
In this the task and mighty labor lies.”
The Works of Virgil (John Dryden)
The impact of AI on society is not all good. AI can encode and
amplify human biases, leading to unfair outcomes at scale
3
Fairness is a hot topic and gaining traction!
4
https://fairmlclass.github.io/
Let’s work through an example
5
Predict if income exceeds $50K/year from 1994 census data
with sensitive attributes
6
Getting a predictive model is easy!
But is it fair?
7
How to measure fairness?
8
Ratio of
• probability of a positive outcome given the sensitive attribute being true;
• probability of a positive outcome given the sensitive attribute being false;
is no less than p:100
p%-rule: measure demographic parity
9
40% : 50% = 80%10% : 50% = 20%
Our model is unfair: low probability of high income for black
people and women
10
How are we getting bias in our systems?
11
• Skewed sample: initial bias that compounds over time
• Tainted examples: bias in the data caused by humans
• Sample size disparity: minority groups not as well represented
• Limited features: less informative or accurate data collected on minority groups
• Proxies: data implicitly encoding sensitive attributes
• …
Many reasons why bias is creeping into our systems
12Barocas & Selbst, 2016
Intermezzo:
Dutch tax authorities in the news
13
Ethnic profiling by the
Dutch tax authorities
14
Profiling people for fraud
“A daycare center in Almere sounded the alarm when
only non-Dutch parents were confronted with
discontinuation of childcare allowances…
…The Tax and Customs Administration says that it
uses the data on Dutch nationality or non-Dutch
nationality in the so-called automatic risk selection
for fraud.”
https://www.nrc.nl/nieuws/2019/05/20/autoriteit
-persoonsgegevens-onderzoekt-mogelijke-
discriminatie-door-belastingdienst-a3960840
Is it enough to leave out
data on (second)
nationality?
15
…In a response, the Tax and Customs Administration
states that the information about the (second) nationality
of parents or intermediary is not used in this
investigation…
“Since 2014, a second nationality with Dutch nationality
is no longer included in the basic registration. This has
been introduced to prevent discrimination for people
with dual nationality.”
Is this enough to
assure that non-Dutch
parents in Almere will
not suffer another tax
injustice?
Towards a fair future?
The model is still unfair
without sensitive data.
Biases are still encoded
by proxies in the dataset!
16
How can we enforce fairness?
Start out with our normal (biased) classifier…
18
Louppe, 2017
… and add an adversarial classifier as a fairness referee
19
Louppe, 2017
Warming up: train classifier to predict income
20
Warming up: train adversarial to detect unfairness
21
Louppe, 2017
Adversarial training: iteratively train classifier and adversarial
22
Louppe, 2017
After enough training rounds the classifier gives fair income
predictions!
23
https://blog.godatadriven.com/fairness-in-ml
Putting in the work
Demographics parity
• Decision uncorrelated with sensitive attribute
Equality of opportunity
• Outcome probability is same for different groups
No mathematical formulation of fairness.
There are many (conflicting) measures
25http://www.ece.ubc.ca/~mjulia/publications/Fairness_Definitions_
Explained_2018.pdf
Many ML fairness approaches
26
https://dzone.com/articles/machine-learning-
models-bias-mitigation-strategies
• Identify product goals
• What? For whom?
• Get the right people in the room
• Identify stakeholders
• Who might be harmed? How?
• Select a fairness approach
• How to measure? What interaction?
• Analyze and evaluate your system
• What decisions are made?
• Mitigates issues
• Intervention needed?
• Monitor continuously and escalation plans
• Auditing and transparency
• Who certifies your system?
Fairness should be part of your product process and larger
strategy
27
https://www.slideshare.net/KrishnaramKenthapadi/
fairnessaware-machine-learning-practical-
challenges-and-lessons-learned-www-2019-tutorial
Fairness is far from being solved and needs active work!
28
Artificial Intelligence at Google: Our Principles - Objectives for AI applications
1. Be socially beneficial.
2. Avoid creating or reinforcing unfair bias.
3. Be built and tested for safety.
4. Be accountable to people.
5. Incorporate privacy design principles.
6. Uphold high standards of scientific excellence.
7. Be made available for uses that accord with these principles.
Even big companies can’t always practice what they preach
29
https://ai.google/principles/
Data Scientists have to put it the work
to not end up in ML hell
30
“The gates of hell are open night and day;
Smooth the descent, and easy is the way:
But to return, and view the cheerful skies,
In this the task and mighty labor lies.”
The Works of Virgil (John Dryden)
An ethics checklist for data scientists
• http://deon.drivendata.org/
Tutorial on fairness for products
• sites.google.com/view/wsdm19-fairness-tutorial
Community concerned with fairness in ML
• www.fatml.org
Our blogs
• blog.godatadriven.com/fairness-in-ml
• blog.godatadriven.com/fairness-in-pytorch
Where to go from here?
31
Fairness in AI (DDSW 2019)

More Related Content

PDF
Fairness in Machine Learning and AI
PDF
Fairness and Bias in Machine Learning
PDF
Responsible AI
PDF
Introduction to Git and Github
PPTX
PDF
The future of AI is hybrid
PPTX
Explainable AI in Industry (WWW 2020 Tutorial)
PDF
Explainable AI (XAI) - A Perspective
Fairness in Machine Learning and AI
Fairness and Bias in Machine Learning
Responsible AI
Introduction to Git and Github
The future of AI is hybrid
Explainable AI in Industry (WWW 2020 Tutorial)
Explainable AI (XAI) - A Perspective

What's hot (20)

PDF
AIF360 - Trusted and Fair AI
PDF
Ethics in the use of Data & AI
PDF
Introduction to AI Ethics
PPTX
Bias in Artificial Intelligence
PPTX
A Tutorial to AI Ethics - Fairness, Bias & Perception
PPTX
Fairness and Privacy in AI/ML Systems
PDF
Algorithmic Bias: Challenges and Opportunities for AI in Healthcare
PPTX
Responsible AI in Industry (ICML 2021 Tutorial)
PDF
Ai Ethics
PDF
Generative AI
PPTX
Generative AI Risks & Concerns
PDF
Artificial Intelligence (AI) in media applications and services
PPTX
The Ethics of AI
PPTX
Artificial intelligence
PPTX
Ethical Considerations in the Design of Artificial Intelligence
PPTX
The Ethics of Artificial Intelligence
PPTX
Implementing Ethics in AI
PPTX
Technology for everyone - AI ethics and Bias
PPTX
Ethical Issues in Machine Learning Algorithms. (Part 3)
PPTX
AI and the Impact on Cybersecurity
AIF360 - Trusted and Fair AI
Ethics in the use of Data & AI
Introduction to AI Ethics
Bias in Artificial Intelligence
A Tutorial to AI Ethics - Fairness, Bias & Perception
Fairness and Privacy in AI/ML Systems
Algorithmic Bias: Challenges and Opportunities for AI in Healthcare
Responsible AI in Industry (ICML 2021 Tutorial)
Ai Ethics
Generative AI
Generative AI Risks & Concerns
Artificial Intelligence (AI) in media applications and services
The Ethics of AI
Artificial intelligence
Ethical Considerations in the Design of Artificial Intelligence
The Ethics of Artificial Intelligence
Implementing Ethics in AI
Technology for everyone - AI ethics and Bias
Ethical Issues in Machine Learning Algorithms. (Part 3)
AI and the Impact on Cybersecurity
Ad

Similar to Fairness in AI (DDSW 2019) (20)

PDF
Fairness in AI - Towards more Ethical predictive models - Big Data Expo 2019
PPTX
Eric van tol
PDF
The Human Side of Data By Colin Strong
PPTX
Data Driven Public Services - Mark Charmer, Akvo
PPTX
Introducing Akvo, an open data startup
PDF
Spring Splash 3.4.2019: When AI Meets Ethics by Meeri Haataja
PDF
The new fundamentals-Seizing opportunities with AI in the cognitive economy
PPTX
AI Ethical Framework.pptx
PDF
From the End of Information Chaos to Contextual Knowledge
PPTX
The Ethics of Structured Information
PPTX
Show & TEL Ethics & Technology-Enhanced Learning
PDF
Have you AI'ed today? A Reality Check
PDF
Solving through Systems Thinking.pdf
PDF
The Anticipatory CPA #PSTECH
PPTX
[DSC Adria 23] MARIJANA ŠAROLIĆ ROBIĆ Everyone is invited to AI enhanced pres...
PPTX
The Future of Advanced Analytics
PDF
Using fairness metrics to solve ethical dilemmas of machine learning
PPTX
The Future of Personalised Education
PDF
How Artificial intelligence is transforming the world.pdf
PDF
Algorithmic Bias - What is it? Why should we care? What can we do about it?
Fairness in AI - Towards more Ethical predictive models - Big Data Expo 2019
Eric van tol
The Human Side of Data By Colin Strong
Data Driven Public Services - Mark Charmer, Akvo
Introducing Akvo, an open data startup
Spring Splash 3.4.2019: When AI Meets Ethics by Meeri Haataja
The new fundamentals-Seizing opportunities with AI in the cognitive economy
AI Ethical Framework.pptx
From the End of Information Chaos to Contextual Knowledge
The Ethics of Structured Information
Show & TEL Ethics & Technology-Enhanced Learning
Have you AI'ed today? A Reality Check
Solving through Systems Thinking.pdf
The Anticipatory CPA #PSTECH
[DSC Adria 23] MARIJANA ŠAROLIĆ ROBIĆ Everyone is invited to AI enhanced pres...
The Future of Advanced Analytics
Using fairness metrics to solve ethical dilemmas of machine learning
The Future of Personalised Education
How Artificial intelligence is transforming the world.pdf
Algorithmic Bias - What is it? Why should we care? What can we do about it?
Ad

More from GoDataDriven (20)

PDF
Streamlining Data Science Workflows with a Feature Catalog
PDF
Visualizing Big Data in a Small Screen
PDF
Building a Scalable and reliable open source ML Platform with MLFlow
PDF
Training Taster: Leading the way to become a data-driven organization
PDF
My Path From Data Engineer to Analytics Engineer
PDF
dbt Python models - GoDataFest by Guillermo Sanchez
PDF
Workshop on Google Cloud Data Platform
PDF
How to create a Devcontainer for your Python project
PDF
Using Graph Neural Networks To Embrace The Dependency In Your Data by Usman Z...
PDF
Common Issues With Time Series by Vadim Nelidov - GoDataFest 2022
PDF
MLOps CodeBreakfast on AWS - GoDataFest 2022
PDF
MLOps CodeBreakfast on Azure - GoDataFest 2022
PDF
Tableau vs. Power BI by Juan Manuel Perafan - GoDataFest 2022
PDF
Deploying a Modern Data Stack by Lasse Benninga - GoDataFest 2022
PPTX
AWS Well-Architected Webinar Security - Ben de Haan
PDF
The 7 Habits of Effective Data Driven Companies
PPTX
DevOps for Data Science on Azure - Marcel de Vries (Xpirit) and Niels Zeilema...
PDF
Artificial intelligence in actions: delivering a new experience to Formula 1 ...
PDF
Smart application on Azure at Vattenfall - Rens Weijers & Peter van 't Hof
PDF
Democratizing AI/ML with GCP - Abishay Rao (Google) at GoDataFest 2019
Streamlining Data Science Workflows with a Feature Catalog
Visualizing Big Data in a Small Screen
Building a Scalable and reliable open source ML Platform with MLFlow
Training Taster: Leading the way to become a data-driven organization
My Path From Data Engineer to Analytics Engineer
dbt Python models - GoDataFest by Guillermo Sanchez
Workshop on Google Cloud Data Platform
How to create a Devcontainer for your Python project
Using Graph Neural Networks To Embrace The Dependency In Your Data by Usman Z...
Common Issues With Time Series by Vadim Nelidov - GoDataFest 2022
MLOps CodeBreakfast on AWS - GoDataFest 2022
MLOps CodeBreakfast on Azure - GoDataFest 2022
Tableau vs. Power BI by Juan Manuel Perafan - GoDataFest 2022
Deploying a Modern Data Stack by Lasse Benninga - GoDataFest 2022
AWS Well-Architected Webinar Security - Ben de Haan
The 7 Habits of Effective Data Driven Companies
DevOps for Data Science on Azure - Marcel de Vries (Xpirit) and Niels Zeilema...
Artificial intelligence in actions: delivering a new experience to Formula 1 ...
Smart application on Azure at Vattenfall - Rens Weijers & Peter van 't Hof
Democratizing AI/ML with GCP - Abishay Rao (Google) at GoDataFest 2019

Recently uploaded (20)

PPT
Quality review (1)_presentation of this 21
PDF
Galatica Smart Energy Infrastructure Startup Pitch Deck
PPTX
Database Infoormation System (DBIS).pptx
PPTX
iec ppt-1 pptx icmr ppt on rehabilitation.pptx
PDF
Fluorescence-microscope_Botany_detailed content
PPTX
ALIMENTARY AND BILIARY CONDITIONS 3-1.pptx
PDF
TRAFFIC-MANAGEMENT-AND-ACCIDENT-INVESTIGATION-WITH-DRIVING-PDF-FILE.pdf
PPTX
Moving the Public Sector (Government) to a Digital Adoption
PPTX
Business Acumen Training GuidePresentation.pptx
PDF
168300704-gasification-ppt.pdfhghhhsjsjhsuxush
PPT
Chapter 2 METAL FORMINGhhhhhhhjjjjmmmmmmmmm
PPTX
IB Computer Science - Internal Assessment.pptx
PDF
Introduction to Business Data Analytics.
PPTX
Global journeys: estimating international migration
PPTX
mbdjdhjjodule 5-1 rhfhhfjtjjhafbrhfnfbbfnb
PPTX
Business Ppt On Nestle.pptx huunnnhhgfvu
PPTX
STUDY DESIGN details- Lt Col Maksud (21).pptx
PPTX
DISORDERS OF THE LIVER, GALLBLADDER AND PANCREASE (1).pptx
PDF
“Getting Started with Data Analytics Using R – Concepts, Tools & Case Studies”
PDF
BF and FI - Blockchain, fintech and Financial Innovation Lesson 2.pdf
Quality review (1)_presentation of this 21
Galatica Smart Energy Infrastructure Startup Pitch Deck
Database Infoormation System (DBIS).pptx
iec ppt-1 pptx icmr ppt on rehabilitation.pptx
Fluorescence-microscope_Botany_detailed content
ALIMENTARY AND BILIARY CONDITIONS 3-1.pptx
TRAFFIC-MANAGEMENT-AND-ACCIDENT-INVESTIGATION-WITH-DRIVING-PDF-FILE.pdf
Moving the Public Sector (Government) to a Digital Adoption
Business Acumen Training GuidePresentation.pptx
168300704-gasification-ppt.pdfhghhhsjsjhsuxush
Chapter 2 METAL FORMINGhhhhhhhjjjjmmmmmmmmm
IB Computer Science - Internal Assessment.pptx
Introduction to Business Data Analytics.
Global journeys: estimating international migration
mbdjdhjjodule 5-1 rhfhhfjtjjhafbrhfnfbbfnb
Business Ppt On Nestle.pptx huunnnhhgfvu
STUDY DESIGN details- Lt Col Maksud (21).pptx
DISORDERS OF THE LIVER, GALLBLADDER AND PANCREASE (1).pptx
“Getting Started with Data Analytics Using R – Concepts, Tools & Case Studies”
BF and FI - Blockchain, fintech and Financial Innovation Lesson 2.pdf

Fairness in AI (DDSW 2019)

  • 1. Fairness in AI “Facilis descensus Averno” Henk Griffioen - 2019-05-07
  • 2. Data Scientists have to put it the work to not end up in ML hell 2 “The gates of hell are open night and day; Smooth the descent, and easy is the way: But to return, and view the cheerful skies, In this the task and mighty labor lies.” The Works of Virgil (John Dryden)
  • 3. The impact of AI on society is not all good. AI can encode and amplify human biases, leading to unfair outcomes at scale 3
  • 4. Fairness is a hot topic and gaining traction! 4 https://fairmlclass.github.io/
  • 5. Let’s work through an example 5
  • 6. Predict if income exceeds $50K/year from 1994 census data with sensitive attributes 6
  • 7. Getting a predictive model is easy! But is it fair? 7
  • 8. How to measure fairness? 8
  • 9. Ratio of • probability of a positive outcome given the sensitive attribute being true; • probability of a positive outcome given the sensitive attribute being false; is no less than p:100 p%-rule: measure demographic parity 9 40% : 50% = 80%10% : 50% = 20%
  • 10. Our model is unfair: low probability of high income for black people and women 10
  • 11. How are we getting bias in our systems? 11
  • 12. • Skewed sample: initial bias that compounds over time • Tainted examples: bias in the data caused by humans • Sample size disparity: minority groups not as well represented • Limited features: less informative or accurate data collected on minority groups • Proxies: data implicitly encoding sensitive attributes • … Many reasons why bias is creeping into our systems 12Barocas & Selbst, 2016
  • 14. Ethnic profiling by the Dutch tax authorities 14 Profiling people for fraud “A daycare center in Almere sounded the alarm when only non-Dutch parents were confronted with discontinuation of childcare allowances… …The Tax and Customs Administration says that it uses the data on Dutch nationality or non-Dutch nationality in the so-called automatic risk selection for fraud.” https://www.nrc.nl/nieuws/2019/05/20/autoriteit -persoonsgegevens-onderzoekt-mogelijke- discriminatie-door-belastingdienst-a3960840
  • 15. Is it enough to leave out data on (second) nationality? 15 …In a response, the Tax and Customs Administration states that the information about the (second) nationality of parents or intermediary is not used in this investigation… “Since 2014, a second nationality with Dutch nationality is no longer included in the basic registration. This has been introduced to prevent discrimination for people with dual nationality.” Is this enough to assure that non-Dutch parents in Almere will not suffer another tax injustice? Towards a fair future?
  • 16. The model is still unfair without sensitive data. Biases are still encoded by proxies in the dataset! 16
  • 17. How can we enforce fairness?
  • 18. Start out with our normal (biased) classifier… 18 Louppe, 2017
  • 19. … and add an adversarial classifier as a fairness referee 19 Louppe, 2017
  • 20. Warming up: train classifier to predict income 20
  • 21. Warming up: train adversarial to detect unfairness 21 Louppe, 2017
  • 22. Adversarial training: iteratively train classifier and adversarial 22 Louppe, 2017
  • 23. After enough training rounds the classifier gives fair income predictions! 23 https://blog.godatadriven.com/fairness-in-ml
  • 25. Demographics parity • Decision uncorrelated with sensitive attribute Equality of opportunity • Outcome probability is same for different groups No mathematical formulation of fairness. There are many (conflicting) measures 25http://www.ece.ubc.ca/~mjulia/publications/Fairness_Definitions_ Explained_2018.pdf
  • 26. Many ML fairness approaches 26 https://dzone.com/articles/machine-learning- models-bias-mitigation-strategies
  • 27. • Identify product goals • What? For whom? • Get the right people in the room • Identify stakeholders • Who might be harmed? How? • Select a fairness approach • How to measure? What interaction? • Analyze and evaluate your system • What decisions are made? • Mitigates issues • Intervention needed? • Monitor continuously and escalation plans • Auditing and transparency • Who certifies your system? Fairness should be part of your product process and larger strategy 27 https://www.slideshare.net/KrishnaramKenthapadi/ fairnessaware-machine-learning-practical- challenges-and-lessons-learned-www-2019-tutorial
  • 28. Fairness is far from being solved and needs active work! 28
  • 29. Artificial Intelligence at Google: Our Principles - Objectives for AI applications 1. Be socially beneficial. 2. Avoid creating or reinforcing unfair bias. 3. Be built and tested for safety. 4. Be accountable to people. 5. Incorporate privacy design principles. 6. Uphold high standards of scientific excellence. 7. Be made available for uses that accord with these principles. Even big companies can’t always practice what they preach 29 https://ai.google/principles/
  • 30. Data Scientists have to put it the work to not end up in ML hell 30 “The gates of hell are open night and day; Smooth the descent, and easy is the way: But to return, and view the cheerful skies, In this the task and mighty labor lies.” The Works of Virgil (John Dryden)
  • 31. An ethics checklist for data scientists • http://deon.drivendata.org/ Tutorial on fairness for products • sites.google.com/view/wsdm19-fairness-tutorial Community concerned with fairness in ML • www.fatml.org Our blogs • blog.godatadriven.com/fairness-in-ml • blog.godatadriven.com/fairness-in-pytorch Where to go from here? 31

Editor's Notes

  • #2: Fairness, privacy, transparency, explainability
  • #3: An earlier version of ‘The road to hell is paved with good intentions’
  • #21: https://www.slon.pics/shop/robot-referee-showing-the-red-card-3d-illustration-isolated-contains-clipping-path/
  • #31: An earlier version of ‘The road to hell is paved with good intentions’