Fairness of machine learning models
an overview and practical considerations
AI & BigData Online Day 2020
Agenda
Disclaimer: The views and opinions expressed in this document are those of the author and do not necessarily reflect the official policy or position of de Volksbank and Tilburg University.
Motivating Examples
Operationalizing Fairness
1
3
Response: high-level frameworks2
1. Motivating examples
 Correctional Offender Management Profiling
for Alternative Sanctions (COMPAS)
 Recidivism Prediction Instruments (RPI) used
by US courts for pretrial decision-making, parole
decisions, and sometimes sentencing.
 On basis of 7,000 cases 2013-2014 in Broward
county, Florida ProPublica’s conclusion is that
algorithm is biased against blacks.
Google (2018). Reducing gender bias in Google Translate.
Zou, J. and L. Schiebinger (2018). AI can be sexist and racist — it’s time to make it fair, Nature 559, pp. 324-326.
1. Motivating examples
1. Motivating examples
Ensign, D., S.A. Friedler, S. Neville, C. Scheidegger, and S. Venkatasubramanian (2018). Runaway Feedback Loops in
Predictive Policing, Proceedings of Machine Learning Research 81, pp. 160-171.
1. Motivating examples
1. Motivating examples
2016 20182016 20182006 2019
2. Response: regulation is coming…
2020
Jobin, Ienca, Vayena (2019):
• Mapped and analyzed the current corpus of
principles and guidelines on ethical AI
• Identified 84 documents containing ethical
principles or guidelines for AI, 88% being after
2016.
• Convergence?
• “In particular, the prevalence of calls for
transparency, justice and fairness points to an
emerging moral priority”.
Transparency & Explainability
Justice and fairness
Non-maleficence
Responsibility
Privacy
2. Meta-analysis: Global Landscape of Ethical Frameworks for AI
and data science
From point-of-view of data scientist: fairness and explainability appear
to most difficult to implement.
3. Non-discrimination and data science
but for data scientist, typical goal is to maximize
discriminatory power!
Models should not yield unintended discrimination
protected features
A
X
Y
data
model
featureslabels3. Avoiding unintended discrimination: forbidden variables
Supervised Learning
Ignoring protected features does not work (in general)!
3. Defining (un)fairness, discrimination
 How to define (un)fairness?
 Note that is actually not a new topic
 see, for example, Hutchinson and
Mitchell (2019)
 rapidly developing literature
 mainly for binary classifiers!
 Barocas, S., M. Hardt, and A. Narayanan (2020). Fairness and machine learning – limitations and opportunities. Book in progress (https://fairmlbook.org/)
 Corbett-Davies, S. and S. Goel (2018). The Measure and Mismeasure of Fairness: A Critical Review of Fair Machine Learning, working paper
 Hutchinson, B. and M. Mitchell (2019). 50 Years of Test (Un)fairness: Lessons for Machine Learning. In: FAT* ’19: Conference on Fairness, Accountability, and Transparency (FAT* ’19), ACM, pp. 49-58.
 Verma, S. and J. Rubin (2018). Fairness Definitions Explained. In: 2018 IEEE/ACM International Workshop on Software Fairness (FairWare). IEEE, pp. 1–7.
statistical
parity
fairness
through
unawareness
equal
opportunity predictive
parity
calibration
conditional
statistical
parity
equalized
odds
counter-
factual
fairness
no proxy
discrimination
Great variety of open-source fairness
dashboarding tools are available
AI Fairness 360
3. Bias/fairness dashboarding
Saleiro, P., B. Kuester, A. Stevens, A. Anisfeld, L. Hinkson, J. London, and R. Ghani (2018). Aequitas: A Bias and Fairness Audit Toolkit, arXiv preprint arXiv:1811.05577.
How to proceed if protected group concerns special personal data?
Implementation requires choices for protected groups and fairness metrics!
3. Why do we need to make a choice: just consider all
fairness metrics
We think ProPublica’s report was based on faulty
statistics and data analysis, and that the report
failed to show that the COMPAS itself is racially
biased, let alone that other risk instruments are
biased. -- Flores et al. (2016)
There’s software used across the country to predict
future criminals. And it’s biased against blacks. --
ProPublica (2016)
source: Chouldekova (2017)
 Chouldechova (2017). Fair Prediction with Disparate Impact: A Study of Bias in Recidivism Prediction Instruments, Big Data 5(2), pp. 153-163.
 Flores, A.W., K. Bechtel, and C.T. Lowenkamp (2016). False Positives, False Negatives, and False Analyses: A Rejoinder to “Machine Bias: There’s Software Used Across the Country to Predict Future Criminals. And It’s
Biased Against Blacks, Federal Probation 80(2), pp.38-46.
 Kleinberg, J., S. Mullainathan, and M. Raghavan (2017). Inherent Trade-Offs in the Fair Determination of Risk Scores. Proc. 8th Conf. on Innovations in Theoretical Computer Science (ITCS).
source: Chouldekova (2017)
In general it is impossible to satisfy multiple fairness definitions at the same time!
A choice is therefore required.
3. Fairness-by-design
“post”
processing
“pre”
processing
“in”
processing
 No changes to model itself
 Alter use of model or insert additional randomness
 No changes to algorithm
 Alter input data
 Tailor-made methods: new algorithms
If we detect problems: how to fix them? And how to create models which are `fair-by-design’?
Google (2018). Reducing gender bias in Google Translate.
 suppose 𝑷 𝒀 = 𝟏 | 𝑨 = 𝒃 = 0.3 and
𝑷 𝒀 = 𝟏 | 𝑨 = 𝒘 = 0.5.
 add additional randomness:
for A = b and 𝒀 = 𝟎 flip a coin with prob.
`heads’ equal to 0.2 and set 𝒀 = 𝟏 if `heads’
 yields statistical parity
 but loss of performance!
Examples Post Processing
3. Fairness-by-design
protected features
A
X
Y
data
model
features
transformed
features
Calmon, F.P., D. Wei, B. Vinzamuri, K.N. Ramamurthy, and K.R. Varshney (2017). Optimized pre-processing for discrimination prevention, NIPS.
 advantage: existing algorithms can still be used in Step 2
 in general there will be a reduction of performance!
Step 1 Step 2
Example Pre-Processing
3. Fairness-by-design
new tailor made algorithms / model
common structure:
max “performance” subject to “fairness
constraints”
max [“performance” + 𝛼“fairness”]
in general there will be a reduction of performance!
see, for example, IBM AI Fairness 360 for (open
source) demo of different algorithms
Zafar, M.B., I. Valera, M. Gomez-Rodriguez, and K.P. Gummadi (2019). Fairness Constraints: A Flexible Approach for Fair Classification, Journal of Machine Learning Research 20 , pp. 1-42
source: Zafar et al. (2019)
3. Fairness-by-design
Example In-Processing
3. Round-up and reflection
 Implementation fairness requires choice of protected groups
and choice for fairness metric
 Law does not seem to offer guidance on the precise metric
 GDPR can yield problems if definition protected group
depends on special personal data.
 Even worse: in some cases it will concern data that most
companies do not (want to) have available
 There is trade-off between performance and fairness. If not
clearly specificied in law: what is the appropriate balance?
Gives rise to new governance questions:
- Who should be responsible for decisions on fairness?
- How to ensure company-wide consistency in decision
making?
About me
Ramon van den Akker
Risk Modelling and AI Center, de Volksbank
Associate Professor, department of Econometrics & OR, Tilburg University
ramon.vandenakker@devolksbank.nl, r.vdnakker@uvt.nl
www.linkedin.com/in/ramonvdakker/

More Related Content

PPTX
Cyworld Jeju 2009 Conference(10 Aug2009)No2(2)
PDF
Using deep learning and Google Street View to estimate the demographic makeup...
PDF
Making sense of text: artificial intelligence-enabled content analysis
PDF
International Journal of Computational Engineering Research (IJCER)
PDF
M2 l10 fairness, accountability, and transparency
PDF
Binary search query classifier
PPT
Who should be nominated to run in the 2012 U.S. Presidential Election?
DOCX
Interpreting the public sentiment variations ons on twitter
Cyworld Jeju 2009 Conference(10 Aug2009)No2(2)
Using deep learning and Google Street View to estimate the demographic makeup...
Making sense of text: artificial intelligence-enabled content analysis
International Journal of Computational Engineering Research (IJCER)
M2 l10 fairness, accountability, and transparency
Binary search query classifier
Who should be nominated to run in the 2012 U.S. Presidential Election?
Interpreting the public sentiment variations ons on twitter

Similar to Ramon van den Akker. Fairness of machine learning models an overview and practical considerations (20)

PDF
Fairness-aware Machine Learning: Practical Challenges and Lessons Learned (KD...
PDF
Algorithmic Bias In Education
PDF
Fairness in Machine Learning @Codemotion
PPTX
Measures and mismeasures of algorithmic fairness
PDF
Regoli fairness deep_learningitalia_20220127
PDF
Don't blindly trust your ML System, it may change your life (Azzurra Ragone, ...
PPTX
Call Center and AI: How Implement Ai and Ethics that Follow
PPTX
Fairness-aware Machine Learning: Practical Challenges and Lessons Learned (WS...
PPTX
Ethical Issues in Machine Learning Algorithms (Part 2)
PDF
1803.09010.pdf
PDF
Beyond-Accuracy Perspectives: Explainability and Fairness
PPTX
Transparency in ML and AI (humble views from a concerned academic)
PPTX
22903-56094---conference-presentation.pptx
PDF
Algorithmic Bias : What is it? Why should we care? What can we do about it?
PPTX
Ai2020 ai and or final
PDF
Predicting property prices with machine learning algorithms.pdf
PDF
Responsible AI
PPTX
Emancipatory Information Retrieval (Invited Talk at UCC)
PDF
Defining the boundary for AI research in Intelligent Systems Dec 2021
PDF
Fairness-aware learning: From single models to sequential ensemble learning a...
Fairness-aware Machine Learning: Practical Challenges and Lessons Learned (KD...
Algorithmic Bias In Education
Fairness in Machine Learning @Codemotion
Measures and mismeasures of algorithmic fairness
Regoli fairness deep_learningitalia_20220127
Don't blindly trust your ML System, it may change your life (Azzurra Ragone, ...
Call Center and AI: How Implement Ai and Ethics that Follow
Fairness-aware Machine Learning: Practical Challenges and Lessons Learned (WS...
Ethical Issues in Machine Learning Algorithms (Part 2)
1803.09010.pdf
Beyond-Accuracy Perspectives: Explainability and Fairness
Transparency in ML and AI (humble views from a concerned academic)
22903-56094---conference-presentation.pptx
Algorithmic Bias : What is it? Why should we care? What can we do about it?
Ai2020 ai and or final
Predicting property prices with machine learning algorithms.pdf
Responsible AI
Emancipatory Information Retrieval (Invited Talk at UCC)
Defining the boundary for AI research in Intelligent Systems Dec 2021
Fairness-aware learning: From single models to sequential ensemble learning a...
Ad

More from Lviv Startup Club (20)

PDF
Maksym Vyshnivetskyi: PMO Maturity and Continuous Improvement (UA)
PPTX
Oleksandr Ivakhnenko: Cold Outreach і Social Selling: просунуті техніки (UA)
PDF
Maksym Vyshnivetskyi: PMO KPIs (UA) - LemBS
PDF
Oleksandr Ivakhnenko: LinkedIn Marketing і Content Marketing: розширений підх...
PDF
Maksym Vyshnivetskyi: PMO Quality Management (UA)
PDF
Oleksandr Ivakhnenko: Вступ до генерації лідів для ІТ-аутсорсингу (UA)
PDF
Oleksandr Osypenko: Поради щодо іспиту та закриття курсу (UA)
PDF
Oleksandr Osypenko: Пробний іспит + аналіз (UA)
PDF
Oleksandr Osypenko: Agile / Hybrid Delivery (UA)
PDF
Oleksandr Osypenko: Стейкхолдери та їх вплив (UA)
PDF
Rostyslav Chayka: Prompt Engineering для проєктного менеджменту (Advanced) (UA)
PPTX
Dmytro Liesov: PMO Tools and Technologies (UA)
PDF
Rostyslav Chayka: Управління командою за допомогою AI (UA)
PDF
Oleksandr Osypenko: Tailoring + Change Management (UA)
PDF
Maksym Vyshnivetskyi: Управління закупівлями (UA)
PDF
Oleksandr Osypenko: Управління ризиками (UA)
PPTX
Dmytro Zubkov: PMO Resource Management (UA)
PPTX
Rostyslav Chayka: Комунікація за допомогою AI (UA)
PDF
Ihor Pavlenko: Комунікація за допомогою AI (UA)
PDF
Maksym Vyshnivetskyi: Управління якістю (UA)
Maksym Vyshnivetskyi: PMO Maturity and Continuous Improvement (UA)
Oleksandr Ivakhnenko: Cold Outreach і Social Selling: просунуті техніки (UA)
Maksym Vyshnivetskyi: PMO KPIs (UA) - LemBS
Oleksandr Ivakhnenko: LinkedIn Marketing і Content Marketing: розширений підх...
Maksym Vyshnivetskyi: PMO Quality Management (UA)
Oleksandr Ivakhnenko: Вступ до генерації лідів для ІТ-аутсорсингу (UA)
Oleksandr Osypenko: Поради щодо іспиту та закриття курсу (UA)
Oleksandr Osypenko: Пробний іспит + аналіз (UA)
Oleksandr Osypenko: Agile / Hybrid Delivery (UA)
Oleksandr Osypenko: Стейкхолдери та їх вплив (UA)
Rostyslav Chayka: Prompt Engineering для проєктного менеджменту (Advanced) (UA)
Dmytro Liesov: PMO Tools and Technologies (UA)
Rostyslav Chayka: Управління командою за допомогою AI (UA)
Oleksandr Osypenko: Tailoring + Change Management (UA)
Maksym Vyshnivetskyi: Управління закупівлями (UA)
Oleksandr Osypenko: Управління ризиками (UA)
Dmytro Zubkov: PMO Resource Management (UA)
Rostyslav Chayka: Комунікація за допомогою AI (UA)
Ihor Pavlenko: Комунікація за допомогою AI (UA)
Maksym Vyshnivetskyi: Управління якістю (UA)
Ad

Recently uploaded (20)

PDF
Journal of Dental Science - UDMY (2020).pdf
PDF
FOISHS ANNUAL IMPLEMENTATION PLAN 2025.pdf
PDF
IP : I ; Unit I : Preformulation Studies
PDF
CRP102_SAGALASSOS_Final_Projects_2025.pdf
PDF
Journal of Dental Science - UDMY (2021).pdf
PDF
Literature_Review_methods_ BRACU_MKT426 course material
PDF
CISA (Certified Information Systems Auditor) Domain-Wise Summary.pdf
PPTX
Introduction to pro and eukaryotes and differences.pptx
PDF
Environmental Education MCQ BD2EE - Share Source.pdf
PPTX
Share_Module_2_Power_conflict_and_negotiation.pptx
PPTX
Core Concepts of Personalized Learning and Virtual Learning Environments
PDF
BP 505 T. PHARMACEUTICAL JURISPRUDENCE (UNIT 1).pdf
PPTX
What’s under the hood: Parsing standardized learning content for AI
PDF
BP 704 T. NOVEL DRUG DELIVERY SYSTEMS (UNIT 1)
PDF
FORM 1 BIOLOGY MIND MAPS and their schemes
PDF
David L Page_DCI Research Study Journey_how Methodology can inform one's prac...
PDF
Hazard Identification & Risk Assessment .pdf
PPTX
Computer Architecture Input Output Memory.pptx
PDF
My India Quiz Book_20210205121199924.pdf
PDF
HVAC Specification 2024 according to central public works department
Journal of Dental Science - UDMY (2020).pdf
FOISHS ANNUAL IMPLEMENTATION PLAN 2025.pdf
IP : I ; Unit I : Preformulation Studies
CRP102_SAGALASSOS_Final_Projects_2025.pdf
Journal of Dental Science - UDMY (2021).pdf
Literature_Review_methods_ BRACU_MKT426 course material
CISA (Certified Information Systems Auditor) Domain-Wise Summary.pdf
Introduction to pro and eukaryotes and differences.pptx
Environmental Education MCQ BD2EE - Share Source.pdf
Share_Module_2_Power_conflict_and_negotiation.pptx
Core Concepts of Personalized Learning and Virtual Learning Environments
BP 505 T. PHARMACEUTICAL JURISPRUDENCE (UNIT 1).pdf
What’s under the hood: Parsing standardized learning content for AI
BP 704 T. NOVEL DRUG DELIVERY SYSTEMS (UNIT 1)
FORM 1 BIOLOGY MIND MAPS and their schemes
David L Page_DCI Research Study Journey_how Methodology can inform one's prac...
Hazard Identification & Risk Assessment .pdf
Computer Architecture Input Output Memory.pptx
My India Quiz Book_20210205121199924.pdf
HVAC Specification 2024 according to central public works department

Ramon van den Akker. Fairness of machine learning models an overview and practical considerations

  • 1. Fairness of machine learning models an overview and practical considerations AI & BigData Online Day 2020
  • 2. Agenda Disclaimer: The views and opinions expressed in this document are those of the author and do not necessarily reflect the official policy or position of de Volksbank and Tilburg University. Motivating Examples Operationalizing Fairness 1 3 Response: high-level frameworks2
  • 3. 1. Motivating examples  Correctional Offender Management Profiling for Alternative Sanctions (COMPAS)  Recidivism Prediction Instruments (RPI) used by US courts for pretrial decision-making, parole decisions, and sometimes sentencing.  On basis of 7,000 cases 2013-2014 in Broward county, Florida ProPublica’s conclusion is that algorithm is biased against blacks.
  • 4. Google (2018). Reducing gender bias in Google Translate. Zou, J. and L. Schiebinger (2018). AI can be sexist and racist — it’s time to make it fair, Nature 559, pp. 324-326. 1. Motivating examples
  • 5. 1. Motivating examples Ensign, D., S.A. Friedler, S. Neville, C. Scheidegger, and S. Venkatasubramanian (2018). Runaway Feedback Loops in Predictive Policing, Proceedings of Machine Learning Research 81, pp. 160-171.
  • 7. 1. Motivating examples 2016 20182016 20182006 2019
  • 8. 2. Response: regulation is coming… 2020
  • 9. Jobin, Ienca, Vayena (2019): • Mapped and analyzed the current corpus of principles and guidelines on ethical AI • Identified 84 documents containing ethical principles or guidelines for AI, 88% being after 2016. • Convergence? • “In particular, the prevalence of calls for transparency, justice and fairness points to an emerging moral priority”. Transparency & Explainability Justice and fairness Non-maleficence Responsibility Privacy 2. Meta-analysis: Global Landscape of Ethical Frameworks for AI and data science From point-of-view of data scientist: fairness and explainability appear to most difficult to implement.
  • 10. 3. Non-discrimination and data science but for data scientist, typical goal is to maximize discriminatory power! Models should not yield unintended discrimination
  • 11. protected features A X Y data model featureslabels3. Avoiding unintended discrimination: forbidden variables Supervised Learning Ignoring protected features does not work (in general)!
  • 12. 3. Defining (un)fairness, discrimination  How to define (un)fairness?  Note that is actually not a new topic  see, for example, Hutchinson and Mitchell (2019)  rapidly developing literature  mainly for binary classifiers!  Barocas, S., M. Hardt, and A. Narayanan (2020). Fairness and machine learning – limitations and opportunities. Book in progress (https://fairmlbook.org/)  Corbett-Davies, S. and S. Goel (2018). The Measure and Mismeasure of Fairness: A Critical Review of Fair Machine Learning, working paper  Hutchinson, B. and M. Mitchell (2019). 50 Years of Test (Un)fairness: Lessons for Machine Learning. In: FAT* ’19: Conference on Fairness, Accountability, and Transparency (FAT* ’19), ACM, pp. 49-58.  Verma, S. and J. Rubin (2018). Fairness Definitions Explained. In: 2018 IEEE/ACM International Workshop on Software Fairness (FairWare). IEEE, pp. 1–7. statistical parity fairness through unawareness equal opportunity predictive parity calibration conditional statistical parity equalized odds counter- factual fairness no proxy discrimination
  • 13. Great variety of open-source fairness dashboarding tools are available AI Fairness 360 3. Bias/fairness dashboarding Saleiro, P., B. Kuester, A. Stevens, A. Anisfeld, L. Hinkson, J. London, and R. Ghani (2018). Aequitas: A Bias and Fairness Audit Toolkit, arXiv preprint arXiv:1811.05577. How to proceed if protected group concerns special personal data? Implementation requires choices for protected groups and fairness metrics!
  • 14. 3. Why do we need to make a choice: just consider all fairness metrics We think ProPublica’s report was based on faulty statistics and data analysis, and that the report failed to show that the COMPAS itself is racially biased, let alone that other risk instruments are biased. -- Flores et al. (2016) There’s software used across the country to predict future criminals. And it’s biased against blacks. -- ProPublica (2016) source: Chouldekova (2017)  Chouldechova (2017). Fair Prediction with Disparate Impact: A Study of Bias in Recidivism Prediction Instruments, Big Data 5(2), pp. 153-163.  Flores, A.W., K. Bechtel, and C.T. Lowenkamp (2016). False Positives, False Negatives, and False Analyses: A Rejoinder to “Machine Bias: There’s Software Used Across the Country to Predict Future Criminals. And It’s Biased Against Blacks, Federal Probation 80(2), pp.38-46.  Kleinberg, J., S. Mullainathan, and M. Raghavan (2017). Inherent Trade-Offs in the Fair Determination of Risk Scores. Proc. 8th Conf. on Innovations in Theoretical Computer Science (ITCS). source: Chouldekova (2017) In general it is impossible to satisfy multiple fairness definitions at the same time! A choice is therefore required.
  • 15. 3. Fairness-by-design “post” processing “pre” processing “in” processing  No changes to model itself  Alter use of model or insert additional randomness  No changes to algorithm  Alter input data  Tailor-made methods: new algorithms If we detect problems: how to fix them? And how to create models which are `fair-by-design’?
  • 16. Google (2018). Reducing gender bias in Google Translate.  suppose 𝑷 𝒀 = 𝟏 | 𝑨 = 𝒃 = 0.3 and 𝑷 𝒀 = 𝟏 | 𝑨 = 𝒘 = 0.5.  add additional randomness: for A = b and 𝒀 = 𝟎 flip a coin with prob. `heads’ equal to 0.2 and set 𝒀 = 𝟏 if `heads’  yields statistical parity  but loss of performance! Examples Post Processing 3. Fairness-by-design
  • 17. protected features A X Y data model features transformed features Calmon, F.P., D. Wei, B. Vinzamuri, K.N. Ramamurthy, and K.R. Varshney (2017). Optimized pre-processing for discrimination prevention, NIPS.  advantage: existing algorithms can still be used in Step 2  in general there will be a reduction of performance! Step 1 Step 2 Example Pre-Processing 3. Fairness-by-design
  • 18. new tailor made algorithms / model common structure: max “performance” subject to “fairness constraints” max [“performance” + 𝛼“fairness”] in general there will be a reduction of performance! see, for example, IBM AI Fairness 360 for (open source) demo of different algorithms Zafar, M.B., I. Valera, M. Gomez-Rodriguez, and K.P. Gummadi (2019). Fairness Constraints: A Flexible Approach for Fair Classification, Journal of Machine Learning Research 20 , pp. 1-42 source: Zafar et al. (2019) 3. Fairness-by-design Example In-Processing
  • 19. 3. Round-up and reflection  Implementation fairness requires choice of protected groups and choice for fairness metric  Law does not seem to offer guidance on the precise metric  GDPR can yield problems if definition protected group depends on special personal data.  Even worse: in some cases it will concern data that most companies do not (want to) have available  There is trade-off between performance and fairness. If not clearly specificied in law: what is the appropriate balance? Gives rise to new governance questions: - Who should be responsible for decisions on fairness? - How to ensure company-wide consistency in decision making?
  • 20. About me Ramon van den Akker Risk Modelling and AI Center, de Volksbank Associate Professor, department of Econometrics & OR, Tilburg University ramon.vandenakker@devolksbank.nl, r.vdnakker@uvt.nl www.linkedin.com/in/ramonvdakker/

Editor's Notes

  • #16: 21+ definitions of fairness No consensus on which one is the best, depends on the case Most are mathematically incompatible