discrimination in decision making humans vs machines
play

Discrimination in Decision Making: Humans vs. Machines Muhammad - PowerPoint PPT Presentation

Discrimination in Decision Making: Humans vs. Machines Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez-Rodriguez, Krishna P. Gummadi Max Planck Institute for Software Systems Machine decision making q Refers to data-driven algorithmic


  1. Discrimination in Decision Making: Humans vs. Machines Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez-Rodriguez, Krishna P. Gummadi Max Planck Institute for Software Systems

  2. Machine decision making q Refers to data-driven algorithmic decision making q By learning over data about past decisions q To assist or replace human decision making q Increasingly being used in several domains q Recruiting: Screening job applications q Banking: Credit ratings / loan approvals q Judiciary: Recidivism risk assessments q Journalism: News recommender systems

  3. The concept of discrimination q Discrimination is a special type of unfairness q Well-studied in social sciences q Political science q Moral philosophy q Economics q Law q Majority of countries have anti-discrimination laws q Discrimination recognized in several international human rights laws q But, less-studied from a computational perspective

  4. Why, a computational perspective? 1. Datamining is increasingly being used to detect discrimination in human decision making q Examples: NYPD stop and frisk, Airbnb rentals

  5. Why, a computational perspective? 2. Learning to avoid discrimination in data-driven (algorithmic) decision making q Aren’t algorithmic decisions inherently objective? q In contrast to subjective human decisions q Doesn’t that make them fair & non-discriminatory? q Objective decisions can be unfair & discriminatory!

  6. Why, a computational perspective? q Learning to avoid discrimination in data-driven (algorithmic) decision making q A priori discrimination in biased training data q Algorithms will objectively learn the biases q Learning objectives target decision accuracy over all users q Ignoring outcome disparity for different sub-groups of users

  7. Our agenda: Two high-level questions How to detect discrimination in decision making? 1. Independently of who makes the decisions q Humans or machines q How to avoid discrimination when learning? 2. q Can we make algorithmic decisions more fair? q If so, algorithms could eliminate biases in human decisions q Controlling algorithms may be easier than retraining people

  8. This talk How to detect discrimination in decision making? 1. Independently of who makes the decisions q Humans or machines q How to avoid discrimination when learning? 2. q Can we make algorithmic decisions more fair? q If so, algorithms could eliminate biases in human decisions q Controlling algorithms may be easier than retraining people

  9. The concept of discrimination q A first approximate normative / moralized definition: wrongfully impose a relative disadvantage on persons based on their membership in some salient social group e.g., race or gender

  10. The concept of discrimination q A first approximate normative / moralized definition: wrongfully impose a relative disadvantage on persons based on their membership in some salient social group e.g., race or gender

  11. The devil is in the details q What constitutes a salient social group? q A question for political and social scientists q What constitutes relative disadvantage? q A question for economists and lawyers q What constitutes a wrongful decision? q A question for moral-philosophers q What constitutes based on? q A question for computer scientists

  12. Discrimination: A computational perspective q Consider binary classification using user attributes A 1 A 2 … A m Decision User 1 x 1,1 x 1,2 … x 1,m Accept x 2,1 x 2,m Reject User 2 User 3 x 3,1 x 3,m Reject … … … … … User n x n,1 x n,2 x n,m Accept

  13. Discrimination: A computational perspective q Consider binary classification using user attributes SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Reject … … … … … User n x n,1 x n,2 x n,m Accept q Some attributes are sensitive, others non-sensitive

  14. Discrimination: A computational perspective q Consider binary classification using user attributes SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Reject … … … … … User n x n,1 x n,2 x n,m Accept q Some attributes are sensitive, others non-sensitive Decisions should not be based on sensitive attributes!

  15. What constitutes “not based on”? q Most intuitive notion: Ignore sensitive attributes q Fairness through blindness or veil of ignorance q When learning, strip sensitive attributes from inputs q Avoids disparate treatment q Same treatment for users with same non-sensitive attributes q Irrespective of their sensitive attribute values q Situational testing for discrimination discovery checks for this condition

  16. Two problems with the intuitive notion When users of different sensitive attribute groups have different non-sensitive feature distributions, we risk Disparate Mistreatment 1. Even when training data is unbiased, sensitive attribute groups q might have different misclassification rates Disparate Impact 2. When labels in training data are biased, sensitive attribute groups q might see different beneficial outcomes to different extents Training data bias due to past discrimination q

  17. Background: Two points about learning To learn, we define & optimize a risk (loss) function 1. q Over all examples in training data q Risk function captures inaccuracy in prediction q So learning is cast as an optimization problem For efficient learning (optimization) 2. q We define loss functions so that they are convex

  18. Origins of disparate mistreatment SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Reject … … … … … User n x n,1 x n,2 x n,m Accept

  19. Origins of disparate mistreatment SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Reject … … … … … User n x n,1 x n,2 x n,m Accept q Suppose users are of two types: blue and pink

  20. Origins of disparate mistreatment SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Reject … … … … … User n x n,1 x n,2 x n,m Accept q Minimizing L(W) , does not guarantee L(W) and L (W) are equally minimized q Blue users might have a different risk / loss than red users!

  21. Origins of disparate mistreatment SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Reject … … … … … User n x n,1 x n,2 x n,m Accept q Minimizing L(W) , does not guarantee L(W) and L (W) are equally minimized q Stripping sensitive attributes does not help!

  22. Origins of disparate mistreatment SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Reject … … … … … User n x n,1 x n,2 x n,m Accept q Minimizing L(W) , does not guarantee L(W) and L (W) are equally minimized q To avoid disp. mistreatment, we need L(W) = L(W)

  23. Origins of disparate mistreatment SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Reject … … … … … User n x n,1 x n,2 x n,m Accept q Minimizing L(W) , does not guarantee L(W) and L (W) are equally minimized q Put differently, we need:

  24. Origins of disparate impact SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Reject … … … … … User n x n,1 x n,2 x n,m Accept

  25. Origins of disparate impact SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Accept … … … … … User n x n,1 x n,2 x n,m Reject q Suppose training data has biased labels!

  26. Origins of disparate impact SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Accept … … … … … User n x n,1 x n,2 x n,m Reject q Suppose training data has biased labels! q Classifier will learn to make biased decisions q Using sensitive attributes (SAs)

  27. Origins of disparate impact SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Accept … … … … … User n x n,1 x n,2 x n,m Reject q Suppose training data has biased labels! q Stripping SAs does not fully address the bias

  28. Origins of disparate impact SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Accept … … … … … User n x n,1 x n,2 x n,m Reject q Suppose training data has biased labels! q Stripping SAs does not fully address the bias q NSAs correlated with SAs will be given more / less weights q Learning tries to compensate for lost SAs

  29. Analogous to indirect discrimination q Observed in human decision making q Indirectly discriminate against specific user groups using their correlated non-sensitive attributes q E.g., voter-id laws being passed in US states q Notoriously hard to detect indirect discrimination q In decision making scenarios without ground truth

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend