What You See Is What You Get? The Impact of Representation Criteria - - PowerPoint PPT Presentation

what you see is what you get the impact of representation
SMART_READER_LITE
LIVE PREVIEW

What You See Is What You Get? The Impact of Representation Criteria - - PowerPoint PPT Presentation

What You See Is What You Get? The Impact of Representation Criteria on Human Bias in Hiring Andi Peng, Besmira Nushi, Emre Kiciman, Kori Inkpen, Sid Suri, Ece Kamar AAAI HCOMP 10.30.2019 Recidivism prediction, bail assessment, proactive


slide-1
SLIDE 1

What You See Is What You Get? The Impact of Representation Criteria

  • n Human Bias in Hiring

AAAI HCOMP 10.30.2019

Andi Peng, Besmira Nushi, Emre Kiciman, Kori Inkpen, Sid Suri, Ece Kamar

slide-2
SLIDE 2

AI-Advised Decision-Making is Everywhere

2

Lending, mortgage risk assessment, quantitative trading Recidivism prediction, bail assessment, proactive policing Drug development, diagnosis, personalized medicine

slide-3
SLIDE 3

Bias from AI is Everywhere

3

Less likely to approve loans to Hispanic applicants More likely to think black defendants to recidivate Under-estimates the necessary amount of care needed for black individuals

slide-4
SLIDE 4

Bias from Humans is Also Everywhere

4

Less likely to approve loans to Hispanic applicants More likely to think black defendants to recidivate Under-estimates the necessary amount of care needed for black individuals

slide-5
SLIDE 5

HIRING

5

slide-6
SLIDE 6

6

slide-7
SLIDE 7

Hiring is a complex workflow

7

Hiring recommendations World distribution Candidate pool Human decision Non-algorithmic decision-making

slide-8
SLIDE 8

8

slide-9
SLIDE 9

Hiring is a complex workflow

9

Hiring recommendations World distribution Candidate pool Human decision Non-algorithmic decision-making World and Societal bias

slide-10
SLIDE 10

Human bias in the workplace

10

Women are:

  • More likely to be employed in low-wage jobs (Tobin, 2017)
  • Less likely to be called back by resume screens (Bertrand and

Mullainathan, 2003)

  • Less likely to be promoted as managers (Koch et al., 2015)
  • Less likely to be recommended as candidates to be promoted as

managers (Work in the Workplace Report, 2019)

  • More likely to face general sexism in the workplace (Masser and

Abrams, 2004)

  • … and all sorts of other bad things
slide-11
SLIDE 11

Hiring is a complex workflow

11

Hiring recommendations World distribution Candidate pool Human decision Non-algorithmic decision-making World and Societal bias Human bias

slide-12
SLIDE 12

Hiring is a complex workflow

12

Hiring recommendations World distribution Candidate pool Screening algorithm Candidate slate Human decision Non-algorithmic decision-making World and Societal bias Algorithmic bias Human bias

slide-13
SLIDE 13

Have we tried fixing it?

13

slide-14
SLIDE 14

Geyik et al., KDD 2019 14

slide-15
SLIDE 15

LinkedIn Representational Ranking

15

Hiring recommendations World distribution Candidate pool Screening algorithm Candidate slate Human decision Non-algorithmic decision-making World and Societal bias Algorithmic bias Human bias

slide-16
SLIDE 16

Does this work?

16

slide-17
SLIDE 17

Does this work? ¯\_(ツ)_/¯

17

slide-18
SLIDE 18

Can we decompose these different sources of biases?

18

slide-19
SLIDE 19

Can we decompose these different sources of biases? Can we mitigate them?

19

slide-20
SLIDE 20

Experimental Design

20

Candidate bios (ex: physician)

slide-21
SLIDE 21

Bucket 1

  • Doctors (dermatologists, neurologists, OBGYNs, orthopedic

surgeons, pediatricians, physicians, urologists), nannies, plumbers, elementary school teachers Bucket 2

  • Software engineers, software engineering managers,

administrative assistants, customer service representatives

We generate controlled candidate bios for different professions

21

slide-22
SLIDE 22

MALE:

  • Dr. Robert Brown, MD, is a board-certified orthopedic surgeon who, since 2002, practices at the

Cleveland Clinic in Beachwood, OH. He is a graduate of the Johns Hopkins School of Medicine and completed his residency in Cleveland. He spends much of his time educating medical students at the Cleveland Clinic Lerner College of Medicine of Case Western Reserve University, where he serves as an Orthopedics Advisor and as Course Director for rotations that integrate bone fracture prevention and healthy living. His practice interests include health maintenance and diet/exercise, in addition to joint

  • replacement. In his free time, Robert enjoys biking and exploring the outdoors.

We generate controlled candidate bios for different professions

22

slide-23
SLIDE 23

FEMALE:

  • Dr. Mary Brown, MD, is a board-certified orthopedic surgeon who, since 2002, practices at the

Cleveland Clinic in Beachwood, OH. She is a graduate of the Johns Hopkins School of Medicine and completed her residency in Cleveland. She spends much of her time educating medical students at the Cleveland Clinic Lerner College of Medicine of Case Western Reserve University, where she serves as an Orthopedics Advisor and as Course Director for rotations that integrate bone fracture prevention and healthy living. Her practice interests include health maintenance and diet/exercise, in addition to joint

  • replacement. In her free time, Mary enjoys biking and exploring the outdoors.

We generate controlled candidate bios for different professions

23

MALE:

  • Dr. Robert Brown, MD, is a board-certified orthopedic surgeon who, since 2002, practices at the

Cleveland Clinic in Beachwood, OH. He is a graduate of the Johns Hopkins School of Medicine and completed his residency in Cleveland. He spends much of his time educating medical students at the Cleveland Clinic Lerner College of Medicine of Case Western Reserve University, where he serves as an Orthopedics Advisor and as Course Director for rotations that integrate bone fracture prevention and healthy living. His practice interests include health maintenance and diet/exercise, in addition to joint

  • replacement. In his free time, Robert enjoys biking and exploring the outdoors.
slide-24
SLIDE 24

Experimental Design

24

Candidate bios (ex: physician) Control distribution of candidate slates

slide-25
SLIDE 25

Representation criteria:

  • World baselines (current world breakdown of the profession)
  • Over/under-representation (25% F, 50% F, 75% F)

Task generation:

  • 8 candidates per slate
  • 100 unique HIT tasks per profession per distribution (100 x 4 x 14)
  • Based on distribution, randomly assign gender
  • Random order

25

We create candidate slates of different distributions

slide-26
SLIDE 26

Experimental Design

26

Candidate bios (ex: physician) Control distribution of candidate slates Human Ranking Task

slide-27
SLIDE 27

27

We ask participants to rank their top 4 candidates

slide-28
SLIDE 28

Experimental Design

28

Decisions (Biased?) Candidate bios (ex: physician) Control distribution of candidate slates Human Ranking Task

slide-29
SLIDE 29

Bias measure

  • We model each outputted set of ranking decisions as a

hypergeometric distribution1

  • If the observed (output) distribution is statistically different from

the expected (input) distribution, the system is biased

  • We ascribe no notion of fairness

We compare expected vs. observed rankings

29

1 This models the discrete probability distribution of binary draws without replacement from a finite population. If you ask me

what that means, I will defer your question to the coffee break so that I have time to re-learn what that means.

slide-30
SLIDE 30

Example: a decision biased towards female candidates

30

slide-31
SLIDE 31

RESULTS

31

slide-32
SLIDE 32

We’ve solved it. No more bias in the world.

32

slide-33
SLIDE 33

Is this a world distribution problem?

33

slide-34
SLIDE 34

Is this a world distribution problem? Can balancing candidate slates mitigate gender bias?

34

slide-35
SLIDE 35

Result 1a: enforcing balanced slates can mitigate bias

35

Profession % Female in World % Female Ranked in Top 4 Plumber 3.5 50.0 (0.513) Orthopedic surgeon 5.3 47.0 (0.086) Software engineer 19.3 53.0 (0.460) Software eng. manager 27.0 48.0 (0.659) Neurologist 29.4 49.0 (0.420) Physician 40.0 51.0 (0.907) Pediatrician 52.8 51.0 (0.171) Customer service rep. 63.7 48.0 (0.301) Administrative assistant 71.7 54.0 (0.301) Elementary teacher 79.8 50.0 (0.391) *Significant at the 0.05 level

slide-36
SLIDE 36

Result 1b: but sometimes, this isn’t enough

36

Profession % Female in World % Female Ranked in Top 4 Urologist 8.7 47.0 (0.005)* Dermatologist 48.9 45.0 (0.013)* OBGYN 57.0 60.0 (<0.000)* Nanny 94.0 58.0 (<0.000)* *Significant at the 0.05 level

slide-37
SLIDE 37

Can over-representation help?

37

slide-38
SLIDE 38

Result 2: no, some professions consistently produce biased decisions

38

slide-39
SLIDE 39

39

Is human preference driving this bias?

slide-40
SLIDE 40

40

Is human preference driving this bias? Do personal features of the decision-maker, such as gender, impact the decision?

slide-41
SLIDE 41

Result 3a: aggregate bias is sometimes driven by one gender

41

slide-42
SLIDE 42

42

Result 3b: aggregate bias is sometimes hidden by

  • pposite effects by each gender
slide-43
SLIDE 43

Limitations

43

  • MTurk generalizability
  • Simulated bios
  • No variance in bios
  • Bias at the group, not individual, level
  • Binary gender
slide-44
SLIDE 44

TAKEAWAYS

44

slide-45
SLIDE 45

45

Look Simba, everything the light touches is

  • ur kingdom.

BIAS

slide-46
SLIDE 46

46

Look Simba, everything the light touches is

  • ur kingdom.

But what about that shadowy place?

BIAS

slide-47
SLIDE 47

47

Look Simba, everything the light touches is

  • ur kingdom.

But what about that shadowy place? That’s beyond our borders. You must never go there, Simba.

INTERPRETABILITY.

BIAS

slide-48
SLIDE 48

Takeaways

48

For many professions, effecting the world distribution can be a successful intervention.

slide-49
SLIDE 49

Takeaways

49

For many professions, effecting the world distribution can be a successful intervention. However, it’s not always feasible.

slide-50
SLIDE 50

Takeaways

50

Generally, hiring and promoting more women is not a bad idea.

slide-51
SLIDE 51

Future Work

51

Continue studying the interaction of algorithmic decision-making, particularly as broken down by human vs. algorithmic features. Deploy real machine learning algorithms to classify real candidate profiles for evaluation.1

1 Perhaps appearing at a conference near you in 2020 *knock on wood*

slide-52
SLIDE 52

Emre Kiciman Besmira Nushi Sid Suri Kori Inkpen Ece Kamar