Five or so actionable tips for building trust and being trustworthy - - PowerPoint PPT Presentation

five or so actionable tips for building trust and being
SMART_READER_LITE
LIVE PREVIEW

Five or so actionable tips for building trust and being trustworthy - - PowerPoint PPT Presentation

Five or so actionable tips for building trust and being trustworthy (in interactive learning) Stefano Teso , University of Trento stefano.teso@unitn.it ML increasingly being used in sensitive domains like criminal justice , hiring , . . .


slide-1
SLIDE 1

Five or so actionable tips for building trust and being trustworthy (in interactive learning)

Stefano Teso, University of Trento stefano.teso@unitn.it

slide-2
SLIDE 2

ML increasingly being used in sensitive domains like criminal justice, hiring, . . . Black-box ML models can be whimsical and hard to control How can we build justifiable trust into black-box ML models?

2

slide-3
SLIDE 3

How do humans establish/reject trust into others?

1R Hoffman et al. “Trust in automation”. In: IEEE Intelligent Systems (2013). 2Luke Chang et al. “Seeing is believing: Trustworthiness as a dynamic belief”. In: Cognitive psychology (2010).

3

slide-4
SLIDE 4

How do humans establish/reject trust into others? Understanding Trust involves understanding the other’s beliefs & intentions; it depends on the perceived understand- ability and competence1 Interaction Trust is updated dynamically, interaction establishes expectations2; it relies on directability (We have “hardware support” for all of this: theory of mind, mirror neurons, . . . )

1R Hoffman et al. “Trust in automation”. In: IEEE Intelligent Systems (2013). 2Luke Chang et al. “Seeing is believing: Trustworthiness as a dynamic belief”. In: Cognitive psychology (2010).

3

slide-5
SLIDE 5

How do humans establish/reject trust into others? Understanding Trust involves understanding the other’s beliefs & intentions; it depends on the perceived understand- ability and competence1 Interaction Trust is updated dynamically, interaction establishes expectations2; it relies on directability (We have “hardware support” for all of this: theory of mind, mirror neurons, . . . ) Alas, explainable AI is passive and interactive ML is opaque

1R Hoffman et al. “Trust in automation”. In: IEEE Intelligent Systems (2013). 2Luke Chang et al. “Seeing is believing: Trustworthiness as a dynamic belief”. In: Cognitive psychology (2010).

3

slide-6
SLIDE 6

Local Explanations with LIME

This helps to identify potential “Clever Hans” behavior3, but it does not provide the means to fix it

3Sebastian Lapuschkin et al. “Unmasking clever hans predictors and assessing what machines really learn”. In:

(2019).

4

slide-7
SLIDE 7

Active Learning

The user a) does not know the model’s beliefs, b) cannot affect them directly, c) has no clue of what his feedback does!

5

slide-8
SLIDE 8

Explanatory Active Learning

slide-9
SLIDE 9

Explanatory Active Learning4

a) Explain predictions (competence, understandability), b) Allow to correct explanations (directability)

4Stefano Teso and Kristian Kersting. “Explanatory interactive machine learning”. In: 2019.

6

slide-10
SLIDE 10

Explanation Corrections

  • 1. User’s correction indicates the false positive segments
  • 2. converts correction to counterexamples, i.e., fill in random

values while keeping the same label Example: husky predicted right for the wrong reasons

7

slide-11
SLIDE 11

Explanatory Guided Learning and Beyond

slide-12
SLIDE 12

On-going work

with Teodora Popordanoska and Mohit Kumar (KU Leuven)

8

slide-13
SLIDE 13

By witnessing that the model’s beliefs improve over time, the human “teacher” builds trust into the “student” model

9

slide-14
SLIDE 14

By witnessing that the model’s beliefs improve over time, the human “teacher” builds trust into the “student” model Problem: nothing prevents the machine from repeatedly choosing instances where it does well. Not so far-fetched: the machine does not know how to choose difficult instances, think of high-loss unknown unknowns

9

slide-15
SLIDE 15

Example: dog vs wolf, machine very certain everywhere

10

slide-16
SLIDE 16

Example: dog vs wolf, machine very certain everywhere What about unknown unknowns?

11

slide-17
SLIDE 17

Example: dog vs wolf, machine very certain everywhere AL doesn’t help with UUs, uncertainty may be wrong too

12

slide-18
SLIDE 18

Idea: let the user choose the challenging instances This is what professors do when testing students

13

slide-19
SLIDE 19

Piggy-back on Guided Learning

machine chooses rare label, user searches for example Useful for tackling class unbalance where AL fails

14

slide-20
SLIDE 20

The human teacher is blind:

  • impossible to establish justifiable trust
  • she may provide examples that teach nothing new to the

machine How can we expect the human to provide useful examples?5

5Interactive machine teaching with black-box models shows that blind teachers cannot do better than random

teachers.

15

slide-21
SLIDE 21

Explanatory Guided Learning

16

slide-22
SLIDE 22

Explanatory Guided Learning

17

slide-23
SLIDE 23

Explanatory Guided Learning

18

slide-24
SLIDE 24

Explanatory Guided Learning

19

slide-25
SLIDE 25

RESULTS

20

slide-26
SLIDE 26

Plan:

  • 1. Polish experiments with “imperfect user”
  • 2. Case study with real users (!)
  • 3. Hook up iterative machine teaching theory

21

slide-27
SLIDE 27

A “Theory of Mind” for Machine-Human Teams

mutual understanding guides trust, learning, and teaching

22

slide-28
SLIDE 28

23

slide-29
SLIDE 29

24