networks with prior knowledge Laura Rieger Chandan Singh W. James - - PowerPoint PPT Presentation

networks with prior knowledge
SMART_READER_LITE
LIVE PREVIEW

networks with prior knowledge Laura Rieger Chandan Singh W. James - - PowerPoint PPT Presentation

Interpretations are useful: penalizing explanations to align neural networks with prior knowledge Laura Rieger Chandan Singh W. James Murdoch Bin Yu DTU UC Berkeley UC Berkeley UC Berkeley overview datasets are biased Benign NNs


slide-1
SLIDE 1

Interpretations are useful: penalizing explanations to align neural networks with prior knowledge

Laura Rieger DTU Chandan Singh UC Berkeley

  • W. James Murdoch

UC Berkeley Bin Yu UC Berkeley

slide-2
SLIDE 2
  • verview
slide-3
SLIDE 3

datasets are biased

  • NNs learn from large datasets
  • often biased
  • we sometimes know the bias

Benign Cancerous

slide-4
SLIDE 4

augmenting the loss function

Prediction True label Explanation Prior knowledge

slide-5
SLIDE 5

using our method improves accuracy

Image Vanilla Our method

Test F1: 0.67 0.73

more focus on skin less focus on band-aid

slide-6
SLIDE 6

details

slide-7
SLIDE 7

Learning from labels (step by step)

 90% accurate

training with biased data

Benign Cancerous

slide-8
SLIDE 8

what did the network learn?

Benign Cancerous

slide-9
SLIDE 9

We know the bias (sometimes)

Gender is not important for job applications! Race shouldn’t determine jail time! Rulers aren’t cancerous! Band aids don’t protect against cancer!

slide-10
SLIDE 10
  • ur method
slide-11
SLIDE 11

augmenting the loss function

Prediction True label

slide-12
SLIDE 12

augmenting the loss function

Prediction True label Explanation Prior knowledge

slide-13
SLIDE 13

Contextual Decomposition Explanation Penalty

13

[1] Singh, Chandan, W. James Murdoch, and Bin Yu. "Hierarchical interpretations for neural network predictions."

any differentiable explanation method works we used contextual decomposition (Singh 2019)

captures interactions computationally lighter

slide-14
SLIDE 14

Contextual Decomposition (Singh 2019)

  • requires partition of input
  • iteratively forward-pass both partitions
  • output contribution of both partitions
slide-15
SLIDE 15

results

slide-16
SLIDE 16

skin cancer (ISIC)

explanations focus more on skin

slide-17
SLIDE 17

mnist variants

slide-18
SLIDE 18

contributions

slide-19
SLIDE 19

contributions

CDEP uses explainability methods to regularize an NN used to incorporate prior knowledge into neural networks usable with more complex knowledge than previous methods

0.67 (f1) 0.73 (f1) unpenalized penalized