Interpreting Interpretations: Organizing Attribution Methods by - - PowerPoint PPT Presentation

interpreting interpretations organizing attribution
SMART_READER_LITE
LIVE PREVIEW

Interpreting Interpretations: Organizing Attribution Methods by - - PowerPoint PPT Presentation

IEEE CVPR2020 WORKSHOP ON FAIR, DATA EFFICIENT AND TRUSTED COMPUTER VISION Interpreting Interpretations: Organizing Attribution Methods by Criteria Zifan Wang, Piotr Mardziel, Anupam Datta, Matt Fredrikson Accountable System Lab


slide-1
SLIDE 1

1

Interpreting Interpretations: Organizing Attribution Methods by Criteria

Zifan Wang, Piotr Mardziel, Anupam Datta, Matt Fredrikson Accountable System Lab (https://fairlyaccountable.org/) Carnegie Mellon University

IEEE CVPR2020 WORKSHOP ON FAIR, DATA EFFICIENT AND TRUSTED COMPUTER VISION

slide-2
SLIDE 2

Introduction

2

Dog Attribution Map : Find the most important features in the input towards the prediction Concept-Based Explanations Attribution Map Activation-Based Explanations Instance-based explanations The lady IS or IS NOT used by the model to predict “dog” class Attribution Maps may not agree with each other

𝑦 = [𝑦!, 𝑦", … , 𝑦#$"] 𝑨 = [𝑨!, 𝑨", … , 𝑨#$"]

Input

Attribution Map

slide-3
SLIDE 3

Background

3

  • Decompose the importance

Importance Necessity Sufficiency Without therse features the model will lose more confidence than others With these features independently, the model will gain more confidence than others A necessary condition is

  • ne without which a

statement is false A sufficient condition is

  • ne which can

independently make a statement true Deep Neural Networks Logical meaning

  • Goal of this paper:

Evaluate different attribution methods with numerical analysis to answer which attribution method is better at what extent ?

  • Recall:

Attribution Map : Find the most important features in the input towards the prediction

slide-4
SLIDE 4

Methods

4

Quantifying Necessity & Sufficiency

  • Criteria One: Ordering

Scores à Rank Order Quantify the ordering: Modify features from top rank orders to the bottom. Modify à Ablate (Necessity) Modify à Add (Sufficiency)

Sufficiency Ordering 𝑔(𝑦) 𝑔(𝑦)

Share of pixel Share of pixel

Necessity Ordering: AOPC [Samek et al. 2017], AUC [Binder et al. 2016], MoRF [Ancona et al. 2017] Sufficiency Ordering:LeRF [Ancona et al. 2017] Average % Dorp [Chattopadhyay et al. 2017] 0.99 0.67 0.95 0.3 0.02 Necessity Ordering Similar Metrics

slide-5
SLIDE 5

Methods

5

Quantifying Necessity & Sufficiency

  • Criteria One: Ordering

However, there is more than rank orders that attribution scores offer, and the actual values are

  • verlooked.
slide-6
SLIDE 6

Methods

6

Quantifying Necessity & Sufficiency

  • Criteria Two: Proportionality

Motivations:

  • Magnitude also captures information
  • Interpret attribution scores linearly:

given 𝑦", 𝑦%, if 𝑨" = 2𝑨%, then 𝑦"is expected to be twice important than 𝑦% is.

Sum ( ) = Sum ( ) Equally important Choose from Top 1 … N Choose from Top N … 1 Randomly choose regions --> Loosing rank order information

Sensitivity-N [Ancona et al. 2017]

Method Necessity Sufficiency Property under a given ordering: TPN TPS Ours:

slide-7
SLIDE 7

Methods

7

Quantifying Necessity & Sufficiency

  • Criteria Two: Proportionality

TPN measures Necessity TPS measures Sufficiency Area between the curves measures disproportionality. Smaller area between the curves, better the Necessity/Sufficiency

slide-8
SLIDE 8

Methods

8

Summary of Evaluation Metrics

Ordering Proportionality Sufficiency Necessity N-Ord S-Ord TPN TPS What it means by importance What should be

  • ffered in the

attribution map

slide-9
SLIDE 9

Results

9

Evaluation performed with ImageNet and pretrainedVGG-16 model

lower scores indicate the better performance Recommended Methods under Ordering Criteria:

  • Necessity: DeepLIFT
  • Sufficiency: GradCAM, LRP

Recommended Methods under Proportionality Criteria:

  • Necessity: SmoothGrad, Saliency Map, Integrated

Gradient

  • Sufficiency: Guided Backpropagation, LRP, DeepLIFT
slide-10
SLIDE 10

Interpret Interpretations

10

The lady IS or IS NOT used by the model to predict “dog” class GradCAM highlights an area of sufficient features (Good S-Ord), the model can make correct prediction without the lady. However, among the sufficient features, higher scores do not strictly mean that those features are a lot more sufficient than features with lower scores (Poor TPS). SmoothGrad highlights necessary features (Poor S-Ord and TPS) , and attribution scores are proportional to actual necessity (Good TPN). Point clouds around the lady are less intensive compared to the dog; therefore, the lady is less necessary compared to the dog in the prediction towards “dog” class. Conclusions of experiments let us impart additional interpretation to these results

How do we use necessity/sufficiency and ordering/proportionality to interpret different attribution maps.

slide-11
SLIDE 11

11

Zifan Wang, Piotr Mardziel, Anupam Datta, Matt Fredrikson Accountable System Lab (https://fairlyaccountable.org/) Carnegie Mellon University

IEEE CVPR2020 WORKSHOP ON FAIR, DATA EFFICIENT AND TRUSTED COMPUTER VISION

Thanks for watching our presentation Contact us: zifan.wang@sv.cmu.edu