Sanity Checks for ‘Saliency’ Maps
1
Julius Adebayo PhD Student, MIT.
Joint work with
Sanity Checks for Saliency Maps Julius Adebayo PhD Student, MIT. - - PowerPoint PPT Presentation
Sanity Checks for Saliency Maps Julius Adebayo PhD Student, MIT. Joint work with 1 Some Motivation [Challenges for Transparency, Weller 2017, & Doshi-Velez & Kim, 2017 ] Developer/Researcher: Model Debugging. Safety
1
Julius Adebayo PhD Student, MIT.
Joint work with
learned from data.
2
[Challenges for Transparency, Weller 2017, & Doshi-Velez & Kim, 2017 ]
3
inputs that a model is most likely to have undesirable performance. [Ribeiro+ 2016]
4
inputs that a model is most likely to have undesirable performance.
Husky
5
inputs that a model is most likely to have undesirable performance.
Explanation
Husky
6
inputs that a model is most likely to have undesirable performance.
Explanation
Husky Fix
for deep neural networks.
8
Predictions
Explanation
Corn
9
Predictions
Explanation
Corn
Attribution maps provide ‘relevance’ scores for each dimension of the input.
10
Predictions
Explanation
Corn
Attribution maps provide ‘relevance’ scores for each dimension of the input.
11
Predictions
Attribution
Corn
Egrad(x) = ∂Si ∂x
[SVZ’13]
12
Predictions Corn
‘Visually noisy’, and can violate sensitivity w.r.t. a baseline input [Sundararajan et. al., Shrikumar et. al., and Smilkov et. al.]
13
Predictions Corn
Sum of ‘interior’ gradients.
[STY’17]
14
Predictions Corn
Average attribution of ‘noisy’ inputs.
[STKVW’17]
15
Predictions Corn
Element-wise product of gradient and input.
16
Predictions Corn
Zero out ‘negative’ gradients and ‘activations’ while back-propagating.
17
Predictions
Explanation
Corn
[FV’17]
Formulate an explanation as through learned patch removal.
18
19
Predictions Corn
initialize) the parameters of a model and now compare attribution maps for a trained model to those derived from a randomized model.
model trained with correct labels to those derived from a model trained with random labels.
Inception V3
Conjecture: If a model captures higher level class concepts, then saliency maps should change as the model is being randomized.
24
Gradient Gradient-SG Gradient Input Guided Back-propagation GradCAM Integrated Gradients Integrated Gradients-SG
Original Explanation
Guided GradCAM
Cascading randomization from top to bottom layers
Original Image
Conjecture: If a model captures higher level class concepts, then saliency maps should change as the model is being randomized.
25
Gradient Gradient-SG Gradient Input Guided Back-propagation GradCAM Integrated Gradients Integrated Gradients-SG
logits Original Explanation
Guided GradCAM
Cascading randomization from top to bottom layers
Original Image
Conjecture: If a model captures higher level class concepts, then saliency maps should change as the model is being randomized.
26
Gradient Gradient-SG Gradient Input Guided Back-propagation GradCAM Integrated Gradients Integrated Gradients-SG
logits mixed_7c mixed_7b mixed_7a mixed_6e mixed_6d mixed_6c mixed_6b Original Explanation
Guided GradCAM
Cascading randomization from top to bottom layers
Original Image
Conjecture: If a model captures higher level class concepts, then saliency maps should change as the model is being randomized.
27
Gradient Gradient-SG Gradient Input Guided Back-propagation GradCAM Integrated Gradients Integrated Gradients-SG
logits conv2d_1a_3x3 mixed_7c mixed_7b conv2d_2a_3x3 conv2d_2b_3x3 conv2d_4a_3x3 mixed_7a mixed_6e mixed_6d mixed_6c mixed_6b mixed_6a mixed_5d mixed_5c mixed_5b conv2d_3b_1x1 Original Explanation
Guided GradCAM
Cascading randomization from top to bottom layers
Original Image
Conjecture: If a model captures higher level class concepts, then saliency maps should change as the model is being randomized.
28
Gradient Gradient-SG Gradient Input Guided Back-propagation GradCAM Integrated Gradients Integrated Gradients-SG
logits conv2d_1a_3x3 mixed_7c mixed_7b conv2d_2a_3x3 conv2d_2b_3x3 conv2d_4a_3x3 mixed_7a mixed_6e mixed_6d mixed_6c mixed_6b mixed_6a mixed_5d mixed_5c mixed_5b conv2d_3b_1x1 Original Explanation
Guided GradCAM
Cascading randomization from top to bottom layers
Original Image
Inception v3 - ImageNet
Mixed
7c 7b 7a 6e 6d 6c 6b 6a 5d 5c 5b 4a 3b 2b 2a 1a
logits
Conv2d
Rank Correlation ABS
Mixed
7c 7b 7a 6e 6d 6c 6b 6a 5d 5c 5b 4a 3b 2b 2a 1a
logits
Conv2d
Rank Correlation No ABS
See Caption Note
those derived from partially randomized models.
attributed.
Successive Randomization of Layers
Gradient Gradient-SG Gradient-VG Guided Backpropagation Guided GradCAM Integrated Gradients Integrated Gradients-SG
conv_hidden1 conv_hidden2 fc2
Original Image Explanation conv_hidden1 conv_hidden2
fc2
Independent Randomization of Layers
CNN MNIST
Guided Backpropagation Skeletal Radiograph
Age
32
CNN - MNIST
True Labels Random Labels
Gradient Gradient-SG Guided BackProp GradCAM Guided GradCAM Integrated Gradients Integrated Gradients-SG Gradient Input
True Labels Random Labels
Gradient Gradient-SG Guided BackProp GradCAM Guided GradCAM Integrated Gradients Integrated Gradients-SG Gradient Input
Rank Correlation - Abs Rank Correlation - No Abs
Absolute-Value Visualization Diverging Visualization
33
True Labels Random Labels True Labels Random Labels Rank Correlation - Abs
Rank Correlation - No Abs MLP - MNIST
Gradient Gradient-SG Guided BackProp Integrated Gradients Integrated Gradients-SG Gradient Input Gradient Gradient-SG Guided BackProp Integrated Gradients Integrated Gradients-SG Gradient Input
Absolute-Value Visualization Diverging Visualization
propagation is doing input reconstruction.
Figure from Nie et. al, 2018.
LIME-5 LIME-10 LIME-20 LIME-50 SHAP Gradient SmoothGrad Guided BackProp PatternNet Pattern Attribution Input-Gradient Integrated Gradients LRP-Z LRP-EPS LRP-SPAF LRP-SPBF VGrad DeepTaylor
Cascading randomization from top to bottom layers for VGG-16
LIME Variants
Not Previously considered in literature.
37
[Figure from Gupta et. al. 2019.]
Neurips 2019) propose to remove and retrain.
‘quantify’ information content.
benchmark (w/ground truth) and other metrics to assess how well a map captures model behavior.
39