1 / 2 8
Interpretable Deep Learning: Towards Understanding & Explaining DNNs
P a r t 3 : V a l i d a t i n g E x p l a n a t i
- n
s
W
- j
c i e c h S a m e k , G r é g
- i
r e M
- n
t a v
- n
, K l a u s
- R
- b
e r t M ü l l e r
Tutorials Interpretable Deep Learning: Towards Understanding & - - PowerPoint PPT Presentation
Tutorials Interpretable Deep Learning: Towards Understanding & Explaining DNNs P a r t 3 : V a l i d a t i n g E x p l a n a t i o n s W o j c i e c h S a m e k , G r g o i r e M o n t
1 / 2 8
W
c i e c h S a m e k , G r é g
r e M
t a v
, K l a u s
e r t M ü l l e r
2 / 2 8
Autonomous Driving Medical Diagnosis Networks (smart grids, etc.)
Visual Reasoning AlphaGo beats Go human champ Deep Net outperforms humans in image classification
3 / 2 8
4 / 2 8
[ B a c h ’ 1 5 ]
5 / 2 8
Q u e s t i
: S u p p
e t h a t w e h a v e p r
a g a t e d t h e r e l e v a n c e u n t i l a g i v e n l a y e r . H
s h
l d i t b e p r
a g a t e d
e l a y e r f u r t h e r ? I d e a : B y p e r f
m i n g a T a y l
e x p a n s i
t h e r e l e v a n c e .
[ M
t a v
’ 1 7 ]
6 / 2 8
R e l e v a n c e n e u r
: T a y l
e x p a n s i
: R e d i s t r i b u t i
:
7 / 2 8
✔ ✔
1 . n e a r e s t r
3 . g e n e r a l i z e d 2 . r e s c a l e d e x c i t a t i
s C h
c e
r
p
n t
( D e e p T a y l
g e n e r i c )
G e n e r a l i z e d r u l e
8 / 2 8
Q u e s t i
: I s t h e r e a c
n e c t i
b e t w e e n t h e t w
e t h
s ? F i n d t h e d i f f e r e n c e . . .
9 / 2 8
w h i c h c a n a l s
e r e w r i t t e n a s : F
n e t w
k s w i t h b i a s z e r
t h e p r
e d u r e b e c
e s e q u i v a l e n t t
r a d x i n p u t
[ s e e a l s
h r i k u m a r ’ 1 7 ]
[Shrikumar’17] Not Just a Black Box: Learning Important Features Through Propagating Activation Differences, ArXiv
1 / 2 8
1 1 / 2 8
1 2 / 2 8
Q u e s t i
: H
t
s s e s s e x p l a n a t i
q u a l i t y ? M
e d i r e c t a p p r
c h : T r y a l l p a r a m e t e r s , a n d s e l e c t t h e
e p r
u c i n g t h e b e s t e x p l a n a t i
s .
1 3 / 2 8
H u m a n a s s e s s m e n t
e s t h e t i c p r
e r t i e s
s a b i l i t y
t h e e x p l a n a t i
( e . g . t
n d e r s t a n d t h e c l a s s i fi e r ) .
→ R e q u i r e s a n e x p e r i me n t a l s t u d y .
1 4 / 2 8
I d e a : T e s t i n g i f e x p l a n a t i
s s a t i s f y c e r t a i n a x i
s / p r
e r t i e s . E x a mp l e s :
x p l a n a t i
m u s t b e s e l f
s i s t e n t ( e . g . c
s e r v a t i
e v i d e n c e )
x p l a n a t i
m u s t b e c
s i s t e n t i n i n p u t d
a i n ( e . g . c
t i n u i t y )
x p l a n a t i
m u s t b e c
s i s t e n t i n t h e s p a c e
m
e l s ( e . g . i m p l e m e n t a t i
i n v a r i a n c e )
1 5 / 2 8
P
s i b l e e x p l a n a t i
s : S i m p l e e x a m p l e :
1 6 / 2 8
1 7 / 2 8
1 8 / 2 8
A n s w e r : N e u r a l n e t w
k d e p t h c a u s e s t h e f u n c t i
t
e c
e s t e e p a n d t h e g r a d i e n t v e r y l a r g e .
[ c f . B e n g i
9 4 , M
t u f a r ’ 1 4 ]
1 2 2
x y x y 1 1
1 2 2
x
1 2 2
y x y 1 1 x y 1 1
1 2 2
x
1 2 2
y
1 2 2
d e p t h 1 d e p t h 2 d e p t h 3
[Bengio’94] Learning long- term dependencies with gradient descent is difficult. IEEE Trans. Neural Networks [Montufar’14] On the Number
NIPS 2014.
1 9 / 2 8
d i v i s i
b y z e r
h i s c a n a l s
e s e e n f r
t h e f
m u l a s :
2 / 2 8
[ M
t a v
’ 1 8 ]
E x p l a n a t i
s c
e s m u s t b e c
t i n u
s i n i n p u t d
a i n .
2 1 / 2 8
v i d e
n p u t
n i m a t i
s a v a i l a b l e a t : h t t p : / / w w w . h e a t m a p p i n g .
g / e v a l u a t i n g
2 2 / 2 8
A n s w e r : A g a i n , b e c a u s e
d e p t h , s p e c i fi c a l l y , b e c a u s e t h e f u n c t i
b e c
e s h i g h l y n
m
h .
[ c f . M
t u f a r ’ 1 4 , B a l d u z z i ’ 1 7 ]
1 2 2
x y x y 1 1
1 2 2
x
1 2 2
y x y 1 1 x y 1 1
1 2 2
x
1 2 2
y
1 2 2
d e p t h 1 d e p t h 2 d e p t h 3
[Montufar’14] On the Number
NIPS 2014. [Balduzzi’17] The Shattered Gradients Problem: If resnets are the answer [...] ICML 2017
2 3 / 2 8
[ S u n d a r a r a j a n ’ 1 7 ]
E x a mp l e : t w
e t w
k s i m p l e m e n t i n g t h e m a x i m u m f u n c t i
:
N e t w
k ( a ) : N e t w
k ( b ) : G r a d i e n t i s i m p l e m e n t a t i
i n v a r i a n t , t h e r e f
e e x p l a n a t i
t
C
n t e r
x a m p l e f
:
[Sundararajan’17] M Sundararajan, A Taly, Q Yan: Axiomatic Attribution for Deep
2 4 / 2 8
n a i v e i m p l e m e n t a t i
s b e t t e r i m p l e m e n t a t i
2 5 / 2 8
r e d i s t r i b u t i n g u n i f
m l y
p i x e l s . I t i s :
s e r v a t i v e
t i n u
s
m p l e m e n t a t i
i n v a r i a n t b u t i t i s a l s
p l e t e l y u n i n f
m a t i v e . C
s i d e r t h e s i m p l e e x p l a n a t i
t e c h n i q u e : → N e e d t
e r i f y s e l e c t i v i t y ( i . e . t h e e x p l a n a t i
s h
l d d i s c r i m i n a t e b e t w e e n r e l e v a n t a n d i r r e l e v a n t v a r i a b l e s . )
2 6 / 2 8
[ B a c h ’ 1 5 , S a m e k ’ 1 7 ]
I d e a : T e s t t h a t r e m
i n g i n p u t v a r i a b l e s w i t h h i g h a s s i g n e d r e l e v a n c e m a k e s t h e f u n c t i
v a l u e d r
q u i c k l y .
c
p u t e h e a t m a p d e s t r
p i x e l s c h e c k n e w f u n c t i
v a l u e
2 7 / 2 8
n p u t e x p l a n a t i
A n i m a t i
s a v a i l a b l e a t : h t t p : / / w w w . h e a t m a p p i n g .
g / e v a l u a t i n g
2 8 / 2 8
Most explanation methods have hyperparameters. As there is no ground-truth explanations available, standard model selection techniques do not apply. The problem of explanation selection can be addressed axiomatically (e.g. conservation, continuity, implementation invariance). Axioms may not suffice in selecting a good explanation. It is also important to design experiments that test the explanation against the model (e.g. pixel-flipping).