the role of mobility and control in the inference of representations
stefano soatto ucla
1
- t. lee, a. ayvaci, j. dong, d. davis, j. balzer, j. hernandez, l. valente
1 Saturday, November 30, 13
the role of mobility and control in the inference of - - PowerPoint PPT Presentation
the role of mobility and control in the inference of representations stefano soatto ucla t. lee, a. ayvaci, j. dong, d. davis, j. balzer, j. hernandez, l. valente 1 Saturday, November 30, 13 1 what is a representation ? why do we need
1
1 Saturday, November 30, 13
what is a “representation”? why do we need it? what does control have to do with it?
keywords: data processing inequality, information bottleneck, lambert-ambient model, sufficient excitation, actionable information gap, active sensing/ perception
2
2 Saturday, November 30, 13
3
3 Saturday, November 30, 13
3
yt . = {y0, . . . , yt}
3 Saturday, November 30, 13
3
yt . = {y0, . . . , yt}
3 Saturday, November 30, 13
3
yt . = {y0, . . . , yt}
3 Saturday, November 30, 13
3
yt . = {y0, . . . , yt} ξ
3 Saturday, November 30, 13
3
yt . = {y0, . . . , yt} ξ ?
3 Saturday, November 30, 13
3
yt . = {y0, . . . , yt} ξ ?
3 Saturday, November 30, 13
3
yt . = {y0, . . . , yt} ξ ˆ ξ = φ(yt)
3 Saturday, November 30, 13
3
yt . = {y0, . . . , yt} ξ ˆ ξ = φ(yt)
3 Saturday, November 30, 13
4
yt . = {y0, . . . , yt}
4 Saturday, November 30, 13
4
I(ξ; yt) ≥ I(ξ; φ(yt)) R(ut|yt) ≤ R(ut|φ(yt)) yt . = {y0, . . . , yt}
4 Saturday, November 30, 13
4
R(ut|yt) ≤ R(ut|φ(yt)) yt . = {y0, . . . , yt} I(ξ; yt) = I(ξ; φ(yt) | {z }
ˆ ξ
)
4 Saturday, November 30, 13
4
R(ut|yt) ≤ R(ut|φ(yt)) yt . = {y0, . . . , yt} I(ξ; yt) = I(ξ; φ(yt) | {z }
ˆ ξ
)
4 Saturday, November 30, 13
4
R(ut|yt) ≤ R(ut|φ(yt)) yt . = {y0, . . . , yt} I(ξ; yt) = I(ξ; φ(yt) | {z }
ˆ ξ
) min I(ξ; yt) − I(ξ; φ(yt) | {z }
ˆ ξ
) + βH(ˆ ξ)
4 Saturday, November 30, 13
4
R(ut|yt) ≤ R(ut|φ(yt)) yt . = {y0, . . . , yt} I(ξ; yt) = I(ξ; φ(yt) | {z }
ˆ ξ
) min I(yt, ˆ ξ) − βI(ˆ ξ; ξ)
4 Saturday, November 30, 13
4
R(ut|yt) ≤ R(ut|φ(yt)) yt . = {y0, . . . , yt} I(ξ; yt) = I(ξ; φ(yt) | {z }
ˆ ξ
) min I(yt, ˆ ξ) − βI(ˆ ξ; ξ)
4 Saturday, November 30, 13
4
R(ut|yt) ≤ R(ut|φ(yt)) yt . = {y0, . . . , yt} I(ξ; yt) = I(ξ; φ(yt) | {z }
ˆ ξ
) min I(yt, ˆ ξ) − βI(ˆ ξ; ξ)
4 Saturday, November 30, 13
4
R(ut|yt) ≤ R(ut|φ(yt)) yt . = {y0, . . . , yt} I(ξ; yt) = I(ξ; φ(yt) | {z }
ˆ ξ
) min I(yt, ˆ ξ) − βI(ˆ ξ; ξ) min H(y∞
t |ˆ
ξ) + 1 β H(ˆ ξ)
4 Saturday, November 30, 13
4
R(ut|yt) ≤ R(ut|φ(yt)) yt . = {y0, . . . , yt} I(ξ; yt) = I(ξ; φ(yt) | {z }
ˆ ξ
) min I(yt, ˆ ξ) − βI(ˆ ξ; ξ) min H(y∞
t |ˆ
ξ) + 1 β H(ˆ ξ)
4 Saturday, November 30, 13
4
R(ut|yt) ≤ R(ut|φ(yt)) yt . = {y0, . . . , yt} I(ξ; yt) = I(ξ; φ(yt) | {z }
ˆ ξ
) min I(yt, ˆ ξ) − βI(ˆ ξ; ξ) min H(y∞
t |ˆ
ξ) + 1 β H(ˆ ξ)
4 Saturday, November 30, 13
4
R(ut|yt) ≤ R(ut|φ(yt)) yt . = {y0, . . . , yt} I(ξ; yt) = I(ξ; φ(yt) | {z }
ˆ ξ
) min I(yt, ˆ ξ) − βI(ˆ ξ; ξ) min H(y∞
t |ˆ
ξ) + 1 β H(ˆ ξ)
4 Saturday, November 30, 13
4
R(ut|yt) ≤ R(ut|φ(yt)) I(ξ; yt) = I(ξ; φ(yt) | {z }
ˆ ξ
) min I(yt, ˆ ξ) − βI(ˆ ξ; ξ) min H(y∞
t |ˆ
ξ) + 1 β H(ˆ ξ)
4 Saturday, November 30, 13
5
5 Saturday, November 30, 13
5
5 Saturday, November 30, 13
5
ˆ ξ
5 Saturday, November 30, 13
account for almost all uncertainty/variability in visual data
some can be removed from the data at the outset:
lossless: canonization (e.g., contrast, planar isometries)* co-variant detection/invariant description
e.g., scale, class-specific deformation
e.g., occlusion “sufficient exploration”
instantiate for a specific data-formation model (LA Model)
6
6 Saturday, November 30, 13
7
SENSING ACTION
REPRESENTATION CONTROL INFERENCE SENSORS NUISANCES CANONIZATION SCENE NUISANCES UNMODELED PHENOMENA TASK QUERIES
IR MS IMU LIDR ...
..
φ∧(yt)
ˆ g(u)
ˆ ν
h(ˆ gˆ ξ, ˆ ν)
u
H(yt+1|ˆ ξt, u)
min
p(ˆ ξ|yt)
H(yt+1|ˆ ξt)
INNOVATION
7 Saturday, November 30, 13
7
SENSING ACTION
REPRESENTATION CONTROL INFERENCE SENSORS NUISANCES CANONIZATION SCENE NUISANCES UNMODELED PHENOMENA TASK QUERIES
IR MS IMU LIDR ...
..
φ∧(yt)
ˆ g(u)
ˆ ν
h(ˆ gˆ ξ, ˆ ν)
u
H(yt+1|ˆ ξt, u)
min
p(ˆ ξ|yt)
H(yt+1|ˆ ξt)
INNOVATION
INFORMATION BOTTLENECK
7 Saturday, November 30, 13
7
SENSING ACTION
REPRESENTATION CONTROL INFERENCE SENSORS NUISANCES CANONIZATION SCENE NUISANCES UNMODELED PHENOMENA TASK QUERIES
IR MS IMU LIDR ...
..
φ∧(yt)
ˆ g(u)
ˆ ν
h(ˆ gˆ ξ, ˆ ν)
u
H(yt+1|ˆ ξt, u)
min
p(ˆ ξ|yt)
H(yt+1|ˆ ξt)
INNOVATION
INFORMATION BOTTLENECK ACTIONABLE INFORMATION INCREMENT
7 Saturday, November 30, 13
8
p
S ⊂ R3
ρ : S → R+
p 7! ρ(p)
x ¯ x
D ⊂ R2
It : D → R+
gt ∈ SE(3)
( yt(x) = κt(ρ(p)) + nt(x) ∈ R+ x = π(gtp), p ∈ S ⊂ R3
8 Saturday, November 30, 13
8
p
S ⊂ R3
ρ : S → R+
p 7! ρ(p)
x ¯ x
D ⊂ R2
It : D → R+
gt ∈ SE(3)
ξ = {ρ, S} ( yt(x) = κt(ρ(p)) + nt(x) ∈ R+ x = π(gtp), p ∈ S ⊂ R3
8 Saturday, November 30, 13
8
p
S ⊂ R3
ρ : S → R+
p 7! ρ(p)
x ¯ x
D ⊂ R2
It : D → R+
gt ∈ SE(3)
ξ = {ρ, S} gt ∈ SE(3) ( yt(x) = κt(ρ(p)) + nt(x) ∈ R+ x = π(gtp), p ∈ S ⊂ R3
8 Saturday, November 30, 13
8
p
S ⊂ R3
ρ : S → R+
p 7! ρ(p)
x ¯ x
D ⊂ R2
It : D → R+
gt ∈ SE(3)
ξ = {ρ, S} gt ∈ SE(3) ( yt(x) = κt(ρ(p)) + nt(x) ∈ R+ x = π(gtp), p ∈ S ⊂ R3 κt : R+ → R+
8 Saturday, November 30, 13
8
p
S ⊂ R3
ρ : S → R+
p 7! ρ(p)
x ¯ x
D ⊂ R2
It : D → R+
gt ∈ SE(3)
ξ = {ρ, S} gt ∈ SE(3) νt = {nt, π} ( yt(x) = κt(ρ(p)) + nt(x) ∈ R+ x = π(gtp), p ∈ S ⊂ R3 κt : R+ → R+
8 Saturday, November 30, 13
9
p
S ⊂ R3
ρ : S → R+
p 7! ρ(p)
x ¯ x
D ⊂ R2
It : D → R+
gt ∈ SE(3)
9 Saturday, November 30, 13
9
p
S ⊂ R3
ρ : S → R+
p 7! ρ(p)
x ¯ x
D ⊂ R2
It : D → R+
gt ∈ SE(3)
yt = h(gt, ξ, νt) + nt
9 Saturday, November 30, 13
yt
ˆ ξt ˆ ξt = φ(yt)
10
10 Saturday, November 30, 13
yt
ˆ ξt ˆ ξt = φ(yt) φ∧(yt) = {h(e, ˆ ξt, νt), e ∈ G, νt ∈ V} . = L(ˆ ξt)
10
10 Saturday, November 30, 13
i.e., a statistic from which the (maximal invariant of the) images can be “hallucinated” up to an “uninformative” residual
yt
ˆ ξt ˆ ξt = φ(yt) φ∧(yt) = {h(e, ˆ ξt, νt), e ∈ G, νt ∈ V} . = L(ˆ ξt)
10
10 Saturday, November 30, 13
i.e., a statistic from which the (maximal invariant of the) images can be “hallucinated” up to an “uninformative” residual L(ˆ ξ) = L(ξ)
yt
ˆ ξt ˆ ξt = φ(yt) φ∧(yt) = {h(e, ˆ ξt, νt), e ∈ G, νt ∈ V} . = L(ˆ ξt)
10
10 Saturday, November 30, 13
i.e., a statistic from which the (maximal invariant of the) images can be “hallucinated” up to an “uninformative” residual
L(ˆ ξ) = L(ξ)
yt
ˆ ξt ˆ ξt = φ(yt) φ∧(yt) = {h(e, ˆ ξt, νt), e ∈ G, νt ∈ V} . = L(ˆ ξt)
10
10 Saturday, November 30, 13
i.e., a statistic from which the (maximal invariant of the) images can be “hallucinated” up to an “uninformative” residual
L(ˆ ξ) = L(ξ) “light field”
yt
ˆ ξt ˆ ξt = φ(yt) φ∧(yt) = {h(e, ˆ ξt, νt), e ∈ G, νt ∈ V} . = L(ˆ ξt)
10
10 Saturday, November 30, 13
actionable information: uncertainty of the maximal invariant (can be computed from finite data) complete information: uncertainty of a minimal sufficient statistic of a (complete) representation actionable information gap (AIG)
11
I . = H(φ∨(ˆ ξ)) G(y) = I − H(y) H(y) . = H(φ∧(y))
11 Saturday, November 30, 13
“visual recognition is difficult in part because of the large variability that images of a particular object exhibit depending
conditions, occlusions and other visibility artifacts.” theorem: for viewpoint and illumination (contrast) nuisances, the quotient (“attributed reeb tree”) is supported on a thin set
12 sundaramoorthi, petersen, varadarajan, soatto, “on the set of images modulo viewpoint and contrast changes”, cvpr 2009
12 Saturday, November 30, 13
“visual recognition is difficult in part because of the large variability that images of a particular object exhibit depending
conditions, occlusions and other visibility artifacts.” theorem: for viewpoint and illumination (contrast) nuisances, the quotient (“attributed reeb tree”) is supported on a thin set
12 sundaramoorthi, petersen, varadarajan, soatto, “on the set of images modulo viewpoint and contrast changes”, cvpr 2009
{I}/W(R2 → R2) × H(R+ → R+) = ART
12 Saturday, November 30, 13
“visual recognition is difficult in part because of the large variability that images of a particular object exhibit depending
conditions, occlusions and other visibility artifacts.” theorem: for viewpoint and illumination (contrast) nuisances, the quotient (“attributed reeb tree”) is supported on a thin set
12 sundaramoorthi, petersen, varadarajan, soatto, “on the set of images modulo viewpoint and contrast changes”, cvpr 2009
ˆ ξt = φ(yt) {I}/W(R2 → R2) × H(R+ → R+) = ART
12 Saturday, November 30, 13
“visual recognition is difficult in part because of the large variability that images of a particular object exhibit depending
conditions, occlusions and other visibility artifacts.” theorem: for viewpoint and illumination (contrast) nuisances, the quotient (“attributed reeb tree”) is supported on a thin set
12 sundaramoorthi, petersen, varadarajan, soatto, “on the set of images modulo viewpoint and contrast changes”, cvpr 2009
ˆ ξt = φ(yt) {I}/W(R2 → R2) × H(R+ → R+) = ART
12 Saturday, November 30, 13
13
13 Saturday, November 30, 13
13
13 Saturday, November 30, 13
13
13 Saturday, November 30, 13
1.how to build the best possible representation given past data
2.how to gather future data to make the representation (as close as possible to) complete
14
14 Saturday, November 30, 13
(1)(Lamber,an(reflec,on (2)(Constant(illumina,on (3)(Co9visibility
small dense large sparse nesterov/split9bregman( w/(weighted(isotropic(TV(for(w
15
15 Saturday, November 30, 13
ˆ c = arg min
c
Z
D
|c(x) − c(y)|dµ(x, y)
16
(
16 Saturday, November 30, 13
ˆ c = arg min
c
Z
D
|c(x) − c(y)|dµ(x, y)
dµ(x, y) = K(x, y)dxdy
16
(
16 Saturday, November 30, 13
ˆ c = arg min
c
Z
D
|c(x) − c(y)|dµ(x, y)
dµ(x, y) = K(x, y)dxdy
16
(
K(x, y) = ( e(It(x)It(y))2 + ⇥ekvt(x)vt(y)k2
2
0,
kx yk2 < ⇤,
16 Saturday, November 30, 13
ˆ c = arg min
c
Z
D
|c(x) − c(y)|dµ(x, y)
dµ(x, y) = K(x, y)dxdy
v(x) . = w(x) − x
16
(
K(x, y) = ( e(It(x)It(y))2 + ⇥ekvt(x)vt(y)k2
2
0,
kx yk2 < ⇤,
16 Saturday, November 30, 13
17
17 Saturday, November 30, 13
18
18 Saturday, November 30, 13
19
SENSING ACTION
REPRESENTATION CONTROL INFERENCE SENSORS NUISANCES CANONIZATION SCENE NUISANCES UNMODELED PHENOMENA TASK QUERIES
IR MS IMU LIDR ...
..
φ∧(yt)
ˆ g(u)
ˆ ν
h(ˆ gˆ ξ, ˆ ν)
u
H(yt+1|ˆ ξt, u)
min
p(ˆ ξ|yt)
H(yt+1|ˆ ξt)
INNOVATION
19 Saturday, November 30, 13
20
ˆ ut = arg max
u
min
ˆ ξt=φ(yt)
H(yt+1|ˆ ξt, u) + λH(ˆ ξt)
20 Saturday, November 30, 13
21
21 Saturday, November 30, 13
unknown environments
φ = φ(·, γ) In the rest,
22 Saturday, November 30, 13
23
M Q ✏ O(MQ) + O( 1 ✏2 )
23 Saturday, November 30, 13
24 Saturday, November 30, 13
1 E x p e r i m e n t
1 . 1 R i s k a n d s e n s
p a r a m e t e r s
i t h 5 , 1 , 3 c l u t t e r
j e c t s ( l
, m e d i u m , h i g h c
p l e x i t y ) . 25 Saturday, November 30, 13
26
26 Saturday, November 30, 13
26
26 Saturday, November 30, 13
26
26 Saturday, November 30, 13
26
26 Saturday, November 30, 13
26
26 Saturday, November 30, 13
26
26 Saturday, November 30, 13
csd100028, september 13, 2010 (also on video lecture, nips tutorial 2010; also icvss lecture notes, ‘07-’09)
vision”, r. cipolla et al. (eds.), 2011
27
27 Saturday, November 30, 13