Likelihood-free gravitational-wave parameter estimation with neural networks
Stephen R. Green Albert Einstein Institute Potsdam based on arXiv:2002.07656 with C. Simpson and J. Gair Gravity Seminar University of Southampton February 27, 2020
1
Likelihood-free gravitational-wave parameter estimation with neural - - PowerPoint PPT Presentation
Likelihood-free gravitational-wave parameter estimation with neural networks Stephen R. Green Albert Einstein Institute Potsdam based on arXiv:2002.07656 with C. Simpson and J. Gair Gravity Seminar University of Southampton February 27, 2020 1
1
2
3
4
I
∞
detector noise power spectral density (PSD)
5
Image: Abbott et al (2016)
6
7
3 5 4 4 5 5 5 5
m2/M
. 1 . 5 3 . 4 . 5 6 .
φ0
. 6 8 2 . 6 8 3 5 . 6 8 5 . 6 8 6 5
tc/s
1 2 1 6 2 2 4
dL/Mpc
− . 3 . . 3 . 6 . 9
χ1z
− 1 . − . 5 . . 5 1 .
χ2z
5 4 6 6 6 7 2 7 8
m1/M
. . 8 1 . 6 2 . 4 3 . 2
θJN
3 5 4 4 5 5 5 5
m2/M
. 1 . 5 3 . 4 . 5 6 .
φ0
. 6 8 2 . 6 8 3 5 . 6 8 5 . 6 8 6 5
tc/s
1 2 1 6 2 2 4
dL/Mpc
− . 3 . . 3 . 6 . 9
χ1z
− 1 . − . 5 . . 5 1 .
χ2z
. . 8 1 . 6 2 . 4 3 . 2
θJN
8
9
Input layer First hidden layer
Second hidden layer
Final hidden layer
x ∈ ℝN
Output layer
σout(Wouthp + bout)
y ∈ ℝNout
10
11
n
ij=1
ij (y)(xj − µj(y))
[First applied to GW by Chua and Vallisneri (2020), Gabbard et al (2019)]
12
Intractable with knowing posterior for each !
Only requires samples from likelihood, not the posterior!
13
N
i=1
Estimate on minibatch of size N Easy to evaluate from neural network Sample parameters from prior Sample strain data from generative process (likelihood)
14
Rezende and Mohamed (2015)
15
1 , . . . , f −1 n )
16
0.0 0.2 0.4 0.6 0.8 1.0
t/s
−8 −6 −4 −2 2 4 6 8
s=h+n
f |
<latexit sha1_base64="ux6p1K1jDQaF96mqdnjK+wNPmcQ=">ACOHicbVBSxtBGJ3V2qZprasexkaGhLQsFsCehGCXqSHNoUmCtkzE6+TYbMzi4z3wphk5/lxZ/hTbz0FJ69Rc4SfbQah8MvHnv+5h5L0ylMOh5d87G5outl69Kr8tv3m6/23F397omyTSHDk9koi9DZkAKBR0UKOEy1cDiUMJFOD1b+hdXoI1I1HecpdCP2ViJSHCGVhq6XwOcADJaDYyIaVpbX+emHgTl6gkNYoYTzmT+ZVHzDvz6QNWiQX7oL4rBen0ejADp5+Fang/ditfwVqDPiV+QCinQHrq3wSjhWQwKuWTG9HwvxX7ONAouYVEOMgMp41M2hp6lisVg+vkq+IJ+tMqIRom2RyFdqX9v5Cw2ZhaHdnIZxDz1luL/vF6G0XE/FyrNEBRfPxRlkmJCly3SkdDAUc4sYVwL+1fKJ0wzjrbrsi3Bfxr5Oel+avjNRvNbs9I6LeokfkA6kRnxyRFjknbdIhnFyTe/KT/HJunB/Ob+fPenTDKXb2yT9wHh4BiSWqdA=</latexit>Papamakarios et al (2017)
17
n
i=1
Papamakarios et al (2017)
1 , …, f −1 n )
n
i=1
18
Papamakarios et al (2017)
19
Papamakarios et al (2017)
20
permute MADE permute MADE permute MADE
[same approach as Gabbard et al (2019)]
21
0.0 0.2 0.4 0.6 0.8 1.0
t/s
−7.5 −5.0 −2.5 0.0 2.5 5.0 7.5 y = h + n h
22
5 5 5 6 6 5
m2/M
− 3 3 6
φ0
. 8 4 8 . 8 5 6 . 8 6 4
tc/s
6 6 7 2 7 8
m1/M
2 4 2 6 2 8 3
dL/Mpc
5 5 5 6 6 5
m2/M
− 3 3 6
φ0
. 8 4 8 . 8 5 6 . 8 6 4
tc/s
2 4 2 6 2 8 3
dL/Mpc
0.0 0.2 0.4 0.6 0.8 1.0
t/s
−7.5 −5.0 −2.5 0.0 2.5 5.0 7.5 y = h + n h
Kingma and Welling (2013)
23
Kingma and Welling (2013)
24
reconstruction loss KL loss
25
Neural network MCMC
5 2 5 6 6 6 4
m2/M
− 2 2 4 6
φ0
. 8 4 4 . 8 5 . 8 5 6 . 8 6 2
tc/s
6 4 6 8 7 2 7 6 8
m1/M
2 4 2 5 5 2 7 2 8 5 3
dL/Mpc
5 2 5 6 6 6 4
m2/M
− 2 2 4 6
φ0
. 8 4 4 . 8 5 . 8 5 6 . 8 6 2
tc/s
2 4 2 5 5 2 7 2 8 5 3
dL/Mpc
26
0.0 0.2 0.4 0.6 0.8 1.0
0.0 0.2 0.4 0.6 0.8 1.0
m1 (0.55) m2 (0.59) φ0 (0.37) tc (0.46) dL (0.56)
27
3 5 4 4 5 5 5 5
m2/M
. 1 . 5 3 . 4 . 5 6 .
φ0
. 6 8 2 . 6 8 3 5 . 6 8 5 . 6 8 6 5
tc/s
1 2 1 6 2 2 4
dL/Mpc
− . 3 . . 3 . 6 . 9
χ1z
− 1 . − . 5 . . 5 1 .
χ2z
5 4 6 6 6 7 2 7 8
m1/M
. . 8 1 . 6 2 . 4 3 . 2
θJN
3 5 4 4 5 5 5 5
m2/M
. 1 . 5 3 . 4 . 5 6 .
φ0
. 6 8 2 . 6 8 3 5 . 6 8 5 . 6 8 6 5
tc/s
1 2 1 6 2 2 4
dL/Mpc
− . 3 . . 3 . 6 . 9
χ1z
− 1 . − . 5 . . 5 1 .
χ2z
. . 8 1 . 6 2 . 4 3 . 2
θJN
0.0 0.2 0.4 0.6 0.8 1.0
0.0 0.2 0.4 0.6 0.8 1.0
m1 (0.60) m2 (0.95) φ0 (0.80) tc (0.21) dL (0.68) χ1z (0.82) χ2z (0.60) θJN (0.55)
28
29
pairs from the data generative process.
standard methods.
30