Course on Inverse Problems Albert Tarantola Lesson XII: - - PowerPoint PPT Presentation
Course on Inverse Problems Albert Tarantola Lesson XII: - - PowerPoint PPT Presentation
Institut de Physique du Globe de Paris & Universit Pierre et Marie Curie (Paris VI) Course on Inverse Problems Albert Tarantola Lesson XII: Optimization and Linear Problems Recall: We have seen that the posterior volumetric probability
Recall: We have seen that the posterior volumetric probability for the model parameters, whose general expression is fpost(m) = 1 ν fprior(m) gobs( o(m) ) . becomes, using the Gaussian model for all uncertainties, fpost(m) = k exp( −S(m) ) , where the misfit function S(m) is the sum of squares 2 S(m) = (m − mprior)t Cprior
- 1 (m − mprior)
+ (o(m) − oobs)t Cobs
- 1 (o(m) − oobs)
. What happens if the forward modeling relation o = o(m) is, in fact, linear?
Linear problems: We have assumed the prior information cor- responds to a Gaussian distribution with mean mprior and co- variance Cprior , and that the observations are described by a Gaussian distribution with mean oobs and covariance Cobs . If the forward relation is linear, o = o(m) = O m , then, the posterior distribution in the model space is also a Gaussian distribution, with the following mean and covariance: mpost = ( Ot C-1
- bs O + C-1
prior )-1 ( Ot C-1
- bs oobs + C-1
prior mprior )
= mprior + ( Ot C-1
- bs O + C-1
prior )-1 Ot C-1
- bs ( oobs − O mprior )
= mprior + Cprior Ot ( O Cprior Ot + Cobs )-1 (oobs − O mprior)
Cpost = ( Ot C-1
- bs O + C-1
prior )-1
= Cprior − Cprior Ot ( O Cprior Ot + Cobs )-1 O Cprior
The posterior distribution in the observable parameter space is also Gaussian, with mean O mpost and covariance O Cpost Ot .
Remark: The five elements of a linear inverse problem are mprior , Cprior , oobs , Cobs , and O . The expressions for mpost and Cpost have been given in the slides above. What these expressions become if, in fact, we don’t have any a priori in- formation (and the forward relation o = O m is redundant)? Cprior → ∞ I ; ( then mprior disappears ) mpost = ( Ot C-1
- bs O )-1 Ot C-1
- bs oobs
Cpost = ( Ot C-1
- bs O )-1
And if the matrix O is square and invertible? mpost = O-1 oobs
4.2 Ray Tomography Using Blocks
❊①❡❝✉t❛❜❧❡ ♥♦t❡❜♦♦❦ ❛t ❤tt♣✿✴✴✇✇✇✳✐♣❣♣✳❥✉ss✐❡✉✳❢r✴⑦t❛r❛♥t♦❧❛✴❡①❡r❝✐❝❡s✴❝❤❛♣t❡r❴✵✹✴❘❛②❚♦♠♦❣r❛♣❤②✳♥❜
This exercise corresponds to a highly idealized version of an X-ray tomography experi-
- ment. A 2D medium is characterized by a parameter m(x, y) whose physical dimension is
the inverse of a length. This parameter may take any real value (including negative values). Assume that when a ray Ri is materialized in the medium (using a source and a receiver), we are able to measure the observable parameter
- i =
- Ri dℓ m(x, y)
, (4.1) where ℓ is the length along the ray. The goal of the exercise is to use some observed values
- i
- bs to infer the values of the function m(x, y) .
❊①❡❝✉t❛❜❧❡ ♥♦t❡❜♦♦❦ ❛t ❤tt♣✿✴✴✇✇✇✳✐♣❣♣✳❥✉ss✐❡✉✳❢r✴⑦t❛r❛♥t♦❧❛✴❡①❡r❝✐❝❡s✴❝❤❛♣t❡r❴✵✹✴❘❛②❚♦♠♦❣r❛♣❤②✳♥❜
01020304 05060708 09101112 13141516
01 02 03 04 19 20 05 07 11 21 06 10 22 18 09 14 16 17 08 12 13 15 Figure 4.1: Geometry of the ‘X-ray experiment’. This drawing is at scale 1:1 (the length of the side of the blocks is 1 cm ). To simplify the problem, the medium is divided into blocks, numbered from 1 to 16 (see figure 4.1), so, instead of evaluating the function m(x, y) , we only need to evaluate
the discrete values {m1, m2, . . . , m16} . With this discretization, the relation (4.1) becomes
- i =
α=16
∑
α=1
Oiα mα , (4.2) where Oiα is the length of the ray i in the block α . We shall use 22 rays, so the index i runs from 1 to 22. For short, equation (4.2) shall be written
- = O m
. (4.3) With the geometry of the 16 blocks and the 22 rays represented in figure 4.1, it is easy to see that the matrix O is (only the nonzero elements are indicated) O = − − − − − − − − − − − − − − − a − − − − − − − − − − − a − − a − − − − − − − − a − − a − − a − − − − − a − − a − − a − − a − − − − − a − − a − − a − − − − − − − − a − − a − − − − − − − − − − − a − − − − − − − − − − − − − − − − − − b − − − b − − − b − − − b − − b − − − b − − − b − − − b − − b − − − b − − − b − − − b − − b − − − b − − − b − − − b − − − − − − a − − − − − − − − − − − − − − a − − − − a − − − − − − − − − a − − − − a − − − − a − − − − a − − − − a − − − − a − − − − a − − − − a − − − − a − − − − a − − − − − − − − − a − − − − a − − − − − − − − − − − − − − a − − − b b b b − − − − − − − − − − − − − − − − b b b b − − − − − − − − − − − − − − − − b b b b − − − − − − − − − − − − − − − − b b b b , (4.4) with a = √ 2 cm and b = 1 cm . In a real-life application, the components of the matrix O should be automatically com-
✭✯ ❚❤❡ ♠❛tr✐① ❖ ✐♥ t❤❡ ♦ ❂ ❖ ♠ r❡❧❛t✐♦♥ ✯✮ ❯♥♣r♦t❡❝t❬❖❪❀ ❖ ❂ ❚❛❜❧❡ ❬✵ ✱ ④✐✱✶ ✱✷✷⑥ ✱ ④❥ ✱✶ ✱✶✻⑥❪❀ sq✷ ❂ ❙qrt ❬✷❪❀ ❖❬❬✵✶ ✱✶✻❪❪ ❂ sq✷❀ ❖❬❬✵✷ ✱✶✷❪❪ ❂ sq✷❀❖❬❬✵✷ ✱✶✺❪❪ ❂ sq✷❀ ❖❬❬✵✸ ✱✵✽❪❪ ❂ sq✷❀❖❬❬✵✸ ✱✶✶❪❪ ❂ sq✷❀❖❬❬✵✸ ✱✶✹❪❪ ❂ sq✷❀ ❖❬❬✵✹ ✱✵✹❪❪ ❂ sq✷❀❖❬❬✵✹ ✱✵✼❪❪ ❂ sq✷❀❖❬❬✵✹ ✱✶✵❪❪ ❂ sq✷❀❖❬❬✵✹ ✱✶✸❪❪ ❂ sq✷❀ ❖❬❬✵✺ ✱✵✸❪❪ ❂ sq✷❀❖❬❬✵✺ ✱✵✻❪❪ ❂ sq✷❀❖❬❬✵✺ ✱✵✾❪❪ ❂ sq✷❀ ❖❬❬✵✻ ✱✵✷❪❪ ❂ sq✷❀❖❬❬✵✻ ✱✵✺❪❪ ❂ sq✷❀ ❖❬❬✵✼ ✱✵✶❪❪ ❂ sq✷❀ ❖❬❬✵✽ ✱✵✹❪❪ ❂ ✶❀❖❬❬✵✽ ✱✵✽❪❪ ❂ ✶❀❖❬❬✵✽ ✱✶✷❪❪ ❂ ✶❀❖❬❬✵✽ ✱✶✻❪❪ ❂ ✶❀ ❖❬❬✵✾ ✱✵✸❪❪ ❂ ✶❀❖❬❬✵✾ ✱✵✼❪❪ ❂ ✶❀❖❬❬✵✾ ✱✶✶❪❪ ❂ ✶❀❖❬❬✵✾ ✱✶✺❪❪ ❂ ✶❀
with a = 2 cm and b = 1 cm . In a real-life application, the components of the matrix O should be automatically com- puted using the positions of the end-points of the ray, but —as writing the associated code is tedious— in this simple exercise, we can enter the components “by hand” as follows1
✭✯ ❚❤❡ ♠❛tr✐① ❖ ✐♥ t❤❡ ♦ ❂ ❖ ♠ r❡❧❛t✐♦♥ ✯✮ ❯♥♣r♦t❡❝t❬❖❪❀ ❖ ❂ ❚❛❜❧❡ ❬✵ ✱ ④✐✱✶ ✱✷✷⑥ ✱ ④❥ ✱✶ ✱✶✻⑥❪❀ sq✷ ❂ ❙qrt ❬✷❪❀ ❖❬❬✵✶ ✱✶✻❪❪ ❂ sq✷❀ ❖❬❬✵✷ ✱✶✷❪❪ ❂ sq✷❀❖❬❬✵✷ ✱✶✺❪❪ ❂ sq✷❀ ❖❬❬✵✸ ✱✵✽❪❪ ❂ sq✷❀❖❬❬✵✸ ✱✶✶❪❪ ❂ sq✷❀❖❬❬✵✸ ✱✶✹❪❪ ❂ sq✷❀ ❖❬❬✵✹ ✱✵✹❪❪ ❂ sq✷❀❖❬❬✵✹ ✱✵✼❪❪ ❂ sq✷❀❖❬❬✵✹ ✱✶✵❪❪ ❂ sq✷❀❖❬❬✵✹ ✱✶✸❪❪ ❂ sq✷❀ ❖❬❬✵✺ ✱✵✸❪❪ ❂ sq✷❀❖❬❬✵✺ ✱✵✻❪❪ ❂ sq✷❀❖❬❬✵✺ ✱✵✾❪❪ ❂ sq✷❀ ❖❬❬✵✻ ✱✵✷❪❪ ❂ sq✷❀❖❬❬✵✻ ✱✵✺❪❪ ❂ sq✷❀ ❖❬❬✵✼ ✱✵✶❪❪ ❂ sq✷❀ ❖❬❬✵✽ ✱✵✹❪❪ ❂ ✶❀❖❬❬✵✽ ✱✵✽❪❪ ❂ ✶❀❖❬❬✵✽ ✱✶✷❪❪ ❂ ✶❀❖❬❬✵✽ ✱✶✻❪❪ ❂ ✶❀ ❖❬❬✵✾ ✱✵✸❪❪ ❂ ✶❀❖❬❬✵✾ ✱✵✼❪❪ ❂ ✶❀❖❬❬✵✾ ✱✶✶❪❪ ❂ ✶❀❖❬❬✵✾ ✱✶✺❪❪ ❂ ✶❀
❖❬❬✶✵ ✱✵✷❪❪ ❂ ✶❀❖❬❬✶✵ ✱✵✻❪❪ ❂ ✶❀❖❬❬✶✵ ✱✶✵❪❪ ❂ ✶❀❖❬❬✶✵ ✱✶✹❪❪ ❂ ✶❀ ❖❬❬✶✶ ✱✵✶❪❪ ❂ ✶❀❖❬❬✶✶ ✱✵✺❪❪ ❂ ✶❀❖❬❬✶✶ ✱✵✾❪❪ ❂ ✶❀❖❬❬✶✶ ✱✶✸❪❪ ❂ ✶❀ ❖❬❬✶✷ ✱✵✹❪❪ ❂ sq✷❀ ❖❬❬✶✸ ✱✵✸❪❪ ❂ sq✷❀❖❬❬✶✸ ✱✵✽❪❪ ❂ sq✷❀ ❖❬❬✶✹ ✱✵✷❪❪ ❂ sq✷❀❖❬❬✶✹ ✱✵✼❪❪ ❂ sq✷❀❖❬❬✶✹ ✱✶✷❪❪ ❂ sq✷❀ ❖❬❬✶✺ ✱✵✶❪❪ ❂ sq✷❀❖❬❬✶✺ ✱✵✻❪❪ ❂ sq✷❀❖❬❬✶✺ ✱✶✶❪❪ ❂ sq✷❀❖❬❬✶✺ ✱✶✻❪❪ ❂ sq✷❀ ❖❬❬✶✻ ✱✵✺❪❪ ❂ sq✷❀❖❬❬✶✻ ✱✶✵❪❪ ❂ sq✷❀❖❬❬✶✻ ✱✶✺❪❪ ❂ sq✷❀ ❖❬❬✶✼ ✱✵✾❪❪ ❂ sq✷❀❖❬❬✶✼ ✱✶✹❪❪ ❂ sq✷❀ ❖❬❬✶✽ ✱✶✸❪❪ ❂ sq✷❀ ❖❬❬✶✾ ✱✵✶❪❪ ❂ ✶❀❖❬❬✶✾ ✱✵✷❪❪ ❂ ✶❀❖❬❬✶✾ ✱✵✸❪❪ ❂ ✶❀❖❬❬✶✾ ✱✵✹❪❪ ❂ ✶❀ ❖❬❬✷✵ ✱✵✺❪❪ ❂ ✶❀❖❬❬✷✵ ✱✵✻❪❪ ❂ ✶❀❖❬❬✷✵ ✱✵✼❪❪ ❂ ✶❀❖❬❬✷✵ ✱✵✽❪❪ ❂ ✶❀ ❖❬❬✷✶ ✱✵✾❪❪ ❂ ✶❀❖❬❬✷✶ ✱✶✵❪❪ ❂ ✶❀❖❬❬✷✶ ✱✶✶❪❪ ❂ ✶❀❖❬❬✷✶ ✱✶✷❪❪ ❂ ✶❀ ❖❬❬✷✷ ✱✶✸❪❪ ❂ ✶❀❖❬❬✷✷ ✱✶✹❪❪ ❂ ✶❀❖❬❬✷✷ ✱✶✺❪❪ ❂ ✶❀❖❬❬✷✷ ✱✶✻❪❪ ❂ ✶❀ ✭✯ ❆ ♣r✐♦r✐ ✐♥❢♦r♠❛t✐♦♥ ✯✮ ♠♣r✐♦r ❂ ④✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳⑥ ❀ s♣r✐♦r ❂ ✶✳✺ ❀ ❈♣r✐♦r ❂ s♣r✐♦r ❫✷ ■❞❡♥t✐t②▼❛tr✐① ❬✶✻❪ ❀
❖❬❬✶✵ ✱✵✷❪❪ ❂ ✶❀❖❬❬✶✵ ✱✵✻❪❪ ❂ ✶❀❖❬❬✶✵ ✱✶✵❪❪ ❂ ✶❀❖❬❬✶✵ ✱✶✹❪❪ ❂ ✶❀ ❖❬❬✶✶ ✱✵✶❪❪ ❂ ✶❀❖❬❬✶✶ ✱✵✺❪❪ ❂ ✶❀❖❬❬✶✶ ✱✵✾❪❪ ❂ ✶❀❖❬❬✶✶ ✱✶✸❪❪ ❂ ✶❀ ❖❬❬✶✷ ✱✵✹❪❪ ❂ sq✷❀ ❖❬❬✶✸ ✱✵✸❪❪ ❂ sq✷❀❖❬❬✶✸ ✱✵✽❪❪ ❂ sq✷❀ ❖❬❬✶✹ ✱✵✷❪❪ ❂ sq✷❀❖❬❬✶✹ ✱✵✼❪❪ ❂ sq✷❀❖❬❬✶✹ ✱✶✷❪❪ ❂ sq✷❀ ❖❬❬✶✺ ✱✵✶❪❪ ❂ sq✷❀❖❬❬✶✺ ✱✵✻❪❪ ❂ sq✷❀❖❬❬✶✺ ✱✶✶❪❪ ❂ sq✷❀❖❬❬✶✺ ✱✶✻❪❪ ❂ sq✷❀ ❖❬❬✶✻ ✱✵✺❪❪ ❂ sq✷❀❖❬❬✶✻ ✱✶✵❪❪ ❂ sq✷❀❖❬❬✶✻ ✱✶✺❪❪ ❂ sq✷❀ ❖❬❬✶✼ ✱✵✾❪❪ ❂ sq✷❀❖❬❬✶✼ ✱✶✹❪❪ ❂ sq✷❀ ❖❬❬✶✽ ✱✶✸❪❪ ❂ sq✷❀ ❖❬❬✶✾ ✱✵✶❪❪ ❂ ✶❀❖❬❬✶✾ ✱✵✷❪❪ ❂ ✶❀❖❬❬✶✾ ✱✵✸❪❪ ❂ ✶❀❖❬❬✶✾ ✱✵✹❪❪ ❂ ✶❀ ❖❬❬✷✵ ✱✵✺❪❪ ❂ ✶❀❖❬❬✷✵ ✱✵✻❪❪ ❂ ✶❀❖❬❬✷✵ ✱✵✼❪❪ ❂ ✶❀❖❬❬✷✵ ✱✵✽❪❪ ❂ ✶❀ ❖❬❬✷✶ ✱✵✾❪❪ ❂ ✶❀❖❬❬✷✶ ✱✶✵❪❪ ❂ ✶❀❖❬❬✷✶ ✱✶✶❪❪ ❂ ✶❀❖❬❬✷✶ ✱✶✷❪❪ ❂ ✶❀ ❖❬❬✷✷ ✱✶✸❪❪ ❂ ✶❀❖❬❬✷✷ ✱✶✹❪❪ ❂ ✶❀❖❬❬✷✷ ✱✶✺❪❪ ❂ ✶❀❖❬❬✷✷ ✱✶✻❪❪ ❂ ✶❀
Let us assume that we have the a priori information that, for all the blocks, m = 5 cm-1 ± 3 cm-1 . (4.5) As we are thinking in using the least-squares theory, we put this a priori information into the form of an a priori model mprior = ( 5 , 5 , 5 , 5 , 5 , 5 , 5 , 5 , 5 , 5 , 5 , 5 , 5 , 5 , 5 , 5 )t (in cm-1 units) , (4.6) and an a priori covariance matrix Cprior = s2
m I
, (4.7) where, interpreting the symbol ± in equation 4.5 as twice the standard deviation of a Gaus- sian distribution, we take sm = 1.5 cm-1 . (4.8) We take a diagonal covariance matrix because we are given no information about some other possible probabilistic model that would better represent the a priori information.
✭✯ ❆ ♣r✐♦r✐ ✐♥❢♦r♠❛t✐♦♥ ✯✮ ♠♣r✐♦r ❂ ④✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳⑥ ❀ s♣r✐♦r ❂ ✶✳✺ ❀ ❈♣r✐♦r ❂ s♣r✐♦r ❫✷ ■❞❡♥t✐t②▼❛tr✐① ❬✶✻❪ ❀
❖❬❬✶✵ ✱✵✷❪❪ ❂ ✶❀❖❬❬✶✵ ✱✵✻❪❪ ❂ ✶❀❖❬❬✶✵ ✱✶✵❪❪ ❂ ✶❀❖❬❬✶✵ ✱✶✹❪❪ ❂ ✶❀ ❖❬❬✶✶ ✱✵✶❪❪ ❂ ✶❀❖❬❬✶✶ ✱✵✺❪❪ ❂ ✶❀❖❬❬✶✶ ✱✵✾❪❪ ❂ ✶❀❖❬❬✶✶ ✱✶✸❪❪ ❂ ✶❀ ❖❬❬✶✷ ✱✵✹❪❪ ❂ sq✷❀ ❖❬❬✶✸ ✱✵✸❪❪ ❂ sq✷❀❖❬❬✶✸ ✱✵✽❪❪ ❂ sq✷❀ ❖❬❬✶✹ ✱✵✷❪❪ ❂ sq✷❀❖❬❬✶✹ ✱✵✼❪❪ ❂ sq✷❀❖❬❬✶✹ ✱✶✷❪❪ ❂ sq✷❀ ❖❬❬✶✺ ✱✵✶❪❪ ❂ sq✷❀❖❬❬✶✺ ✱✵✻❪❪ ❂ sq✷❀❖❬❬✶✺ ✱✶✶❪❪ ❂ sq✷❀❖❬❬✶✺ ✱✶✻❪❪ ❂ sq✷❀ ❖❬❬✶✻ ✱✵✺❪❪ ❂ sq✷❀❖❬❬✶✻ ✱✶✵❪❪ ❂ sq✷❀❖❬❬✶✻ ✱✶✺❪❪ ❂ sq✷❀ ❖❬❬✶✼ ✱✵✾❪❪ ❂ sq✷❀❖❬❬✶✼ ✱✶✹❪❪ ❂ sq✷❀ ❖❬❬✶✽ ✱✶✸❪❪ ❂ sq✷❀ ❖❬❬✶✾ ✱✵✶❪❪ ❂ ✶❀❖❬❬✶✾ ✱✵✷❪❪ ❂ ✶❀❖❬❬✶✾ ✱✵✸❪❪ ❂ ✶❀❖❬❬✶✾ ✱✵✹❪❪ ❂ ✶❀ ❖❬❬✷✵ ✱✵✺❪❪ ❂ ✶❀❖❬❬✷✵ ✱✵✻❪❪ ❂ ✶❀❖❬❬✷✵ ✱✵✼❪❪ ❂ ✶❀❖❬❬✷✵ ✱✵✽❪❪ ❂ ✶❀ ❖❬❬✷✶ ✱✵✾❪❪ ❂ ✶❀❖❬❬✷✶ ✱✶✵❪❪ ❂ ✶❀❖❬❬✷✶ ✱✶✶❪❪ ❂ ✶❀❖❬❬✷✶ ✱✶✷❪❪ ❂ ✶❀ ❖❬❬✷✷ ✱✶✸❪❪ ❂ ✶❀❖❬❬✷✷ ✱✶✹❪❪ ❂ ✶❀❖❬❬✷✷ ✱✶✺❪❪ ❂ ✶❀❖❬❬✷✷ ✱✶✻❪❪ ❂ ✶❀
At this point, we have the choice to keep in mind the vector mprior and the matrix Cprior , and proceed directly to the examination of the measurements, or, alternatively, we may pause to examine exactly which kid of a priori information have we introduced. Least-squares formulas are only interesting when all uncertainties are of the Gaussian
- type. This means the we are, in fact, assuming that we have an a priori probability density
in the model space that is a Gaussian distribution, with mean mprior and covariance Cprior . The best we can do to grasp exactly the a priori information we are using is to generate a few random models from such a distribution. After introducing mprior and Cprior
✭✯ ❆ ♣r✐♦r✐ ✐♥❢♦r♠❛t✐♦♥ ✯✮ ♠♣r✐♦r ❂ ④✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳ ✱✺✳⑥ ❀ s♣r✐♦r ❂ ✶✳✺ ❀ ❈♣r✐♦r ❂ s♣r✐♦r ❫✷ ■❞❡♥t✐t②▼❛tr✐① ❬✶✻❪ ❀
we perform the Cholesky decomposition of the prior covariance matrix (in this case, this is trivial, as Cprior is diagonal)
✭✯ ❙q✉❛r❡ r♦♦t ♦❢ t❤❡ ❝♦✈❛r✐❛♥❝❡ ♠❛tr✐① ✯✮ ✭✯ ❖♥❡ ❤❛s ❝♦✈ ❂ ▲✳❚r❛♥s♣♦s❡❬▲❪ ✯✮ ▲♣r✐♦r ❂ ❚r❛♥s♣♦s❡❬ ❈❤♦❧❡s❦②❉❡❝♦♠♣♦s✐t✐♦♥ ❬❈♣r✐♦r ❪❪ ❀
With the matrix L at hand, we generate nine pseudo-random realizations of the prior (Gaus- sian) distribution as explained in section 1.8.2:
✭✯ ●❡♥❡r❛t✐♦♥ ♦❢ r❛♥❞♦♠ ♠♦❞❡❧s ✭♣r✐♦r ❞✐str✐❜✉t✐♦♥✮ ✯✮ ❙❡❡❞❘❛♥❞♦♠ ❬✶❪ r♥❞ ✿❂ ❙qrt ❬✷❪ ■♥✈❡rs❡❊r❢❬❘❛♥❞♦♠❬❘❡❛❧ ✱ ④✲✶✱ ✶⑥❪❪ ❉♦❬ ④ ♠r❛♥❞♦♠✵ ❂ ❚❛❜❧❡❬r♥❞ ✱ ④✶✻⑥❪ ✱ ♠r❛♥❞♦♠ ❂ ♠♣r✐♦r ✰ ▲♣r✐♦r✳♠r❛♥❞♦♠✵ ✱ P▲❖❚❬♠r❛♥❞♦♠❪ ⑥✱ ④✐✱✶✱✾⑥ ❪
Using the color scale in figure 4.2, the nine models so obtained are displayed in figure 4.3
✭✯ ❙q✉❛r❡ r♦♦t ♦❢ t❤❡ ❝♦✈❛r✐❛♥❝❡ ♠❛tr✐① ✯✮ ✭✯ ❖♥❡ ❤❛s ❝♦✈ ❂ ▲✳❚r❛♥s♣♦s❡❬▲❪ ✯✮ ▲♣r✐♦r ❂ ❚r❛♥s♣♦s❡❬ ❈❤♦❧❡s❦②❉❡❝♦♠♣♦s✐t✐♦♥ ❬❈♣r✐♦r ❪❪ ❀ ✭✯ ●❡♥❡r❛t✐♦♥ ♦❢ r❛♥❞♦♠ ♠♦❞❡❧s ✭♣r✐♦r ❞✐str✐❜✉t✐♦♥✮ ✯✮ ❙❡❡❞❘❛♥❞♦♠ ❬✶❪ r♥❞ ✿❂ ❙qrt ❬✷❪ ■♥✈❡rs❡❊r❢❬❘❛♥❞♦♠❬❘❡❛❧ ✱ ④✲✶✱ ✶⑥❪❪ ❉♦❬ ④ ♠r❛♥❞♦♠✵ ❂ ❚❛❜❧❡❬r♥❞ ✱ ④✶✻⑥❪ ✱ ♠r❛♥❞♦♠ ❂ ♠♣r✐♦r ✰ ▲♣r✐♦r✳♠r❛♥❞♦♠✵ ✱ P▲❖❚❬♠r❛♥❞♦♠❪ ⑥✱ ④✐✱✶✱✾⑥ ❪
Figure 4.2: The scale used for the model parameter values. All mod- els represented in this section use this scale.
2 cm-1 9 cm-1
Figure 4.3: Least-squares formulas are only interest- ing when all uncertainties are of Gaussian nature. This is true, in particular, for the prior distribution in the model space. This prior distribution is com- pletely described by a model vector mprior and a co- variance matrix Cprior . At the right, nine pseudo- random samples of the prior Gaussian distribution. To set the exercise, let us now generate “artificial data”, to be later use as “observations”. To do that, let us invent an arbitrary model mtrue = ( 7 , 3 , 3 , 3 , 7 , 3 , 5 , 3 , 7 , 3 , 3 , 3 , 7 , 3 , 5 , 5 )t (in cm-1 units) . (4.9) Using the scale in figure 4.2, this “true model” can be represented as in figure 4.4 The reader Figure 4.4: The “true model” invented for artificially generating some “data”, to be later used in the inversion exercice. may perhaps read the two letters “IP” in this model. This shows that I am violating the
assumption that the true model is a random model from the prior Gaussian distribution, i.e.,
- ne model that could be amid models like those in figure 4.3, while it is extremely unlikely
that one of those models shows any “readable letter”. So, when using this “true model” we are certainly violating the a priori assumption. But the least-squares method is so practical that we often use it, even when we know that other, more expensive, methods should be used instead. So, this (mild) violation of the a priori assumptions is part of the exercise. Computing the artificial data associated to the model mtrue just amounts to implement the computation dtrue = O mtrue . These true values can then be converted into pseudo-
- bserved values by adding some noise, here independent random Gaussian values with
zero mean and with standard deviation sobs = 0.15 cm-1 . (4.10) When implementing this,
✭✯ ❈❛❧❝✉❧❛t✐♥❣ ✧tr✉❡ ♦❜s❡r✈❛t✐♦♥s✧✱ t❤❡♥ ❛❞❞✐♥❣
- ❛✉ss✐❛♥
r❛♥❞♦♠ ❡rr♦rs ✯✮ ♦tr✉❡ ❂ ❖✳♠tr✉❡❀ ♦♦❜s ❂ ♦tr✉❡❀ ❙❡❡❞❘❛♥❞♦♠ ❬✶✷✸❪ r♥❞ ✿❂ ❙qrt ❬✷❪ ■♥✈❡rs❡❊r❢❬❘❛♥❞♦♠❬❘❡❛❧ ✱ ④✲✶✱ ✶⑥❪❪ s♦❜s ❂ ✵✳✶✺❀ ❉♦❬ ♦♦❜s ❬❬✐❪❪ ❂ ♦♦❜s ❬❬✐❪❪ ✰ s♦❜s r♥❞ ✱ ④✐✱ ✶✱ ✷✷⑥ ❪
we obtain the 22 values
- obs = ( 6.93 , 11.29 , 12.89 , 25.72 , 18.28 , 14.09 , 10.06 , 14.05 , 15.91 , 12.07 , 28.14 ,
4.21 , 8.32 , 15.52 , 25.50 , 21.10 , 14.12 , 9.98 , 15.95 , 17.89 , 16.18 , 20.07)t . (4.11) This vector of “observed values” is accompanied by the 22 × 22 diagonal2 covariance matrix Cobs = s2
- bs I
, (4.12) that describes our “experimental” uncertainties. The value of sobs is that in equation 4.10. As explained in the introduction to this chapter, in least-squares theory, one demonstrates
✭✯ ❈❛❧❝✉❧❛t✐♥❣ ✧tr✉❡ ♦❜s❡r✈❛t✐♦♥s✧✱ t❤❡♥ ❛❞❞✐♥❣
- ❛✉ss✐❛♥
r❛♥❞♦♠ ❡rr♦rs ✯✮ ♦tr✉❡ ❂ ❖✳♠tr✉❡❀ ♦♦❜s ❂ ♦tr✉❡❀ ❙❡❡❞❘❛♥❞♦♠ ❬✶✷✸❪ r♥❞ ✿❂ ❙qrt ❬✷❪ ■♥✈❡rs❡❊r❢❬❘❛♥❞♦♠❬❘❡❛❧ ✱ ④✲✶✱ ✶⑥❪❪ s♦❜s ❂ ✵✳✶✺❀ ❉♦❬ ♦♦❜s ❬❬✐❪❪ ❂ ♦♦❜s ❬❬✐❪❪ ✰ s♦❜s r♥❞ ✱ ④✐✱ ✶✱ ✷✷⑥ ❪
sobs As explained in the introduction to this chapter, in least-squares theory, one demonstrates that the posterior probability density for the model parameters is a Gaussian, whose center is given by any of the three equivalent equations mpost = ( Ot C-1
- bs O + C-1
prior )-1 ( Ot C-1
- bs oobs + C-1
prior mprior )
= mprior + ( Ot C-1
- bs O + C-1
prior )-1 Ot C-1
- bs ( oobs − O mprior )
= mprior + Cprior Ot ( O Cprior Ot + Cobs )-1 (oobs − O mprior) , (4.13) and whose covariance is given by any of the two equivalent expressions Cpost = ( Ot C-1
- bs O + C-1
prior )-1
= Cprior − Cprior Ot ( O Cprior Ot + Cobs )-1 O Cprior . (4.14)
2The matrix is diagonal because the noise values generated are independent from each other.
In this small-scale problem there is not much computational difference between these equiv- alent expressions, so let us use for mpost the expression mpost = ( Ot C-1
- bs O + C-1
prior )-1 ( Ot C-1
- bs oobs + C-1
prior mprior )
, (4.15) and for Cpost the expression Cpost = ( Ot C-1
- bs O + C-1
prior )-1
. (4.16) After introducing the vector oobs and the matrix Cobs ,
♦♦❜s ❂ ④✻✳✾✸ ✱ ✶✶✳✷✾ ✱ ✶✷✳✽✽ ✱ ✷✺✳✻✺ ✱ ✶✽✳✷✽ ✱ ✶✹✳✵✾ ✱ ✶✵✳✵✺ ✱ ✶✹✳✵✻ ✱ ✶✺✳✾✵ ✱ ✶✷✳✵✼ ✱ ✷✽✳✶✸ ✱ ✹✳✷✵✱ ✽✳✸✸✱ ✶✺✳✺✶ ✱ ✷✺✳✺✶ ✱ ✷✶✳✵✾ ✱ ✶✹✳✶✶ ✱ ✾✳✾✾✱ ✶✺✳✾✹ ✱ ✶✼✳✽✾ ✱ ✶✻✳✶✻ ✱ ✷✵✳✵✽⑥ ❀ ❈♦❜s ❂ s♦❜s ❫✷ ■❞❡♥t✐t②▼❛tr✐① ❬✷✷❪ ❀
we compute Cpost and mpost using the commands
❈♣♦st ❂ ■♥✈❡rs❡❬ ❚r❛♥s♣♦s❡❬❖❪✳ ■♥✈❡rs❡❬❈♦❜s ❪✳❖ ✰ ■♥✈❡rs❡❬ ❈♣r✐♦r❪ ❪ ❀ ♠♣♦st ❂ ❈♣♦st ✳✭ ❚r❛♥s♣♦s❡❬❖❪✳ ■♥✈❡rs❡❬❈♦❜s ❪✳ ♦♦❜s ✰ ■♥✈❡rs❡❬❈♣r✐♦r ❪✳ ♠♣r✐♦r ✮ ✭✯ ●❡♥❡r❛t✐♦♥ ♦❢ r❛♥❞♦♠ ♠♦❞❡❧s ✭♣♦st❡r✐♦r ❞✐str✐❜✉t✐♦♥✮ ✯✮ ❈♣♦st ❂ ✭✶✴✷✮✭❈♣♦st ✰ ❚r❛♥s♣♦s❡❬❈♣♦st ❪✮❀ ▲♣♦st ❂ ❚r❛♥s♣♦s❡❬ ❈❤♦❧❡s❦②❉❡❝♦♠♣♦s✐t✐♦♥ ❬❈♣♦st ❪❪❀ ❙❡❡❞❘❛♥❞♦♠ ❬✶❪ r♥❞ ✿❂ ❙qrt ❬✷❪ ■♥✈❡rs❡❊r❢❬❘❛♥❞♦♠❬❘❡❛❧ ✱ ④✲✶✱ ✶⑥❪❪ ❉♦❬④ ♠r❛♥❞♦♠✵ ❂ ❚❛❜❧❡❬r♥❞ ✱ ④✶✻⑥❪ ✱ ♠r❛♥❞♦♠ ❂ ♠♣♦st ✰ ▲♣♦st✳♠r❛♥❞♦♠✵ ✱ P▲❖❚❬♠r❛♥❞♦♠❪ ⑥✱ ④✐✱ ✶✱ ✾⑥❪
♦♦❜s ❂ ④✻✳✾✸ ✱ ✶✶✳✷✾ ✱ ✶✷✳✽✽ ✱ ✷✺✳✻✺ ✱ ✶✽✳✷✽ ✱ ✶✹✳✵✾ ✱ ✶✵✳✵✺ ✱ ✶✹✳✵✻ ✱ ✶✺✳✾✵ ✱ ✶✷✳✵✼ ✱ ✷✽✳✶✸ ✱ ✹✳✷✵✱ ✽✳✸✸✱ ✶✺✳✺✶ ✱ ✷✺✳✺✶ ✱ ✷✶✳✵✾ ✱ ✶✹✳✶✶ ✱ ✾✳✾✾✱ ✶✺✳✾✹ ✱ ✶✼✳✽✾ ✱ ✶✻✳✶✻ ✱ ✷✵✳✵✽⑥ ❀ ❈♦❜s ❂ s♦❜s ❫✷ ■❞❡♥t✐t②▼❛tr✐① ❬✷✷❪ ❀ ❈♣♦st ❂ ■♥✈❡rs❡❬ ❚r❛♥s♣♦s❡❬❖❪✳ ■♥✈❡rs❡❬❈♦❜s ❪✳❖ ✰ ■♥✈❡rs❡❬ ❈♣r✐♦r❪ ❪ ❀ ♠♣♦st ❂ ❈♣♦st ✳✭ ❚r❛♥s♣♦s❡❬❖❪✳ ■♥✈❡rs❡❬❈♦❜s ❪✳ ♦♦❜s ✰ ■♥✈❡rs❡❬❈♣r✐♦r ❪✳ ♠♣r✐♦r ✮
Impatient people may now rush to look the the model mpost , and to examine the stan- dard deviations and correlations in Cpost , but I prefer we proceed orderly. The philosophy
- f our approach is to pass from the prior Gaussian distribution in the model space to the pos-
terior Gaussian distribution. In the same way we obtained samples of the prior distribution (figure 4.3), the commands3
✭✯ ●❡♥❡r❛t✐♦♥ ♦❢ r❛♥❞♦♠ ♠♦❞❡❧s ✭♣♦st❡r✐♦r ❞✐str✐❜✉t✐♦♥✮ ✯✮ ❈♣♦st ❂ ✭✶✴✷✮✭❈♣♦st ✰ ❚r❛♥s♣♦s❡❬❈♣♦st ❪✮❀ ▲♣♦st ❂ ❚r❛♥s♣♦s❡❬ ❈❤♦❧❡s❦②❉❡❝♦♠♣♦s✐t✐♦♥ ❬❈♣♦st ❪❪❀ ❙❡❡❞❘❛♥❞♦♠ ❬✶❪ r♥❞ ✿❂ ❙qrt ❬✷❪ ■♥✈❡rs❡❊r❢❬❘❛♥❞♦♠❬❘❡❛❧ ✱ ④✲✶✱ ✶⑥❪❪ ❉♦❬④ ♠r❛♥❞♦♠✵ ❂ ❚❛❜❧❡❬r♥❞ ✱ ④✶✻⑥❪ ✱ ♠r❛♥❞♦♠ ❂ ♠♣♦st ✰ ▲♣♦st✳♠r❛♥❞♦♠✵ ✱ P▲❖❚❬♠r❛♥❞♦♠❪ ⑥✱ ④✐✱ ✶✱ ✾⑥❪
produce the 9 models displayed in figure 4.5. This figure gives a quite clear idea of the resolution allowed by our data. The expert eye can directly guess what the posterior un- certainties and the posterior correlations are (and, of course, a larger number of posterior realization would allow to compute these with great accuracy). The average of a large enough set of posterior sample would be prior , and an evaluation
- f the statistical covariances would give Cpost . But they are already available to us, so we
- nly need to plot them. The vector mpost (the output of the least-squares formulas) is plotted
in figure 4.6.
3Although the matrix Cpost is, in principe symmetric, rounding errors may cause the Cholesky decomposi-
tion to work. For this reason, it is better to “resymmetrize” it.
Figure 4.5: Nine pseudorandom models that are sam- ples of the posterior Gaussian distribution in the model space obtained used the data of 22 rays. Figure 4.6: Mean posterior model mpost . This is the output of the least- squares formulas. We would also obtain this model if taking the average of a large number of models like those in figure 4.5.
To examine the matrix Cpost we can first compute the associated standard deviations, that are the square roots of the diagonal elements, σi =
- Cii
. (4.17) Using the color scale in figure 4.7 the posterior uncertainties can be represented as in fig- ure 4.8. Figure 4.7: The scale used for the model parameter uncertainties. The same scale is used for all the display of uncertainties in this sec- tion.
0.08 cm-1 1.5 cm-1
The off-diagonal elements of the posterior covariance matrix Cpost are better analyzed by introducing the correlations ρij = Cij σi σj (4.18) whose values satisfy the constraints −1 ≤ ρij ≤ +1 , (4.19) with the value +1 meaning perfect correlation, and the value −1 meaning perfect anticor-
- relation. Using the color scale in figure 4.9, the correlations can be displayed as in figure 4.10.
It is not of much interest to represent a crude 16 × 16 matrix. Rather, I have plotted a 4 × 4 array of 4 × 4 matrices, each array representing the correlation between the uncertainty of a given block with all other blocks. In each submatrix, the correlation of the block to which the submatrix is associated with the block itself is always 1, per definition. So, what matters are the other correlations. They are quite large in absolute value, this being due to the fact
Figure 4.8: The posterior uncertainties, obtained as the square root of the diag-
- nal elements of the matrix Cpost . They could also be obtained as the statisti-
cal standard deviations from a large number of models like those in figure 4.5. Note that the uncertainties in the blocks that are in the corner and in the cen- ter are smaller than those of the other blocks, due to the special geometry of the rays: the blocks at the corners are well resolved because there is, for each block, one ray that only explores the block. As a consequence, the central blocks also are well resolved (because of the existence of the diagonal rays).
Figure 4.9: The scale used for the correlation values. The value +1 means perfect correlation, and the value −1 means perfect anticor- relation.
- 1
+1
Figure 4.10: Posterior correlations. This 4 × 4 array corresponds to the 4 × 4 blocks in the model: at position (i, j) of this array, there are the correlations of the posterior value (i, j) with all other pos- terior values. As an example, the green cross corresponds to the correlation be- tween the value for block (1, 4) and it- self (necessarily +1 ), while the red cross corresponds to the correlation between the value for block (1, 4) and the value for block (3, 4) . Although these corre- lations have been obtained from the ma- trix Cpost , they could have been obtained as the statistical correlations from a large number of models like those in figure 4.5.
that we only have ‘22 data for 16 unknowns’. The patterns of covariances in the blocks that habe large posterior uncertainties is particularly interesting, with correlations both, close to +1 and close to −1 . Once again, this results from the particular geometry of the rays. One may anticipate that, adding more and more data, with paths that are not only the horizontal- vertical-diagonal of this exercise, would make (i) the uncertainties tend to zero, and (ii) the correlations between distinct blocks tend to zero. To further explore this exercise, let us suppress the data corresponding to the rays 07, 11, 15, and 19, solve the problem again. When the data corresponding to these four rays is suppressed, one is left with the geometry indicated in figure 4.11. There are two ways of Figure 4.11: Geometry of the ‘X-ray experiment’ when the rays 07, 11, 15, and 19, are suppressed. Figure at scale 1:3.
01020304 05060708 09101112 13141516
suppressing these data: setting a new problem with less rays (redefining the matrix O , sup- pressing 4 of its rows) or artificially putting enormous uncertainties on the data associated to these rays. In the accompanying computer code both possibilities are implemented, and they give practically undistinguishable results. To implement the second possibility we only need to replace a few of the lines in the
✭✯ ❚❤❡ ♦❜s❡r✈❛t✐♦♥s ❛❣❛✐♥ ✱ ❢♦✉r ♦❢ t❤❡ ❞❛t❛ ✇✐t❤ ❤✉❣❡ ✉♥❝❡rt❛✐♥t✐❡s ✯✮ ♦♦❜s ❂ ④✻✳✾✸ ✱ ✶✶✳✷✾ ✱ ✶✷✳✽✽ ✱ ✷✺✳✻✺ ✱ ✶✽✳✷✽ ✱ ✶✹✳✵✾ ✱ ✶✵✵✳✱ ✶✹✳✵✻ ✱ ✶✺✳✾✵ ✱ ✶✷✳✵✼ ✱ ✶✵✵✳✱ ✹✳✷✵✱ ✽✳✸✸✱ ✶✺✳✺✶ ✱ ✶✵✵✳✱ ✷✶✳✵✾ ✱ ✶✹✳✶✶ ✱ ✾✳✾✾✱ ✶✵✵✳✱ ✶✼✳✽✾ ✱ ✶✻✳✶✻ ✱ ✷✵✳✵✽⑥ ❀ ❈♦❜s ❂ s❞❫✷ ■❞❡♥t✐t②▼❛tr✐① ❬✷✷❪ ❀ ❈♦❜s ❬❬✵✼✱ ✵✼❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀ ❈♦❜s ❬❬✶✶✱ ✶✶❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀ ❈♦❜s ❬❬✶✺✱ ✶✺❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀ ❈♦❜s ❬❬✶✾✱ ✶✾❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀
To implement the second possibility we only need to replace a few of the lines in the previous code, those defining the observations:
✭✯ ❚❤❡ ♦❜s❡r✈❛t✐♦♥s ❛❣❛✐♥ ✱ ❢♦✉r ♦❢ t❤❡ ❞❛t❛ ✇✐t❤ ❤✉❣❡ ✉♥❝❡rt❛✐♥t✐❡s ✯✮ ♦♦❜s ❂ ④✻✳✾✸ ✱ ✶✶✳✷✾ ✱ ✶✷✳✽✽ ✱ ✷✺✳✻✺ ✱ ✶✽✳✷✽ ✱ ✶✹✳✵✾ ✱ ✶✵✵✳✱ ✶✹✳✵✻ ✱ ✶✺✳✾✵ ✱ ✶✷✳✵✼ ✱ ✶✵✵✳✱ ✹✳✷✵✱ ✽✳✸✸✱ ✶✺✳✺✶ ✱ ✶✵✵✳✱ ✷✶✳✵✾ ✱ ✶✹✳✶✶ ✱ ✾✳✾✾✱ ✶✵✵✳✱ ✶✼✳✽✾ ✱ ✶✻✳✶✻ ✱ ✷✵✳✵✽⑥ ❀ ❈♦❜s ❂ s❞❫✷ ■❞❡♥t✐t②▼❛tr✐① ❬✷✷❪ ❀ ❈♦❜s ❬❬✵✼✱ ✵✼❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀ ❈♦❜s ❬❬✶✶✱ ✶✶❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀ ❈♦❜s ❬❬✶✺✱ ✶✺❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀ ❈♦❜s ❬❬✶✾✱ ✶✾❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀
The data corresponding to the missing rays has been set to an arbitrary value (100) and the associated standard deviations has been set to a large value (10 000). Note: explain here that we obtain the posterior models are displayed in figure 4.12, the mean posterior model is displayed in figure 4.13, the posterior uncertainties are in figure 4.14, and the posterior correlations in figure 4.15. When comparing the two figures 4.13 and 4.14 to the two figures 4.6 and 4.8, we observe the following. The posterior model is similar, excepted for the block (1, 1) , that now takes the a priori value ( 5 cm-1 ), with a posterior uncertainty equal to the prior one ( 1.5 cm-1 ). This is because the suppression of the four rays makes that the block (1, 1) is not explored by any ray. The least-squares solution in this case gives a consistent a posteriori value for this block, identical to the a priori value, with a posterior uncertainty identical to the prior one. In addition we see in the central panel that the posterior uncertainty for this block is also identical to the prior one, as it should. Note also, in figure 4.15, that the uncertainty of this block is completely uncorrelated with the uncertainties of the other blocks.
Figure 4.12: Posterior models. Figure 4.13: Mean posterior model.
Figure 4.14: Posterior uncertainties. Figure 4.15: Posterior correlations.
As a further example, let us now suppress the data corresponding to the rays 02, 05, 08, 10, 13, 16, 19, and 22. When the data corresponding to these eight rays is suppressed, one is left with the geometry indicated in figure 4.16. As above, we simply replace a few of the Figure 4.16: Geometry of the ‘X-ray experiment’ when the rays 02, 05, 08, 10, 13, 16, 19, and 22, are suppressed. Figure at scale 1:3.
01020304 05060708 09101112 13141516
lines in the previous code, those defining the observations, simulating the absence of data as data with huge uncertainties:
✭✯ ❚❤❡ ♦❜s❡r✈❛t✐♦♥s ❛❣❛✐♥ ✱ ❡✐❣❤t ❞❛t❛ ✇✐t❤ ❤✉❣❡ ✉♥❝❡rt❛✐♥t✐❡s ✯✮ ♦♦❜s ❂ ④✻✳✾✸ ✱ ✶✵✵✳✱ ✶✷✳✽✽ ✱ ✷✺✳✻✺ ✱ ✶✵✵✳✱ ✶✹✳✵✾ ✱ ✶✵✳✵✺ ✱ ✶✵✵✳✱ ✶✺✳✾✵ ✱ ✶✵✵✳✱ ✷✽✳✶✸ ✱ ✹✳✷✵✱ ✶✵✵✳✱ ✶✺✳✺✶ ✱ ✷✺✳✺✶ ✱ ✶✵✵✳✱ ✶✹✳✶✶ ✱ ✾✳✾✾✱ ✶✵✵✳✱ ✶✼✳✽✾ ✱ ✶✻✳✶✻ ✱ ✶✵✵✳⑥ ❀ ❈♦❜s ❂ s❞❫✷ ■❞❡♥t✐t②▼❛tr✐① ❬✷✷❪ ❀ ❈♦❜s ❬❬✵✷✱ ✵✷❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀ ❈♦❜s ❬❬✵✺✱ ✵✺❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀ ❈♦❜s ❬❬✵✽✱ ✵✽❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀ ❈♦❜s ❬❬✶✵✱ ✶✵❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀ ❈♦❜s ❬❬✶✸✱ ✶✸❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀ ❈♦❜s ❬❬✶✻✱ ✶✻❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀ ❈♦❜s ❬❬✶✾✱ ✶✾❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀ ❈♦❜s ❬❬✷✷✱ ✷✷❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀
As above, the data corresponding to the missing rays has been set to an arbitrary value (100) and the associated standard deviations has been set to a large value (10 000). Note: explain here that we obtain the posterior models are displayed in figure 4.17, the
✭✯ ❚❤❡ ♦❜s❡r✈❛t✐♦♥s ❛❣❛✐♥ ✱ ❡✐❣❤t ❞❛t❛ ✇✐t❤ ❤✉❣❡ ✉♥❝❡rt❛✐♥t✐❡s ✯✮ ♦♦❜s ❂ ④✻✳✾✸ ✱ ✶✵✵✳✱ ✶✷✳✽✽ ✱ ✷✺✳✻✺ ✱ ✶✵✵✳✱ ✶✹✳✵✾ ✱ ✶✵✳✵✺ ✱ ✶✵✵✳✱ ✶✺✳✾✵ ✱ ✶✵✵✳✱ ✷✽✳✶✸ ✱ ✹✳✷✵✱ ✶✵✵✳✱ ✶✺✳✺✶ ✱ ✷✺✳✺✶ ✱ ✶✵✵✳✱ ✶✹✳✶✶ ✱ ✾✳✾✾✱ ✶✵✵✳✱ ✶✼✳✽✾ ✱ ✶✻✳✶✻ ✱ ✶✵✵✳⑥ ❀ ❈♦❜s ❂ s❞❫✷ ■❞❡♥t✐t②▼❛tr✐① ❬✷✷❪ ❀ ❈♦❜s ❬❬✵✷✱ ✵✷❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀ ❈♦❜s ❬❬✵✺✱ ✵✺❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀ ❈♦❜s ❬❬✵✽✱ ✵✽❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀ ❈♦❜s ❬❬✶✵✱ ✶✵❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀ ❈♦❜s ❬❬✶✸✱ ✶✸❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀ ❈♦❜s ❬❬✶✻✱ ✶✻❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀ ❈♦❜s ❬❬✶✾✱ ✶✾❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀ ❈♦❜s ❬❬✷✷✱ ✷✷❪❪ ❂ ✶✵✵✵✵✵✵✵✵ ❀
Note: explain here that we obtain the posterior models are displayed in figure 4.17, the mean posterior model is displayed in figure 4.18, the posterior uncertainties are in figure 4.19, and the posterior correlations in figure 4.20. Figure 4.17: Posterior models. The most remarkable fact on this solution is that the a posteriori model is very similar to that obtained using the whole 22 data. We are computing the value of 16 parameters, and we are only using 14 data, so we have ‘less data than unknowns’. Of course, it is the use of
Figure 4.18: Mean posterior model. Figure 4.19: Posterior uncertainties.
Figure 4.20: Posterior correlations.
the a priori information that allows to obtain reasonable results even in this situation. Note that two of the blocks have large posterior uncertainties (note: I have to explain this better).
First comment on last lesson: Consider a vector x whose probability distribution is Gaus- sian, with mean x0 and covariance C , f (x) = const. exp
- 1
2 (x − x0)t C-1 (x − x0)
- To generate a random realization of x ,
- make the Cholesky decomposition C = L Lt ,
- generate a vector w , each component being a random num-
ber generated from a 1D Gaussian with zero mean and unit variance, w[i] = Sqrt[2] InverseErf[Random[Real,-1,1]],
- take x = x0 + L w .