Model-based adaptive spatial sampling for occurrence map - - PowerPoint PPT Presentation

model based adaptive spatial sampling for occurrence map
SMART_READER_LITE
LIVE PREVIEW

Model-based adaptive spatial sampling for occurrence map - - PowerPoint PPT Presentation

Model-based adaptive spatial sampling for occurrence map construction N. Peyrard and R. Sabbadin CompSust09 - Cornell University - june 2009 p. 1 Mapping spatial processes in environmental management Mapping pest occurrence


slide-1
SLIDE 1

CompSust’09 - Cornell University - june 2009

Model-based adaptive spatial sampling for occurrence map construction

  • N. Peyrard and R. Sabbadin

– p. 1

slide-2
SLIDE 2

CompSust’09 - Cornell University - june 2009

Mapping spatial processes in environmental management

P2001 200 400 600 100 200 300 400 500 P2002 200 400 600 100 200 300 400 500 P2003 200 400 600 100 200 300 400 500 P2004 200 400 600 100 200 300 400 500 200 400 600 100 200 300 400 500 Y2002 200 400 600 100 200 300 400 500 Y2003 200 400 600 100 200 300 400 500 Y2004 200 400 600 100 200 300 400 500

Mapping pest occurrence

  • Building pest occurrence map

in order to eradicate

  • Observations costly
  • Errors in mapping also costly

– p. 2

slide-3
SLIDE 3

CompSust’09 - Cornell University - june 2009

Mapping spatial processes in environmental management

Different problems depending on observations nature

  • Data visualization
  • Complete observations (everywhere)
  • Perfect observations (No errors/missing data)

⇒ How to visualize data?

  • Map reconstruction
  • Complete observations
  • Noisy observations

⇒ How to reconstruct the “true” map?

  • Sampling and map construction
  • Incomplete observations (not everywhere)
  • Noisy observations

⇒ Where to observe? / How to reconstruct?

– p. 3

slide-4
SLIDE 4

CompSust’09 - Cornell University - june 2009

Mapping spatial processes in environmental management

How to design an efficient spatial sampling method to estimate an occurrence (0/1) map when process to map has spatial structure

  • bservations are

imperfect/incomplete sampling is costly process does not evolve during the sampling period

– p. 4

slide-5
SLIDE 5

CompSust’09 - Cornell University - june 2009

Overview of the proposed approach

Optimization approach for designing spatial sampling policies The Hidden Markov Random Field model is used for:

  • Representing current uncertain knowledge about map to

reconstruct

  • Updating knowledge after observations
  • Defining a unique criterion for
  • map reconstruction from observed data
  • spatial sampling actions selection

– p. 5

slide-6
SLIDE 6

CompSust’09 - Cornell University - june 2009

Optimal sampling problem

X Y a

Hidden variable X Sampling action a Observation model p(Y = o|x, a) Question: How to reconstruct hidden variable X using sampling actions?

  • 1. Hidden variable model
  • 2. Updated model after sampling result
  • 3. Hidden variable reconstruction
  • 4. Sampling action optimization

– p. 6

slide-7
SLIDE 7

CompSust’09 - Cornell University - june 2009

Spatial sampling optimization

The hidden variable x is a map ⇒ The sampling

  • ptimization

problem has to be revisited Question: How to reconstruct hidden map x using sampling actions?

  • 1. Hidden map model
  • 2. Updated model after sampling result
  • 3. Hidden map reconstruction
  • 4. Sampling action optimization

– p. 7

slide-8
SLIDE 8

CompSust’09 - Cornell University - june 2009

Pairwise Markov random field (1)

  • Multiple interacting variables
  • Independence given

neighborhood ⇒ Pairwise Markov random field Question: How to reconstruct hidden map x using sampling actions?

  • 1. Hidden map model
  • 2. Updated model after sampling result
  • 3. Hidden map reconstruction
  • 4. Sampling action optimization

– p. 8

slide-9
SLIDE 9

CompSust’09 - Cornell University - june 2009

Pairwise Markov random field (2)

  • Multiple interacting variables
  • Independence given

neighborhood ⇒ Pairwise Markov random field

  • Interaction graph G = (V, E)
  • ψi: “weights” on states of vertex i
  • ψij: correlations “strength” between neighbor vertices
  • Z: normalizing constant / partition function

P(x) = 1 Z

i∈V

ψi(xi)

  • (i,j)∈E

ψij(xi, xj)

  • – p. 9
slide-10
SLIDE 10

CompSust’09 - Cornell University - june 2009

Hidden Markov random field (1)

Hidden variables Observations

  • a ∈ {0, 1}|V |: subset of V

selected for sampling

  • Independent observations:

P(o|x, a) =

  • i∈V

Pi(oi|xi, ai) Question: How to reconstruct hidden map x using sampling actions?

  • 1. Hidden map model
  • 2. Updated model after sampling result
  • 3. Hidden map reconstruction
  • 4. Sampling action optimization

– p. 10

slide-11
SLIDE 11

CompSust’09 - Cornell University - june 2009

Hidden Markov random field (2)

Hidden variables Observations

  • a ∈ {0, 1}|V |: subset of V

selected for sampling

  • Independent observations:

P(o|x, a) =

  • i∈V

Pi(oi|xi, ai) Updated Markov random field (Bayes’ theorem) P(x|o, a) = 1 Z

i∈V

ψ′

i(xi, oi, ai)

  • (i,j)∈E

ψij(xi, xj)

  • where

ψ′

i(xi, oi, ai)

= ψi(xi)Pi(oi|xi, ai)

– p. 11

slide-12
SLIDE 12

CompSust’09 - Cornell University - june 2009

Hidden map reconstruction (1)

  • Hidden variables

Observations Reconstruction

Local (MPM): x∗

i = arg maxxi Pi(xi|o, a), ∀i ∈ V

Question: How to reconstruct hidden map x using sampling actions?

  • 1. Hidden map model
  • 2. Updated model after sampling result
  • 3. Hidden map reconstruction
  • 4. Sampling action optimization

– p. 12

slide-13
SLIDE 13

CompSust’09 - Cornell University - june 2009

Hidden map reconstruction (2)

  • Hidden variables

Observations Reconstruction

Local (MPM): x∗

i = arg maxxi Pi(xi|o, a)

Value of reconstructed map Expected number of well classified sites in x∗ V MPM(o, a) = f

i∈V

max

xi Pi(xi|o, a)

  • – p. 13
slide-14
SLIDE 14

CompSust’09 - Cornell University - june 2009

Sampling action optimization (1)

Hidden variables Observations

  • a ∈ {0, 1}|V | selected for

sampling

  • Independent observations
  • ∈ {0, 1}|V |

⇒ How to optimize the choice

  • f a?

Question: How to reconstruct hidden map x using sampling actions?

  • 1. Hidden map model
  • 2. Updated model after sampling result
  • 3. Hidden map reconstruction
  • 4. Sampling action optimization

– p. 14

slide-15
SLIDE 15

CompSust’09 - Cornell University - june 2009

Sampling action optimization (2)

Hidden variables Observations

  • a ⊆ V selected for sampling
  • Independent observations o

result ⇒ How to optimize the choice

  • f a?

U(a) = −c(a) +

  • P(o|a)V (o, a)

a∗ = arg max

a

U(a)

  • The computation of a∗ is hard! (NP-hard)
  • Only feasible for small problems or needs approximation!

– p. 15

slide-16
SLIDE 16

CompSust’09 - Cornell University - june 2009

Approximate spatial sampling (1)

Approximate the computation of a∗ = arg max

a

−c(a) +

  • P(o|a)V MPM(o, a)
  • Explore cells where initial knowledge is the most uncertain:

marginal Pi(xi|o, a) closest to 1

2

˜ a = arg max

a

−c(a) + f  

i,ai=1

min

  • Pi(Xi = 1), Pi(Xi = 0)

  • Marginals computation is itself NP-hard

⇒ approximation using belief propagation (sum prod) algorithm

– p. 16

slide-17
SLIDE 17

CompSust’09 - Cornell University - june 2009

Approximate spatial sampling (2)

The approximation results from simplifying assumptions:

  • Sampling actions are reliable
  • No passive observations
  • Joint probability approximated by one with idependent

factors

– p. 17

slide-18
SLIDE 18

CompSust’09 - Cornell University - june 2009

Adaptive spatial sampling (1)

  • Idea:
  • Sampling locations not chosen once for all before the

sampling campaign

  • Intermediate observations are taken into account to design

next sampling step

  • Possibility to visit a cell more than once

– p. 18

slide-19
SLIDE 19

CompSust’09 - Cornell University - june 2009

Adaptive spatial sampling (2)

  • a sampling strategy δ is a tree
  • a

trajectory in δ: τ = (a1, o1, . . . , aK, oK)

Value of a leaf

U(τ) = −

K

  • k=1

c(ak) + V MPM(o0, o1, . . . , oK, a0, a1, . . . , aK)

Value of a strategy V (δ) =

τ U(τ)P(τ | δ)

– p. 19

slide-20
SLIDE 20

CompSust’09 - Cornell University - june 2009

Heuristic adaptive spatial sampling

  • Exact computation is PSPACE-hard !

⇒ Heuristic algorithm

  • on line computation
  • approximate method for static sampling at each step

– p. 20

slide-21
SLIDE 21

CompSust’09 - Cornell University - june 2009

Concluding remarks

  • A framework for spatial sampling optimization:
  • based on Hidden Markov random fields
  • different map quality criteria
  • extended to “adaptive” sampling
  • Problems too complex for exact resolution

⇒ Heuristic solution based on approximate marginals computation

  • Empirical validation on simulated problems:
  • Comparison of SSS, ASS and classical sampling

methods (random sampling, ACS)

  • Markov random fields parameters learned from real data
  • ASS > SSS > classical methods

– p. 21

slide-22
SLIDE 22

CompSust’09 - Cornell University - june 2009

Ongoing work

  • Exact algorithms for small problems (Usman Farrokh):

combining variable elimination and tree search

  • “Random sets + kriging” approach (Mathieu Bonneau):

development of a dedicated approximate method and comparison to the HMRF approach

  • PhD thesis on adaptive spatial sampling for weeds mapping

at the scale of an agricultural area (Sabrina Gaba, INRA-Dijon).

  • Future?

⇒ Spatial partially observed Markov decision processes

– p. 22

slide-23
SLIDE 23

CompSust’09 - Cornell University - june 2009

Questions?

Thanks for listening

– p. 23

slide-24
SLIDE 24

CompSust’09 - Cornell University - june 2009

Contents

1- Optimal sampling of a hidden random variable 2- Defining optimal spatial sampling problems 3- Approximate computation of an optimal strategy 4- Evaluation of proposed method on simulated data

– p. 24

slide-25
SLIDE 25

CompSust’09 - Cornell University - june 2009

Optimal sampling problem

Hidden variable model

X Y a

Prior model P(x) Question: How to reconstruct hidden variable X using sampling actions?

  • 1. Hidden variable model
  • 2. Updated model after sampling result
  • 3. Hidden variable reconstruction
  • 4. Sampling action optimization

– p. 25

slide-26
SLIDE 26

CompSust’09 - Cornell University - june 2009

Optimal sampling problem

Updated model

X Y a

Posterior: P(x|o, a) = P(o|x, a)P(x) P(o|a) Question: How to reconstruct hidden variable X using sampling actions?

  • 1. Hidden variable model
  • 2. Updated model after sampling result
  • 3. Hidden variable reconstruction
  • 4. Sampling action optimization

– p. 26

slide-27
SLIDE 27

CompSust’09 - Cornell University - june 2009

Optimal sampling problem

Hidden variable reconstruction

X Y a

x∗(o, a) = arg max

x

P(x|o, a) V (o, a) = f(P(x∗|o, a)) Question: How to reconstruct hidden variable X using sampling actions?

  • 1. Hidden variable model
  • 2. Updated model after sampling result
  • 3. Hidden variable reconstruction
  • 4. Sampling action optimization

– p. 27

slide-28
SLIDE 28

CompSust’09 - Cornell University - june 2009

Optimal sampling problem

Hidden variable reconstruction

X Y a

x∗(o, a) = arg max

x

P(x|o, a) V (o, a) = f(P(x∗|o, a)) Question: How to reconstruct hidden variable X using sampling actions?

  • x∗(o, a) is the best reconstruction given sampling result

(o, a)

  • V (o, a) is the value of reconstructed variable after sampling

result (o, a)

– p. 28

slide-29
SLIDE 29

CompSust’09 - Cornell University - june 2009

Optimal sampling problem

Sampling action optimization

X Y a

U(a) = −c(a) +

  • P(o|a)V (o, a)

a∗ = arg max

a

U(a) Question: How to reconstruct hidden variable X using sampling actions?

  • 1. Hidden variable model
  • 2. Updated model after sampling result
  • 3. Hidden variable reconstruction
  • 4. Sampling action optimization

– p. 29

slide-30
SLIDE 30

CompSust’09 - Cornell University - june 2009

Optimal sampling problem

Sampling action optimization

X Y a

U(a) = −c(a) +

  • P(o|a)V (o, a)

a∗ = arg max

a

U(a) Question: How to reconstruct hidden variable X using sampling actions? The value of an action is a tradeoff between

  • The cost c(a) of the action and
  • The expected quality of the reconstructed variable

(over all possible sample results)

– p. 30

slide-31
SLIDE 31

CompSust’09 - Cornell University - june 2009

Contents

1- Optimal sampling of a hidden random variable 2- Defining optimal spatial sampling problems 3- Approximate computation of an optimal strategy 4- Evaluation of proposed method on simulated data

– p. 31

slide-32
SLIDE 32

CompSust’09 - Cornell University - june 2009

Contents

1- Optimal sampling of a hidden random variable 2- Defining optimal spatial sampling problems 3- Approximate computation of an optimal strategy 4- Evaluation of proposed method on simulated data

– p. 32

slide-33
SLIDE 33

CompSust’09 - Cornell University - june 2009

HMRF model for fire ants problem (1)

Eradicated cells, year 2001 10 20 30 40 50 60 70 80 90 100 10 20 30 40 50 60 70 80 90 100 Searched cells, year 2002 10 20 30 40 50 60 70 80 90 100 10 20 30 40 50 60 70 80 90 100 Observations, year 2002 10 20 30 40 50 60 70 80 90 100 10 20 30 40 50 60 70 80 90 100

Eradication Search actions Observations (e) (a) (o)

  • eradication (at previous year): ei ∈ {0, 1}, i = 1, . . . n
  • search actions: passive search or active search,

ai ∈ {0, 1}, i = 1, . . . n

  • observations: no nest detected / at least one nest detected,
  • i ∈ {0, 1}, i = 1, . . . n

– p. 33

slide-34
SLIDE 34

CompSust’09 - Cornell University - june 2009

HMRF model for fire ants problem (2)

  • Distribution on maps = Potts model

Pe(x | α, β) = 1 Z exp

i∈V

αei eq(xi, 1) + β

  • (i,j)∈E

eq(xi, xj)

  • Distribution of observation given map, Pai(oi | xi, θ)
  • i \ xi

1 1 1 − θai 1 θai with θ0 < θ1

– p. 34

slide-35
SLIDE 35

CompSust’09 - Cornell University - june 2009

HMRF model for fire ants problem (3)

An initial arbitrary sampling (a0, o0) is used for:

  • Parameters estimation: λ = (α, β, θ)

approximate version of EM for HMRF (Simul field EM)

  • identification problem between α and θ
  • OK if θ known: use of expert values
  • Marginals computation: Pi(xi|o0

i , a0 i )

5 10 15 20 25 30 35 40 45 50 5 10 15 20 25 30 35 40 45 50

– p. 35

slide-36
SLIDE 36

CompSust’09 - Cornell University - june 2009

Heuristic sampling methods evaluation (1)

  • Evaluation on simulated data
  • Comparison of behavior of
  • random sampling (RS)
  • adaptive cluster sampling (ACS)
  • static heuristic sampling (SHS)
  • adaptive heuristic sampling (AHS)

– p. 36

slide-37
SLIDE 37

CompSust’09 - Cornell University - june 2009

Heuristic sampling methods evaluation (2)

  • Procedure: repeat 10 times
  • simulate hidden map x from P(x | α, β) (50 × 50 cells)
  • apply regular sampling (about 10% of area): a0
  • simulate o0 from Pai(oi | xi, θ) (regular sampling plus

passive search)

  • estimate initial knowledge
  • apply RS, ACS, SHS, AHS, 10 times

– p. 37

slide-38
SLIDE 38

CompSust’09 - Cornell University - june 2009

Rate of misclassified cells

500 1000 1500 2000 2500 0.02 0.04 0.06 0.08 0.1 0.12 0.14 0.16 0.18

Number of sampled cells Proportion of misclassified cells

Configuration 2: total classification errors

Static Adaptive Cluster Random 500 1000 1500 2000 2500 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45

Number of sampled cells Proportion of misclassified cells

Configuration 6: total classification errors

Static Adaptive Cluster Random 500 1000 1500 2000 2500 0.02 0.04 0.06 0.08 0.1 0.12 0.14

Number of sampled cells Proportion of misclassified cells

Configuration 8: total classification errors

Static Adaptive Cluster Random

α = (0, −2), β = 0.8 α = (0, 0), β = 0.5 α = (1 − 1), β = 0.4 θ = (0, 0.8)

legend: SHS AHA ACS RS

– p. 38

slide-39
SLIDE 39

CompSust’09 - Cornell University - june 2009

Per color error rate

misclassified empty cells misclassified

  • ccupied

cells

legend: SHS AHA ACS RS

500 1000 1500 2000 2500 0.02 0.04 0.06 0.08 0.1 0.12 0.14 Number of sampled cells Proportion of misclassified empty cells Configuration 2: misclassified empty cells Static Adaptive Cluster Random 500 1000 1500 2000 2500 0.1 0.2 0.3 0.4 0.5 0.6 0.7 Number of sampled cells Proportion of misclassified empty cells Configuration 6: misclassified empty cells Static Adaptive Cluster Random 500 1000 1500 2000 2500 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.1 0.11 0.12 Number of sampled cells Proportion of misclassified empty cells Configuration 8: misclassified empty cells Static Adaptive Cluster Random 500 1000 1500 2000 2500 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 Number of sampled cells Proportion of misclassified occupied cells Configuration 2: misclassified occupied cells Static Adaptive Cluster Random 500 1000 1500 2000 2500 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 Number of sampled cells Proportion of misclassified occupied cells Configuration 6: misclassified occupied cells Static Adaptive Cluster Random 500 1000 1500 2000 2500 0.02 0.04 0.06 0.08 0.1 0.12 0.14 0.16 0.18 Number of sampled cells Proportion of misclassified occupied cells Configuration 8: misclassified occupied cells Static Adaptive Cluster Random

α = (0, −2) α = (0, 0) α = (1 − 1) β = 0.8 β = 0.5 β = 0.4

– p. 39

slide-40
SLIDE 40

CompSust’09 - Cornell University - june 2009

General behavior

  • ACS is not adapted (as expected): poor results
  • Adaptive HS ≥ Static HS ≥ Random S
  • Discrepancy between Adaptive HS and Static HS increases

with

  • sampling ressources
  • hidden map structure

– p. 40

slide-41
SLIDE 41

CompSust’09 - Cornell University - june 2009

Where do we sample?

Hidden map

5 10 15 20 25 30 35 40 45 50 5 10 15 20 25 30 35 40 45 50

α = (1, −1), β = 0.4, θ = (0, 0.8)

– p. 41

slide-42
SLIDE 42

CompSust’09 - Cornell University - june 2009

Where do we sample?

20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 Static sampling: A and O

– p. 42

slide-43
SLIDE 43

CompSust’09 - Cornell University - june 2009

Where do we sample?

20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 Static sampling:marginals

– p. 43

slide-44
SLIDE 44

CompSust’09 - Cornell University - june 2009

Where do we sample?

20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 Adaptive sampling: A and O (cumul)

– p. 44

slide-45
SLIDE 45

CompSust’09 - Cornell University - june 2009

Where do we sample?

20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 20 40 Adaptive sampling: marginals

– p. 45

slide-46
SLIDE 46

CompSust’09 - Cornell University - june 2009

Where do we sample?

  • No sampling in large empty areas
  • Sampling preferably near detected occupied sites within low

density areas

  • If sampling ressources increase
  • SHS complete exploration until the whole area is

covered

  • AHA can visit several times a site before extending

exploration to another area

– p. 46