Reconst nstruct ruct Radio o Map with Automatic atically ally - - PowerPoint PPT Presentation

reconst nstruct ruct radio o map with automatic atically
SMART_READER_LITE
LIVE PREVIEW

Reconst nstruct ruct Radio o Map with Automatic atically ally - - PowerPoint PPT Presentation

Reconst nstruct ruct Radio o Map with Automatic atically ally Constru tructed cted Gaussia sian n Proces ess s for Localiz lizatio ation 01 | Background C O N T E N T 02 | Gaussian Process 03 | Kernel Selection 04 | Model


slide-1
SLIDE 1

Reconst nstruct ruct Radio

  • Map with Automatic

atically ally Constru tructed cted Gaussia sian n Proces ess s for Localiz lizatio ation

slide-2
SLIDE 2

目 录

C O N T E N T

01 | Background 02 | Gaussian Process 03 | Kernel Selection 04 | Model Ensemble 05 | Experiment

slide-3
SLIDE 3

Page . Page . 3

1

P A R T O N E

Background

slide-4
SLIDE 4

GPS: time consuming Power consuming Turn on meter Base station Signal strength

slide-5
SLIDE 5
  • indoor localization use

fingerprinting

  • creating a radio map
  • Received Signal Strength

Indicators (RSSI) values

  • btained from multiple

access points (APs)

  • a large outdoor

environment

  • sample thousands of

survey sites to construct a fine grain radio map

  • a university usually

needs hundred thousand training data

slide-6
SLIDE 6
slide-7
SLIDE 7

SJTU 3G 4G Yindu Road 3G 4G About one million sample data 20 square kilometers

slide-8
SLIDE 8
slide-9
SLIDE 9

2

P A R T T W O

Gaussian Process

slide-10
SLIDE 10

Given a training set 𝐸 = 𝒚𝑗, 𝑧𝑗 𝑗 = 1, … , 𝑜} How to calculate the output 𝑧∗ for a new input 𝑦∗ Linear regression? – Least Square Method Nonlinear regression?

  • - Gaussian Process
slide-11
SLIDE 11

Rela elation

  • nship

ip to Lin inear ar Regr egression

  • n
  • In logistic regression, the input to the sigmoid

function is where are parameters.

  • A Gaussian process places a prior on the space of

functions f directly, without parameterizing f.

  • Therefore, Gaussian processes are non-parametric
  • more general than standard regression the form

not limited by a parametric form

T

f x b   

slide-12
SLIDE 12

Assume 𝒛 = {𝑧1, 𝑧2, … , 𝑧𝑜} obey multivariate Gaussian Distribution 𝑧1 ⋮ 𝑧𝑜 ~𝑂 0, 𝐿 → 𝒛~𝑂 0, 𝐿 Where Given a training set 𝐸 = 𝒚𝑗, 𝑧𝑗 𝑗 = 1, … , 𝑜} How to calculate the output 𝑧∗(RSS) for a new input 𝑦∗(longitude/latitude)

slide-13
SLIDE 13
  • A Gaussian process any finite number of which have

joint Gaussian distributions.

  • A Gaussian process is fully specified by its mean

function m(x) and covariance function k(x,x’).

  • Two things to define our GP:
  • choose a form for the mean function.
  • choose a form for the covariance function

De Definiti nition n of GP GP

( , ) f gp m k

slide-14
SLIDE 14

For new input data y* joint distribution defined as:

slide-15
SLIDE 15

Get conditional distribution: Mean and variance:

slide-16
SLIDE 16

Maxlimum: Hyper-parameters::

Tun une Hyper-Par aram ameters

Maxlimum the log likelihood: (conjugate gradients)

slide-17
SLIDE 17

SE kernel:

Hype per-parameters

𝜏

𝑔 2:overall vertical scale of variation of the latent value.

𝑚: characteristic length-scale

  • short means the error bars can grow rapidly away from the

data points.

  • large implies irrelevant features .

𝜏𝑜

2: noise variance

The covariance function defines how smoothly the (latent) function f varies from a given x.

slide-18
SLIDE 18

Hype per-parameters

Different hyper parameters Bias & Variance trade off

left with small 𝑚 right with large 𝑚

slide-19
SLIDE 19

3

P A R T t h r e e

Kernel Selection

slide-20
SLIDE 20

Squared Exponential Kernel Periodic Kernel Linear Kernel

slide-21
SLIDE 21

(1)ka + kb = ka(x, x′) + kb(x, x′) (Summation) (2)ka × kb = ka(x, x′) × kb(x, x′) (Multiplication)

slide-22
SLIDE 22

Automatic Mo Model Construction

choosing the structural form of the kernel: a black art

  • Search over sums and

products of kernels

  • Maximizing the BIC(M)
  • Show how any model be

decomposed into different parts BIC trades off model fit and complexity

slide-23
SLIDE 23

4

P A R T F O U R

Model Ensemble

slide-24
SLIDE 24

1)average out biases 2)reduce the variance 3)unlikely to overfit

Don’t Overfit! Averaging multiple different green lines should bring us closer to the black line.

slide-25
SLIDE 25

Linear regression SVR Gradient Tree Boosting Xgboost GP (RQ) GP (Compose) Rate Averaging Blending

slide-26
SLIDE 26

5

P A R T F I V E

Experiment Results

slide-27
SLIDE 27
slide-28
SLIDE 28
slide-29
SLIDE 29

6

P A R T S I X

Q&A