Intrinsic Structure Study on Whale Vocalizations Yin Xian 1 , - - PowerPoint PPT Presentation

intrinsic structure study on whale vocalizations
SMART_READER_LITE
LIVE PREVIEW

Intrinsic Structure Study on Whale Vocalizations Yin Xian 1 , - - PowerPoint PPT Presentation

2015 DCLDE Conference Intrinsic Structure Study on Whale Vocalizations Yin Xian 1 , Xiaobai Sun 2 , Yuan Zhang 3 , Wenjing Liao 3 Doug Nowacek 1,4 , Loren Nolte 1 , Robert Calderbank 1,2,3 1 Department of Electrical and Computer Engineering, Duke


slide-1
SLIDE 1

1

1Department of Electrical and Computer Engineering, Duke University 2Department of Computer Science, Duke University 3Department of Mathematics, Duke University 4Duke Marine Lab, Duke University

Intrinsic Structure Study on Whale Vocalizations

Yin Xian1, Xiaobai Sun2, Yuan Zhang3, Wenjing Liao3 Doug Nowacek1,4, Loren Nolte1, Robert Calderbank1,2,3

2015 DCLDE Conference

slide-2
SLIDE 2

2

Problem Statement

Goal: classify the whale signal from the hydrophone.

  • Passive acoustic;
  • Challenge: variation of whale vocalizations, background noise
slide-3
SLIDE 3

3

Variation of Whale Vocalizations

Bowhead whale calls [1] Humpback whale calls [1]

[1] Mobysound data. http://www.mobysound.org/mysticetes.html.

slide-4
SLIDE 4

4

Overview

  • Many whale vocalizations frequency modulated and can be

modeled as polynomial phase signals[2,3].

  • The intrinsic dimension can be described and estimated by the

number of polynomial phase parameters.

  • Use low dimension representation for the signals and classify

them.

[2] I. R. Urazghildiiev, and C. W. Clark. “Acoustic detection of North Atlantic right whale contact calls using the generalized likelihood ratio test,” J. Acoust. Soc. Am. 120, 1956-1963 (2006). [3] M. D. Beecher. “Spectrographic analysis of animal vocalizations: implications of the “uncertainty principle”,” Bioacoustics 1, 187-208 (1988).

slide-5
SLIDE 5

5

Dimension reduction Linear methods:

  • PCA
  • MDS (Multidimensional Scaling)

Non-linear methods:

  • Laplacian-Eigenmap
  • Isomap
slide-6
SLIDE 6

6

Road map

p n Linear model Non-linear model PCA MDS 1 𝑜 − 1 𝑌 𝑌 𝑈 𝑌 𝑈𝑌

….

Isomap Laplacian-Eigenmap

X

𝑀 = exp⁡ (𝐸, 𝑢)

….

𝐻 = 𝑉Σ𝑉𝑈, then spectral embedding to 𝑙 dimensions Classifier

slide-7
SLIDE 7

7 7

  • Denote the data by 𝑌 = [𝑦1, … , 𝑦𝑜] ∈ 𝑆𝑞×𝑜,
  • Covariance matrix: Σ𝑜 =

1 𝑜−1 𝑌

𝑌 𝑈, where 𝑌 = 𝑌 −

1 𝑜 𝑌𝑓𝑓𝑈

  • The eigenvalue decomposition: Σ𝑜 = 𝑉Λ𝑉𝑈
  • Choose the top k eigenvalues and the corresponding eigenvectors for Σ𝑜,

and compute 𝑍

𝑙 = 𝑉𝑙(Λ𝑙)−1/2

  • The PCA compute the top k right singular vectors for 𝑌

.

PCA

[4] H. Hotelling. Analysis of a complex of statistical variables into principal components. Journal

  • f Educational Psychology, 24(4):17–441,498–520 (1933).
slide-8
SLIDE 8

8 8

  • Denote the data by 𝑌 = [𝑦1, … , 𝑦𝑜] ∈ 𝑆𝑞×𝑜,
  • The distance matrix: 𝐸𝑗𝑘 = 𝑒𝑗𝑘

2 = ||𝑦𝑗 − 𝑦𝑘||2.

  • Compute 𝐶 = −

1 2 𝐼𝐸𝐼𝑈, where 𝐼 = 𝐽 − 𝑓e𝑈/𝑜, the centering matrix.

  • Compute eigenvalue decomposition 𝐶 = 𝑉Λ𝑉𝑈.
  • Choose top k nonzero eigenvalue and corresponding eigenvector for 𝐶⁡,

𝑌 𝑙 = 𝑉𝑙(Λ𝑙)−1/2

MDS

[5] J. B. Kruskal, and M. Wish. Multidimensional scaling. Vol. 11. Sage (1978).

slide-9
SLIDE 9

9 9

  • Denote the data by 𝑌 = [𝑦1, … , 𝑦𝑜] ∈ 𝑆𝑞×𝑜,
  • The distance matrix: 𝐸𝑗𝑘 = 𝑒𝑗𝑘

2 = ||𝑦𝑗 − 𝑦𝑘||2.

  • Compute 𝐶 = −

1 2 𝐼𝐸𝐼𝑈, where 𝐼 = 𝐽 − 𝑓e𝑈/𝑜, the centering matrix.

  • Compute eigenvalue decomposition 𝐶 = 𝑉Λ𝑉𝑈.
  • Choose top k nonzero eigenvalue and corresponding eigenvector for 𝐶⁡,

𝑌 𝑙 = 𝑉𝑙(Λ𝑙)−1/2

  • The relationship between 𝐶⁡and the covariance matrix via 𝑌

: 𝐶 = 𝑌 𝑈𝑌 , Σ𝑜 =

1 𝑜−1 𝑌

𝑌 𝑈 so the MDS is actually compute the top left singular vectors of 𝑌 .

MDS

[5] J. B. Kruskal, and M. Wish. Multidimensional scaling. Vol. 11. Sage (1978).

slide-10
SLIDE 10

10

Idea of Isomap

[6] J. B. Tenenbaum et al. "A global geometric framework for nonlinear dimensionality reduction." Science 290, 2319-2323 (2000).

  • For two arbitrary points on a nonlinear manifold, their Euclidean distance in the

high-dimensional input space may not accurately reflect their intrinsic similarity, as measured by geodesic distance along the low-dimensional manifold.

  • The two-dimensional embedding recovered by Isomap, which best preserves the

shortest path distances in the neighborhood graph. The “Swiss roll” data set, illustrating Isomap exploits geodesic paths for nonlinear dimensionality reduction.

slide-11
SLIDE 11

Isomap

  • Construct a neighborhood graph G=(X, E, D), based on k nearest

neighborhood, or ε-neighborhood.

  • Compute 𝐸,
  • Compute 𝐿 = −

1 2 𝐼𝐸𝐼𝑈, where 𝐼 = 𝐽 − 𝑓𝑓𝑈/𝑜 is the centering matrix.

  • Compute eigenvalue decomposition 𝐿 = 𝑉Λ𝑉𝑈.
  • Choose the top k eigenvalues and eigenvectors and compute 𝑌

𝑙 = 𝑉𝑙(Λ𝑙)−1

2.

11 [6] J. B. Tenenbaum et al. "A global geometric framework for nonlinear dimensionality reduction." Science 290, 2319-2323 (2000).

slide-12
SLIDE 12

Laplacian-Eigenmap

  • Construct a neighborhood graph G=(X, E, W), based on k nearest

neighborhood, or ε-neighborhood.

  • Choose the weight:

𝑥𝑗𝑘 = 𝑓−||𝑦𝑗−𝑦𝑘||2

𝑢

, 𝑗, 𝑘⁡𝑑𝑝𝑜𝑜𝑓𝑑𝑢𝑓𝑒 0, ⁡⁡⁡⁡⁡⁡𝑝𝑢ℎ𝑓𝑠𝑥𝑗𝑡𝑓

  • Eigenmap:

– Construct Laplacian matrix L=D-W, where D=diag( 𝑥𝑗𝑘

𝑘∈𝑂𝑗

) – Compute eigenvalues and eigenvectors:

𝑀𝒈 = 𝜇𝐸𝒈

𝒈 = [𝑔

0, … , 𝑔 𝑙] corresponds to 𝐸 = diag(𝜇1, …, 𝜇k), 𝜇i <= 𝜇i + 1

  • Leave out the eigenvector 𝑔

0.

  • The m dimensional embedding with (𝑔

1, … , 𝑔 𝑛).

12 [7] M. Belkin, and P. Niyogi. “Laplacian Eigenmaps for dimensionality reduction and data representation.” Neural computation, 15, 1373-1396, (2003).

slide-13
SLIDE 13

Laplacian-Eigenmap

  • Construct a neighborhood graph G=(X, E, W), based on k nearest

neighborhood, or ε-neighborhood.

  • Choose the weight:

𝑥𝑗𝑘 = 𝑓−||𝑦𝑗−𝑦𝑘||2

𝑢

, 𝑗, 𝑘⁡𝑑𝑝𝑜𝑜𝑓𝑑𝑢𝑓𝑒 0, ⁡⁡⁡⁡⁡⁡𝑝𝑢ℎ𝑓𝑠𝑥𝑗𝑡𝑓

  • Eigenmap:

– Construct Laplacian matrix L=D-W, where D=diag( 𝑥𝑗𝑘

𝑘∈𝑂𝑗

) – Compute eigenvalues and eigenvectors:

𝑀𝒈 = 𝜇𝐸𝒈

𝒈 = [𝑔

0, … , 𝑔 𝑙] corresponds to 𝐸 = diag(𝜇1, …, 𝜇k), 𝜇i <= 𝜇i + 1

  • Leave out the eigenvector 𝑔

0.

  • The m dimensional embedding with (𝑔

1, … , 𝑔 𝑛).

  • For normalized Laplacian-Eigenmap, we compute Ф:

𝐸−1

2 𝐸 − 𝑋 𝐸−1 2Ф = 𝜇Ф

13

slide-14
SLIDE 14

DCLDE 2015 Data

14

Blue whale (# 851) Fin whale (# 244)

[8] DCLDE conference data. http://www.cetus.ucsd.edu/dclde/dataset.html.

slide-15
SLIDE 15

Mapping Data to two dimensions

15

PCA Laplacian Eigenmap Isomap Normalized Laplacian Eigenmap

slide-16
SLIDE 16

Mapping Data to three dimensions

16

PCA Laplacian Eigenmap Isomap Normalized Laplacian Eigenmap

slide-17
SLIDE 17

Eigenvalues (energy) distributions

17

PCA Laplacian Eigenmap Isomap Normalized Laplacian Eigenmap

slide-18
SLIDE 18

ROCs comparisons (2D)

18

KNN Logistic regression Naïve Bayes

  • We use 5-folds cross validation to

generate the ROC. That is, 681 blue whale sounds and 196 fin whale sounds for training, and 170 blue whale sounds and 48 fin whale sounds for testing.

  • We use k=7 for Isomap, and k=7, t=1

for Laplacian Eigenmap.

slide-19
SLIDE 19

AUCs comparisons

19

KNN Logistic regression Naïve Bayes

slide-20
SLIDE 20

Plots of adjacency matrix of Laplacian-Eigenmap

20

By systematic spectral re-ordering, the blue whale data and fin whale data are well separated in the adjacency matrix (we use 851 blue whale data, and 244 fin whale data).

slide-21
SLIDE 21

Summary

21

  • Efficient classification of the whale vocalizations from low-

dimensional intrinsic structure.

  • The intrinsic dimension of whale vocalizations can be

recovered from the eigenvalues energy distribution.

  • The nonlinear dimensional reduction methods work better

with data of nonlinear structure.

slide-22
SLIDE 22

Future topics

22

  • Further develop efficient manifold mappings for more

complex whale vocalizations, and other acoustic signals.

  • Apply optimization methods to enhance computational

efficiency for nonlinear dimensional mappings.

slide-23
SLIDE 23

Reference

  • [1] Mobysound data. http://www.mobysound.org/mysticetes.html.
  • [2] I. R. Urazghildiiev, and C. W. Clark. “Acoustic detection of North Atlantic right whale

contact calls using the generalized likelihood ratio test,” J. Acoust. Soc. Am. 120, 1956- 1963 (2006).

  • [3] M. D. Beecher. “Spectrographic analysis of animal vocalizations: implications of the

“uncertainty principle”,” Bioacoustics 1, 187-208 (1988).

  • [4] H. Hotelling. Analysis of a complex of statistical variables into principal components.

Journal of Educational Psychology, 24(4):17–441,498–520 (1933).

  • [5] J. B. Kruskal, and M. Wish. Multidimensional scaling. Vol. 11. Sage (1978).
  • [6] J. B. Tenenbaum, V. D. Silva, and J. C. Langford. "A global geometric framework for

nonlinear dimensionality reduction." Science 290, 2319-2323 (2000).

  • [7] M. Belkin, and P. Niyogi. “Laplacian Eigenmaps for dimensionality reduction and

data representation.” Neural computation, 15, 1373-1396, (2003).

  • [8] DCLDE conference data. http://www.cetus.ucsd.edu/dclde/dataset.html.
  • [9] Y. Xian, X. Sun, Y. Zhang, W. Liao, D. Nowacek, L. Nolte, and R. Calderbank.

“Intrinsic structure study of whale vocalization.” J. Acoust. Soc. Am. (in preparation)

23

slide-24
SLIDE 24

24

Backup slides

slide-25
SLIDE 25

Mobysound data

Humpback whale (#2310) Bowhead whale (#446)

25 [1] Mobysound data. http://www.mobysound.org/mysticetes.html.

slide-26
SLIDE 26

Mapping to two dimensions

26

PCA Isomap Laplacian Eigenmap

slide-27
SLIDE 27

Mapping to three dimensions

27

PCA Isomap Laplacian-Eigenmap

slide-28
SLIDE 28

Eigenvalues energy distributions

28

PCA Laplacian Eigenmap Isomap Normalized Laplacian Eigenmap

slide-29
SLIDE 29

ROCs comparisons (2D)

29

KNN Logistic regression Naïve Bayes

  • We use 5-folds cross validation to

generate the ROC. That is, 358 bowhead whale sounds and 1848 humpback whale sounds for training, and 88 bowhead whale sounds and 462 humpback whale sounds for testing.

  • We use k=7 for Isomap, and k=7, t=1

for Laplacian Eigenmap

slide-30
SLIDE 30

AUCs Comparisons

30

KNN Naïve Bayes Logistic regression

slide-31
SLIDE 31

Plots of adjacency matrix W of Laplacian-Eigenmap

31

By systematic spectral re-ordering, the bowhead whale data and humpback whale data are well separated in the adjacency matrix (we use 446 bowhead whale data, and 2310 humpback whale data).

slide-32
SLIDE 32

Local Linear Embedding

  • Construct a neighborhood graph G=(V, E, W),

– where V is the vertex {𝑦𝑗: 𝑗 = 1, … , 𝑜}; E is the edge { 𝑗, 𝑘 : if 𝑘 is a neighbor

  • f 𝑗}, k-nearest neighbors, 𝜁-neighbors; W is the Euclidean distance: 𝑒(𝑦𝑗, 𝑦𝑘)
  • Local fitting:

– Compute the weights – Solve the equation according to Lagrange multipler method: Let 𝑥𝑗 = [𝑥𝑗𝑘1, … , 𝑥𝑗𝑘𝑙]𝑈∈⁡𝑆𝑙 , 𝑌 𝑗 = [𝑦𝑘1 − 𝑦𝑗, … , 𝑦𝑘𝑙 − 𝑦𝑗], the local covariance matrix , we have:

32

slide-33
SLIDE 33

Local Linear Embedding

  • Global alignment

– Define a n by n weight matrix W: – Compute the global embedding matrix Y: That is construct a semi-definite matrix 𝐶 = 𝐽 − 𝑋 𝑈(𝐽 − 𝑋) and find the d+1 smallest eigenvectors of B: 𝑤0, 𝑤1, … , 𝑤𝑒, and the corresponding eigenvalues: 𝜇0, … , 𝜇𝑒,⁡drop the smallest eigenvector which is the constant vector, we have: Y=

𝑤1 𝜇1⁡, … , 𝑤𝑒 𝜇𝑒 .

  • Advantage of LLE

– Neighbor graph: k nearest neighbors is of O(kn). – W is sparse; – 𝐶 = 𝐽 − 𝑋 𝑈(𝐽 − 𝑋) is positive semi-definite.

33

𝑋 = 𝑥𝑗𝑘, 𝑘 ∈ 𝑂𝑗 0, 𝑝𝑢ℎ𝑓𝑠𝑥𝑗𝑡𝑓