Steganalysis in high dimensions: Fusing classifiers built on random - - PowerPoint PPT Presentation

steganalysis in high dimensions fusing classifiers built
SMART_READER_LITE
LIVE PREVIEW

Steganalysis in high dimensions: Fusing classifiers built on random - - PowerPoint PPT Presentation

Steganalysis in high dimensions: Fusing classifiers built on random subspaces Jan Kodovsk, Jessica Fridrich January 25, 2011 / SPIE Steganalysis in high dimensions:, Fusing classifiers built on random subspaces 1 / 14 Motivation Modern


slide-1
SLIDE 1

Steganalysis in high dimensions: Fusing classifiers built on random subspaces

Jan Kodovský, Jessica Fridrich January 25, 2011 / SPIE

1 / 14 Steganalysis in high dimensions:, Fusing classifiers built on random subspaces

slide-2
SLIDE 2

Motivation

Modern steganography

– Minimizing a distortion function in a high dimensional feature space Example: HUGO [Pevný-2010] (spatial domain) – 107 dimensions – Preserving complex models Example: Optimized ±1 embedding (JPEG domain) [Filler-Yesterday]

Modern approach to steganalysis

– Needs to follow the suit and capture more and more statistics – Cartesian calibration [2009] – doubles dimensionality – Merging of existing features together – ±1 embedding − → SPAM features (686) [Pevný-2009] – YASS algorithm (JPEG domain) − → CDF features (1,234) [2010]

2 / 14 Steganalysis in high dimensions:, Fusing classifiers built on random subspaces

slide-3
SLIDE 3

Curse of dimensionality

Growing complexity of training Limited training data / no access to the cover source Degradation of generalization abilities (overtraining) ⇒ model assumptions / regularization Problems with data / memory management Saturation of performance below its potential Features are designed to have low dimensionality

3 / 14 Steganalysis in high dimensions:, Fusing classifiers built on random subspaces

slide-4
SLIDE 4

Our goals

Challenge the low-dimensional limitation for a feature design Replace human design of features with an automatized procedure Rethink machine learning approach to steganalysis Classify in very high dimensions with low complexity and without compromising the performance Improve state-of-the-art steganalysis

4 / 14 Steganalysis in high dimensions:, Fusing classifiers built on random subspaces

slide-5
SLIDE 5

What are the options?

  • 1. Apply a classification tool of choice directly
  • 2. Reduce dimensionality and then classify

Unsupervised techniques (PCA) Supervised techniques (feature extraction / selection) Can be thought of as part of the feature design

  • 3. Reduce dimensionality and simultaneously classify

Minimize an appropriately defined objective function (SVDM) Iterative process with a classification feedback (embedded methods)

  • 4. Ensemble methods

Reduce dimensionality randomly and construct a simple classifier Repeat L times and aggregate the individual decisions

5 / 14 Steganalysis in high dimensions:, Fusing classifiers built on random subspaces

slide-6
SLIDE 6

The proposed framework

Step 1 – Form high-dimensional prefeatures

Capture as many dependencies among cover elements as possible Don’t be restricted by a dimensionality Emphasize diversity of individual features

Step 2 – Classify in high dimensions using an ensemble approach

prefeatures high dimension

  • dim. d
  • dim. k ≪ d

repeat L times random subspace random subspace random subspace classification classification classification classifier fusion 6 / 14 Steganalysis in high dimensions:, Fusing classifiers built on random subspaces

slide-7
SLIDE 7

Specific implementation

Random subspace = random selection (without repetition)

⇒ The complexity does not depend on the dimensionality d

Individual classifiers (base learners)

– Need to be sufficiently diverse (need to make different errors) – Weak and unstable classifiers preferable – Our choice: Fisher Linear Discriminants (FLDs)

Fusion = majority voting scheme L

i=1decision(i) > threshold

Parameters k ≈ 300 – 3000, L ≈ 30 – 150

Relation to previous art:

[Freund-1999] – Boosting (aggregation of weak classifiers) [Breiman-2001] – Random forests (base learners = trees)

7 / 14 Steganalysis in high dimensions:, Fusing classifiers built on random subspaces

slide-8
SLIDE 8

Comparison with SVM

JPEG domain, algorithm nsF5, database of 6500 images State-of-the-art feature sets

– CC-PEV (2×274 = 548) – [Pevný-2007] + Cartesian calibration – CC-SHI (2×324 = 648) [Shi-2006] Cartesian calibration 0.05 0.10 0.15 0.20 0.1 0.2 0.3 0.4 Relative payload (bpac) Testing error

G-SVM Ensemble

CC-PEV (548)

– k = 400, L = 31 – Ensemble: 70 sec – G-SVM: 250 sec (3.5 × longer) Full training: 8 hrs!

8 / 14 Steganalysis in high dimensions:, Fusing classifiers built on random subspaces

slide-9
SLIDE 9

Comparison with SVM

JPEG domain, algorithm nsF5, database of 6500 images State-of-the-art feature sets

– CC-PEV (2×274 = 548) – [Pevný-2007] + Cartesian calibration – CC-SHI (2×324 = 648) – [Shi-2006] + Cartesian calibration 0.05 0.10 0.15 0.20 0.1 0.2 0.3 0.4 Relative payload (bpac) Testing error

G-SVM Ensemble

CC-PEV (548) CC-SHI (648)

– k = 400, L = 31 – Ensemble: 70 sec – G-SVM: 250 sec (3.5 × longer) Full training: 8 hrs!

8 / 14 Steganalysis in high dimensions:, Fusing classifiers built on random subspaces

slide-10
SLIDE 10

Generating high-dimensional prefeatures (in JPEG domain)

DCT Plane 8×8 grid

intra-block dependencies inter-block dependencies combination of both

– 2D co-occurence matrices – Driven by mutual information – N matrices in total – Truncated to [−T, T] – Cartesian calibration – Dimension 2×N×(2×T+1)2 – T = 4, N = 300 → dim = 48,600 CC-CF features

9 / 14 Steganalysis in high dimensions:, Fusing classifiers built on random subspaces

slide-11
SLIDE 11

Steganalysis of nsF5

Influence of parameters L and k

30 60 90 120 150 0.31 0.34 0.37 0.4 0.43 Number of fused classifiers L Testing error

k = 1000 k = 2000 k = 3000

CC-CF (48,600)

– Payload 0.05 bpac – k = 2000, L = 149 → 30 min – G-SVM: 7.5 hrs (15 × longer) Full training > month – Performance quickly saturates as L grows – Choice of k is important (1D search may be conducted)

10 / 14 Steganalysis in high dimensions:, Fusing classifiers built on random subspaces

slide-12
SLIDE 12

Steganalysis of nsF5

Can we improve state-of-the-art?

0.05 0.10 0.15 0.20 0.1 0.2 0.3 0.4 Relative payload (bpac) Testing error

CC-PEV (548) CC-CF (48,600)

– CC-PEV: G-SVM – Rest: Ensemble k = 2000, L = 149

10 / 14 Steganalysis in high dimensions:, Fusing classifiers built on random subspaces

slide-13
SLIDE 13

Steganalysis of nsF5

Can we improve state-of-the-art?

0.05 0.10 0.15 0.20 0.1 0.2 0.3 0.4 Relative payload (bpac) Testing error

CC-PEV (548) CC-CF (48,600) ALL (49,796)

– CC-PEV: G-SVM – Rest: Ensemble k = 2000, L = 149 – ALL (49,796) = CC-PEV (548) + CC-SHI (648) + CC-CF (48,600)

10 / 14 Steganalysis in high dimensions:, Fusing classifiers built on random subspaces

slide-14
SLIDE 14

Steganalysis of nsF5

Can we improve state-of-the-art?

0.05 0.10 0.15 0.20 0.1 0.2 0.3 0.4 Relative payload (bpac) Testing error

CC-PEV (548) CC-CF (48,600) ALL (49,796) ALL+ (49,796)

– CC-PEV: G-SVM – Rest: Ensemble k = 2000, L = 149 – ALL (49,796) = CC-PEV (548) + CC-SHI (648) + CC-CF (48,600) – ALL+ = ALL with 300/2000 always chosen from CC-PEV

10 / 14 Steganalysis in high dimensions:, Fusing classifiers built on random subspaces

slide-15
SLIDE 15

Generating high-dimensional prefeatures (in SPATIAL domain)

Modeling the joint distribution of higher order local residuals Horizontal residual Hij = xij − Pred(N h

ij)

N h

ij

xij

Order

Hij 2

1 2(−xi,j−1 + 2xij − xi,j+1)

3

1 3(−xi,j−1 + 3xij − 3xi,j+1 + xi,j+2)

4

1 6(xi,j−2 − 4xi,j−1 + 6xij − 4xi,j+1 + xi,j+2)

5

1 10(xi,j−2 − 5xi,j−1 + 10xi,j − 10xi,j+1 + 5xi,j+2 − xi,j+3)

6

1 20(−xi,j−3 + 6xi,j−2 − 15xi,j−1 + 20xij − 15xi,j+1 + 6xi,j+2 − xi,j+3) 11 / 14 Steganalysis in high dimensions:, Fusing classifiers built on random subspaces

slide-16
SLIDE 16

Generating high-dimensional prefeatures (in SPATIAL domain)

Modeling the joint distribution of higher order local residuals Horizontal residual Hij = xij − Pred(N h

ij)

N h

ij

N v

ij

N d

ij

N m

ij

xij Hij = xij − Pred(N h

ij)

Vij = xij − Pred(N v

ij)

Dij = xij − Pred(N d

ij)

Mij = xij − Pred(N m

ij )

Hij, Vij, Dij, Mij − → MARKOV min{Hij, Vij, Dij, Mij} max{Hij, Vij, Dij, Mij} − → MINMAX 3D co-occurences, dimension 20×(2×T+1)3 (T = 4 → dim = 14,580)

11 / 14 Steganalysis in high dimensions:, Fusing classifiers built on random subspaces

slide-17
SLIDE 17

Steganalysis of HUGO

G-SVM − → CDF (1,234) = CC-PEV (548) + SPAM (686) Ensemble − → MINMAX+MARKOV (14,580), k = 1600, L = 51

0.1 0.2 0.3 0.4 0.5 0.1 0.2 0.3 0.4 0.5 Relative payload (bpp) Testing error G-SVM (CDF) Ensemble (MINMAX+MARKOV) BOSSbase (9074 images) size: 512×512, resized

12 / 14 Steganalysis in high dimensions:, Fusing classifiers built on random subspaces

slide-18
SLIDE 18

Summary

The main contributions for future steganalysis

High dimensionality doesn’t have to be a restriction for the feature design Proposed scalable, fast, and simple classification methodology based

  • n ensemble classifiers

One step further towards automatization of steganalysis Showed that state-of-the-art steganalysis can be improved by a large margin

Open problems

How to design prefeatures? How to define random projections?

13 / 14 Steganalysis in high dimensions:, Fusing classifiers built on random subspaces

slide-19
SLIDE 19

The power of random projections

Shigeo Fukuda, Lunch With a Helmet On (1987)