On Attacking S tatistical S pam Filters Greg Wittel & S . - - PowerPoint PPT Presentation

on attacking s tatistical s pam filters
SMART_READER_LITE
LIVE PREVIEW

On Attacking S tatistical S pam Filters Greg Wittel & S . - - PowerPoint PPT Presentation

On Attacking S tatistical S pam Filters Greg Wittel & S . Felix Wu U.C. Davis CEAS 2004 1 Outline Introduction Attack Classes Testing A New Attack Conclusions & Future 2 Attack Classes Attempted attack


slide-1
SLIDE 1

1

On Attacking S tatistical S pam Filters

Greg Wittel & S . Felix Wu U.C. Davis

CEAS 2004

slide-2
SLIDE 2

2

Outline

  • Introduction
  • Attack Classes
  • Testing A New Attack
  • Conclusions & Future
slide-3
SLIDE 3

3

Attack Classes

  • Attempted attack methods:

– Tokenization

  • Works against feature selection by splitting or

modifying key message features

  • e.g. S

plitting up words with spaces, HTML tricks

– Obfuscation

  • Use encoding or misdirection to hide contents

from filter

  • e.g. HTML/ URL encoding, letter substitution
slide-4
SLIDE 4

4

Attack Classes cont.

– Weak Statistical

  • S

kew message statistics by adding in random data

  • e.g. Add in random words, fake HTML tags,

random text excerpts

– Strong Statistical

  • Differentiated from ‘ weak’ attacks by using

more intelligence in the attack

  • Guessing v. educated guessing
  • e.g. Graham-Cumming Attack
slide-5
SLIDE 5

5

Attack Classes cont.

– Misc:

  • S

parse Data attack

  • Hash breaking attacks
slide-6
SLIDE 6

6

Testing A New Attack

  • Tested two types of attacks:

– Dictionary word attack (old) – Common word attack (new)

  • Both attacks add n random words to a

base message.

  • Tested against two filters:

– CRM114 - S parse binary poly. + Naïve Bayesian – SpamBayes (S B) - Naïve bayesian

slide-7
SLIDE 7

7

Procedure

  • Training data

– 3000 hams from S pamAssassin corpus – 3000 spams from S pamArchive-mod corpus – CRM114 trained on errors – SB using bulk training

slide-8
SLIDE 8

8

Procedure cont.

  • Test data

– Started with a base ‘ picospam’ not in training data:

From: Kelsey Stone <bouhooh@entitlement.com> To: submit@spamarchive.org Subject: Erase hidden Spies or Trojan Horses from your computer Erase E-Spyware from your computer http://boozofoof.spywiper.biz

slide-9
SLIDE 9

9

Procedure cont.

  • Test data cont.

– Base picospam is detectable by filters – Generated 1000 variations with n words added.

  • Words selected with and without replacement
  • n = 10, 25, 50, 100, 200, 300, 400

– Recorded classifications, effect on score

slide-10
SLIDE 10

10

Results

  • Using 10,000 variants didn’ t effect results
  • S

election with/ without replacement had no effect

  • Mixed results
slide-11
SLIDE 11

11

CRM114 Results

  • Both attacks failed; 0 false negatives
  • S

pam score was effected...

slide-12
SLIDE 12

12

CRM114 Results cont.

0.75 0.8 0.85 0.9 0.95 1 400 300 200 100 50 25 10 Spam probability Words added Base score Dictionary Common

slide-13
SLIDE 13

13

SpamBayes Results

  • Baseline Dictionary attack: mild success
  • Common word attack...
slide-14
SLIDE 14

14

S pamBayes Results cont.

0.2 0.4 0.6 0.8 1 400 300 200 100 50 25 10 Spam probability Words added Ham Thresh. Spam Thresh. Dictionary Common

slide-15
SLIDE 15

15

S pamBayes Results cont.

  • Common word attack reduces attack size

by up to 4x

  • What Happened?

Why such poor performance on either attack?

  • Hypothesis: Basis picospam was not in

training data.

  • Added the basis spam to S

B’ s training data…

slide-16
SLIDE 16

16

S pamBayes Results Part 2

  • Retrained filter offered greater resistance

to ‘ weak’ dictionary attack.

  • S

mall performance gain against common word attack.

  • Gains not big enough to resist attack
slide-17
SLIDE 17

17

S pamBayes Results Part 2 cont.

Dict ionary Word Attack

0.2 0.4 0.6 0.8 1 400 300 200 100 50 25 10 Spam probability Words added Ham Thresh. Spam Thresh. Before After

slide-18
SLIDE 18

18

S pamBayes Results Part 2 cont.

Common Word At tack

0.2 0.4 0.6 0.8 1 400 300 200 100 50 25 10 Spam probability Words added Ham Thresh. Spam Thresh. Before After

slide-19
SLIDE 19

19

Conclusion & Future...

  • Mixed success of common word attack

shows need for further study

  • Other filters

– Bogofilter shows similar vulnerability

  • Effect of re-training on attack msgs v.

– False negative, false positive rate

  • Testing other basis picospams
slide-20
SLIDE 20

20

Future cont.

  • What makes a filter hard to distract?
  • Relevance of independence assumption
  • More advanced attacks

– Natural language generation

  • Traditional software flaws

– Exploitable buffer overflows – Remote code execution

slide-21
SLIDE 21

21

Colophon

  • Contact information:

– Greg Wittel ( wittel at cs . ucdavis . edu ) – S. Felix Wu ( wu at cs . ucdavis . edu )

  • Questions?