Scalability, Fidelity, and Containment in the Potemkin Virtual - - PowerPoint PPT Presentation

scalability fidelity and containment in the potemkin
SMART_READER_LITE
LIVE PREVIEW

Scalability, Fidelity, and Containment in the Potemkin Virtual - - PowerPoint PPT Presentation

Introduction Design Architecture & Evaluation Scalability, Fidelity, and Containment in the Potemkin Virtual Honeyfarm Michael Vrable , Justin Ma, Jay Chen, David Moore, Erik Vandekieft, Alex C. Snoeren, Geoffrey M. Voelker, Stefan Savage


slide-1
SLIDE 1

Introduction Design Architecture & Evaluation

Scalability, Fidelity, and Containment in the Potemkin Virtual Honeyfarm

Michael Vrable, Justin Ma, Jay Chen, David Moore, Erik Vandekieft, Alex C. Snoeren, Geoffrey M. Voelker, Stefan Savage

Collaborative Center for Internet Epidemiology and Defenses (CCIED) University of California, San Diego

The Potemkin Virtual Honeyfarm 1 / 20

slide-2
SLIDE 2

Introduction Design Architecture & Evaluation

Background

◮ Large-scale host exploitation a serious problem

◮ Worms, viruses, bots, spyware. . . ◮ Supports an emerging economic criminal enterprise ◮ SPAM, DDoS, phishing, piracy, ID theft. . . ◮ Two weeks ago, one group arrested—controlled 1.5 M hosts!

◮ Quality and sophistication of malware increasing rapidly

The Potemkin Virtual Honeyfarm 2 / 20

slide-3
SLIDE 3

Introduction Design Architecture & Evaluation

Motivation

◮ Intelligence about new threats is critical for defenders ◮ Principal tool is the network honeypot

◮ Monitored system deployed for the purpose of being attacked

◮ Honeyfarm: Collection of honeypots

◮ Provide early warning, accurate inference of global activity,

cover wide range of software

◮ Design issues

◮ Scalability: How many honeypots can be deployed ◮ Fidelity: How accurately systems are emulated ◮ Containment: How well innocent third parties are protected

◮ Challenge: tension between scalability and fidelity

The Potemkin Virtual Honeyfarm 3 / 20

slide-4
SLIDE 4

Introduction Design Architecture & Evaluation

Honeyfarm Scalability/Fidelity Tradeoff

High Scalability High Fidelity

The Potemkin Virtual Honeyfarm 4 / 20

slide-5
SLIDE 5

Introduction Design Architecture & Evaluation

Honeyfarm Scalability/Fidelity Tradeoff

High Scalability High Fidelity VM-based Honeyfarms

(Collapsar, Symantec)

Physical Honeypots

(Honeynet Project)

Execute real code

The Potemkin Virtual Honeyfarm 4 / 20

slide-6
SLIDE 6

Introduction Design Architecture & Evaluation

Honeyfarm Scalability/Fidelity Tradeoff

High Scalability High Fidelity Network Telescopes Lightweight Responders

(iSink, IMS, honeyd)

Millions of addresses VM-based Honeyfarms

(Collapsar, Symantec)

Physical Honeypots

(Honeynet Project)

Execute real code

The Potemkin Virtual Honeyfarm 4 / 20

slide-7
SLIDE 7

Introduction Design Architecture & Evaluation

Honeyfarm Scalability/Fidelity Tradeoff

High Scalability High Fidelity Network Telescopes Lightweight Responders

(iSink, IMS, honeyd)

Millions of addresses VM-based Honeyfarms

(Collapsar, Symantec)

Physical Honeypots

(Honeynet Project)

Execute real code Our Goal: Both

The Potemkin Virtual Honeyfarm 4 / 20

slide-8
SLIDE 8

Introduction Design Architecture & Evaluation Approach Network-Level Multiplexing Host-Level Multiplexing

Approach

◮ Dedicated honeypot systems are overkill ◮ Can provide the illusion of dedicated systems via aggressive

resource multiplexing at network and host levels

The Potemkin Virtual Honeyfarm 5 / 20

slide-9
SLIDE 9

Introduction Design Architecture & Evaluation Approach Network-Level Multiplexing Host-Level Multiplexing

Network-Level Multiplexing

◮ Most addresses don’t receive traffic most of the time

⇒ Apply late binding of IP addresses to honeypots

◮ Most traffic that is received causes no interesting effects

⇒ Allocate honeypots only long enough to identify interesting behavior ⇒ Recycle honeypots as soon as possible

◮ How many honeypots are required?

◮ For a given request rate, depends upon recycling rate The Potemkin Virtual Honeyfarm 6 / 20

slide-10
SLIDE 10

Introduction Design Architecture & Evaluation Approach Network-Level Multiplexing Host-Level Multiplexing

Effectiveness of Network-Level Multiplexing

0.001 0.01 0.1 1 1 10 100 Active Machines / Total Addresses Recycling Time (s)

The Potemkin Virtual Honeyfarm 7 / 20

slide-11
SLIDE 11

Introduction Design Architecture & Evaluation Approach Network-Level Multiplexing Host-Level Multiplexing

Effectiveness of Network-Level Multiplexing

0.001 0.01 0.1 1 1 10 100 Active Machines / Total Addresses Recycling Time (s)

2–3 orders of magnitude improvement!

The Potemkin Virtual Honeyfarm 7 / 20

slide-12
SLIDE 12

Introduction Design Architecture & Evaluation Approach Network-Level Multiplexing Host-Level Multiplexing

Host-Level Multiplexing

◮ CPU utilization in each honeypot quite low (milliseconds to

process traffic)

⇒ Use VMM to multiplex honeypots on a single physical machine

◮ Few memory pages actually modified when handling network

data

⇒ Share unmodified pages among honeypots within a machine

◮ How many virtual machines can we support?

◮ Limited by unique memory required per VM The Potemkin Virtual Honeyfarm 8 / 20

slide-13
SLIDE 13

Introduction Design Architecture & Evaluation Approach Network-Level Multiplexing Host-Level Multiplexing

Effectiveness of Host-Level Multiplexing

0.5 1 1.5 2 2.5 3 3.5 4 20 40 60 80 100 Memory Pages Modified (MB) Time (s) HTTP Telnet Ping

The Potemkin Virtual Honeyfarm 9 / 20

slide-14
SLIDE 14

Introduction Design Architecture & Evaluation Approach Network-Level Multiplexing Host-Level Multiplexing

Effectiveness of Host-Level Multiplexing

0.5 1 1.5 2 2.5 3 3.5 4 20 40 60 80 100 Memory Pages Modified (MB) Time (s) HTTP Telnet Ping

Further 2–3 orders of magnitude improvement

The Potemkin Virtual Honeyfarm 9 / 20

slide-15
SLIDE 15

Introduction Design Architecture & Evaluation Overview Potemkin VMM Containment Challenges

The Potemkin Honeyfarm Architecture

◮ Two components:

◮ Gateway ◮ VMM The Potemkin Virtual Honeyfarm 10 / 20

slide-16
SLIDE 16

Introduction Design Architecture & Evaluation Overview Potemkin VMM Containment Challenges

The Potemkin Honeyfarm Architecture

◮ Two components:

◮ Gateway ◮ VMM

◮ Basic operation:

◮ Packet received by

gateway

The Potemkin Virtual Honeyfarm 10 / 20

slide-17
SLIDE 17

Introduction Design Architecture & Evaluation Overview Potemkin VMM Containment Challenges

The Potemkin Honeyfarm Architecture

◮ Two components:

◮ Gateway ◮ VMM

◮ Basic operation:

◮ Packet received by

gateway

◮ Dispatched to

honeyfarm server

The Potemkin Virtual Honeyfarm 10 / 20

slide-18
SLIDE 18

Introduction Design Architecture & Evaluation Overview Potemkin VMM Containment Challenges

The Potemkin Honeyfarm Architecture

◮ Two components:

◮ Gateway ◮ VMM

◮ Basic operation:

◮ Packet received by

gateway

◮ Dispatched to

honeyfarm server

◮ VM instantiated ◮ Adopts IP

address

The Potemkin Virtual Honeyfarm 10 / 20

slide-19
SLIDE 19

Introduction Design Architecture & Evaluation Overview Potemkin VMM Containment Challenges

Potemkin VMM Requirements

◮ VMs created on

demand

◮ VM creation must

be fast enough to maintain illusion

The Potemkin Virtual Honeyfarm 11 / 20

slide-20
SLIDE 20

Introduction Design Architecture & Evaluation Overview Potemkin VMM Containment Challenges

Potemkin VMM Requirements

◮ VMs created on

demand

◮ VM creation must

be fast enough to maintain illusion

◮ Many VMs created

◮ Must be

resource-efficient

The Potemkin Virtual Honeyfarm 11 / 20

slide-21
SLIDE 21

Introduction Design Architecture & Evaluation Overview Potemkin VMM Containment Challenges

Potemkin VMM Overview

◮ Modified version of Xen 3.0 (pre-release) ◮ Flash cloning

◮ Fork copies from a reference honeypot VM ◮ Reduces VM creation time—no need to boot ◮ Applications all ready to run

◮ Delta virtualization

◮ Copy-on-write sharing (between VMs) ◮ Reduces per-VM state—only stores unique data ◮ Further reduces VM creation time The Potemkin Virtual Honeyfarm 12 / 20

slide-22
SLIDE 22

Introduction Design Architecture & Evaluation Overview Potemkin VMM Containment Challenges

Flash Cloning Performance

Time required to clone a 128 MB honeypot: Control tools overhead 124 ms Low-level clone 11 ms Device setup 149 ms Other management overhead 79 ms Networking setup & overhead 158 ms Total 521 ms 0.5 s already imperceptible to external observers unless looking for delay, but we can do even better

The Potemkin Virtual Honeyfarm 13 / 20

slide-23
SLIDE 23

Introduction Design Architecture & Evaluation Overview Potemkin VMM Containment Challenges

Flash Cloning Performance

Time required to clone a 128 MB honeypot: Control tools overhead 124 ms Low-level clone 11 ms Device setup 149 ms Other management overhead 79 ms Networking setup & overhead 158 ms Total 521 ms 0.5 s already imperceptible to external observers unless looking for delay, but we can do even better

The Potemkin Virtual Honeyfarm 13 / 20

slide-24
SLIDE 24

Introduction Design Architecture & Evaluation Overview Potemkin VMM Containment Challenges

Delta Virtualization Performance

◮ Deployed using 128 MB Linux honeypots ◮ Using servers with 2 GB RAM, have memory available to

support ≈ 1000 VMs per physical host

◮ Currently tested with ≈ 100 VMs per host

◮ Hits artificial resource limit in Xen, but this can be fixed The Potemkin Virtual Honeyfarm 14 / 20

slide-25
SLIDE 25

Introduction Design Architecture & Evaluation Overview Potemkin VMM Containment Challenges

Containment Policies

◮ Must also care about traffic going out ◮ We deliberately run unpatched, insecure software in honeypots ◮ Containment: Should not permit attacks on third parties ◮ As with scalability, there is a tension between containment

and fidelity

◮ Various containment policies we support:

◮ Allow no traffic out ◮ Allow traffic over established connections ◮ Allow traffic back to original host ◮ . . . The Potemkin Virtual Honeyfarm 15 / 20

slide-26
SLIDE 26

Introduction Design Architecture & Evaluation Overview Potemkin VMM Containment Challenges

Containment Implementation in Gateway

◮ Containment policies implemented in network gateway ◮ Tracks mappings between IP addresses, honeypots, and past

connections

◮ Modular implementation in Click ◮ Gateway adds insignificant overhead (≪ 1 ms)

The Potemkin Virtual Honeyfarm 16 / 20

slide-27
SLIDE 27

Introduction Design Architecture & Evaluation Overview Potemkin VMM Containment Challenges

Traffic Reflection

Example gateway policy: Redirect traffic back to honeyfarm

The Potemkin Virtual Honeyfarm 17 / 20

slide-28
SLIDE 28

Introduction Design Architecture & Evaluation Overview Potemkin VMM Containment Challenges

Traffic Reflection

Example gateway policy: Redirect traffic back to honeyfarm

◮ Packets sent out to

third parties. . .

The Potemkin Virtual Honeyfarm 17 / 20

slide-29
SLIDE 29

Introduction Design Architecture & Evaluation Overview Potemkin VMM Containment Challenges

Traffic Reflection

Example gateway policy: Redirect traffic back to honeyfarm

◮ Packets sent out to

third parties. . .

◮ . . . may be redirected

back into honeyfarm Reuses honeypot creation functionality

The Potemkin Virtual Honeyfarm 17 / 20

slide-30
SLIDE 30

Introduction Design Architecture & Evaluation Overview Potemkin VMM Containment Challenges

Challenges

◮ Honeypot detection

◮ If malware detects it is in a honeypot, may act differently ◮ How easy it is to detect virtualization? ◮ VMware detection code used in the wild ◮ Open arms race between honeypot detection and camouflage

◮ Resource exhaustion

◮ Under high load, difficult to maintain accurate illusion ◮ Large-scale outbreak ◮ Honeypot denial-of-service ◮ Challenge is intelligently shedding load The Potemkin Virtual Honeyfarm 18 / 20

slide-31
SLIDE 31

Introduction Design Architecture & Evaluation Overview Potemkin VMM Containment Challenges

Summary

◮ Can achieve both high fidelity and scalability

◮ Sufficient to provide the illusion of scale

◮ Potemkin prototype: 65k addresses → 10 physical hosts

◮ Largest high-fidelity honeypot that we are aware of

◮ Provides important tool for study of and defenses against

malware

The Potemkin Virtual Honeyfarm 19 / 20

slide-32
SLIDE 32

Introduction Design Architecture & Evaluation Overview Potemkin VMM Containment Challenges

For more information: http://www.ccied.org/

The Potemkin Virtual Honeyfarm 20 / 20

slide-33
SLIDE 33

Windows on Xen Camouflage Honeypot Monitoring

Windows on Xen

The Potemkin Virtual Honeyfarm 21 / 20

slide-34
SLIDE 34

Windows on Xen Camouflage Honeypot Monitoring

Camouflage

Malware may detect honeypot environment in various ways:

◮ Detect virtualization

◮ Via incomplete x86 virtualization ◮ Searching for characteristic hardware configurations ◮ More complete virtualization can mitigate these leaks

◮ Detect monitoring tools

◮ Network, VM-instrospection tools harder to detect

◮ Detect network environment

◮ Containment requirement places some limits on camouflage

effectiveness

◮ Network security trends may be in our favor here The Potemkin Virtual Honeyfarm 22 / 20

slide-35
SLIDE 35

Windows on Xen Camouflage Honeypot Monitoring

Honeypot Monitoring

Various means to monitor honeypots for interesting activity

◮ Network-level monitoring: Network intrusion detection

systems, Earlybird-like detectors, . . .

◮ Host-level intrusion detection ◮ Virtual machine introspection

The Potemkin Virtual Honeyfarm 23 / 20