NetMicroscope: Passive Measurements of Residential Internet - - PowerPoint PPT Presentation

netmicroscope passive measurements of residential
SMART_READER_LITE
LIVE PREVIEW

NetMicroscope: Passive Measurements of Residential Internet - - PowerPoint PPT Presentation

NetMicroscope: Passive Measurements of Residential Internet Performance Renata Teixeira with Francesco Bronzino, Sara Ayoubi, Israel Salinas (Inria) Paul Schmitt, Guilherme Martins,Joon Kim, Nick Feamster (Princeton) Who cares about


slide-1
SLIDE 1

NetMicroscope: Passive Measurements of Residential Internet Performance

Renata Teixeira with Francesco Bronzino, Sara Ayoubi, Israel Salinas (Inria) Paul Schmitt, Guilherme Martins,Joon Kim, Nick Feamster (Princeton)

slide-2
SLIDE 2

Who cares about residential Internet performance?

§ Home users § ISPs, content providers

1

§ Regulators, policymakers

slide-3
SLIDE 3

Current approach: Active measurements

2

Access ISP Monitoring Server

slide-4
SLIDE 4

Active measurements are reaching their limits

§Access link may not be the bottlenecks §“Filling up” path is disruptive §Measured paths != application paths §Per-application active measurements != user

experience

3

slide-5
SLIDE 5

NetMicroscope

§Measure traffic, infer application performance

– Passive measurements to infer application quality – Targeted active probes to pinpoint bottlenecks

4

slide-6
SLIDE 6

Challenges

§Infer application quality from network traffic

– Applications have different communication patterns – Application traffic is often encrypted

§Passive measurements at increasing line rates §Distinguish performance per network segments

5

slide-7
SLIDE 7

6

slide-8
SLIDE 8

Use case: IP video

7

slide-9
SLIDE 9

Video delivery is complex

8

Home Network ISP Local Caches IXP Interconnect Caches Service Servers

How to monitor video quality for encrypted video traffic?

slide-10
SLIDE 10

Challenges of video quality inference

§Identify video streams within network traffic §Online monitoring at increasing line rates §Large diversity of video streaming services

9

slide-11
SLIDE 11

Our approach

§Identification of video streams

– DNS request/response

§Inference of video quality

– Rely on statistical learning – Can we rely only on lightweight features? – Do models generalize across video services?

§Deployed in home networks

– Between modem and WiFi router – Implemented for low-cost devices

  • Raspberry Pi, Odroid

10

slide-12
SLIDE 12

Statistical learning to infer video quality

§Inference goal: Video quality metrics

– Startup delay – Video resolution – Resolution changes – Rebuffering

§Training data with ground truth from browser

– Services: Netflix, Youtube, Twitch, Amazon Prime – Controlled and in-home experiments

  • Over 11K video sessions

11

slide-13
SLIDE 13

Input: Encrypted video traffic

12

Network layer Transport layer Application layer throughput up/down #flags up/down

  • seg. sizes (all previous, last-

10, cumulative) throughput down diff rcv window size up/down

  • seg. request interarrivals

pkt count up/down idle time up/down

  • seg. completions interarrivals

byte count up/down goodput up/down #pending requests pkt interarrivals up/down bytes per pkt up/down #downloaded seg. #parallel flows round trip time #requested seg. bytes in flight up/down #retransmissions up/down #out of order pks up/down

slide-14
SLIDE 14

Modeling approach

§Startup delay

– Random forest regressor

§Video resolution

– Random forest multi-class classifier

  • Classes: 240p, 360p, 480p, 720p, and 1080p

§Resolution changes

– Random forest binary classifier

§Rebuffering

– Random forest binary classifier

13

slide-15
SLIDE 15

CAN WE RELY ONLY ON LIGHTWEIGHT FEATURES?

14

slide-16
SLIDE 16

Feature importance: Video resolution

15

Most important features based on video segment size and interarrival times

slide-17
SLIDE 17

DO MODELS GENERALIZE ACROSS VIDEO SERVICES?

16

slide-18
SLIDE 18

General vs. specific models for video resolution

17

training & testing Netflix training & testing Youtube training All & testing Netflix training All & testing Youtube

slide-19
SLIDE 19

Deployment

§Instrumented homes

– ~10 in Paris – ~50 in the US

18

5 10 15 20 25 0 < x < 20 20 <= x < 100 100 <= x < 200 200 <= x <= 1000

Speed (mbps) Count

(0,20) [20,100) [100,200) [200,1000]

slide-20
SLIDE 20

Preliminary lessons

§Identification of video sessions

– Auto-play merges sessions – DNS method fails for some devices

§Inference of video quality

– Harder to model rebuffering and resolution switches – Resolution model needs adjustment for more

diverse set of devices

19

slide-21
SLIDE 21

20

slide-22
SLIDE 22

Tracing paths of application flows

21

video flow traceroute flow

§Problem

– Traceroute may not capture application paths

slide-23
SLIDE 23

Service traceroute

§Basics

– Listen to application traffic – Embeds traceroute probes within application flow

§New features

– Signature DB to identify flows of given applications – Support for UDP – Support to trace multiple concurrent flows

22

“Service Traceroute: Tracing Paths of Application Flows”.

  • I. Morandi et al., to appear in PAM 19
slide-24
SLIDE 24

Looking ahead

§How does speed relate to application quality? §How to generalize quality inference to other

applications?

§How to preserve privacy? §How to regulate Internet access using

application quality inference?

23