Passive NFS Tracing of Email and Research Workloads Daniel Ellard, - - PowerPoint PPT Presentation

passive nfs tracing of email and research workloads
SMART_READER_LITE
LIVE PREVIEW

Passive NFS Tracing of Email and Research Workloads Daniel Ellard, - - PowerPoint PPT Presentation

Passive NFS Tracing of Email and Research Workloads Daniel Ellard, Jonathan Ledlie, Pia Malkani, Margo Seltzer FAST 2003 - April 1, 2003 Talk Outline Motivation Tracing Methodology Trace Summary New Findings Conclusion


slide-1
SLIDE 1

Passive NFS Tracing of Email and Research Workloads

Daniel Ellard, Jonathan Ledlie, Pia Malkani, Margo Seltzer FAST 2003 - April 1, 2003

slide-2
SLIDE 2

4/1/2003 2 Daniel Ellard - FAST 2003

Talk Outline

  • Motivation
  • Tracing Methodology
  • Trace Summary
  • New Findings
  • Conclusion
slide-3
SLIDE 3

4/1/2003 3 Daniel Ellard - FAST 2003

Motivation - Why Gather Traces?

  • Our research agenda: build file systems that

– Tune themselves for their workloads – Can adapt to diverse workloads

  • Underlying assumptions:

– There is a significant variation between workloads. – There are workload-specific optimizations that we can apply on-the fly.

  • We must test whether these assumptions

hold for contemporary workloads.

slide-4
SLIDE 4

4/1/2003 4 Daniel Ellard - FAST 2003

Why Use Passive NFS Traces?

  • Passive: no changes to the server or client

– Sniff packets from the network – Non-invasive trace methods are necessary for real-world data collection

  • NFS is ubiquitous and important

– Many workloads to trace – Analysis is useful to real users

  • Captures exactly what the server sees

– Matches our research needs

slide-5
SLIDE 5

4/1/2003 5 Daniel Ellard - FAST 2003

Difficulties of Analyzing NFS Traces

  • Underlying file system details are hidden

– Disk activity – File layout

  • The NFS interface is different from a native

file system interface

– No open/close, no seek – Client-side caching can skew the operation mix

  • Some NFS calls and responses are lost
  • NFS calls may arrive out-of-order
slide-6
SLIDE 6

4/1/2003 6 Daniel Ellard - FAST 2003

Our Tracing Software

  • Based on tcpdump and libpcap (a packet-

capture library)

– Captures more information than tcpdump – Handles RPC over TCP and jumbo frames

  • Anonymizes the traces

– Very important for real-world data collection – Tunable to remove/preserve specific information

  • Open source, freely available
slide-7
SLIDE 7

4/1/2003 7 Daniel Ellard - FAST 2003

Overview of the Traced Systems

CAMPUS

  • Central college facility
  • Almost entirely email:

SMTP, POP/IMAP, pine

  • No R&D
  • “Normal” users
  • Digital UNIX
  • 53G of storage (1 of 14

home directory disks) EECS

  • EE/CS facility
  • No email
  • Research: software

projects, experiments

  • Research users
  • Network Appliances filer
  • 450G of storage
slide-8
SLIDE 8

4/1/2003 8 Daniel Ellard - FAST 2003

Summary of Average Daily Activity

0.69 3.01 R/W Ops 4.4 Million 26.7 Million Total Ops 10% (5.1 GB) 65% (119.6 GB) Read Ops 75% - getattr, lookup, access 14% Other 15% (9.1 GB) 21% (44.6 GB) Write Ops

EECS CAMPUS

10/21/2001 - 10/27/2001

slide-9
SLIDE 9

4/1/2003 9 Daniel Ellard - FAST 2003

Workload Characteristics

CAMPUS

  • Data-Oriented
  • 95%+ of reads/writes

are to large mailboxes

  • For newly created files:

– 96%+ are zero-length – Most of the remainder are < 16k – < 1% are “write-only”

EECS

  • Metadata-Oriented
  • Mix of applications, mix
  • f file sizes
  • For newly created files:

– 5% are zero-length – Less than half of the remainder are < 16k – 57% are “write-only”

slide-10
SLIDE 10

4/1/2003 10 Daniel Ellard - FAST 2003

How File Data Blocks Die

  • CAMPUS:

– 99.1% of the blocks die by overwriting – Most blocks live in “immortal” mailboxes

  • EECS:

– 42.4% of the blocks die by overwriting – 51.8% die because their file is deleted

  • Overwriting is common, and a potential
  • pportunity to relocate/reorganize blocks on

disk

slide-11
SLIDE 11

4/1/2003 11 Daniel Ellard - FAST 2003

File Data Block Life Expectancy

  • CAMPUS:

– More than 50% live longer than 15 minutes

  • EECS:

– Less than 50% live longer than 1 second – Of the rest, only 50% live longer than two minutes

  • Most blocks die in the cache on EECS, but on

CAMPUS blocks are more likely to die on disk

slide-12
SLIDE 12

4/1/2003 12 Daniel Ellard - FAST 2003

Talk Outline

  • Motivation
  • Tracing Methodology
  • Trace Summary
  • New Findings
  • Conclusion
slide-13
SLIDE 13

4/1/2003 13 Daniel Ellard - FAST 2003

Variation of Load Over Time

  • EECS: load has detectable patterns
  • CAMPUS: load is quite predictable

– Busiest 9am-6pm and evenings Monday - Friday – Quiet in the late night / early morning

  • Each system has idle times, which could be

used for file system tuning or reorganization.

  • Analyses of workload must include time.
slide-14
SLIDE 14

4/1/2003 14 Daniel Ellard - FAST 2003

The Daily Rhythm of CAMPUS

slide-15
SLIDE 15

4/1/2003 15 Daniel Ellard - FAST 2003

New Finding: File Names Predict File Properties

  • For most files, there is a strong relationship

between the file name and its properties

– Many filenames are chosen by applications – Applications are predictable

  • The filename suffix is useful by itself, but the

entire name is better

  • The relationships between filenames and file

properties vary from system to another

slide-16
SLIDE 16

4/1/2003 16 Daniel Ellard - FAST 2003

Name-Based Hints for CAMPUS

  • Files named “inbox” are large, live forever, are
  • verwritten frequently, and read sequentially.
  • Files with names starting with “inbox.lock” or ending

with the name of the client host are zero-length lock files and live for a fraction of a second.

  • Files with names starting with # are temporary

composer files. They always contain data, but are usually short and are deleted after a few minutes.

  • Dot files are read-only, except .history.
slide-17
SLIDE 17

4/1/2003 17 Daniel Ellard - FAST 2003

Name-Based Hints for EECS

  • On EECS the patterns are harder to see.
  • We wrote a program to detect relationships

between file names and properties, and make predictions based upon them

– Developed this tool on CAMPUS and EECS data – Successful on other later traces as well

  • We can automatically build a model to

accurately predict important attributes of a file based on its name.

slide-18
SLIDE 18

4/1/2003 18 Daniel Ellard - FAST 2003

Accuracy of the Models

Prediction Accuracy 91.8% Length < 16K 99.4% Length = 0

Accuracy of predictions for EECS, for the model trained on 10/22/2001 for the trace from 10/23/2001.

slide-19
SLIDE 19

4/1/2003 19 Daniel Ellard - FAST 2003

Accuracy of the Models

64.8% 87.5% Accuracy w/o names 35.2% 12.5% % of files Prediction Accuracy 91.8% Length < 16K 99.4% Length = 0

Accuracy of predictions for EECS, for the model trained on 10/22/2001 for the trace from 10/23/2001.

slide-20
SLIDE 20

4/1/2003 20 Daniel Ellard - FAST 2003

Accuracy of the Models

64.8% 87.5% Accuracy w/o names 76.6% 94.9% % Error Reduction 35.2% 12.5% % of files Prediction Accuracy 91.8% Length < 16K 99.4% Length = 0

Accuracy of predictions for EECS, for the model trained on 10/22/2001 for the trace from 10/23/2001.

slide-21
SLIDE 21

4/1/2003 21 Daniel Ellard - FAST 2003

New Finding: Out-of-Order Requests

  • On busy networks, requests can be delivered

to the server in a different order than they were generated by the client

– nfsiods can re-order requests – Network effects can also contribute

  • This can break fragile read-ahead heuristics
  • n the server
  • We investigated this for FreeBSD and found

that read-ahead was affected

slide-22
SLIDE 22

4/1/2003 22 Daniel Ellard - FAST 2003

Conclusions

  • Workloads do vary, sometimes enormously
  • New traces are valuable

– We gain new insights from almost every trace

  • We have identified several possible areas for

future research:

– Name-based file system heuristics – Handling out-of-order requests

slide-23
SLIDE 23

4/1/2003 23 Daniel Ellard - FAST 2003

The Last Word

Please contact me if you are interested in exchanging traces or using our tracing software or anonymizer:

http://www.eecs.harvard.edu/sos ellard@eecs.harvard.edu

Another resource:

www.snia.org