Wide Area Distributed File Systems
Tevfik Kosar, Ph.D. CSE 710 Seminar
Week 1: January 16, 2013
Wide Area Distributed File Systems Tevfik Kosar, Ph.D. Week 1: - - PowerPoint PPT Presentation
CSE 710 Seminar Wide Area Distributed File Systems Tevfik Kosar, Ph.D. Week 1: January 16, 2013 Data Deluge Big Data in Science Scientific data outpaced Moores Law! Demand for data brings demand for computational power: ATLAS and CMS
Tevfik Kosar, Ph.D. CSE 710 Seminar
Week 1: January 16, 2013
ATLAS and CMS applications alone require more than 100,000 CPUs!
Demand for data brings demand for computational power:
Scientific data outpaced Moore’s Law!
ATLAS: High Energy Physics project
Generates 10 PB data/year --> distributed to and processed by 1000s of researchers at 200 institutions in 50 countries.
many science applications today
across several sites A survey among 106 organizations
primary data center
more sites Science Industry
Phillip B. Gibbons, Data-Intensive Computing Symposium
6
Particle Physics Large Hadron Collider
(15PB)
Human Genomics
(7000PB)
1GB / person 200PB+ captured
http://www.int ttp://www.inte tp://www.intel p://www.intel. ://www.intel.c //www.intel.co
World Wide Web
(10PB)
Wikipedia
400K Articles/ Year
Internet Archive
(1PB+)
Typical Oil Company
(350TB+)
Estimated On-line RAM in Google
(8PB)
Personal Digital Photos
(1000PB+)
Total digital data to be created this year 270,000PB (IDC)
200 of London’s Traffic Cams
(8TB/day)
Walmart Transaction DB
(500TB)
Annual Email Traffic, no spam
(300PB+)
Merck Bio Research DB
(1.5TB/qtr)
One Day of Instant Messaging
(1TB)
Terashake Earthquake Model
(1PB)
MIT Babytalk Speech Experiment
(1.4PB)
UPMC Hospitals Imaging Data
(500TB/yr)
“In the future, U.S. international leadership in science and engineering will increasingly depend upon our ability to leverage this reservoir of scientific data captured in digital form.”
“In the future, U.S. international leadership in science and engineering will increasingly depend upon our ability to leverage this reservoir of scientific data captured in digital form.”
9
TB TB PB PB
10
Carl Kesselman
ISI/USC
They have coined the term “Grid Computing” in 1996!
Ian Foster
Uchicago/Argonne
In 2002, “Grid Computing” selected one of the Top 10 Emerging Technologies that will change the world!
11
– Availability – Standards – Interface – Distributed – Heterogeneous
12
and Grid computing
– Implementation of Distributed computing – A common set of interfaces, tools and APIs – inter-institutional, spanning multiple administrative domains – “The Virtualization of Resources” abstraction of resources
13
According to Foster & Kesselman:
“coordinated resource sharing and problem solving in dynamic, multi-institutional virtual
2001)
14
10,000s processors PetaBytes of storage
15
SETI@home:
radio telescope
generated from the telescope Others: Folding@home, FightAids@home
run screensaver on home PC
– TeraGrid: 40 TeraFlop/src
– TeraGrid: > $100M
16
17
18
– 150,000 machines – Growth rate of 10,000 per month – Largest datacenter: 48,000 machines – 80,000 total running Bing
– 25,000 machines – Split into clusters of 4000
– 40,000 machines – 8 cores/machine
– (Rumored) several hundreds of thousands of machines
19
– Local-area vs Wide area DFS – Fully Distributed FS vs DFS requiring central coordinator
20
nine days (assuming 100% efficiency) -- too optimistic!
S3 cost $100 and took several weeks [Garfinkel 2007]
and sending them to Sandia Lab via Fedex [Feng 2003]
disks, and sending them as packages through UPS or FedEx [Cho et al 2011].
CPU CPU Memory Memory NIC NIC DISK Tnetwork TSmem->network TSdisk->mem Tnetwork -> Network Throughput TSmem->network -> Memory-to-network Throughput on source TSdisk->mem -> Disk-to-memory Throughput on source TDnetwork->mem -> Network-to-memory Throughput on Destination
TDmem->disk -> Memory-to-disk Throughput on
destination DISK TDnetwork->mem TDmem->disk
Data flow Control flow
CPU Memory NIC CPU CPU CPU CPU Memory NIC CPU CPU CPU CPU Memory NIC CPU CPU CPU CPU Memory NIC CPU CPU CPU DISK2 CPU Memory NIC CPU CPU CPU 10G Network 10Gbps 1Gbps 1Gbps 1Gbps 1Gbps Headnode Worker Nodes DISK1 DISK3 DISKn Parallel StreamsParameters to be
protocol tuning disk I/O
CPU
24
efforts ¡in ¡wide-‑area ¡distributed ¡9ile ¡systems ¡on ¡clustered, ¡ grid, ¡and ¡cloud ¡infrastructures.
– NFS (Sun) – AFS (CMU) – Coda (CMU) – xFS (UC Berkeley)
– GPFS (IBM) – Panasas (CMU/Panasas) – PVFS (Clemson/Argonne) – Lustre (Cluster Inc) – Nache (IBM) – Panache (IBM)
25
– OceanStore (UC Berkeley) – Ivy (MIT) – WheelFS (MIT) – Shark (NYU) – Ceph (UC-Santa Cruz) – Giga+ (CMU) – BlueSky (UC-San Diego) – Google FS (Google) – Hadoop DFS (Yahoo!) – Farsite (Microsoft) – zFS (IBM)
26
27
http://www.cse.buffalo.edu/faculty/tkosar/cse710_spring13/ reading_list.htm
– Presenting 1 paper – Reading and contributing the discussion of all the other papers (ask questions, make comments etc)
28
presentation to show their slides!
night before the presentation:
29
Design and implementation of a Distributed Metadata Server for Global Name Space in a Wide-area File System [3-student teams] Design and implementation of a serverless Distributed File System (p2p) for smartphones [3-student teams] Design and implementation of a Cloud-hosted Directory Listing Service for lightweight clients (i.e. web clients, smartphones) [2-student teams] Design and implementation of a Fuse-based POSIX Wide-area File System interface to remote GridFTP servers [2-student teams]
30
31
32
Any Questions? Hmm..