parallel file systems
play

Parallel File Systems John White Lawrence Berkeley National Lab - PowerPoint PPT Presentation

Parallel File Systems John White Lawrence Berkeley National Lab Topics Defining a File System Our Specific Case for File Systems Parallel File Systems A Survey of Current Parallel File Systems Implementation What is a


  1. Parallel File Systems John White Lawrence Berkeley National Lab

  2. Topics • Defining a File System • Our Specific Case for File Systems • Parallel File Systems • A Survey of Current Parallel File Systems • Implementation

  3. What is a File System? • Simply, a method for ensuring ▪ A Unified Access Method to Data ▪ Organization (in a technical sense…) ▪ Data Integrity ▪ Efficient Use of Hardware

  4. The HPC Application (our application) • Large Node Count • High IO Code (small file operations) • High Throughput Code (large files fast) • You Can Never Provide Too Much Capacity

  5. What’s the Problem With Tradition? • NFS/CIFS/AFP/NAS is slow Single point of contact for both data and metadata ▪ Protocol Overhead ▪ File based locking ▪ We want parallelism from the application to disk ▪ • We Need a Single Namespace • We Need Truly Massive Aggregate Throughput (stop thinking MB/s) • Bottlenecks are Inherent to Architecture • Most Importantly:

  6. • Researchers Just Don’t Care ▪ They want their data available everywhere ▪ They hate transferring data (this bears repeating) ▪ Their code wants the data several cycles ago ▪ If they have to learn new IO APIs, they commonly won't use it, period ▪ An increasing number aren’t aware their code is inefficient

  7. Performance in Aggregate: A Specific Case • File System capable of Performance of 5GB/s • Researcher running an analysis of past stock ticker data ▪ 10 independent processes per node, 10+ nodes, sometimes 1000+ processes ▪ Was running into “performance issues” • In Reality, code was hitting 90% of peak performance ▪ 100s of processes choking each other ▪ Efficiency is key

  8. Parallel File Systems • A File System That Provides ▪ Access to Massive Amounts of Data at Large Client Counts ▪ Simultaneous Client Access at Sub-File Levels ▪ Striping at Sub-File Levels ▪ Massive Scalability ▪ A Method to Aggregate Large Numbers of Disks

  9. Popular Parallel File Systems • Lustre Purchased by Intel ▪ Support offerings from Intel, Whamcloud and numerous ▪ vendors Object based ▪ Growing feature list ▪ ∼ Information Lifecycle Management ∼ “Wide Area” mounting support ∼ Data replication and Metadata clustering planned Open source ▪ ∼ Large and growing install base, vibrant community ∼ “Open” compatibility

  10. Popular Parallel File Systems • GPFS IBM, born around 1993 as Tiger Shark multimedia file ▪ system Support direct from vendor ▪ AIX, Linux, some Windows ▪ Ethernet and Infiniband support ▪ Wide Area Support ▪ ILM ▪ Distributed metadata and locking ▪ Matured storage pool support ▪ Replication ▪

  11. Licensing Landscape • GPFS (A Story of a Huge Feature Set at a Huge Cost) ▪ Binary ∼ IBM licensing • Per Core ▪ Site-Wide • Lustre ▪ Open ▪ Paid Licensing available tied to support offerings

  12. Striping Files

  13. SAN – All nodes have access to storage fabric, all LUNs

  14. Direct Connect – A separate storage cluster hosts and exports via common fabric

  15. Berkeley Research Computing • Current Savio Scratch File System ▪ Lustre 2.5 ▪ 210TB of DDN 9900 ∼ ~10GB/s ideal throughput ▪ Accessible on all nodes • Future ▪ Lustre 2.5 or GPFS 4.1 ▪ ~1PB+ Capacity ▪ ~20GB/s throughput ▪ Vendor yet to be determined

  16. Berkeley Research Computing • Access Methods ▪ Available on every node ∼ POSIX ∼ MPIIO ▪ Data Transfer ∼ Globus Online • Ideal for large transfers • Restartable • Tuned for large networks and long distance • Easy to use graphical interface online ∼ SCP/SFTP • Well known • Suitable for quick and dirty transfers

  17. Current Technological Landscape • Tiered Storage (Storage Pools) ▪ When you have multiple storage needs within a single namespace ∼ SSD/FC for for jobs, metadata (Tier0) ∼ SATA for capacity (Tier1) ∼ Tape for long-term/archival (Tier2) • ILM ▪ Basically, perform actions on data per a rule set ∼ Migration to Tape ∼ Fast Tier 0 storage use case ∼ Purge Policies • Replication ▪ Dangers of metadata operations ▪ Long term storage

  18. Further Information Berkeley Research Computing http://research-it.berkeley.edu/brc HPCS At LBNL http://scs.lbl.gov/ Email: jwhite@lbl.gov

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend