Superfacility: How new workflows in the DOE Office of Science are - - PowerPoint PPT Presentation

superfacility
SMART_READER_LITE
LIVE PREVIEW

Superfacility: How new workflows in the DOE Office of Science are - - PowerPoint PPT Presentation

Superfacility: How new workflows in the DOE Office of Science are influencing storage system requirements Katie Antypas Department Head Scientific Computing and Data Services May 3, 2016 - 1 - NERSC is the mission HPC computing center


slide-1
SLIDE 1

Katie Antypas

Department Head Scientific Computing and Data Services

May 3, 2016

Superfacility: How new workflows in the DOE Office of Science are influencing storage system requirements

  • 1 -
slide-2
SLIDE 2

NERSC is the mission HPC computing center for the DOE Office of Science

  • NERSC deploys advanced HPC and data

systems for the broad Office of Science community

  • NERSC staff provide advanced

application and system performance expertise to users

  • Approximately 6000 users and 750

projects

slide-3
SLIDE 3

NERSC has been supporting data intensive science for a long time

Ice Cube Neutrinos Planck Satellite Cosmic Microwave Background Radiation Alice Large Hadron Collider Atlas Large Hadron Collider Dayabay Neutrinos Joint Genome Institute Bioinformatics

  • 3 -
slide-4
SLIDE 4

Historically NERSC has deployed separate Compute Intensive and Data Intensive Systems

Compute Intensive

Data Intensive Carver Genepool PDSF

  • 4 -
slide-5
SLIDE 5

What has changed? Coupling of experiments with large scale simulations

Nyx simulation of Lyman alpha forest Kitt Peak National Observatory’s Mayall 4-meter telescope, planned site of the DESI experiment

  • 5 -

Genomes to watersheds New climate modeling methods, produce new understanding of ice

slide-6
SLIDE 6

data rates and new sensing capabilities

LCLS Light Source

new accumulat

  • r ring

Advanced Lightsource Upgrade Sequencers that fit into the palm of your hand Environmental sensors

  • In the next 5 years, data rates will be approaching

Tb/sec for many instruments

  • Infeasible to put a supercomputer at the site of every

data generator

Next generation electron microscope

slide-7
SLIDE 7
  • 7 -

Gigabytes Data Retention Time

Forever Temp

F E Sino Ring Correct Image- Magick Norm Recon 2 A E B B C C D D D

Initial Data (HDF5 and extracted tif stack)

Ring Correct Recon 1 tif stack

Web Thumbnails and tifs packed as HDF5 for Visit Vis.

Optimizing workflows becomes as important as optimizing computational kernels

Work by: Chris Daley, NERSC Based on workflow diagram format created by David Montoya, LANL

  • This workflow consists of many dependent tasks which read and write files

– Files are either discarded (in yellow layer - bottom) or saved forever (in blue layer - top)

  • Helps us understand how the scientist wants to use storage

Reading an input file Writing a temporary file

Time 

slide-8
SLIDE 8

Experimental Facilities New mathematical models Unified Computing Facilities Network for Big Data Science Fast Implementations

  • n latest

computers Real-time analysis and Data management

Superfacility Vision: A network of connected facilities, software and expertise to enable new modes of discovery

  • 8 -
slide-9
SLIDE 9

Some thoughts on how storage requirements will be influenced by experimental data

  • Seamless data movement and management

from experiment through memory/storage hierarchy will require more coordinated software stacks, data models and metadata

  • The same data will need to be accessed by

different users and groups during a workflow

  • Components of workflows outside a compute

system, (web gateways and databases), will need equal access to data and storage

  • 9 -
slide-10
SLIDE 10

Some thoughts on how storage requirements will be influenced by experimental data

  • Scheduling will need to expand to more than

just compute -- to include storage, bandwidth and experiment allowing guaranteed QoS

  • Analyzing streaming data will require high

bandwidth networking to storage and compute nodes

  • Authentication and identity management across

facilities and storage systems will need to be robust and coordinated

  • 10 -