DYNES: DYnamic NEtwork System Artur Barczyk California Institute - - PowerPoint PPT Presentation

dynes dynamic network system
SMART_READER_LITE
LIVE PREVIEW

DYNES: DYnamic NEtwork System Artur Barczyk California Institute - - PowerPoint PPT Presentation

DYNES: DYnamic NEtwork System Artur Barczyk California Institute of Technology / US LHCNet TERENA e2e Workshop TERENA e2e Workshop Prague, November 29 th , 2010 1 2 Deployment Plan What is DYNES OUTLINE Status DYNES Overview What is


slide-1
SLIDE 1

DYNES: DYnamic NEtwork System

Artur Barczyk California Institute of Technology / US LHCNet TERENA e2e Workshop TERENA e2e Workshop Prague, November 29th, 2010

1

slide-2
SLIDE 2

OUTLINE

What is DYNES Status Deployment Plan

2

slide-3
SLIDE 3

DYNES Overview

  • What is DYNES?

– A U.S-wide dynamic network “cyber-instrument” spanning ~40 US universities and ~14 Internet2 connectors – Extends Internet2’s dynamic network service “ION” into U.S. regional networks and campuses; Aims to support LHC traffic (also internationally) – Based on the implementation of the Inter-Domain Circuit protocol developed by ESnet and Internet2; Cooperative development also with GEANT, GLIF

  • Who is it?

– Collaborative team: Internet2, Caltech, Univ. of Michigan, Vanderbilt – The LHC experiments, astrophysics community, WLCG, OSG, other VOs – The community of US regional networks and campuses

  • What are the goals?

– Support large, long-distance scientific data flows in the LHC, other programs (e.g. LIGO, Virtual Observatory), & the broader scientific community – Build a distributed virtual instrument at sites of interest to the LHC but available to R&E community generally

3

slide-4
SLIDE 4

DYNES Team

  • Internet2,

Caltech, V d bilt Vanderbilt,

  • Univ. of Michigan

PI E i B d

  • PI: Eric Boyd

(Internet2)

  • Co-PIs:

– Harvey Newman (Caltech) – Paul Sheldon (V d bilt) (Vanderbilt) – Shawn McKee (Univ. of Michigan) Michigan)

4

http://www.internet2.edu/dynes

slide-5
SLIDE 5

The Problem to be Addressed

  • Sustained throughputs at 1-10 Gbps (and some > 10 Gbps) are in

production use today by some Tier2s as well as Tier1s

  • LHC data volumes and transfer rates are expected to expand by an order
  • f magnitude over the next several years

– As higher capacity storage and regional, national and transoceanic 40G d 100 Gb t k li k b il bl d ff d bl 40G and 100 Gbps network links become available and affordable.

  • Network usage on this scale can only be accommodated with planning,

an appropriate architecture, and nationwide and Int’l community i l t b involvement by – The LHC groups at universities and labs – Campuses, regional and state networks connecting to Internet2 – ESnet, US LHCNet, NSF/IRNC, other major networks in US & Europe

  • Network resource allocation and data operations need to be consistent

– DYNES will help provide standard services and low cost equipment – DYNES will help provide standard services and low cost equipment to help meet the needs

5

slide-6
SLIDE 6

Addressing the The Problem with Dynamic Network Circuits t y a c et o C cu ts

  • DYNES will deliver the needed capabilities to the LHC, and to the broader

scientific community at all the campuses served, by coupling to their l i t analysis systems: – Dynamic network circuit provisioning: IDC Controller – Data transport: Low Cost IDC-capable Ethernet Switch; FDT Server for high th h t L t t h d d ( l LHC) throughput, Low cost storage array where needed (also non-LHC) – End-to-end monitoring services

  • DYNES does not fund more bandwidth, but provides access to Internet2’s

dynamic circuit network (“ION”), plus the standard mechanisms, tools and equipment needed – To build circuits with bandwidth guarantees across multiple network domains, th U S d t E i th f t across the U.S. and to Europe in the future

  • In a manageable way, with fair-sharing
  • Will require scheduling services at some stage

T b ild i i h hi h h h bili i d di d – To build a community with high throughput capability, using standardized, common methods

6

slide-7
SLIDE 7

Leveraging on the Internet2 Dynamic Circuit Network y a c C cu t et o

7

slide-8
SLIDE 8

DYNES System Description

  • AIM: extend hybrid & dynamic capabilities to campus & regional networks.

– A DYNES instrument must provide two basic capabilities at the Tier 2S, Tier3s and regional networks: and regional networks:

  • 1. Network resource allocation such as

bandwidth to ensure transfer performance 2 M it i f th t k d d t t f

  • 2. Monitoring of the network and data transfer

performance

  • All networks in the path require the ability

to allocate network resources and monitor to allocate network resources and monitor the transfer. This capability currently exists

  • n backbone networks such as Internet2 and

ESnet but is not widespread at the campus ESnet, but is not widespread at the campus and regional level. – In addition Tier 2 & 3 sites require: 3 Hardware at the end sites capable of making

Two typical transfers that DYNES supports: one Tier2 - Tier3 and

  • 3. Hardware at the end sites capable of making
  • ptimal use of the available network resources

8

supports: one Tier2 - Tier3 and another Tier1-Tier2. The clouds represent the network domains involved in such a transfer.

slide-9
SLIDE 9

DYNES: Regional Network - Instrument Design st u e t es g

  • Regional networks require

1. An Ethernet switch 2. An Inter-domain Controller (IDC)

  • The configuration of the IDC

consists of OSCARS, DRAGON, and perfSONAR. This allows the regional network to provision resources on-demand through i t ti ith th th interaction with the other instruments

  • A regional network does not

require a disk array or FDT server

At the network level, each regional connects the incoming campus connection to the Ethernet switch provided. Optionally, if a regional network already has a qualified switch compatible with the dynamic software that they prefer they

require a disk array or FDT server because they are providing transport for the Tier 2 and Tier 3 data transfers not initiating them

compatible with the dynamic software that they prefer, they may use that instead, or in addition to the provided

  • equipment. The Ethernet switch provides a VLAN dynamically

allocated by OSCARS & DRAGON. The VLAN has quality of service (QoS) parameters set to guarantee the bandwidth requirements of the connection as defined in the VLAN These

data transfers, not initiating them.

9 requirements of the connection as defined in the VLAN. These parameters are determined by the original circuit request from the researcher / application. through this VLAN, the regional provides transit between the campus IDCs connected in the same region or to the global IDC infrastructure.

slide-10
SLIDE 10

DYNES: Tier2 and Tier3 Instrument Design st u e t es g

  • Each DYNES (sub-)instrument

at a Tier2 or Tier3 site consists

  • f the following hardware
  • f the following hardware,

combining low cost & high performance: 1 An Inter-domain Controller (IDC)

  • 1. An Inter-domain Controller (IDC)
  • 2. An Ethernet switch
  • 3. A Fast Data Transfer (FDT)

server Sites with 10GE

  • server. Sites with 10GE

throughput capability will have a dual-port Myricom 10GE network interface in the server.

The Fast Data Transfer (FDT) server connects to the disk array via the SAS controller and runs FDT software developed by Caltech. FDT i h l i h d d h i ll

network interface in the server.

  • 4. An optional attached disk array

with a Serial Attached SCSI (SAS) controller capable of

FDT is an asynchronous multithreaded system that automatically adjusts I/O and network buffers to achieve maximum network

  • utilization. The disk array stores datasets to be transferred among

the sites in some cases. The FDT server serves as an aggregator/ throughput optimizer in this case, feeding smooth flows over the networks directly to the Tier2 or Tier3 clusters The IDC server

( ) p several hundred MBytes/sec to local storage.

10 networks directly to the Tier2 or Tier3 clusters. The IDC server handles the allocation of network resources on the switch, inter- actions with other DYNES instruments related to network pro- visioning, and network performance monitoring. The IDC creates virtual LANs (VLANs) as needed.

slide-11
SLIDE 11

Local Connectivity Options

DYNES offers several connectivity options for the local sites and the RONs. Two examples:

11

Connector and campus incorporating DYNES as part of production infrastructure Campus using 1 of 2 connections to the regional for DYNES and the other for general purpose IP connectivity

slide-12
SLIDE 12

Fast Data Transfer (FDT)

  • DYNES instrument includes a storage element, with FDT as

data transfer application pp

  • FDT is an open source application for efficient data transfers
  • Easy to use: similar syntax with SCP, iperf/netperf

W itt i j d ll j l tf

  • Written in java and runs on all major platforms
  • Single .jar file (~800 KB)
  • Based on an asynchronous, multithreaded system

Based on an asynchronous, multithreaded system

  • Uses the New I/O (NIO) interface and is able to:

– stream continuously a list of files i d d t th d t d d it h h i l d i – use independent threads to read and write on each physical device – transfer data in parallel on multiple TCP streams, when necessary – use appropriate size of buffers for disk IO and networking – resume a file transfer session

12

slide-13
SLIDE 13

FDT Main Features

  • User defined loadable modules for Pre and Post Processing to provide

support for dedicated Mass Storage system, compression, dynamic circuit setup circuit setup, …

  • Pluggable file systems “providers” (e.g. non-POSIX FS: FDT/Hadoop, etc.)
  • Dynamic bandwidth allocation

U i IDC API – Using IDC API – Limiting transfer rate on end-host

  • Different transport strategies:

– blocking (1 thread per channel) – non-blocking (selector + pool of threads)

  • On the fly MD5 checksum on the reader side
  • Configurable number of streams and threads per physical device (useful

for distributed FS)

  • Automatic updates
  • Can be used as network testing tool (/dev/zero → /dev/null memory

transfers, or –nettest flag)

13

slide-14
SLIDE 14

DYNES SC10 Demo

Successful demonstration setup between Internet2, Caltech and Vanderbilt booths, through SCinet

5

FDT disk-disk throughput on 5 Gbps ION channel (1 host/booth)

5 4

s]

FDT used to automatically request circuit 3 2

[Gbps

provisioning between booths 1

14

slide-15
SLIDE 15

DYNES Roadmap and Status

  • ICO and Council notification (August, 2010)

– Briefings to AOAC, RAC, and ICO Briefings to AOAC, RAC, and ICO

  • Community announcement

– Internet2 Letter, and Website (August, 2010) – 2 Community Calls Held (August 5th and August 17th, 2010)

  • Set up external review panel (September, 2010)
  • Call for Proposals (September 2010)

Call for Proposals (September, 2010)

  • Calls with individual regionals and with individual campuses

(plus associated regional network) upon request

  • DYNES BoF (Internet2 Fall MM in Atlanta in Nov.)
  • Site applications deadline December 15th, 2010
  • Selection done/announcements by January 31st 2011
  • Selection done/announcements by January 31st, 2011
  • Distributed virtual Instrument building begins in 2011

15

slide-16
SLIDE 16

THANK YOU!

Artur.Barczyk@cern.ch

16