Software Defined Networking for big-data science Eric Pouyoul Chin - - PowerPoint PPT Presentation

software defined networking for big data science
SMART_READER_LITE
LIVE PREVIEW

Software Defined Networking for big-data science Eric Pouyoul Chin - - PowerPoint PPT Presentation

Software Defined Networking for big-data science Eric Pouyoul Chin Guok Inder Monga (presenting) TERENA Network Architects meeting, Copenhagen November 21 st , 2012 ESnet: Worlds Leading Science Network ASIA-PACIFIC RUSSIA US R&E


slide-1
SLIDE 1

Software Defined Networking for big-data science

Eric Pouyoul Chin Guok Inder Monga (presenting)

TERENA Network Architects meeting, Copenhagen November 21st, 2012

slide-2
SLIDE 2

Lawrence Berkeley National Laboratory U.S. Department of Energy | Office of Science

ESnet: World’s Leading Science Network

ASIA-PACIFIC (ASGC/Kreonet2/ TWAREN) ASIA-PACIFIC (KAREN/KREONET2/ NUS-GP/ODN/ REANNZ/SINET/ TRANSPAC/TWAREN) AUSTRALIA (AARnet) LATIN AMERICA CLARA/CUDI CANADA (CANARIE) RUSSIA AND CHINA (GLORIAD) US R&E (DREN/Internet2/NLR) US R&E (DREN/Internet2/ NASA) US R&E (NASA/NISN/ USDOI) ASIA-PACIFIC (BNP/HEPNET) ASIA-PACIFIC (ASCC/KAREN/ KREONET2/NUS-GP/ ODN/REANNZ/ SINET/TRANSPAC) AUSTRALIA (AARnet) US R&E (DREN/Internet2/ NISN/NLR) US R&E (Internet2/ NLR) CERN US R&E (DREN/Internet2/ NISN) CANADA (CANARIE) LHCONE CANADA (CANARIE) FRANCE (OpenTransit) RUSSIA AND CHINA (GLORIAD) CERN (USLHCNet) ASIA-PACIFIC (SINET) EUROPE (GÉANT/ NORDUNET) EUROPE (GÉANT) LATIN AMERICA (AMPATH/CLARA) LATIN AMERICA (CLARA/CUDI)

HOUSTON ALBUQUERQUE El PASO SUNNYVALE BOISE SEATTLE KANSAS CITY NASHVILLE WASHINGTON DC NEW YORK BOSTON CHICAGO DENVER SACRAMENTO ATLANTA

PNNL SLAC AMES PPPL BNL ORNL JLAB FNAL ANL LBNL

Major R&E and International peering connections 4x10G IP Hub 100G IP Hubs

slide-3
SLIDE 3

Lawrence Berkeley National Laboratory U.S. Department of Energy | Office of Science

Problems = Opportunities for innovation (1)

Elephant Flows: ‘big-data’ movement for Science, end-to-end

slide-4
SLIDE 4

CERN →T1 mile s kms France 350 565 Italy 570 920 UK 625 1000 Netherlands 625 1000 Germany 700 1185 Spain 850 1400 Nordic 1300 2100 USA – New York 3900 6300 USA - Chicago 4400 7100 Canada – BC 5200 8400 Taiwan 6100 9850

CERN Computer Center

The LHC Optical Private Network (LHCOPN) LHC Tier 1 Data Centers LHC Tier 2 Analysis Centers

Universities/ physics groups Universities/ physics groups Universities/ physics groups Universities/ physics groups Universities/ physics groups Universities/ physics groups Universities/ physics groups Universities/ physics groups Universities/ physics groups Universities/ physics groups Universities/ physics groups Universities/ physics groups Universities/ physics groups Universities/ physics groups Universities/ physics groups Universities/ physics groups Universities/ physics groups

The LHC Open Network Environment (LHCONE)

~50 Gb/s (25Gb/s ATLAS, 25Gb/s CMS)

Level 1 and 2 triggers

O(1-10) meter O(10-100) meters O(1) km 1 PB/s 500-10,000 km

LHC Tier 0 Deep archive and send data to Tier 1 centers

detector Level 3 trigger

Complexity = Opportunity (2): Global Multi-Domain Collaborations like LHC

Source: Bill Johnston

slide-5
SLIDE 5

Lawrence Berkeley National Laboratory U.S. Department of Energy | Office of Science

Science DMZ, perfSONAR, NSI, OSCARS, 100G network is the current reality – still a lot of work to be done The following slides are forward-looking perspective

slide-6
SLIDE 6

Lawrence Berkeley National Laboratory U.S. Department of Energy | Office of Science

Software-Defined Networking

slide-7
SLIDE 7

Lawrence Berkeley National Laboratory U.S. Department of Energy | Office of Science

What is Software-Defined Networking?

(as defined by Scott Shenker, October 2011)

“The ability to master complexity is not the same as the ability to extract simplicity” “Abstractions key to extracting simplicity” “SDN is defined precisely by these three abstractions

  • Distribution, forwarding, configuration “

http://opennetsummit.org/talks/shenker-tue.pdf

slide-8
SLIDE 8

Lawrence Berkeley National Laboratory U.S. Department of Energy | Office of Science

Fundamental Network Abstraction: a end-to-end circuit

A Z Wavelength, PPP, MPLS, L2TP, GRE, NSI-CS… Switching points, store and forward, transformation …

Simple, Point-to-point, Provisonable

slide-9
SLIDE 9

Lawrence Berkeley National Laboratory U.S. Department of Energy | Office of Science

New Network Abstraction: “WAN Virtual Switch” WAN Virtual Switch WAN Virtual Switch

Simple, Multipoint, Programmable Configuration abstraction:

  • Expresses desired behavior
  • Hides implementation on physical infrastructure

It is not only about the concept, but implementation

slide-10
SLIDE 10

Lawrence Berkeley National Laboratory U.S. Department of Energy | Office of Science

Simple Example: One Virtual Switch per Collaboration

ASIA-PACIFIC (ASGC/Kreonet2/ TWAREN) ASIA-PACIFIC (KAREN/KREONET2/ NUS-GP/ODN/ REANNZ/SINET/ TRANSPAC/TWAREN) AUSTRALIA (AARnet) LATIN AMERICA CLARA/CUDI CANADA (CANARIE) RUSSIA AND CHINA (GLORIAD) US R&E (DREN/Internet2/NLR) US R&E (DREN/Internet2/ NASA) US R&E (NASA/NISN/ USDOI) ASIA-PACIFIC (BNP/HEPNET) ASIA-PACIFIC (ASCC/KAREN/ KREONET2/NUS-GP/ ODN/REANNZ/ SINET/TRANSPAC) AUSTRALIA (AARnet) US R&E (DREN/Internet2/ NISN/NLR) US R&E (Internet2/ NLR) CERN US R&E (DREN/Internet2/ NISN) CANADA (CANARIE) LHCONE CANADA (CANARIE) FRANCE (OpenTransit) RUSSIA AND CHINA (GLORIAD) CERN (USLHCNet) ASIA-PACIFIC (SINET) EUROPE (GÉANT/ NORDUNET) EUROPE (GÉANT) LATIN AMERICA (AMPATH/CLARA) LATIN AMERICA (CLARA/CUDI) HOUSTON ALBUQUERQUE El PASO SUNNYVALE BOISE SEATTLE KANSAS CITY NASHVILLE WASHINGTON DC NEW YORK BOSTON CHICAGO DENVER SACRAMENTO ATLANTA PNNL SLAC AMES PPPL BNL ORNL JLAB FNAL ANL LBNL

WAN Virtual Switch WAN Virtual Switch

NERSC ALCF OLCF

slide-11
SLIDE 11

Lawrence Berkeley National Laboratory U.S. Department of Energy | Office of Science

Programmability WAN Virtual Switch WAN Virtual Switch

OpenFlow Controller OF protocol Site Domain WAN Domain Expose ‘flow’ programming interface leveraging standard OF protocol

slide-12
SLIDE 12

Lawrence Berkeley National Laboratory U.S. Department of Energy | Office of Science

“Programmable” by end-sites

Multi-Domain Wide Area Network

WAN Virtual Switch WAN Virtual Switch

Program flows: Science Flow1: Science Flow2: Science Flow3: OF Ctrl.

App 1 App 2

OF Ctrl.

App 1 App 2

slide-13
SLIDE 13

Lawrence Berkeley National Laboratory U.S. Department of Energy | Office of Science

Many collaborations, Many Virtual Switches

Multi-Domain Wide Area Network

WAN Virtual Switch WAN Virtual Switch WAN Virtual Switch WAN Virtual Switch WAN Virtual Switch WAN Virtual Switch WAN Virtual Switch WAN Virtual Switch WAN Virtual Switch WAN Virtual Switch

slide-14
SLIDE 14

Lawrence Berkeley National Laboratory U.S. Department of Energy | Office of Science

SRS Demonstration Physical Topology

Ciena 5410 @Ciena booth NEC IP8800 @ LBL

SRS Brocade @SCinet

@ANL @BNL DTNs DTNs: Data Transfer Nodes OSCARS virtual circuits

slide-15
SLIDE 15

Lawrence Berkeley National Laboratory U.S. Department of Energy | Office of Science

Virtual Switch Implementation: Mapping abstract model to the physical

OF Switch OF Switch OF Switch OF Switch

SRS Virtual Switch SRS Virtual Switch

Virtual Physical

Create Virtual switch:

  • Specify edge OF ports
  • Specify backplane topology and

bandwidth

  • Policy constraints like flowspace
  • Store the switch into a topology

service

A B C D A B C D

slide-16
SLIDE 16

Lawrence Berkeley National Laboratory U.S. Department of Energy | Office of Science

WAN Virtual Switch: Deploying it as a service

Wide Area Network

Network Flowvisor OSCARS

Virtual SW Controller

Virtual Switch Software stack

OSCARS Client

Virtual Switch Application OpenFlow API Customer Flowvisor

OF Switch OF Switch OF Switch OF Switch OF OF OSCARS API OF OF

End-site OF controller

App 1 App 2

Infrastructure Software, Slicing and provisioning

Policy/Isolation of customer OF control

slide-17
SLIDE 17

Lawrence Berkeley National Laboratory U.S. Department of Energy | Office of Science

Example of ping across WAN virtual switch

OF Switch

  • 1. Ping H2

H1

OF Switch OF Switch

SRS Virtual Switch SRS Virtual Switch

  • 2. Packet_in
  • 3. Need to ARP
  • 4. ARP
  • 4. ARP
  • 5. ARP response
  • 6. virtual-mac-addresses learned and mapped to real flow
  • 7. flow_mod
  • 8. Ping H2

H2

  • 8. Ping H2
slide-18
SLIDE 18

Lawrence Berkeley National Laboratory U.S. Department of Energy | Office of Science

What does this mean for networking?

Customer/User Control Plane Policy and Isolation End-to-End Dataplane Legacy and OpenFlow control plane User SDN Controller OpenFlow Programmable service provisioning plane

Network Interface Network Service Interface

Multi-domain

Network Interface Network Service Interface

WAN Virtual Switch

  • Creation of a programmable network provisioning layer
  • Sits on top of the “network OS”
slide-19
SLIDE 19

Lawrence Berkeley National Laboratory U.S. Department of Energy | Office of Science

Summary

  • Powerful network abstraction
  • Files / Storage
  • Benefits
  • Simplicity for the end-site
  • Works with off-the-shelf, open-source controller
  • Topology simplification
  • Generic code for the network provider
  • Virtual switch can be layered over optical, routed or switched network

elements

  • OpenFlow support needed on edge devices only, core stays same
  • Programmability for applications
  • Allows end-sites to innovate and use the WAN effectively
slide-20
SLIDE 20

Lawrence Berkeley National Laboratory U.S. Department of Energy | Office of Science

Future Work

Harden the architecture and software implementation

  • Move from experiment to test service

Verify scaling of the model

  • Using virtual machines, other emulation environments

Automation and Intelligent provisioning

  • Work over multi-domain
  • Wizards for provisioning
  • Dynamic switch backplane

Create recurring abstractions

  • Virtual switch in campus
  • How do we deal with a “network” of virtual switches
slide-21
SLIDE 21

Lawrence Berkeley National Laboratory U.S. Department of Energy | Office of Science

Acknowledgements

Many folks at ESnet who helped with the deployment and planning

  • Sanjay Parab (CMU), Brian Tierney, John Christman, Mark

Redman, Patrick Dorn among other ESnet NESG/OCS folks Ciena Collaborators:

  • Rodney Wilson, Marc Lyonnais, Joshua Foster, Bill Webb

SRS Team

  • Andrew Lee, Srini Seetharaman

DOE ASCR research funding that has made this work possible

slide-22
SLIDE 22

Lawrence Berkeley National Laboratory U.S. Department of Energy | Office of Science

Thank you!

Questions – please contact imonga at es.net www.es.net

slide-23
SLIDE 23

Lawrence Berkeley National Laboratory U.S. Department of Energy | Office of Science

Computer virtualization