Next Generation Cloud Computing: Advanced Services, Architecture, - - PowerPoint PPT Presentation

next generation cloud computing advanced services
SMART_READER_LITE
LIVE PREVIEW

Next Generation Cloud Computing: Advanced Services, Architecture, - - PowerPoint PPT Presentation

Next Generation Cloud Computing: Advanced Services, Architecture, and Technologies Joe Mambretti, Director, (j-mambretti@northwestern.edu) International Center for Advanced Internet Research (www.icair.org) Northwestern University Director,


slide-1
SLIDE 1

Next Generation Cloud Computing: Advanced Services, Architecture, and Technologies

Joe Mambretti, Director, (j-mambretti@northwestern.edu) International Center for Advanced Internet Research (www.icair.org) Northwestern University Director, Metropolitan Research and Education Network (www.mren.org) Partner, StarLight/STAR TAP, PI-OMNINet (www.icair.org/omninet)

Technische Universit Carolo-Wilhelmina zu Braunschweig Braunschweig, July 1-3, 2009

slide-2
SLIDE 2
  • Creation and Early Implementation of Advanced Networking

Technologies - The Next Generation Internet All Optical Networks, Terascale Networks, Networks for Petascale Science

  • Advanced Applications, Middleware, Large-Scale Infrastructure, NG

Optical Networks and Testbeds, Public Policy Studies and Forums Related to NG Networks

  • Three Major Areas of Activity: a) Basic Research b) Design and

Implementation of Prototypes c) Operations of Specialized Communication Facilities (e.g., StarLight)

Accelerating Leading Edge Innovation and Enhanced Global Communications through Advanced Internet Technologies, in Partnership with the Global Community Introduction to iCAIR:

slide-3
SLIDE 3

Invisible Nodes, Elements, Hierarchical, Centrally Controlled, Fairly Static

Traditional Provider Services: Invisible, Static Resources, Centralized Management, Highly Layered Distributed Programmable Resources, Dynamic Services, Visible & Accessible Resources, Integrated As Required, Non-Layered

Limited Services, Functionality, Flexibility Unlimited Services, Functionality, Flexibility

Paradigm Shift – Ubiquitous Services Based on Large Scale Distributed Facility vs Isolated Services Based on Separate Component Resources

slide-4
SLIDE 4

A Next Generation Architecture: Distributed Facility Enabling Many Types Network/Services

Commodity Internet

Environment: VO Environment: International Gaming Fabric Environment: Control Plane TransLight Environment: Real Org Environment: Sensors Environment: Intelligent Power Grid Control Environment: Real Org1 Environment: Large Scale System Control Environment: Global App Environment: Financial Org Environment: Gov Agency Environment: RFIDNet Environment: Bio Org Environment: Lab Environment: Real Org2

SensorNet FinancialNet HPCNet MediaGridNet R&DNet RFIDNet BioNet PrivNet GovNet1 MedNet

slide-5
SLIDE 5

Cloud Context (1)

  • In General, Clouds Are A Means To Support Large Scale Computing

and Data Capabilities for Distributed On-Demand Resources Using Data Networks (WANs)

  • There Are Many Different Types of Clouds
  • Some Are Oriented Toward Services (e.g., Web 2.0 Based)
  • Some Are Oriented Toward Resources

– For Example, On-Demand Computing Instances Using Infrastructure As A Service Techniques (IaaS, Amazon EC2, S3, etc., Eucalyptus)

  • Some provide Large Scale On-Demand Computing Capacity (G

FS/MapReduce/Bigtable, Hadoop, Sector, etc

  • Some Support Public Services Provided By Global Corporations,

Some Support Private Enterprises Through External Resources, Some Support Private Organizations Through Internal Resources – and There Are Many Variations and Hybrids

5

slide-6
SLIDE 6

Cloud Context (2)

  • To Date, Clouds Have Been Successful
  • However, Current Clouds Have Limitations
  • For Example, They Do Not Necessarily Provide

Optimal Performance

  • Also, Current Clouds Are Oriented Toward

Providing Support For Many Billions of Small Data Over Commodity “Best Effort” Routed Networks

  • Current Clouds Do Not Provide Optimal Support for

Large Capacity Data Flows and Extremely Large Amounts of Individual Data Components

  • They Have Not Been Integrated Into Next

Generation Networking Capabilities

  • They Do Not Handle Specialized Data Well – e.g.,

Digital Media

slide-7
SLIDE 7

Cloud Context (3)

  • If Clouds Are Successful, Why Improve Them

Despite Limitations?

  • They Are Successful for Today’s Consumer and

Enterprise Services –

  • However, They Will Not Meet Future Challenges

Using Current Architectures and Technologies

  • Illustration – And Also Motivation For Improving

the State-of-the-Art – Large Scale Science

  • Why Large Scale Science?
  • Large Scale Science Provides a Looking Glass Into

the Future

slide-8
SLIDE 8

A Scientific Perspective

  • Scientific Research Requires The Resolution of

Extremely Complex Problems

  • Scientific Research Requires The Design And

Creation Of Specialized Tools

  • Increasingly, These Tools Are Being Created

Using Digital Technologies

  • Because of the Complexity and Scale of Major

Scientific Problems, Many Areas of Research Encounter Technical Barriers Years Before They Are Recognized By Other Domains

  • Technical Solutions That Are Created Later

Migrate To Wider Communities

  • Can Large Scale Science Use Clouds? Not Until

the Limitations Described Earlier Are Addressed

slide-9
SLIDE 9

Motivation: Data-Intensive Science & Engineering-e-Science Community Resources

ATLAS

Sloan Digital Sky Survey

LHC

ALMA

slide-10
SLIDE 10

Magnetic Fusion Energy

Source: DOE Source: DOE New Sources Of Power

slide-11
SLIDE 11

Spallation Neutron Source (SNS) at ORNL

Source: DOE

slide-12
SLIDE 12

USGS Images 10,000 Times More Data than Landsat7

Shane DeGross, Telesis

USGS

Landsat7 Imagery 100 Foot Resolution Draped on elevation data New USGS Aerial Imagery At 6-inch Resolution

Source: EVL, UIC

slide-13
SLIDE 13

Today’s Aerial Imaging is >500,000 Times More Detailed than Landsat7

30 meter pixels 4 centimeter pixels

Shane DeGross Laurie Cooper, SDSU

SDSU Campus

Source: Eric Frost, SDSU

slide-14
SLIDE 14

iGrid 2005 UCSD

4K Digital Media Ultra High Definition Digital Communications

  • NTT, Japan

NTT’s digital communications using SHD transmits extra-high-quality, digital, full-color, full motion images. 4k pixels horizontal, 2k vertical 4 * HDTV – 24 * DVD .

www.onlab.ntt.co.jp/en/mn/shd

slide-15
SLIDE 15
  • Center

nter for

  • r Computat

putatio ion n and d Te Tech chno nolo logy gy, , Lou

  • uisiana

isiana St State te Unive vers rsity ity (LSU SU), ), USA SA

  • Northwester

rthwestern n Unive versity rsity

  • MCNC, USA

SA

  • NCSA

SA, USA SA

  • Lawrence

awrence Be Berkeley keley Natio iona nal l Lab abor

  • rato

atory, ry, USA SA

  • Masaryk

saryk Unive vers rsity ity/CES ESNET, ET, Czech ch Re Repub public ic

  • Zu

Zuse se Inst stitute itute Be Berlin, in, German many

  • Vr

Vrije je Unive versiteit, rsiteit, NL www.cct.lsu.edu/Visualization/iGrid2005 http://sitola.fi.muni.cz/sitola/igrid/

  • Interactive visualization coupled with

computing resources and data storage archives over optical networks enhance the study of complex problems, such as the modeling of black holes and other sources of gravitational waves.

  • HD video teleconferencing is used to

stream the generated images in real time from Baton Rouge to Brno and other locations

High-Performance Digital Media

For Interactive Remote Visualization (2006)

slide-16
SLIDE 16

OptIPuter JuxtaView Software for Viewing High Resolution BioImages on Tiled Displays

30 Million Pixel Display NCMIR Lab UCSD

Source: David Lee, Jason Leigh, EVL, UIC

slide-17
SLIDE 17

Components Comprising Environment

  • Overall Architecture
  • Compute Nodes
  • Storage Performance
  • Network Architecture, Protocols,

Performance and Technology Note=>Ultra High Performance Networks Can Make Remote Data Appear To Be Local

  • Proof of Concept – Large Scale

Testbed/Prototype

  • Integration With Emerging Technologies, e.g.,

Massive Multicore, FPGAs, Customized Integrated Components, etc.

slide-18
SLIDE 18

Architecture (1)

Storage Compute Data

Super-Computer Model:

  • Expensive
  • IO is a bottleneck

Alternative Model:

  • Inexpensive,
  • Parallel data IO
  • Examples: Hadoop
  • Sphere/Sector

Source: NCDM, UIC

slide-19
SLIDE 19

Architecture (2)

Parallel/Distributed Programming With MPI, etc.:

  • Flexible and powerful.
  • BUT Very Complicated
  • No Data Locality

Sector/Sphere Model:

  • Very Simple to Apply UDF to

All Data in Parallel;

  • Exploits Data Locality
  • Limited to Certain Data

Parallel Applications.

Source: NCDM, UIC

slide-20
SLIDE 20

What is Sector/Sphere?

  • Sector/Sphere = Wide Area Cloud Providing

On-Demand Computing Capacity

  • Sector: Distributed Storage System
  • Sphere: Run-time Middleware Applies User

Defined Functions (UDF) to Sector Datasets.

  • Open Source Software, GPL/LGPL, written in

C++.

  • Initiated 2006, Current Version 1.19
  • http://sector.sf.net

Source: NCDM, UIC

slide-21
SLIDE 21

Sector

  • Sector: Provides Long Term Persistent Storage to

Large Datasets Managed as Distributed Indexed Files.

  • File Segments Are Placed Throughout Distributed

Storage Managed by Sector.

  • Sector Generally Replicates Data To
  • Ensure Longevity,
  • Decrease the Latency When Retrieving It,
  • Provide Opportunities for Parallelism.
  • Sector is Designed to Take Advantage of Wide Area

High Performance Networks When Available.

  • Sector Can Address Issues Of Extremely Large Data Sets,

Including Very Large Scale Science Data Sets

Source: NCDM, UIC

slide-22
SLIDE 22

Sphere

  • Sphere: Designed To Execute User Defined Functions

(UDF) In Parallel Using a Stream Processing Pattern for The Data That Is Managed By Sector

  • UDFs Are Applied To Every Data Record In a Data Set

Managed by Sector

  • Each Data Segment Is Processed Independently

Providing a Natural Parallelism

  • The Sector/Sphere Design Results in Allowing Data

To Be Frequently Processed in Place Without Moving It

  • If Data Must be Moved, It Can Be Transported Over High

Performance Channels With High Performance Protocols

Source: NCDM, UIC

slide-23
SLIDE 23

Comparing Hadoop and Sector

Hadoop Sector Storage Cloud Block-based file system File-based Programming Model MapReduce MapReduce& UDF Protocol TCP UDP-based protocol (UDT) Replication At time of writing Periodically Security Within 6 months Security (HIPAA) Language Java C++

23

slide-24
SLIDE 24

Proof of Concept: Large Scale Testbed/Prototype

  • Theories Must Be Proven in Practice Using Real Facilities
  • Questions: Can This Concept Scale?

Will It Work At Extreme Scales? Can High Performance Be Achieved?

  • Lab Modeling and Simulation Cannot Substitute for

Real Empirical Studies

  • Experimental Research testing Is Required
  • - Using Real World Large Scale Facilities
  • Distributed Environments and Infrastructure
  • With National Science Foundation Funding A Research

Consortium (NCDM, iCAIR, Et Al) Has Created:

  • An International Scale TeraFlow Testbed Using NLR and

the Global Lambda Integrated Facility (GLIF)

  • A National Scale Open Cloud Testbed, Based on the NLR
slide-25
SLIDE 25

Source: NLR

slide-26
SLIDE 26

Teraflow 1 & 2 Testbed

Teraflow Network is Built Over the National Lambda Rail and GLIF

Seoul Asia EU GLIF

slide-27
SLIDE 27

StarLight – “By Researchers For Researchers”

Abbott Hall, Northwestern University’s Chicago downtown campus View from StarLight

StarLight is an experimental

  • ptical infrastructure and

proving ground for network services optimized for high-performance applications GE+2.5+10GE Exchange Soon: Multiple 10GEs Over Optics – World’s “Largest” 10GE Exchange First of a Kind Enabling Interoperability At L1, L2, L3

slide-28
SLIDE 28

iCAIR: Founding Partner of the Global Lambda Integrated Facility Available Advanced Customizable Network Resources

Visualization courtesy of Bob Patterson, NCSA; data compilation by Maxine Brown, UIC.

www.glif.is

GLIF is a consortium of institutions, organizations, consortia and country National Research & Education Networks who voluntarily share optical networking resources and expertise to develop the Global LambdaGrid for the advancement of scientific collaboration and discovery.

Enables the Creation of Distributed Virtual Environments

slide-29
SLIDE 29
slide-30
SLIDE 30
slide-31
SLIDE 31
slide-32
SLIDE 32

Open Cloud Testbed – 2009

6 Locations

  • 8 racks
  • 256 Nodes
  • 1024 Cores
  • 10+ Gb/s

32

MREN CENIC Dragon

  • Hadoop
  • Sector/Sphere
  • Thrift
  • Eucalyptus

C-Wave

slide-33
SLIDE 33

Example: Sorting a TeraByte

  • Data is Split Into Multiple Small Files,

Scattered On All Nodes

  • Stage 1: On Each Node, an SPE Scans

Local Files, Sends Each Record To a “Bucket File” On a Remote Node According To The key, So That All Buckets Are Sorted.

  • Stage 2: On eEach Destination Node, an

SPE Sorts All Data Inside Each Bucket.

Source: NCDM, UIC

slide-34
SLIDE 34

TeraSort Using Sector & Data-Parallel UDFs

  • n Open Cloud Testbed

10-byte 90-byte Key Value 10-bit Bucket-0 Bucket-1 Bucket-1023 0-1023 Stage 1: Hash based on the first 10 bits Bucket-0 Bucket-1 Bucket-1023 Stage 2: Sort each bucket

  • n local node

Binary Record 100 bytes Source: NCDM, UIC

slide-35
SLIDE 35

Performance Results: TeraSort on Open Cloud Testbed

Data Size Sphere Hadoop (3 replicas) Hadoop (1 replica) UIC 300GB 1265 2889 2252 UIC + StarLight 600GB 1361 2896 2617 UIC + StarLight + Calit2 900GB 1430 4341 3069 UIC + StarLight + Calit2 + JHU 1200GB 1526 6675 3702

Run time: seconds Sector v1.16 vs Hadoop 0.17 Source: NCDM, UIC

slide-36
SLIDE 36

Terasort on Open Cloud Testbed

Source: NCDM, UIC

slide-37
SLIDE 37

Testbed Demonstrations With National Science Foundation at the Annual Conference of The American Association for the Advancement of Science February 2009

Using An Optical Fiber Extension from StarLight/GLIF

slide-38
SLIDE 38

NSF Director Former NSF Director NCDM Director

slide-39
SLIDE 39

Future: On-Going Expansion

  • Expansion Using More Resources
  • Additional Enhancements
  • Integration With Emerging Technologies
  • Expansion Across National Fabrics
  • Expansion Across International Fabrics,

Using GLIF, StarLight

  • Additional Communities
slide-40
SLIDE 40
slide-41
SLIDE 41
slide-42
SLIDE 42

For More Information

  • iCAIR: www.icair.org
  • NCDM: www.ncdm.uic.edu
  • Open Cloud Testbed:

www.opencloudconsortium.org

  • Sector: sector.sourceforget.net