Service-Oriented Science: Scaling eScience Impact Ian Foster - - PowerPoint PPT Presentation

service oriented science scaling escience impact
SMART_READER_LITE
LIVE PREVIEW

Service-Oriented Science: Scaling eScience Impact Ian Foster - - PowerPoint PPT Presentation

Service-Oriented Science: Scaling eScience Impact Ian Foster Computation Institute Argonne National Lab & University of Chicago Acknowledgements Carl Kesselman, with whom I developed many ideas (& slides) Bill Allcock, Charlie


slide-1
SLIDE 1

Ian Foster

Computation Institute Argonne National Lab & University of Chicago

Service-Oriented Science: Scaling eScience Impact

slide-2
SLIDE 2

2

Acknowledgements

Carl Kesselman, with whom I developed

many ideas (& slides)

Bill Allcock, Charlie Catlett, Kate Keahey,

Jennifer Schopf, Frank Siebenlist, Mike Wilde @ ANL/ UC

Ann Chervenak, Ewa Deelman, Laura

Pearlman @ USC/ ISI

Karl Czajkowski, Steve Tuecke @ Univa Numerous other fine colleagues in NESC,

EGEE, OSG, TeraGrid, etc.

NSF & DOE for research support

slide-3
SLIDE 3

3

Context: System-Level Science

Problems too large &/ or complex to tackle alone …

slide-4
SLIDE 4

4

Two Perspectives

  • n System-Level Science

System-level problems require integration

Of expertise Of data sources (“data deluge”) Of component models Of experimental modalities Of computing systems

Internet enables decom position

“When the network is as fast as the

computer's internal links, the machine disintegrates across the net into a set of special purpose appliances” (George Gilder)

slide-5
SLIDE 5

5

Integration & Decomposition: A Two-Dimensional Problem

Decom pose across network

Clients integrate dynamically

Select & compose services Select “best of breed” providers Publish result as new services

Decouple resource & service providers

Function Resource

Data Archives Analysis tools Discovery tools Users

Fig: S. G. Djorgovski

slide-6
SLIDE 6

6

A Unifying Concept: The Grid

“Resource sharing & coordinated

problem solving in dynamic, multi- institutional virtual organizations”

1.

Enable integration of distributed resources

2.

Using general-purpose protocols & infrastructure

3.

To deliver required quality of service

“The Anatomy of the Grid”, Foster, Kesselman, Tuecke, 2001

slide-7
SLIDE 7

Facilities Computers Storage Networks Services Software People

Implementation System-Level Problem Grid technology Decomposition

  • U. Colorado

Experimental Model

NCSA

Computational Model

COORD. COORD.

UIUC

Experimental Model

slide-8
SLIDE 8

8

Provisioning

Service-Oriented Systems: Applications vs. Infrastructure

Service-oriented Grid

infrastructure

Provision physical

resources to support application workloads Appln Service Appln Service Users Workflows Composition Invocation

Service-oriented applications

Wrap applications as

services

Compose applications

into workflows

“The Many Faces of IT as Service”, ACM Queue, Foster, Tuecke, 2005

slide-9
SLIDE 9

9

Scaling eScience: Forming & Operating Communities

Define membership & roles; enforce laws &

community standards

I.e., policy for service-oriented architecture Addressing dynamic membership & policy

Build, buy, operate, & share infrastructure

Decouple consumer & provider For data, programs, services, computing,

storage, instruments

Address dynamics of community demand

slide-10
SLIDE 10

10

Defining Community: Membership and Laws

Identify VO participants and roles

For people and services

Specify and control actions of members

Empower members delegation Enforce restrictions federate policy

A 1 2 B 1 2 A B 1 10 1 10 1 16 Access granted by community to user Site admission- control policies

Effective Access

Policy of site to community

slide-11
SLIDE 11

11

Evolution of Grid Security & Policy

1) Grid security infrastructure

Public key authentication & delegation Access control lists (“gridmap” files)

Limited set of policies can be expressed

2) Utilities to simplify operational use, e.g.

MyProxy: online credential repository VOMS, ACL/ gridmap management

Broader set of policies, but still ad-hoc

3) General, standards-based framework for authorization & attribute management

slide-12
SLIDE 12

12

Core Security Mechanisms

Attribute Assertions

C asserts that S has attribute A with value V

Authentication and digital signature

Allows signer to assert attributes

Delegation

C asserts that S can perform O on behalf of C

Attribute mapping

{ A1, A2… An} vo1 { A’1, A’2… A’m} vo2

Policy

Entity with attributes A asserted by C may

perform operation O on resource R

slide-13
SLIDE 13

13

Security Services for VO Policy

Attribute Authority (ATA)

Issue signed attribute assertions

(incl. identity, delegation & mapping)

Authorization Authority (AZA)

Decisions based on assertions & policy VO A Service VO ATA VO AZA VO User A VO User B

slide-14
SLIDE 14

14

Security Services for VO Policy

Attribute Authority (ATA)

Issue signed attribute assertions

(incl. identity, delegation & mapping)

Authorization Authority (AZA)

Decisions based on assertions & policy VO A Service VO ATA VO AZA VO User A Delegation Assertion User B can use Service A VO User B Resource Admin Attribute

slide-15
SLIDE 15

15

Security Services for VO Policy

Attribute Authority (ATA)

Issue signed attribute assertions

(incl. identity, delegation & mapping)

Authorization Authority (AZA)

Decisions based on assertions & policy VO A Service VO ATA VO AZA VO User A Delegation Assertion User B can use Service A VO User B Resource Admin Attribute VO Member Attribute VO Member Attribute

slide-16
SLIDE 16

16

Security Services for VO Policy

Attribute Authority (ATA)

Issue signed attribute assertions

(incl. identity, delegation & mapping)

Authorization Authority (AZA)

Decisions based on assertions & policy VO A Service VO ATA VO AZA Mapping ATA VO B Service VO User A Delegation Assertion User B can use Service A VO-A Attr VO-B Attr VO User B Resource Admin Attribute VO Member Attribute VO Member Attribute

slide-17
SLIDE 17

17

Closing the Loop: GT4 Security Toolkit

VO Rights Users Rights’ Compute Center Access Services (running

  • n user’s behalf)

Rights

Local policy

  • n VO identity
  • r attribute

authority CAS or VOMS issuing SAML

  • r X.509 ACs

SSL/ WS-Security with Proxy Certificates

Authz Callout: SAML, XACML

KCA

MyProxy

Shib

slide-18
SLIDE 18

18

Security Needn’t Be Hard: Earth System Grid

Purpose

Access to large data

Policies

Per-collection control Different user classes

Implementation (GT)

Portal-based User

Registration Service

PKI, SAML assertions

Experience

> 2000 users > 100 TB downloaded

PURSE User Registration

Optional review

www.earthsystemgrid.org

See also: GAMA (SDSC), Dorian (OSU)

slide-19
SLIDE 19

19

Scaling eScience: Forming & Operating Communities

Define membership & roles; enforce laws &

community standards

I.e., policy for service-oriented architecture Addressing dynamics of membership & policy

Build, buy, operate, & share infrastructure

Decouple consum er & provider For data, program s, services, com puting,

storage, instrum ents

Address dynam ics of com m unity dem and

slide-20
SLIDE 20

20

Community Services Provider Content Services Capacity

Bootstrapping a VO by Assembling Services

1) Integrate services from other sources

Virtualize external services as VO services

2) Coordinate & compose

Create new services from existing ones

Capacity Provider

“Service-Oriented Science”, Science, 2005

slide-21
SLIDE 21

21

Providing VO Services: (1) Integration from Other Sources

Negotiate service

level agreements

Delegate and deploy

capabilities/ services

Provision to deliver

defined capability

Configure environment Host layered functions Community A Community Z

slide-22
SLIDE 22

22

Virtualizing Existing Services into a VO

Establish service agreement with service

E.g., WS-Agreement

Delegate use to VO user

User A

VO Admin

User B

VO User

Existing Services

slide-23
SLIDE 23

23

Deploying New Services

Policy Client Environment Activity Allocate/provision Configure Initiate activity Monitor activity Control activity Interface Resource provider

WSRF (or WS-Transfer/ WS-Man, etc.), Globus GRAM, Virtual Workspaces

slide-24
SLIDE 24

24

Available in High-Quality Open Source Software …

Data Mgm t Security Com m on Runtim e Execution Mgm t I nfo Services GridFTP

Authentication Authorization

Reliable File Transfer

Data Access & Integration

Grid Resource Allocation & Management

Index

Community Authorization

Data Replication Community Scheduling Framework Delegation Replica Location Trigger Java Runtime C Runtime Python Runtime WebMDS Workspace Management Grid Telecontrol Protocol

Globus Toolkit v4 w w w .globus.org

Credential Mgmt

Globus Toolkit Version 4: Software for Service-Oriented Systems, LNCS 3779, 2-13, 2005

slide-25
SLIDE 25

25

http: / / dev.globus.org

Guidelines (Apache) Infrastructure (CVS, email, bugzilla, Wiki) Projects Include …

slide-26
SLIDE 26

26

Virtual Workspaces (Kate Keahey et al.)

GT4 service for the creation, monitoring, &

management of virtual w orkspaces

High-level workspace description Web Services interfaces for monitoring &

managing

Multiple implementations

Dynamic accounts Xen virtual machines (VMware virtual machines)

Virtual clusters as a higher-level construct

slide-27
SLIDE 27

27

deploy, suspend

How do Grids and VMs Play Together?

Client request VM EPR inspect & manage use existing VM image

Create VM image

VM Factory VM Repository VM Manager create new VM image Resource VM start program

slide-28
SLIDE 28

28

Virtual OSG Clusters

OSG cluster Xen hypervisors TeraGrid cluster

OSG

“Virtual Clusters for Grid Communities,” Zhang et al., CCGrid 2006

slide-29
SLIDE 29

29

Dynamic Service Deployment (Argonne + China Grid)

Interface

Upload-push Upload-pull Deploy Undeploy Reload “HAND: Highly Available Dynamic Deployment Infrastructure for GT4,” Li Qi et al., 2006

slide-30
SLIDE 30

30

Providing VO Services: (2) Coordination & Composition

Take a set of provisioned services …

… & compose to synthesize new behaviors

This is traditional service composition

But must also be concerned with emergent

behaviors, autonomous interactions

See the work of the agent & PlanetLab

communities

“Brain vs. Brawn: Why Grids and Agents Need Each Other," Foster, Kesselman, Jennings, 2004.

slide-31
SLIDE 31

31

Birmingham•

The Globus-Based LIGO Data Grid

Replicating > 1 Terabyte/ day to 8 sites > 40 million replicas so far MTBF = 1 month

LIGO Gravitational Wave Observatory www.globus.org/ solutions

Cardiff

AEI/Golm

slide-32
SLIDE 32

32

Pull “missing” files to a storage system

GridFTP Reliable File Transfer Service GridFTP

Data Replication Service

“Design and Implementation of a Data Replication Service Based on the Lightweight Data Replicator System,” Chervenak et al., 2005

Data Movement

slide-33
SLIDE 33

33

Pull “missing” files to a storage system

GridFTP Local Replica Catalog Replica Location Index Reliable File Transfer Service Local Replica Catalog GridFTP

Data Replication Service

“Design and Implementation of a Data Replication Service Based on the Lightweight Data Replicator System,” Chervenak et al., 2005 Replica Location Index

Data Movement Data Location

slide-34
SLIDE 34

34

Pull “missing” files to a storage system List of required Files

GridFTP Local Replica Catalog Replica Location Index Data Replication Service Reliable File Transfer Service Local Replica Catalog GridFTP

Data Replication Service

“Design and Implementation of a Data Replication Service Based on the Lightweight Data Replicator System,” Chervenak et al., 2005 Replica Location Index

Data Movement Data Location Data Replication

slide-35
SLIDE 35

35

Hypervisor/ OS Deploy hypervisor/ OS

Composing Resources … Composing Services

Physical machine Procure hardware VM Deploy virtual machine

State exposed & access uniformly at all levels Provisioning, management, and monitoring at all levels

JVM Deploy container Deploy service GridFTP

slide-36
SLIDE 36

36

Hypervisor/ OS Deploy hypervisor/ OS

Composing Resources … Composing Services

Physical machine Procure hardware VM VM Deploy virtual machine

State exposed & access uniformly at all levels Provisioning, management, and monitoring at all levels

JVM Deploy container DRS Deploy service GridFTP LRC

VO Services

GridFTP

slide-37
SLIDE 37

37

Decomposition Enables Separation of Concerns & Roles

User Service Provider “Provide access to data D at S1, S2, S3 with performance P” Resource Provider “Provide storage with performance P1, network with P2, … ” D S1 S2 S3 D S1 S2 S3

Replica catalog, User-level multicast, …

D S1 S2 S3

slide-38
SLIDE 38

38

Another Example: Astro Portal Stacking Service

Purpose

On-demand “stacks”

  • f random locations

within ~ 10TB dataset

Challenge

Rapid access to 10-

10K “random” files

Time-varying load

Solution

Dynamic acquisition

  • f compute, storage

+ + + + + + = +

S4

Sloan Data

Web page

  • r Web

Service

slide-39
SLIDE 39

39

Astro Portal Stacking Performance (LAN GPFS)

slide-40
SLIDE 40

40

Summary

Com m unity based science will be the norm

Requires collaborations across sciences—

including computer science

Many different types of com m unities

Differ in coupling, membership, lifetime, size

Must think beyond science stovepipes

Community infrastructure will increasingly

become the scientific observatory

Scaling requires a separation of concerns

Providers of resources, services, content

Small set of fundam ental m echanism s

required to build communities

slide-41
SLIDE 41

41

For More Information

Globus Alliance

www.globus.org

Dev.Globus

dev.globus.org

Open Science Grid

www.opensciencegrid.org

TeraGrid

www.teragrid.org

Background

www.mcs.anl.gov/ ~ foster

2nd Edition www.mkp.com/grid2