Tutorial 3 Tutorial 3 Sherbrooke Lundi 12 Mai 2003 Tutorial 3 An - - PowerPoint PPT Presentation

tutorial 3 tutorial 3
SMART_READER_LITE
LIVE PREVIEW

Tutorial 3 Tutorial 3 Sherbrooke Lundi 12 Mai 2003 Tutorial 3 An - - PowerPoint PPT Presentation

The 17th Annual International Symposium on The 17th Annual International Symposium on High Performance Computing Systems and High Performance Computing Systems and Applications Applications The First Annual The First Annual OSCAR OSCAR


slide-1
SLIDE 1

Tutorial 3 Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

The 17th Annual International Symposium on The 17th Annual International Symposium on High Performance Computing Systems and High Performance Computing Systems and Applications Applications The First Annual The First Annual OSCAR OSCAR Symposium Symposium

slide-2
SLIDE 2

Tutorial 3

An Evaluation of An Evaluation of Globus Globus and Legion and Legion Software Environments Software Environments

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

slide-3
SLIDE 3

Tutorial 3

  • Prof. Mario Dantas

Department of I nformatics and Statistics Federal University of Santa Catarina (UFSC) Florianopolis – Brazil E-mail : mario@inf.ufsc.br http://www.inf.ufsc.br/~mario

Sherbrooke – Lundi 12 Mai 2003

slide-4
SLIDE 4

Tutorial 3

PART I

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of An Evaluation of Globus Globus and Legion and Legion Software Environments Software Environments

slide-5
SLIDE 5

Tutorial 3

An Evaluation of Globus and Legion Software Environments

Dans cet article nous présentons une étude comparative descaractéristiques d'implémentation de deux environnements logiciels bienconnus dans le monde du calcul distribué sur une "Grille" (GRI Dcomputing). Nous évaluons la performance de chacun de ces environnementspendant l'exécution en parallèle de tâches MPI distribuées. Uneintroduction des concepts entourant les calculs distribués sur une"Grille" est présentée, suivie de l'étude comparative des deuxenvironnements logiciels, Globus et Legion, ces derniers étant les plusavancés dans le domaine. Nos résultats expérimentaux montrent quel'utilisation de la "Grille" peux s'avérer intéressante pour l'exécutiond'applications parallèles MPI avec une certaine amélioration desperformances.

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

slide-6
SLIDE 6

Tutorial 3

An Evaluation of Globus and Legion Software Environments

I n this article we present a case study comparison of the I mplementation characteristics of two software environm which are well known in grid computing configurations. W evaluate the performance of these environments during t execution of parallel distributed MPI tasks.Therefore, firs consider some concepts of the grid paradigm and then w present a comparison between the two software environ Our case study is based on the Globus and Legion environments, because these two research projects are in more developed stage when compared to other research

  • initiatives. Our experimental results indicate that the grid

computing approach can be interesting to execute paralle distributed MPI applications with a performance improve

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

slide-7
SLIDE 7

Tutorial 3

Agenda

Sherbrooke – Lundi 12 Mai 2003

  • Computing Paradigms and Applications
  • Users
  • Grid Architecture
  • Grid Computing Environments
  • Experimental Results
  • Conclusions and Future Work

Tutorial 3 Part I

slide-8
SLIDE 8

Tutorial 3

Agenda

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

  • Computing Paradigms and Applications

Computing Paradigms and Applications

  • Users
  • Grid Architecture
  • Grid Computing Environments
  • Experimental Results
  • Conclusions and Future Work
slide-9
SLIDE 9

Tutorial 3 What is a Grid Computing ?

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

slide-10
SLIDE 10

Tutorial 3

What is a Grid Computing ?

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

A Grid is a computational high-performance environment which is characterized by resource sharing providing services for organizations geographically distributed.

slide-11
SLIDE 11

Tutorial 3

What is a Grid Computing ?

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

A Grid can also be view under the following physical aspects:

  • Better utilization of bandwidth
  • The use of a great computational power
  • Fast access to data, software and remote facilities

with QoS.

  • Better utilization of remote CPUs, memories and

disk spaces.

slide-12
SLIDE 12

Tutorial 3

What is a Grid Computing ?

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

A Grid is parallel and distributed computational system that enables the sharing, selection and aggregation of geographically distributed autonomous resources dynamically at runtime depending on their availability, capability, performance, cost, and providing users´ applications with their requirements

  • f quality-of-service.
slide-13
SLIDE 13

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

We can also say : Grids target to exploit synergies that result from cooperation-ability to share and aggregate distributed computational capabilities and deliver them as service services s. What is a Grid Computing ?

slide-14
SLIDE 14

Tutorial 3

What is the difference between Cluster and Grid Computing ?

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

slide-15
SLIDE 15

Tutorial 3

What is the difference between Cluster and Grid Computing ?

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

A important difference between clusters and grids is mainly based in the way resources are managed. In the clusters, the resource allocation is performed by a centralized resource manager and all nodes cooperatively work together as a single unified resource. Inside the Grids, each node has its own resource manager and do not target for providing a single system view.

slide-16
SLIDE 16

Tutorial 3

Computing Paradigms and Applications Computing Paradigms and Applications

Sherbrooke – Lundi 12 Mai 2003

The experimental research with the I-WAY, first large scale Grid effort, bring to us that there were five classes of applications using a specific computing paradigm.

Tutorial 3 Part I

slide-17
SLIDE 17

Tutorial 3

Computing Paradigms and Applications Computing Paradigms and Applications

Sherbrooke – Lundi 12 Mai 2003

Computing paradigms Computing paradigms and applications applications can be classify as following :

  • Distributed Supercomputing
  • High-Throughput Computing
  • On-Demand Computing
  • Computing for Large Amount of Data
  • Collaborative Computing

Tutorial 3 Part I

slide-18
SLIDE 18

Tutorial 3

Computing Paradigms and Applications Computing Paradigms and Applications

Sherbrooke – Lundi 12 Mai 2003

  • 1. Distributed Supercomputing

Tutorial 3 Part I

Applications that use this approach can be characterized by the fact that it is not possible to solve these applications in a single computational system. The aggregation environment which we are referring to can be represented by all the supercomputers of a specific country or all the workstation inside of an

  • rganization.
slide-19
SLIDE 19

Tutorial 3

Distributed Supercomputing

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

Examples of applications using the distributed supercomputing approach are :

  • Distributed Interactive Simulation (DIS) : this is a

simulation technique used to model the behaviour and movement of hundred (or thousand) of entities which are usually employed for military planning and teaching.

slide-20
SLIDE 20

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

  • Simulation of complex models such as those in

weather forecast and cosmology. Distributed Supercomputing

slide-21
SLIDE 21

Tutorial 3

Computing Paradigms and Applications Computing Paradigms and Applications

Sherbrooke – Lundi 12 Mai 2003

  • 2. High-Throughput Computing

Tutorial 3 Part I

The main objective of this approach it solve the problem

  • f applications that require a transfer of a large amount
  • f data.

The computational environment is used for scheduling a large number of loosely couple tasks and enhance the utilization of machines with a low workload.

slide-22
SLIDE 22

Tutorial 3

High-Throughput Computing

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

Classical examples for high-throughput computing are :

  • Condor High-Throughput – this software environment

from the University of Wisconsin is used to manage pools

  • f hundreds workstations in the university and other labs

around the world.

slide-23
SLIDE 23

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

  • The Platform Computing software - used by AMD

during the projects of K6 e K7 processors. It is reported that the company has used all the desktops which were not in use by the engineers in a specific period of time. High-Throughput Computing

slide-24
SLIDE 24

Tutorial 3

Computing Paradigms and Applications Computing Paradigms and Applications

Sherbrooke – Lundi 12 Mai 2003

  • 3. On-Demand Computing

Tutorial 3 Part I

This class of applications usually can be characterized by the use of resources that can not be used in the local site, because it is not available. The resources can be computing, data streams, software, archives and for examples experimental results.

slide-25
SLIDE 25

Tutorial 3

Computing Paradigms and Applications Computing Paradigms and Applications

Sherbrooke – Lundi 12 Mai 2003

  • 3. On-Demand Computing

Tutorial 3 Part I

Difference between this approach and distributed Supercomputing is related to the cost of performance then the complete performance behaviour.

slide-26
SLIDE 26

Tutorial 3

Computing Paradigms and Applications Computing Paradigms and Applications

Sherbrooke – Lundi 12 Mai 2003

  • 4. Computing for Large Amount of Data

Tutorial 3 Part I

This class of application and computing paradigm covers the requirement for processing large amount

  • f data stored in a geographic distributed fashion.

Examples are large databases and digital libraries that are available for access in a distributed way. The Digital Sky Survey and modern weather forecast Systems are applications examples.

slide-27
SLIDE 27

Tutorial 3

Computing Paradigms and Applications Computing Paradigms and Applications

Sherbrooke – Lundi 12 Mai 2003

  • 5. Collaborative Computing

Tutorial 3 Part I

Examples for this class are those which are oriented to the improvement the relation between humans. Many collaborative applications allow the share use

  • f computational resources.
slide-28
SLIDE 28

Tutorial 3

Computing Paradigms and Applications Computing Paradigms and Applications

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

NICE is a collaborative learning environment for young children (approximately 6-8 years

  • f age). The environment depicts a virtual

island in which the children can tend a virtual garden and learn about environmental concepts.

slide-29
SLIDE 29

Tutorial 3

Computing Paradigms and Applications Computing Paradigms and Applications

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

Cave5D Cave5D

slide-30
SLIDE 30

Tutorial 3

Agenda

Sherbrooke – Lundi 12 Mai 2003

  • Computing Paradigms and Applications
  • Users

Users

  • Grid Architecture
  • Grid Computing Environments
  • Experimental Results
  • Conclusions and Future Work

Tutorial 3 Part I

slide-31
SLIDE 31

Tutorial 3

Users Users

Sherbrooke – Lundi 12 Mai 2003

Another approach used to understand what is a Grid, is to understand who is going to use. A Grid is above of the mechanisms of resource sharing therefore we can image two questions :

Tutorial 3 Part I

A - Which kind of entity is going to invest in the infrastructure for a Grid ? B - Which kind of resources each community of the entity will be share ?

slide-32
SLIDE 32

Tutorial 3

Users Users

Sherbrooke – Lundi 12 Mai 2003

Answers for the two questions should be based on costs and benefits for sharing resources. Therefore it is usually presented in the academic and commercial reports efforts for the following groups of grid environments :

  • National Grid
  • Private Grid
  • Virtual Grid
  • Public Grid

Tutorial 3 Part I

slide-33
SLIDE 33

Tutorial 3

Users Users

Sherbrooke – Lundi 12 Mai 2003

  • National Grid – the target of this group is to be

a strategic computational resource and serve as a bridge between national sharing facilities.

  • Private Grid – the heath community it is an

example of private grid organization. This group, was identified to benefit from grid configurations because of the strategic utilization of computational power.

Tutorial 3 Part I

slide-34
SLIDE 34
slide-35
SLIDE 35

Network for EarthquakeEngineering Simulation

♦ NEESgrid: national

infrastructure to couple earthquake engineers with experimental facilities, databases, computers, & each

  • ther

♦ On-demand access to

experiments, data streams, computing, archives, collaboration

NEESgrid: Argonne, Michigan, NCSA, UIUC, USC

slide-36
SLIDE 36

Tutorial 3

Users Users

Sherbrooke – Lundi 12 Mai 2003

  • Virtual Grid – this community is formed by

researches and scientists which require the use of expensive equipments and a great computational power.

  • Public Grid – this group is basically characterized

by those which the main activity includes services using a great quantity of computational power.

Tutorial 3 Part I

slide-37
SLIDE 37

DOE X-ray grand challenge: ANL, USC/ ISI, NIST, U.Chicago tomographic reconstruction real-time collection wide-area dissemination desktop & VR clients with shared controls

Advanced Photon Source

Online Access to Scientific Instruments

archival storage

slide-38
SLIDE 38

Data Grids for High Energy Physics

Tier2 Centre ~1 TIPS Online System Offline Processor Farm ~20 TIPS CERN Computer Centre FermiLab ~4 TIPS France Regional Centre Italy Regional Centre Germany Regional Centre Institute Institute Institute Institute ~0.25TIPS Physicist workstations ~100 MBytes/sec ~100 MBytes/sec ~622 Mbits/sec ~1 MBytes/sec

There is a “bunch crossing” every 25 nsecs. There are 100 “triggers” per second Each triggered event is ~1 MByte in size Physicists work on analysis “channels”. Each institute will have ~10 physicists working on one or more channels; data for these channels should be cached by the institute server

Physics data cache

~PBytes/sec

~622 Mbits/sec

  • r Air Freight (deprecated)

Tier2 Centre ~1 TIPS Tier2 Centre ~1 TIPS Tier2 Centre ~1 TIPS Caltech ~1 TIPS ~622 Mbits/sec

Tier 0 Tier 0 Tier 1 Tier 1 Tier 2 Tier 2 Tier 4 Tier 4

1 TIPS is approximately 25,000 SpecInt95 equivalents

Image courtesy Harvey Newman, Caltech

slide-39
SLIDE 39

Tutorial 3

Agenda

Sherbrooke – Lundi 12 Mai 2003

  • Computing Paradigms and Applications
  • Users
  • Grid Architecture

Grid Architecture

  • Grid Computing Environments
  • Experimental Results
  • Conclusions and Future Work

Tutorial 3 Part I

slide-40
SLIDE 40

Tutorial 3

Grid Architecture Grid Architecture

Sherbrooke – Lundi 12 Mai 2003

Before we start to study the Grid architecture it is interesting to know about Virtual Organizations (VO). Virtual organizations are the entities that share resources

  • f the Grid under a specific policy .

Examples of VO are :

  • Providers of applications, data storage and

computational power.

  • Research organizations
  • Universities

Tutorial 3 Part I

slide-41
SLIDE 41

Tutorial 3

Virtual Organizations Virtual Organizations

Sherbrooke – Lundi 12 Mai 2003

Virtual Organizations are different from each other considering the following parameters :

  • Main objective
  • Geographic extension
  • Size (or physical dimensions)
  • Time to use the facilities
  • Structure
  • Community

Tutorial 3 Part I

slide-42
SLIDE 42

Tutorial 3

Grid Architecture Grid Architecture

Sherbrooke – Lundi 12 Mai 2003

Similar to the experience with Internet, researches involved with the Grid established an architecture aiming the interoperability between VOs.

Tutorial 3 Part I

slide-43
SLIDE 43

Tutorial 3

Grid Architecture Grid Architecture

Sherbrooke – Lundi 12 Mai 2003

Aspects such as :

  • authentication,
  • authorization,
  • mechanism of message passing,
  • resource sharing,
  • scheduling and
  • load balancing of tasks

are some of issues which a Grid architecture should provide.

Tutorial 3 Part I

slide-44
SLIDE 44

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

A standard Grid architecture was proposed as : A standard Grid architecture was proposed as :

slide-45
SLIDE 45

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Application Application Collective Collective Resource Connectivity Connectivity Fabric Fabric

Five Layers Grid Architecture Five Layers Grid Architecture

slide-46
SLIDE 46

Tutorial 3

Grid Architecture Grid Architecture - LAYERS

  • LAYERS

Sherbrooke – Lundi 12 Mai 2003

Fabric – Fabric – Components of this layer implement local operations which occurs in each resource mainly because of the sharing provided by the above layers.

Tutorial 3 Part I

Application Application

Collective Collective

Resource

Con Conn nectivity ectivity

Fabric Fabric

slide-47
SLIDE 47

Tutorial 3

Grid Architecture Grid Architecture - LAYERS

  • LAYERS

Sherbrooke – Lundi 12 Mai 2003

Fabric – Fabric – Mechanisms are necessary to obtain information about the structure, state and available resources. On the other hand, it is also important techniques to management the QoS (Quality of Service) for each query.

Tutorial 3 Part I

slide-48
SLIDE 48

Tutorial 3

Grid Architecture Grid Architecture - LAYERS

  • LAYERS

Sherbrooke – Lundi 12 Mai 2003

Connectivity Connectivity In this layer exists the definition of the basic protocols necessary for communication and authentication for a specific transaction of the Grid. The communication protocols allow the data exchange between the Fabric layers. This service includes the transport, routing and name services.

Tutorial 3 Part I

Application Application

Collective Collective

Resource

Con Conn nectivity ectivity

Fabric Fabric

slide-49
SLIDE 49

Tutorial 3

Grid Architecture Grid Architecture - LAYERS

  • LAYERS

Sherbrooke – Lundi 12 Mai 2003

Connectivity Connectivity The authentication protocols are responsible for building the communication services which are way to prove secure mechanism to verify the identity of users and resources

Tutorial 3 Part I

slide-50
SLIDE 50

Tutorial 3

Grid Architecture Grid Architecture - LAYERS

  • LAYERS

Sherbrooke – Lundi 12 Mai 2003

Resource Resource This layer uses the connectivity protocols(communication and authentication) to define protocols and APIs to provide security during the negotiation, starting, control, monitoring, creating reports and details involved during the individual resources operations.

Tutorial 3 Part I

Application Application

Collective Collective

Resource

Con Conn nectivity ectivity

Fabric Fabric

slide-51
SLIDE 51

Tutorial 3

Grid Architecture Grid Architecture - LAYERS

  • LAYERS

Sherbrooke – Lundi 12 Mai 2003

Resource Resource Protocol implementations of this layer utilizes calls from the Fabric to access and control local resources.

Tutorial 3 Part I

Application Application

Collective Collective

Resource

Con Conn nectivity ectivity

Fabric Fabric

slide-52
SLIDE 52

Tutorial 3

Grid Architecture Grid Architecture - LAYERS

  • LAYERS

Sherbrooke – Lundi 12 Mai 2003

Collective Collective The resource layer treats the scope of individual resource operations. On the other hand, in the collective layer collective layer components work with the interaction of resource collections.

Tutorial 3 Part I

Application Application

Collective Collective

Resource

Con Conn nectivity ectivity

Fabric Fabric

slide-53
SLIDE 53

Tutorial 3

Grid Architecture Grid Architecture - LAYERS

  • LAYERS

Sherbrooke – Lundi 12 Mai 2003

Collective Collective The elements from this layer use the resource resource and application application layers to implement a variety of services, such as :

  • Directory service : this facility allows members
  • f virtual organization

virtual organization to discover which are the resources available .

Tutorial 3 Part I

Application Application

Collective Collective

Resource

Con Conn nectivity ectivity

Fabric Fabric

slide-54
SLIDE 54

Tutorial 3

Grid Architecture Grid Architecture - LAYERS

  • LAYERS

Sherbrooke – Lundi 12 Mai 2003

Collective Collective

  • Common Authorization Servers : this facility is also

design to implement a better policy to access resources.

Tutorial 3 Part I

Application Application

Collective Collective

Resource

Con Conn nectivity ectivity

Fabric Fabric

slide-55
SLIDE 55

Tutorial 3

Grid Architecture Grid Architecture - LAYERS

  • LAYERS

Sherbrooke – Lundi 12 Mai 2003

Application Application This layer is related to the users´ applications in their virtual organizations virtual organizations The previous commented layers provide services for this layer.

Tutorial 3 Part I

Application Application

Collective Collective

Resource

Con Conn nectivity ectivity

Fabric Fabric

slide-56
SLIDE 56

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

Collective Collective

Resource Resource Connectivity Connectivity Fabric Fabric Application Application Grid Grid Application Application Transport Transport Internet Internet Link Link Internet Internet

Equivalence between the Gird and Internet Models

slide-57
SLIDE 57

Tutorial 3

Agenda

Sherbrooke – Lundi 12 Mai 2003

  • Applications and Computing Paradigms
  • Users
  • Grid Architecture
  • Grid Computing Environments

Grid Computing Environments

  • Experimental Results
  • Conclusions and Future Work

Tutorial 3 Part I

slide-58
SLIDE 58

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Grid Computing Environments

Tutorial 3 Part I

Grid Consortiums and Open Forums

  • C3CA

C3CA

  • Global Grid Forum

Global Grid Forum

  • Australian Grid Forum

Australian Grid Forum

  • Peer-to-Peer (P2P) Working Group

Peer-to-Peer (P2P) Working Group

  • eGrid

eGrid: European Grid Computing Initiative : European Grid Computing Initiative

  • Asia Pacific Grid

Asia Pacific Grid

1/3

slide-59
SLIDE 59

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Grid Computing Environments

Tutorial 3 Part I

Grid Consortiums and Open Forums

  • GridForum

GridForum Korea Korea

  • EuroTools

EuroTools SIG on SIG on Metacomputing Metacomputing

  • IEEE Task Force on Cluster Computing

IEEE Task Force on Cluster Computing

  • New Productivity Initiative (NPI)

New Productivity Initiative (NPI)

  • The Distributed Coalition

The Distributed Coalition

  • Content Alliance: About Content Peering

Content Alliance: About Content Peering

  • The Brazilian

The Brazilian .... ....

2/3

slide-60
SLIDE 60

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Grid Computing Environments

Tutorial 3 Part I

3/3

Our Our Brazilian Brazilian .... ....

slide-61
SLIDE 61

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

Grid Computing Environments

Grid Middleware

  • Cosm P2P Toolkit
  • Globus

Globus

  • GRACE: GRid Architecture for Computational

Economy

  • Gridbus

1/2

slide-62
SLIDE 62

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

Grid Computing Environments

Grid Middleware

  • Grid

Grid Datafarm Datafarm

  • GridSim

GridSim: Toolkit for Grid Resource Modeling : Toolkit for Grid Resource Modeling and Scheduling and Scheduling Simultation Simultation

  • Simgrid

Simgrid

  • Jxta

Jxta Peer to Peer Network Peer to Peer Network

  • Legion: A Worldwide Virtual Computer

Legion: A Worldwide Virtual Computer

2/2

slide-63
SLIDE 63

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

Grid Computing Environments

DataGrid Initiatives

  • Virtual Laboratory: Tools for Data Intensive

Science on Grid

  • EU DataGrid
  • DIDC Data Grid work
  • GriPhyN (Grid Physics Network)
  • HEPGrid (High Energy Physics and Grid Networks)
  • Particle Physics Data Grid (PPDG)
  • Datacentric Grid
slide-64
SLIDE 64

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

Grid Computing Environments

Grid Systems

  • Compute Power Market
  • Global Operating Systems
  • XtremWeb
  • JAVELIN: Java-Based Global Computing
  • MILAN: Metacomputing In Large Asynchronous

Networks

  • Harness Parallel Virtual Machine Project
  • Management System for Heterogeneous Networks
  • PUNCH - Network Computing Hub
  • MOBIDICK
  • MetaNEOS

1/2

slide-65
SLIDE 65

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

Grid Computing Environments

Grid Systems

  • Amica
  • MultiCluster
  • Poland Metacomputing
  • Echelon: Agent Based Grid Computing
  • Bayanihan
  • NeuroGrid
  • GridLab
  • DAMIEN
  • CrossGrid
  • DIET

2/2

slide-66
SLIDE 66

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

Grid Computing Environments

Computational Economy

  • GRACE: GRid Architecture for Computational

Economy

  • Compute Power Market (CPM)
  • G-Commerce
  • Mariposa: A New Approach to Distributed Data
  • The Information Economy
  • FORTH Information Economies
  • Share Meta
  • D'Agent
  • Program for Research on the Information Economy

1/3

slide-67
SLIDE 67

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

Grid Computing Environments

Computational Economy

  • Xenoservers - Accountable Execution of

Untrusted Programs

  • Electricity Trading Over the Internet Begins in Six

New England States

  • POPCORN
  • CSAR: Resource Tokens and Trading Pool
  • OCEAN - The Open Computation Exchange &

Arbitration Network

  • Spawn: A Distributed Computational Economy
  • Market-Based Computing
  • Multiagent systems

2/3

slide-68
SLIDE 68

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

Grid Computing Environments

Computational Economy

  • W3C effort: Common Markup for micropayment

per-fee-links

  • Agent-Based Computational Economics
  • Electronic Brokerage
  • Society for Computational Economics
  • Internet Ecologies

3/3

slide-69
SLIDE 69

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

Grid Computing Environments

Grid Schedulers

  • Nimrod/G Grid Resource Broker
  • AppLeS
  • SILVER Metascheduler
  • ST-ORM
  • Condor/G
  • NetSolve
  • DISCWorld
  • Computing Centre Software (CCS)
slide-70
SLIDE 70

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

Grid Computing Environments

Grid Portals

  • ActiveSheets
  • UNICORE - Uniform Interface to Computing

Resources

  • SDSC GridPort Toolkit
  • Enginframe
  • Lecce GRB Portal
  • Grid Enabled Desktop Environments
  • Interactive Control and Debugging of Distribution-

IC2D

  • NLANR Grid Portal Development Kit
slide-71
SLIDE 71

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

Grid Computing Environments

Grid Programming Environments

  • Nimrod - A tool for distributed parametric modeling
  • Ninf
  • Cactus Code
  • MetaMPI - Flexible Coupling of Heterogeneous

MPI Systems

  • Virtual Distributed Computing Environment
  • GrADS: Grid Application Development Software

Project

  • Jave-based CoG Kit
  • GAF3J - Grid Application Framework for Java
  • ProActive PDC
  • REDISE - Remote and Distributed Software

Engineering

  • Albatross: Wide Area Cluster Computing
slide-72
SLIDE 72

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

Grid Computing Environments

Grid Performance Monitoring and Forecasting

  • Network Weather Service
  • NetLogger
  • Remos
slide-73
SLIDE 73

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

Grid Computing Environments

Grid Testbeds and Developments

  • World Wide Grid (WWG)
  • Polder Metacomputer
  • NASA Information Power Grid (IPG)
  • NPACI: Metasystems
  • Asia Pacific Bioinformatics Network
  • The Distributed ASCI Supercomputer (DAS)
  • G-WAAT
  • Micro Grid
  • Alliance Grid Technologies
  • The Alliance Virtual Machine Room
  • EuroGrid

1/2

slide-74
SLIDE 74

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

Grid Computing Environments

Grid Testbeds and Developments

  • Internet Movie Project
  • Nordic Grid
  • ThaiGrid
  • TeraGrid
  • Irish Computational Grid (ICG)
  • GrangeNet
  • LHC Grid
  • I-Grid

2/2

slide-75
SLIDE 75

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

Grid Computing Environments

Grid Applications

  • Molecular Modelling for Drug Design
  • Neuro Science - Brain Activity Analysis
  • Cellular Microphysiology
  • HEPGrid: High Energy Physics and the Grid

Network

  • Access Grid
  • Globus Applications
  • The International Grid (iGrid)
  • UK Grid Apps Working Group
  • NLANR Distributed Applications
  • DataGRID - WP9: Earth Observation Science

Application

1/3

slide-76
SLIDE 76

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

Grid Computing Environments

Grid Applications

  • Particle Physics Data Grid
  • DREAM project: Evolutionary Computing and

Agents Applications

  • Knowledge Grid
  • Fusion Collaboratory
  • APEC Cooperation for Earthquake Simulation
  • Australian Computational Earth Systems Simulator
  • EarthSystemGrid

2/3

slide-77
SLIDE 77

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

Grid Computing Environments

Grid Applications

  • Australian Virtual Observatory
  • US Virtual Observatory
  • Distributed Proofreaders
  • NEESgrid: Earthquake Engineering Virtual

Collaboratory

  • Geodise: Aerospace Design Optimisation
  • Japanese BioGrid

3/3

slide-78
SLIDE 78

Tutorial 3

Agenda

Sherbrooke – Lundi 12 Mai 2003

  • Applications and Computing Paradigms
  • Users
  • Grid Architecture
  • Grid Computing Environments
  • Experimental Results

Experimental Results

  • Conclusions and Future Work

Tutorial 3 Part I

slide-79
SLIDE 79

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

Globus The Globus software environment is a project developed by Argonne National Laboratory (ANL) and University

  • f Southern California. In our work we use the version

1.1.4 of the Globus software package because this release provides support to MPI applications. The Globus environment is composed by a set of components implementing basic services to resource allocation, communication, security, process management and access to remote data .

slide-80
SLIDE 80

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

The resource allocation component of the Globus environment (GRAM - Globus Resource Allocation Manager) is the element that acts as an interface between global and local services. Application programmers use the GRAM element, through the gatekeeper software portion which is responsible for the user authentication and association with a local computer account.

slide-81
SLIDE 81

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

The mechanism to identify users of the grid is based on a file called map-file. In this file exists information about authorized users of the grid configuration. Any requirement for resource should be translated to the Resource Specification Language (RSL).

slide-82
SLIDE 82

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

GRAM GRAM GRAM LSF EASY-LL NQE Application RSL Sim ple ground RSL Information Service Local resource m anagers RSL specialization Broker Ground RSL Co-allocator Queries & I nfo

slide-83
SLIDE 83

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

slide-84
SLIDE 84

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

Communication in the Globus environment is performed using a communication library called Nexus. This component defines low a level API to support high level programming paradigms. Examples of high level programming paradigms supported are message passing, remote procedure call and remote I/O procedures. The information about the system and the grid configuration are management by a component called Metacomputing Directory Service (MDS).

slide-85
SLIDE 85

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

An important aspect of the Globus software environment is the security. This software tool employs the certificate approach, which is carried by a CA (Certificate Authority) using the protocol Secure Socket Layer (SSL)

slide-86
SLIDE 86

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

Legion The Legion software environment is a system object

  • riented which is being developed since 1993 at

University of Virginia. This environment has an architecture concept of grid computing providing a unique virtual machine for users´ applications. The approach of the Legion is to have some important concepts of a grid configuration (e.g. scalability, easy to program, fault tolerance and security) transparent to final users.

slide-87
SLIDE 87

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

In the Legion, every entity such as processing power, RAM memory and storage capacity is represented as

  • bjects. Objects communicate with each other using

services calls to a remote mechanism.

slide-88
SLIDE 88

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

The security component of the Legion, as the others elements of this software, is based on an object. The application programmer specifies the security related to an object, where it is defined which type of mechanism is allowed. In addition, the Legion provides some extra basic mechanism to ensure more security. The May I method is an example. Every class should define the method May I, which check for a called

  • bject the related allowed access.
slide-89
SLIDE 89

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Tutorial 3 Part I

slide-90
SLIDE 90

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

The traditional system file is emulated in the Legion environment through the combination of persistent objects with the global information of object identification. This approach simplifies the manipulation of files to application programmers. In addition, it is allow to users to add fault tolerance characteristics to applications using rollback and recovery mechanisms

slide-91
SLIDE 91

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

Grid Environment Legion Globus Software requirement

  • OpenSSL 0.9.5
  • bin/ksh
  • SSLeay 0.9.0
  • OpenLDAP 1.2.7

Minimum Disk space De 250MB a 300 MB 200 MB Minimum Memory RAM 256 MB Not specified

slide-92
SLIDE 92

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

slide-93
SLIDE 93

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

slide-94
SLIDE 94

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

slide-95
SLIDE 95

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

slide-96
SLIDE 96

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

slide-97
SLIDE 97

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

slide-98
SLIDE 98

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

slide-99
SLIDE 99

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

slide-100
SLIDE 100

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

slide-101
SLIDE 101

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

slide-102
SLIDE 102

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

slide-103
SLIDE 103

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

slide-104
SLIDE 104

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

After providing some characteristics of the Globus and Legion software tools, in this section we present our grid configuration environment. It is important to mention that all the machines were in the same laboratory. However, using a Ethernet Layer 3 Switch we were able to have the abstraction of a WAN (Wide Area Network) inside this box. In other words, this equipment could prove the abstraction

  • f a distributed resource environment for our experiments.
slide-105
SLIDE 105

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Computer Name AIX 1 AIX 2 AIX 3 AIX 4 Operating System AIX 4.3 AIX 4.3 AIX 4.3 AIX 4.3 Processor PowerPC_ 604 233 MHz PowerPC_ 604 233 MHz PowerPC_ 604 233 MHz PowerPC_ 604 233 MHz Memory RAM 256 MB 128 MB 128 MB 512 MB Hard disk Two disks

  • f

9 GB Two disks

  • f

4 GB Two disks

  • f 4 GB

and one 2 GB disk Two disks

  • f 4 GB

and one 2 GB disk Software Environment Legion Globus Globus Legion

Table I: The grid environment configuration

Tutorial 3 Part I

slide-106
SLIDE 106

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

The Legion software provides a homogeneous view

  • f the grid to the application programmer.

The environment uses its own tools to create the

  • homogeneity. The procedure to install the software does

not represent any problem, because the application programmer needs only to uncompress binary files and execute some script files. However, for the AIX environment it is necessary more information then those available from the software documents.

slide-107
SLIDE 107

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

We fixed some problems using our background on AIX and exchanging several e-mails with other AIX systems

  • managers. The Legion concept of file system represents

an advantage of the environment. The Legion file system presents a unique identifier for each object.

slide-108
SLIDE 108

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

This approach provides application programmers the facility to access files widely distributed only using their names. In other words, the users only use the name of the file, which can be storage in a local or remote machine. On the other hand, we have verified some problems with the package. As a first problem, we can mention the necessary installation of the entire environment when the bootstrap host has a power failure.

slide-109
SLIDE 109

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

The bootstrap host is responsible for the domain control. Another drawback of the environment is the low communication rate between objects. The paradigm of the Legion is to be a framework environment, where users can develop their own tools, such as security and fault tolerance facilities. This freedom can represent some flexibility to any developers group. However, it does not allow the use external tools.

slide-110
SLIDE 110

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

The Globus approach allows users to use existing system available tools and have a uniform interface to the gird

  • environment. Interesting features of the Globus

environment are related to the security and to the autonomy of the configuration.

slide-111
SLIDE 111

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

The system has an infrastructure based on X509 certificate and the use the mutual authentification. On the other hand,

  • ne drawback of the software is the scalability, which can

be understood as the capability to add new resources and new sites. When considering new facilities application programmers are required to have account into all new hosts.

slide-112
SLIDE 112

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

MATRIX MULTIPLICATION (500x500)

100 200 300 400 500 600 700 800 2 4 8 16 NUMBER OF PROCESSES ELAPSED-TIME Legion Globus

Tutorial 3 Part I

slide-113
SLIDE 113

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

MATRIX MULTIPLICATION (500x500)

100 200 300 400 500 600 700 800 2 4 8 16 NUMBER OF PROCESSES ELAPSED-TIME Legion Globus MPI Nativo

Tutorial 3 Part I

slide-114
SLIDE 114

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

slide-115
SLIDE 115

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Metric Multicast 1:3 x FTP "Multicast" 1:3

Buffer 1024 bytes (Timing)

1000 2000 3000 4000 5000 64k 128k 256k 512k 1M 2M 4M 8M 16M 32M 64M 128M 256M 512M 1G B Tamanho (bytes) T e m p o (s ) Metric Multicast 1:3 FTP "Multicast" 1:3

Tutorial 3 Part I

slide-116
SLIDE 116

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

15.000 30.000 45.000 60.000 75.000 90.000 105.000 120.000 135.000 150.000 1M 4M 16M 64M Tamanho do Arquivo T e m p o ( m s ) XTP multicast 1:2 TCP p/ c4 TCP p/ c5 1 2 3 4 5 6 7 8 9 1M 4M 16M 64M Tamanho do Arquivo T h r o u g h p u t ( M b it s /s )

TCP p/ c4 TCP p/ c5 XTP multicast 1:2

Tutorial 3 Part I

slide-117
SLIDE 117

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

20.000 40.000 60.000 80.000 100.000 120.000 140.000 160.000 180.000 200.000 220.000 1M 4M 16M 64M Tamanho do Arquivo T e m p o ( m s )

XTP multicast 1:3 TCP p/ c2 TCP p/ c4 TCP p/ c5

1 2 3 4 5 6 7 8 9 1M 4M 16M 64M Tamanho do Arquivo T h r o u g h p u t ( M b it s / s ) TCP p/ c2 TCP p/ c4 TCP p/ c5 XTP multicast 1:3

Tutorial 3 Part I

slide-118
SLIDE 118

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

20.000 40.000 60.000 80.000 100.000 120.000 140.000 160.000 180.000 200.000 220.000 240.000 260.000 280.000 1M 4M 16M 64M Tamanho do Arquivo T e m p o ( m s )

XTP multicast 1:4 TCP p/ c2 TCP p/ c3 TCP p/ c4 TCP p/ c5

1 2 3 4 5 6 7 8 9 1M 4M 16M 64M Tamanho do Arquivo T h r o u g h p u t ( M b its / s )

TCP p/ c2 TCP p/ c3 TCP p/ c4 TCP p/ c5 XTP multicast 1:4

Tutorial 3 Part I

slide-119
SLIDE 119

Tutorial 3

Agenda

Sherbrooke – Lundi 12 Mai 2003

  • Applications and Computing Paradigms
  • Users
  • Grid Architecture
  • Grid Computing Environments
  • Experimental Results
  • Conclusions and Future Work

Conclusions and Future Work

Tutorial 3 Part I

slide-120
SLIDE 120

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

In this research work we have presented the implementation characteristics of the Globus and Legion software environments. In addition, we have evaluated these two environments when executing a parallel distributed MPI application.

slide-121
SLIDE 121

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

Globus and Legion are important tools to configure grid

  • configurations. The Globus environment has presented

a more robust features, because the software includes security and fault monitoring mechanisms together with many others services.

slide-122
SLIDE 122

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

An Evaluation of Globus and Legion Software Environments

Tutorial 3 Part I

On the other hand, because it is an object-oriented package the Legion environment is more efficient to present the grid abstraction. This software is a framework and it is not a finished tool. However, we believe that Legion can address those users who are expecting a grid configuration that can be customized for their

  • rganisations.
slide-123
SLIDE 123

Tutorial 3

An Evaluation of An Evaluation of Globus Globus and Legion and Legion Software Environments Software Environments

Sherbrooke – Lundi 12 Mai 2003

Questions ? Questions ?

slide-124
SLIDE 124

Tutorial 3

PART I I

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

slide-125
SLIDE 125

Tutorial 3 Part II

Les grappes d'ordinateurs sont généralement peu dispendieuses, etdonnent accès à une performance intéressante lorsqu'on les compare auxordinateurs parall

  • classiques. Mais, sous certains aspects, leprocessus de

configuration de ces grappes doit être amélioré pour ype l'exécution d'applications courantes. Dans cet article, nousprésentons un cadre qui combine des fonctions de h disponibilitéavec l'utilisation d'un serveur virtuel, dans le d'améliorer lesservices accessibles aux programmeurs en environnement grappe. Lesrésultats que nous obtenons indiquent que cette façon de faire peutaméliorer sensible la capacité d'une grappe à exécuter desapplications cour

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

slide-126
SLIDE 126

Cluster environments are usually provided at a low cost w an interesting performance when compared to parallel

  • machines. However,some aspects of clusters configuratio

should be tackled to improve the environment to execute

  • applications. I n this paper we present a framework in wh

join functions of high-availability and virtual server to en services for application programmers using a cluster

  • environment. Our results indicate that this approach can

I mprove successfully this distributed system to execute r applications.

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

slide-127
SLIDE 127

Tutorial 3

Agenda

Sherbrooke – Lundi 12 Mai 2003

  • Introduction to the problem
  • High-Availability
  • Virtual Server
  • Integrating High-Availability and

Virtual Server

  • Experimental Results
  • Conclusions and Future Work

Tutorial 3 Part II

slide-128
SLIDE 128

Tutorial 3

Agenda

Sherbrooke – Lundi 12 Mai 2003

  • Introduction to the problem

Introduction to the problem

  • High-Availability
  • Virtual Server
  • Integrating High-Availability and

Virtual Server

  • Experimental Results
  • Conclusions and Future Work

Tutorial 3 Part II

slide-129
SLIDE 129

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

Nowadays we cannot image any organization working without its computational systems even for a few minutes. A computational system not working for any fraction of time can represent an enormous lost for the organization. A redundancy approach it is necessary to avoid any risk to stop the computational system. The concern with this issue can be exemplified by IBM and Microsoft, the two companies are already providing some features of reliability for their cluster solutions.

slide-130
SLIDE 130

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

Cluster environments are usually provided at a low cost with an interesting performance when compared to parallel machines. However, some aspects of clusters configurations should be tackled to improve the environment to execute real applications.

slide-131
SLIDE 131

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

The virtual server addresses the load balancing of a cluster configuration providing an even distribution

  • f the workload among all machines of the environment.

During an ordinary period of the cluster environment execution many requests are received to execute many tasks.

slide-132
SLIDE 132

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

Therefore, an element of the cluster could be in charge to redirect all the incoming tasks to the appropriate computers. An appropriate computer is a machine that has a low workload index. However, this approach has a drawback on the central element, which is responsible for redirects the tasks. This central function cannot work properly if the computer presents a failure.

slide-133
SLIDE 133

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

The goal of the high-availability project is to provide a software package where a user can define one machine as a shadow of another computer. In other words, if any problem occurs with the first computer the second machine acts as a mirror, working in the same fashion as the original machine. This approach provides to programmers a more reliable environment to execute their applications.

slide-134
SLIDE 134

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

slide-135
SLIDE 135

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

slide-136
SLIDE 136

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

slide-137
SLIDE 137

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

In this second part of the tutorial we present a framework in which we join functions of high-availability and virtual server to enhance services for application programmers using a cluster environment. Our results indicate that this approach can improve successfully this distributed system to execute real applications.

slide-138
SLIDE 138

Tutorial 3

Agenda

Sherbrooke – Lundi 12 Mai 2003

  • Introduction to the problem
  • High-Availability

High-Availability

  • Virtual Server
  • Integrating High-Availability and

Virtual Server

  • Experimental Results
  • Conclusions and Future Work

Tutorial 3 Part II

slide-139
SLIDE 139

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

The high-availability project targets to provide reliability in distributed environment providing a shadow of one computer on a secondary machine. There are two important concepts when we consider especially the Linux High-Availability (LHA), these are the virtual address IP and the heartbeat.

slide-140
SLIDE 140

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

The first concept (virtual address IP) means that the IP address is linked to a service and not to a certain host. During an interval of time, for example, services can be provided in computer A. In other interval of time, computer B, which is the backup service of computer A, can provide the services because computer A presents hardware problems.

slide-141
SLIDE 141

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

In a cluster we can have many virtual IPs services, where each IP can be linked to one or more services.

slide-142
SLIDE 142

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

The heartbeat concept is a process responsible for monitoring all services in a high-availability environment. The concept also provides the intranet services communications executing the message passing among servers .

slide-143
SLIDE 143

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

Therefore, the heartbeat is the element that decides which server will be responsible for assuming a certain virtual IP address.

slide-144
SLIDE 144

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Client Switch Primary Server Secondary Server Virtual IP Service Heartbeat Message

Tutorial 3 Part II

slide-145
SLIDE 145

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

The LHA works with deamons on both servers. After initializing the Linux, the heartbeat is initialized and checks if the servers are working. The virtual IP is created on the primary server and the machine exchange continuously messages. The procedure of message exchanging is used for check the availability of the servers.

slide-146
SLIDE 146

Tutorial 3

Agenda

Sherbrooke – Lundi 12 Mai 2003

  • Introduction to the problem
  • High-Availability
  • Virtual Server

Virtual Server

  • Integrating High-Availability and

Virtual Server

  • Experimental Results
  • Conclusions and Future Work

Tutorial 3 Part II

slide-147
SLIDE 147

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

The project call Linux Virtual Server (LVS) is designed as an abstraction in which inside a cluster environment

  • nly one computer can provides services to all incoming

requests. This central host is called virtual server. An external host requesting services to this host suppose that all the tasks will be execute by this central node. However, the virtual server has two parts: one load balancer and others n computers.

slide-148
SLIDE 148

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

The function of the load balancer is to receive the external work requests and then to distribute among the

  • thers n computers of the virtual server.

Reading the IP datagram, the load balancer, decides in which computer the task will execute. The external node does not know that another computer is executing the incoming request .

slide-149
SLIDE 149

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

The main target of the virtual server approach is to prove high performance and high availability for distributed applications executing in the cluster. These two aspects are reached by the load balancing and redundancy functions. The first function executes a workload among the n computers of the configuration. The second feature it is provided by the n computers available in the cluster.

slide-150
SLIDE 150

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

S w itc h R e a l S e r v e r R e a l S e r v e r L o a d B a la n c e r R e a l S e r v e r Q u e r ie s f o r S e r v ic e s

Tutorial 3 Part II

slide-151
SLIDE 151

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Processors Pentium 100-100- 233-233-233 Memory (Mbytes) 32-32-64-64-64 Operating system Linux 6.0 kernel 2.2.17

Table I: Cluster configuration

Tutorial 3 Part II

slide-152
SLIDE 152

Tutorial 3

Agenda

Sherbrooke – Lundi 12 Mai 2003

  • Introduction to the problem
  • High-Availability
  • Virtual Server
  • Integrating High-Availability and

Integrating High-Availability and Virtual Server Virtual Server

  • Experimental Results
  • Conclusions and Future Work

Tutorial 3 Part II

slide-153
SLIDE 153

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

After understanding interesting features of the high-availability and virtual server projects we decided to extend the functionality of these environments, therefore we decide to integrate the two approaches. It is clear that the load balancer as a central element of the cluster receiving requests, it is a point that can occurs a failure. In this case, all the interesting functions of the cluster environment cannot be translated as a useful work.

slide-154
SLIDE 154

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

R e a l S e r v e r R e a l S e r v e r R e a l S e r v e r P r i m a r y L o a d B a l a n c e r S e c o n d a r y L o a d B a l a n c e r S w i t c h Tutorial 3 Part II

slide-155
SLIDE 155

Tutorial 3

Agenda

Sherbrooke – Lundi 12 Mai 2003

  • Introduction to the problem
  • High-Availability
  • Virtual Server
  • Integrating High-Availability and

Virtual Server

  • Experimental Results

Experimental Results

  • Conclusions and Future Work

Tutorial 3 Part II

slide-156
SLIDE 156

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

All experimental results present in this section are based on our change of the cluster configuration. Therefore, we have created several study case situations to verify if our approach works as expected.

slide-157
SLIDE 157

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

The next figure shows the IP addresses and names used for our experiments. The integrated environment is form by the load balancer. The primary balancer was called Laico Z and the secondary as Laico Y. On the other hand, the n computers used were called as Laico 2, Laico 3 and Laico 4.

slide-158
SLIDE 158

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

R e a l S e r v e r R e a l S e r v e r R e a l S e r v e r S w i t c h P r i m a r y L o a d B a l a n c e r S e c o n d a r y L o a d B a l a n c e r

Tutorial 3 Part II

slide-159
SLIDE 159

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

slide-160
SLIDE 160

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

R e a l S e r v e r R e a l S e r v e r R e a l S e r v e r S e c o n d a ry L o a d B a la n c e r P r im a r y L o a d B a la n c e r

Tutorial 3 Part II

slide-161
SLIDE 161

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

slide-162
SLIDE 162

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

slide-163
SLIDE 163

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

slide-164
SLIDE 164

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

R eal S erver R ea l S erver R ea l S erver S w itch Secon d ary L oa d B alan cer P rim a ry L oa d B alan cer

Tutorial 3 Part II

slide-165
SLIDE 165

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

slide-166
SLIDE 166

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

slide-167
SLIDE 167

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

slide-168
SLIDE 168

Tutorial 3

Agenda

Sherbrooke – Lundi 12 Mai 2003

  • Introduction to the problem
  • High-Availability
  • Virtual Server
  • Integrating High-Availability and

Virtual Server

  • Experimental Results
  • Conclusions and Future Work

Conclusions and Future Work

Tutorial 3 Part II

slide-169
SLIDE 169

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Providing Cluster Environments with Providing Cluster Environments with High-Availability and Load-Balancing High-Availability and Load-Balancing

Tutorial 3 Part II

The several experiments illustrated in our study cases indicate the integration success of HA and LVS. Integrating the high-availability and the virtual server functions for a distributed cluster can represent to this environment an interesting reliable configuration to execute several applications.

slide-170
SLIDE 170

Tutorial 3

Sherbrooke – Lundi 12 Mai 2003

Questions ? Questions ?

slide-171
SLIDE 171

Tutorial 3

  • Prof. Mario Dantas

Department of I nformatics and Statistics Federal University of Santa Catarina (UFSC) Florianopolis – Brazil E-mail : mario@inf.ufsc.br http://www.inf.ufsc.br/~mario

Sherbrooke – Lundi 12 Mai 2003