Multi-scale Analysis of Large Distributed Computing Systems Lucas - - PowerPoint PPT Presentation

multi scale analysis of large distributed computing
SMART_READER_LITE
LIVE PREVIEW

Multi-scale Analysis of Large Distributed Computing Systems Lucas - - PowerPoint PPT Presentation

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion Multi-scale Analysis of Large Distributed Computing Systems Lucas M. Schnorr, Arnaud Legrand, Jean-Marc Vincent INRIA Mescal, CNRS LIG


slide-1
SLIDE 1

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Multi-scale Analysis of Large Distributed Computing Systems

Lucas M. Schnorr, Arnaud Legrand, Jean-Marc Vincent INRIA Mescal, CNRS LIG Grenoble, France

LSAP’2011 Workshop San Jose, CA June 8th, 2011

1 / 27

slide-2
SLIDE 2

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Introduction

Distributed systems to monitoring tools

Large scale distributed systems

Heterogeneous & Shared Examples: grids, clouds, volunteer systems Composed of several thousands components Context: observation and analysis

Monitoring tools

Provide high-level statistics Basic resource usage traces through time Examples: Ganglia, NWS, Monalisa Scalable techniques, but often lack the details to

understand unexpected behavior correlate application behavior to resource utilization

2 / 27

slide-3
SLIDE 3

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Introduction

Tracing techniques to large-scale distributed systems?

Tracing techniques (from parallel computing)

Precise and fine grain application-level events Complex behavior patterns can be detected Interactive Analysis with visualization tools Examples: TAU, Vampir, Jumpshot, MPE

Some challenges remain

visualization scalability intrusion control large trace files in large-scale scenarios

Question

Can we use tracing techniques to collect precise and fine grain in large-scale distributed systems?

3 / 27

slide-4
SLIDE 4

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Approach

Multi-scale analysis of detailed traces

Aggregation technique applied to

time: long periods of observation space: a considerable amount of observed entities

Analysis through data visualization Identify and understand non-trivial behavior Applied to BOINC client resource analysis

4 / 27

slide-5
SLIDE 5

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Outline

1

Introduction

2

Multi-scale Analysis Approach

3

Experimental Framework

4

Results and Analysis

5

Conclusion

5 / 27

slide-6
SLIDE 6

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Analysis approach

Detailed observation over long periods of time

Results in large data sets to be analyzed Difficult to extract useful information

Example: BOINC availability traces

One volunteer during an eigth-month period Graphical zoom in an arbitrary twelve-day period

6 / 27

slide-7
SLIDE 7

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Analysis approach: aggregation

Using temporal aggregation

Integrate using a time frame defined by the analyst Analyze integrated data instead of raw traces

Example: BOINC availability traces

  • ne-hour integration

easier to understand patterns

7 / 27

slide-8
SLIDE 8

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Analysis approach: space dimension

Large number of resources

Spatial integration of their behavior Neighborhood definition

8 / 27

slide-9
SLIDE 9

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Framework: Distributed Application Scenario

Scheduling of bag-of-tasks in volunteer computing BOINC architecture used as example

Several projects, each with a number of tasks Volunteers decide to which project work for

1 Fair sharing

Volunteers define how much to work for each project Check if a fair share is locally and globally achieved

2 Response time

BOINC not designed to give good response time How it behaves if projects have bursts of small tasks

9 / 27

slide-10
SLIDE 10

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Framework: BOINC Simulator (with SimGrid)

Raw traces are needed to validate our approach Real world large-scale platforms

Hard to measure, time consuming Most of traces are already reduced in time Sometimes also in space

Simulation using the BOINC Simulator

Already validated against the real BOINC client Supports the most important features of the real one

deadline scheduling, long term debt fair sharing, exponential back-off

Considers real availability traces from the FTA

SimGrid toolkit

Fast simulation of several thousands hosts Able to trace categorized resource utilization

10 / 27

slide-11
SLIDE 11

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Framework: Triva Visualization Tool

Implements the time & space data aggregation Analyst interaction

Time frame to be considered Select/Filter a set of resources Animation of variables through time

Two visualization techniques

Treemap view: addresses scalability Topology-based view: network correlation

11 / 27

slide-12
SLIDE 12

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Results and Analysis

Overview

Two scenarios

1 Fair sharing 2 Response time

For each scenario

Setting the scenario, initial configuration Expected behavior based on previous knowledge Observed results and visualization analysis

12 / 27

slide-13
SLIDE 13

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Fair Sharing – Setting

Server side

Two BOINC projects servers with continuous jobs

Project Task Size Deadline Continuous-0 30 300 Continuous-1 1 15 Volunteer side

Number of volunteers: 65 Evenly share its resource

Simulation

Simulated time of ten weeks Volunteer hosts power and availability from FTA

13 / 27

slide-14
SLIDE 14

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Fair Sharing – Expected

Long term period Equal global and local share

50% for Continuous-0 50% for Continuous-1

14 / 27

slide-15
SLIDE 15

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Fair Sharing – Observed 1/3

Aggregation: whole time – all clients Reasonable fair

15 / 27

slide-16
SLIDE 16

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Fair Sharing – Observed 2/3

Aggregation: whole time – per-client view Size occupied by a volunteer indicates its donation

16 / 27

slide-17
SLIDE 17

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Fair Sharing – Observed 3/3

Anomalies detected on fair sharing mechanism Correlation with small volunteer contribution

17 / 27

slide-18
SLIDE 18

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Fair Sharing – Final

Early development of BOINC simulator

Counting of time worked for each project Bug fixed

18 / 27

slide-19
SLIDE 19

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Response Time

Overview

BOINC is for CPU-bound throughput computing Jobs in BOINC have a deadline attribute Volunteer clients try to respect the deadline

Earliest deadline first mode

Projects interested in response time

Play with the job deadline parameter May gain short-term priority over other projects

19 / 27

slide-20
SLIDE 20

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Response Time – Setting

Server side

Two BOINC projects servers / different job policies

Project Task Size Deadline Period Replicas Continuous 30 300 always 1 Burst 1 6 10 days 5 Volunteer side

Number of volunteers: 65 Evenly share its resource, respecting deadlines

Simulation

Simulated time of ten weeks Volunteer hosts power and availability from FTA

20 / 27

slide-21
SLIDE 21

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Response Time – Expected

BOINC has a pull architecture

Volunteers ask jobs to the project servers Projects wait for volunteers to distribute jobs

Slow-start effect for jobs that belong to burst project Higher priority to burst jobs Observe the shape of the effect in the visualization

21 / 27

slide-22
SLIDE 22

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Response Time – Observed 1/2

Aggregated work for all volunteers

Continuous (light gray) Burst (dark gray)

22 / 27

slide-23
SLIDE 23

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Response Time – Observed 2/2

Anomalies

Wasted computations (replica/deadline parameters) Low priority of the Burst project (dark gray)

23 / 27

slide-24
SLIDE 24

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Response Time – Further Investigation 1/2

Graph-based view

Each host is represented by a square Size represents CPU-power of volunteers The two servers

white if inactive black if active

Volunteers

burst: dark gray continuous: light gray

One-hour aggregated

Continuous Burst 24 / 27

slide-25
SLIDE 25

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Response Time – Further Investigation 2/2

Graph-based view

Behavior of each volunteer in time Each screenshot – one-hour temporal aggregation

Continuous Burst

Timeline

12h 13h 14h 15h 16h 17h

Cyclic behavior in all volunteers Reason: short time fairness of BOINC

25 / 27

slide-26
SLIDE 26

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Final Remarks

Monitoring systems are common

Scalable techniques Difficult to perform a profound analysis

This paper

Detailed traces on large-scale distributed systems Elaborated visualization techniques for analysis

Main contributions

Identification of anomalies and unexpected behavior

Fair sharing issue - BOINC simulator fixed Identification of wasted computations Suprinsingly low priority of response time projects

Study of response time

Shape of slow-start BOINC’s short time fairness is negative

26 / 27

slide-27
SLIDE 27

Introduction Multi-scale Analysis Approach Experimental Framework Results and Analysis Conclusion

Thank you

Simgrid Framework – http://simgrid.gforge.inria.fr Triva Visualization Tool – http://triva.gforge.inria.fr Questions?

27 / 27