DAY 2 Agenda for Today Introduce the workload characterization - - PowerPoint PPT Presentation

day 2 agenda for today
SMART_READER_LITE
LIVE PREVIEW

DAY 2 Agenda for Today Introduce the workload characterization - - PowerPoint PPT Presentation

DAY 2 Agenda for Today Introduce the workload characterization problem. Discuss a simple example of characterizing the workload for an intranet. What is workload characterization? What workload we want to characterize?


slide-1
SLIDE 1

DAY 2

slide-2
SLIDE 2

Agenda for Today

Introduce the workload characterization problem. Discuss a simple example of characterizing the workload for an intranet.

– What is workload characterization? – What workload we want to characterize?

Present a workload characterization methodology.

slide-3
SLIDE 3

References

  • Thyagaraj Thanalapati, Sivarama P. Dandamudi:

An Efficient Adaptive Scheduling Scheme for Distributed Multicomputers. IEEE Trans. Parallel

  • Distrib. Syst. 12(7): 758-768 (2001)
  • A. Iosup and D.H.J. Epema, Grid Computing

Workloads, IEEE Internet Computing 15(2): 19-26 (2011)

  • Feitelson DG. Workload modeling for computer

systems perfMemory ormance evaluation. Cambridge University Press; 2015.

slide-4
SLIDE 4

Introduction

  • Workloads appear in many contexts and therefore

have many different types.

  • Different workloads have a significant effect on

performance.

  • Different workloads have different features
  • Ensuring that systems meet desired performance

goals without excessive (and expensive) over- provisioning is of great importance.

  • Therefor, reliable workload models are needed.
slide-5
SLIDE 5

What is workload?

  • The workload of a system can be defined as the set
  • f all inputs that the system receives from its

environment during any given period of time.

A variety of jobs with different characteristics are submitted to the cloud by different users. A static workload is one in which a certain amount of work is given, and when it is done that is it. A dynamic workload, in contrast, is one in which work continues to arrive all the time; it is never “done”. Need to capture

slide-6
SLIDE 6

HP Engineering

Sample Cloud Computing Applications

Demand Variability Demand Volume Taxes, @Home Low Low High High Web Server Space Survey Comet Detected Sky Survey Pharma Research Social Gaming Online Gaming Social Networking Big Data Analytics SW Dev/Test Office Tools Tsunami Prediction Epidemic Simulation Exp. Research

From Iosup

slide-7
SLIDE 7

Batch workloads

  • Typically tend to process huge volumes of data.

– Example: An application that produce monthly bills for utilities (e.g., power, water, mobile phone bills, etc.) – Require considerable compute and storage resources. – Batch workloads are rarely time sensitive and can be scheduled when few real-time tasks are running.

slide-8
SLIDE 8

High-performance workloads

  • High-performance workloads typically require

significant compute capabilities.

– next day weather prediction application

  • High-throughput workloads Examples of these

applications include

– different patient data in large scale biomedical trials – different parts of a genome or protein sequence in bioinformatics applications – different random numbers in a simulations based on Monte Carlo methods – different model parameters in ensemble simulations or explorations of parameter spaces

slide-9
SLIDE 9

Bag-of-Tasks

  • Bag-of-tasks refers to parallel jobs with many

independent tasks

  • The tasks can be executed out of the submission
  • rder.
  • The following is not bag of task

T1 T2 T3 T5 T4 T7 T8 T6

slide-10
SLIDE 10

Example of BoT Applications

  • Parameter sweeps

– Comprehensive, possibly exhaustive investigation of a model – Very useful in engineering and simulation-based science

  • Monte Carlo simulations

– Simulation with random elements: fixed time yet limited inaccuracy – Very useful in engineering and simulation-based science

  • Many other types of batch processing

– Periodic computation, Cycle scavenging – Very useful to automate operations and reduce waste

slide-11
SLIDE 11

Workflow Applications

  • Workflow is a set of jobs

with precedence (think Direct Acyclic Graph)

  • Complex applications

– Complex filtering of data – Complex analysis of instrument measurements

Maslina Abdul Aziz, Jemal H. Abawajy, Tutut Herawan: Layered workflow scheduling algorithm. FUZZ-IEEE 2015: 1-7

slide-12
SLIDE 12

ADDRESSING BIG DATA ENERGY CONSUMPTION

Collection Station

Emergency Monitoring Centre Medical staff

Observation Station Data Center

Fog Node Push Pull Push Pull Push Pull Physical Machines Sensor Data Computational resources Virtual (Sensor) Machines VM Manager Service Scheduler Analytics Engine Visualization Enigne

Access Control Data Sharing Policy

Application 1 Application N Medical Application Services E-Health Data Middleware Publish Alert

Patient

IoT+Fog

Jemal H. Abawajy, Mohammad Mehedi Hassan, Federated Internet of Things and Cloud Computing Pervasive Patient Health Monitoring System. IEEE Communications Magazine 55(1): 48-53 (2017)

Sara Ghanavati, Jemal Abawajy and Davood Izadi (2017), The Internet of Things: Foundation for Smart Cities, eHealth, and Ubiquitous Computing, Ricardo Armentano, Robin Singh Bhadoria, Parag Chatterjee, Ganesh Chandra Deka (eds.), CRC Press.

  • Integrated Fog computing (FC) and Internet of

Everything (IoE)

slide-13
SLIDE 13

Workflow Applications

  • Workflow applications are jobs with precedence (

represented by a directed acyclic task graph).

  • 75%+ WFs are sized 40 jobs or less, 95% are sized 200

jobs or less (Ostermann et al., 2008)

T1 T2 T3 T5 T4 T7 T8 T6

Structure of divide-and-conquer programs

Maslina Abdul Aziz, Jemal H. Abawajy, Tutut Herawan: Layered workflow scheduling algorithm. FUZZ-IEEE 2015: 1-7

  • Example applications

– Complex filtering of data – Complex analysis of instrument measurements

slide-14
SLIDE 14

Web Workload

  • Example of execution of HTTP requests (sec)

Characteristics of Web workloads:

– burstiness – heavy-tailed distributions

Request No. CPU time I/O time Elapsed time

1 0.0095 0.04 0.071 2 0.0130 0.11 0.145 3 0.0155 0.12 0.156 4 0.0088 0.04 0.065 5 0.0111 0.09 0.114 6 0.0171 0.14 0.163 7 0.2170 1.20 4.380 8 0.0129 0.12 0.151 9 0.0091 0.05 0.063 10 0.0170 0.14 0.189 Average 0.0331 0.205 0.550

slide-15
SLIDE 15

Performance Analysis

  • There are three common methodologies to

evaluate performance

– Use actual workload – Use traced workload directly to drive a simulation. – Create a model from the trace and use the model for either analysis or simulation.

slide-16
SLIDE 16

Trace Data

  • Trace is recorded data as a trace, or log, of

workload-related events that happened in a certain system.

  • Example

– The Failure Trace Archive (FTA) is a centralized public repository of availability traces of parallel and distributed systems, and tools for their analysis. – The purpose of this archive is to facilitate the design, validation, and comparison of fault- tolerant models and algorithms.

slide-17
SLIDE 17

Trace-data simulation

  • Trace-driven simulations are common in parallel and

distributed computing

  • Shortcomings

– Collecting data may be inconvenient or costly, and instrumentation may introduce overhead. – Data collection is inherently limited as the conditions under which data is gathered may restrict one to a small sample of the space of interest. – For example, given a server with 128 nodes, it is not possible to collect data about systems with 64 or 256 nodes.

  • We need models to transcend these limitations.
slide-18
SLIDE 18

Workload Modeling

  • Workload modeling is used to generate synthetic

workloads based on real-life job execution observations.

  • The goal is typically to be able to create workloads that

can be used in performance evaluation studies

Workload Submission

Workload description (may include QoS)

Workload Generation

J1 J1 Jn

Analysis SLA

Consolidate result

Negotiation

slide-19
SLIDE 19

Framework

Cloud Computing Environment Performance Measures Preal Cloud Computing Environment Performance Measures Pmodel

Workload Model

The synthetic workload is supposed to be similar to those that occur in practice on real systems

End User

Real Workload Synthetic Workload

The result is supposed to align with real application Replicate in experimental setting cloud environment

slide-20
SLIDE 20

Workload Resource Usage Observation

Transaction Classes Frequency Maximum CPU time

(msec)

Maximum I/O time

(msec)

Trivial 40% 8 120 Light 30% 20 300 Medium 20% 100 700 Heavy 10% 900 1200

slide-21
SLIDE 21

Workload Features

  • Workloads are characterized by many attributes that

have a great effect on scheduling

– I/O can also have a great effect on scheduling – memory can lead to swapping. – for parallel jobs, the number of processors used is an additional parameter, which influences how well jobs pack together.

  • The level of detail needed in workload

characterization depends on the goal of the evaluation.

Jemal H. Abawajy, Sivarama P. Dandamudi: Scheduling Parallel Jobs with CPU and I/O Resource Requirements in Cluster Computing Systems. MASCOTS 2003: 336-343

slide-22
SLIDE 22

Workload Features

  • Workloads are characterized by many attributes

– arrival times – running times – total memory usage (maybe locality of reference) – distribution of I/O sizes and how they interleave with the computation – Communication – number of processors (VMs)

slide-23
SLIDE 23

Adapted from Menascé & Almeida. 52

Selection of characterizing parameters

  • Each workload component is characterized by two groups of

information:

  • Workload intensity

– arrival rate – number of clients and think time – number of processes or threads in execution simultaneously

  • Service demands (Di1, Di2, … DiK), where Dij is the service

demand of component i at resource j.

slide-24
SLIDE 24

Exponential Distribution

  • The Exponential:

– 𝜇 = measures how many things happen per unit 𝑔 𝑦 = 𝜇 ∙ 𝑓−𝜇∙𝑦

0.5 1 1.5 2 2.5 1 2 3 4 5 6 7 probability density functions X

Lambda

0.5 1 2

slide-25
SLIDE 25

Hyper-exponential distribution

  • Hyper-exponential distributions is used in

service demand modeling

  • The hyper-exponential distribution is obtained

by selecting from a mixture of several exponential distributions.

  • The simplest variant has only two stages:

𝜇1 𝜇2 𝑄 1 − 𝑄 𝜇1 ≠ 𝜇2

𝑔 𝑦 = 𝑞𝑗 ∙ 𝜇𝑗 ∙ 𝑓𝜇𝑗𝑦

𝑜 𝑗=1

0 ≤ 𝑞𝑗 ≤ 1 0 ≤ 𝜇1, 𝜇2𝑗 ≤ 1

slide-26
SLIDE 26

Coefficient of Variation

  • Coefficient of variation is defined as follows

𝐷𝑤 = 𝜏 𝜈

  • Where

– 𝜏 is the standard deviation – 𝜈 is the mean

slide-27
SLIDE 27

Long-lived vs short-lived Jobs

  • Long-lived processes
  • short-lived processes

𝑄 𝑈 > 𝜐 ∝ 𝛽2, 𝛽 ≈ 1

  • This means that most processes are short, but

a small number are very long.

slide-28
SLIDE 28

Selecting VM for Migration

  • A VM running on an overloaded physical

machine will receive worse service than a VM running on an unloaded machine.

  • Load balancing is the activity of migrating a

running VM from an overloaded machine to an underloaded one.

slide-29
SLIDE 29

Selecting VM for Migration

  • It would be most beneficial if we could

identify VM that may be expected to continue to run for a long time, and select them for migration.

  • But how can we know in advance whether a

VM is going to finish executing soon or not?

– The answer is that it depends on the statistics of VM runtimes.

slide-30
SLIDE 30

Selecting VM for Migration

  • Suppose we have the following as an example

VM Start End Now Remain 𝑊𝑁1 0 16 12 4 𝑊𝑁2 0 64 20 44 𝑊𝑁3 0 14 4 10 Time unit 2 4 6 8 10 12 14 16 18 20 22 24 𝑊𝑁1 𝑊𝑁2 𝑊𝑁3 64

  • The challenge is deciding the VM to migrate.
slide-31
SLIDE 31

Selecting VM for Migration

  • VM migration involves considerable overhead.
  • If the VM terminates soon after being

migrated, that overhead has been wasted.

  • In addition, the VM cannot run during the

time it is being migrated.

  • Again, if it terminates soon after the

migration, it would have been better off staying in its place.

slide-32
SLIDE 32

Performance Measurements

  • Classical performance metrics:

– Response time, – throughput, – scalability, – resource/cost/energy, – efficiency, – Elasticity – Availability, – reliability, and – security – SLA violation

slide-33
SLIDE 33

Energy Consumption

Feature Extraction Feature Selection Applications Resources Features Subset Features Regression Method Result Evaluation

slide-34
SLIDE 34

Makespan

– Traditional: Makespan, Job Slowdown – Workload Speedup One (SU1) – Workload Slowdown Infinite (SUinf)

slide-35
SLIDE 35

Cost Metrics

  • Cost Metrics

– Actual Cost (Ca) – Charged Cost (Cc)

slide-36
SLIDE 36

Compound Metrics

  • Compound Metrics

– Cost Efficiency (Ceff) – Utility

slide-37
SLIDE 37

Adapted from Menascé & Almeida. 66

Summary

Workload Characterization

what is it? basic concepts workload description and modeling representativeness of a workload model

Methodology (1)

Choice of an analysis standpoint Identification of the basic component Choice of the characterizing parameters Data collection

slide-38
SLIDE 38

Application

slide-39
SLIDE 39

Communication

  • Given

– 𝑢0 is the start up time or latency, – n is the number of nodes involved in the communication, – 𝑠

∞ = is the asymptotic bandwidth, which is defined as

the maximal bandwidth achievable when the message length approaches infinity, – m denotes the number of bytes of data to be transmitted.

slide-40
SLIDE 40

Workload Parameters

  • We can use Sevcik's execution time function to

model a wide range of representative applications

– 𝜚: is an indication of the load imbalance in the computation, – W is the cumulative service demand for the job, – p is the number of processors allocated to the job, – 𝛽: represents the amount of sequential computation,

– 𝛾:represents the amount of communication and congestion overhead that increases with the number of processors.

– represents the ideal division of an application's basic work across p processors

slide-41
SLIDE 41

Maximum Parallelism

slide-42
SLIDE 42

Values

slide-43
SLIDE 43

Acknowledgement

  • The slides have been adapted from

– Menascé & Almeida.