URSA: Precise Capacity Planning and Fair Scheduling based on - - PowerPoint PPT Presentation

ursa precise capacity planning and fair scheduling based
SMART_READER_LITE
LIVE PREVIEW

URSA: Precise Capacity Planning and Fair Scheduling based on - - PowerPoint PPT Presentation

URSA: Precise Capacity Planning and Fair Scheduling based on Low-level Statistics for Public Clouds Wei Zhang, Ningxin Zheng, Quan Chen, Yong Yang, Zhuo Song, Tao Ma, Jingwen Leng, Minyi Guo Shanghai Jiao Tong University & Alibaba Cloud


slide-1
SLIDE 1

URSA: Precise Capacity Planning and Fair Scheduling based on Low-level Statistics for Public Clouds

Wei Zhang, Ningxin Zheng, Quan Chen, Yong Yang, Zhuo Song, Tao Ma, Jingwen Leng, Minyi Guo Shanghai Jiao Tong University & Alibaba Cloud

slide-2
SLIDE 2

1

Background & Motivation

2

The methodology of URSA

3

Evaluation

4

Conclusion

slide-3
SLIDE 3

1

Background & Motivation

2

The Methodology of URSA

3

Evaluation

4

Conclusion

slide-4
SLIDE 4

Problem:Datacenter Underutilization

4

dbPaaS ■ The excessive purchase of resources on the cloud 2x-5x

Low resource utilization!

Reserved vs Used Resources:Twitter: up to 5x CPU & memory overprovisioning

Capacity planning Overprovisioned reservations by users

slide-5
SLIDE 5

Problems in Capacity Planning

Utilization Performance

?

Improve utilization while guaranteeing the performance goals of users.

5

slide-6
SLIDE 6

Solutions for Private Datacenter

workload1

Private data centers

Bubble-Flux (ISCA’13) Paragon (ASPLOS’13) Bubble-up (MICRO’11) Quasar (ASPLOS’14)

6

workloadn

slide-7
SLIDE 7

Problems in dbPaaS public clouds

7

New challenges?

  • Poor Resource Utilization

Ø Heuristic search will get stuck in local optima

Ø Extensive profiling is not applicable due to privacy problem Ø Unawareness of shared resource contention and pressure

  • Performance unfairness

Prior work is not applicable for Database platform- as-a-service(dbPaaS) in public Clouds!

slide-8
SLIDE 8

1

Background & Motivation

2

The methodology of URSA

3

Evaluation

4

Conclusion

slide-9
SLIDE 9

Main Idea of URSA

  • Predicting the scaling surface of the target workload based on the

low level statistics and adjusting the resource specification

  • accordingly. (A online capacity planner)
  • Designing a contention-aware scheduling engine at the Cloud level.
  • Quantifying the interference “pressure” and its “tolerance” to the

contention on shared resources using low-level statistics. (An

performance interference estimator)

9

slide-10
SLIDE 10

Overview

  • nline capacity

planner

10

contention- aware scheduler

The contention on shared resources.

performance Interference estimator

Predicting workload performance scaling pattern based on low-level statistics URSA Predicting the scaling surface

slide-11
SLIDE 11

The Design of URSA

11

slide-12
SLIDE 12

Construct capacity planner

12

Se Sele lected system-le level l in indexes

  • How to construct the capacity planner.
slide-13
SLIDE 13

Online capacity planning

  • How to perform capacity planning for an online workload.
slide-14
SLIDE 14

Interference estimator

14

  • Interference due to LLC
  • Interference due to Memory Bandwidth

𝑙𝑛𝑞𝑡 = !!"!#$%&'(($(

"

(1)

slide-15
SLIDE 15

Contention-aware Scheduler

15

Based on the quantified pressures and tolerances of each database workload

  • n all the shared resources, the contention-aware scheduling engine carefully

places the workloads for enforcing the performance fairness. Each node is given a Schedule Score(SS). CS quantifies the contention score of the node (smaller is better) and RS quantifies the resource score of the node (smaller is better). For a node, RS is calculated to be the average percentage of the used CPUs and memory of the node. CS is calculated in the upon formula.

slide-16
SLIDE 16

1

Background & Motivation

2

The methodology of URSA

3

Evaluation

4

Conclusion

slide-17
SLIDE 17

Experimental setup

17

Benchmarks

  • Generating database workloads using two widely-used workload generators:

Sysbench and OLTPBench that includes YCSB , TPC-C, LinkBench and SiBench workloads.

  • We adjust the configurations of Sysbench, YCSB, TPC-C, LinkBench, SiBench, and

generate 11 variations for each of them. The 55 workloads are randomly divided into a training set containing 44 workloads and a validation set containing 11 workloads.

slide-18
SLIDE 18

Evaluation

18

18/22 is the optimal resource specification

  • Efffectiveness of the Capacity Planning

Ø Scenario 1: Achieving Performance Target. Ø Scenario 1: Cutting Down Rent Cost.

5/11 is the optimal resource specification

slide-19
SLIDE 19

Evaluation

Overhead The main overhead of URSA is from scheduling. URSA identifies the appropriate node for a workload on our 7-node Cloud in 0.12ms using a single thread.

19

  • Efffectiveness of improving Resource utilization and Fairness
slide-20
SLIDE 20

1

Background & Motivation

2

The methodology of URSA

3

Evaluation

4

Conclusion

slide-21
SLIDE 21

Conclusion

  • Our work
  • An online capacity planner
  • A performance interference estimator
  • A contention-aware scheduling engine
  • Results
  • URSA reduces up to 25.9% of CPU usage, 53.4% of memory and

reduces the performance unfairness between the co-located workloads by 47.6% usage without hurting their performance.

  • Propose
  • Automatically suggest the just-enough resource specification that

fulfills the performance requirement of dbPaaS in Public Clouds

slide-22
SLIDE 22

Thanks for attention! Q&A

zhang-w@sjtu.edu.cn