Position: Synergetic Effects of Software and Hardware Parameters on - - PowerPoint PPT Presentation

position synergetic effects of software and hardware
SMART_READER_LITE
LIVE PREVIEW

Position: Synergetic Effects of Software and Hardware Parameters on - - PowerPoint PPT Presentation

Position: Synergetic Effects of Software and Hardware Parameters on the LSM System Authors: Jinghuan Yu, Heejin Yoon* Sam H. Noh*, Young-ri Choi*, Chun Jason Xue * Log Structured Merge-tree (LSM) Specific designs for HDD and write-intensive


slide-1
SLIDE 1

Position: Synergetic Effects of Software and Hardware Parameters on the LSM System

Authors: Jinghuan Yu, Heejin Yoon* Sam H. Noh*, Young-ri Choi*, Chun Jason Xue *

slide-2
SLIDE 2

Log Structured Merge-tree (LSM)

What is the critical factor deciding performance? Does the working principle of LSM still fit these new mediums?

Specific designs for HDD and write-intensive workload. Periodical compaction with various resource occupation.

slide-3
SLIDE 3

Performance Feature of Each Media

PMM

  • Relatively low latency
  • Stable performance in any

parallelism

  • Strong wear lifetime without GC

NVMe SSD

  • Highest sequential bandwidth
  • Strong parallel performance, with

higher requirement for CPU

  • Performance is strongly affected

by write granularity SATA SSD

  • Limited bandwidth
  • Poor parallel performance
  • Unstable performance due to

foreground garbage collection Media Type Average Access Latency (µs) SATA SSD 37.78 NVMe SSD 11.77 PMM 2.61

slide-4
SLIDE 4

Performance Comparison of Devices in RocksDB

  • Increasing number of

CPUs causes IO congestion, decreasing performance

  • With fixed CPUs,

benefits from larger batch size

  • Best throughput
  • Performance increase

tends to be stable as the number of CPUs increases

  • Suffers from larger batch

size

  • Not sensitive to

the number of CPUs or batch size

  • Throughput

difference is far from bandwidth comparison

50 100 150 200 250 300 350 400 450 2 CPUs 4 CPUs 8 CPUs 2 CPUs 4 CPUs 8 CPUs 2 CPUs 4 CPUs 8 CPUs SATA SSD NVMe SSD PMM Throughput (kOps/Sec) 16 MB 32 MB 64 MB 128 MB

Operation Batch Size

slide-5
SLIDE 5

Existing Solutions Our Targets

VS

Resource Utilization Heterogenous Storage Parameter Tuning

  • Both CPU and disk utilization
  • Smooth, effective, and

predictable

  • Auto scaling
  • Driven by workload
  • Device oriented
  • Online tuning
  • Based on quantitative

analysis

  • Disk utilization first
  • Lazy scheduling
  • Devices features based
  • Rule-based selection
  • Size-based scaling
  • Unified configuration
  • Offline tuning
  • Based on statistics data
slide-6
SLIDE 6

Characteristic analysis and design points

slide-7
SLIDE 7

Performance Traits of SATA SSD

Strength

  • Effective for bulk single-thread write workload

Weakness

  • Serious IO congestion during multi-thread

compaction

Design Opportunities

  • Single queue continuous write
  • Large-grained operations

10000 20000 30000 40000 2 CPUs 4 CPUs 8 CPUs SATA SSD IO Time(s)

Cumulative IO Time in Compaction Runs

16 MB 32 MB 64 MB 128 MB Operation Batch Size

In SATA SSD, IO time increases dramatically as number of CPUs increases

slide-8
SLIDE 8

Performance Traits of NVMe SSD

Strength

  • High bandwidth

Weaknesses

  • Larger batch size decreases the

performance

Design Opportunities

  • Quicksand effect: quicker devices make

the data sink too quickly and decreases the performance.

  • Improve the pipeline of compaction

works

Performance of NVMe SSD with different bandwidth (limited by cgroup).

Bandwidth = 400MB Bandwidth = 800MB Bandwidth = 1200MB Bandwidth = 1600 MB Unlimited bandwidth Operation Batch Size

slide-9
SLIDE 9

Performance Traits of PMM

Strength

  • Stable parallel performance

Weaknesses

  • More sensitive to the slow L0 compactions,

which can be solved by changing the size ratio between L0 and L1 files.

Design Opportunities

  • More flexible data structures
  • Can be used directly as a memory expansion
  • Non-volatile, free of consistency overhead

such as WAL

Size Ratio here means the (total size of L0 files) / (total size of L1 files), controlled by compaction scheduling parameters Operation Batch Size

slide-10
SLIDE 10

Challenges Solutions DOTA: Device Oriented Tuning Advisor

Workload adapting Global resource management Online Tuning Thread pool and resource allocation Amplification reduction and data reuse Environment detecting and monitoring Data placement and migration Online modeling

slide-11
SLIDE 11

Email Address: jinghuayu2-c@my.cityu.edu.hk Github link: https://github.com/supermt/ utils_for_lsm.git