Improving Agility and Elasticity in Bare-metal Clouds Yushi Omote , - - PowerPoint PPT Presentation

improving agility and elasticity in bare metal clouds
SMART_READER_LITE
LIVE PREVIEW

Improving Agility and Elasticity in Bare-metal Clouds Yushi Omote , - - PowerPoint PPT Presentation

Improving Agility and Elasticity in Bare-metal Clouds Yushi Omote , Takahiro Shinagawa , Kazuhiko Kato University of Tsukuba, The University of Tokyo 1 Bare-metal Clouds An IaaS for high performance and device functionality


slide-1
SLIDE 1

Improving Agility and Elasticity in Bare-metal Clouds

Yushi Omote†, Takahiro Shinagawa‡, Kazuhiko Kato†

†University of Tsukuba, ‡The University of Tokyo 1
slide-2
SLIDE 2

Bare-metal Clouds

2

OS Provider User

Physical Machine

An IaaS for high performance and device functionality OS transparent

No Virtual Machine

slide-3
SLIDE 3

OS-deployment Problem

(1) Image Copy (2) Reboot from Local Disk

User OS Image

(Tens of minutes) (A few minutes)

OS Installer User OS Image Image Server

3

Long wait time sacrifices agility and elasticity

slide-4
SLIDE 4

Existing Approach 1 OS Streaming Deployment ✗

Agility and Elasticity OS-specific drivers are required. Network Boot + Background Copy OS transparency

User OS Special Driver Image Server

4

Performance

[Clerc et al. IPCCC’10]

slide-5
SLIDE 5

VMM User OS Image Server

5

Existing Approach 2 Conventional VMMs

OS transparency Agility and Elasticity

Streaming deployment with VMMs

Performance

[VMware’01, Xen’03, KVM’07]

Continuous virtualization overhead

slide-6
SLIDE 6

OS Deployment with a Special-purpose VMM

VMM User OS VMM User OS

1) Streaming deployment 2) Seamless de-virtualization OS transparency Agility and Elasticity Performance

6
slide-7
SLIDE 7

Challenge Expose & Control Physical Devices

7

Virtual Devices? Direct I/O? Control I/Os Expose physical interface

OS VMM

✗ ✗

slide-8
SLIDE 8

Device-interface-level I/O mediation

(1) I/O interpretation

to understand I/O context

(2) I/O redirection

to perform network booting

(3) I/O multiplexing

to perform background install

8

Physical device interface A device mediator performs:

OS VMM Device Mediator Device Driver

slide-9
SLIDE 9

I/O Interpretation

Determine when/how to mediate I/O requests

Understand state transitions based on monitoring I/O

9

Device State Transitions

OS VMM Device Mediator Device Driver

slide-10
SLIDE 10

I/O Redirection

VMM OS Data Small Request

Interrupt (1) Interpret

10

Image Server

(2) Redirect (3) Restart

Disk LBA=4 NUM=8 LBA=4 NUM=8

slide-11
SLIDE 11

I/O Multiplexing

VMM OS Status Check

(2) Emulate

11

Idle State

(3) Queue

Disk VMM Request

(1) Request

OS Request Image Server

slide-12
SLIDE 12

CPU/Memory Virtualization for De-virtualizable VMM

12

CPU

VMM

No indirection

VMM runs passively with VMX No guest scheduling

Memory

OS Guest Physical Address

=

VMM Physical Address

Identity Mapping

VMM exposes physical memory Mark VMM regions as reserved (via BIOS INT15/e802)

slide-13
SLIDE 13

De-virtualization

VMM OS VMM OS VMM OS

H/W (1) Turns off IO VM exits (2) Turns off nested paging (3) Turns off CPU virtualization

Device Driver

Find safe I/O timing Unsynchronized TLB flush Ease VM exits condition (VMXOFF Issue)

13
slide-14
SLIDE 14

Performance Evaluation

  • Deployed 32-GB OS Image (Ubuntu 14.04 64-bit)
  • OS-startup Time
  • Cassandra Throughput
  • Storage Throughput
  • InfiniBand Latency
14

Intel Xeon X5680 (3.33 GHz) / 96GB RAM HDD 500GB/7200 RPM SATA Mellanox InfiniBand (4X QDR) Intel 82575 EM GbE Network Card

A HPC Cluster

Interconnected by A Mellanox Grid Director InfiniBand Switch & A FUJITSU SR- S348TC1 GbE Switch

slide-15
SLIDE 15

OS-startup Time

Elapsed Time (sec)

0.00 150.00 300.00 450.00 600.00

Image Copy Reboot+Firminit. VMM Boot OS Boot

370 145 29 5+58 30+42 49

15

VM Streaming (KVM/NFS) OS Streaming (NFSRoot) Proposed Image Copy

Quick start up (8.6 times faster)

slide-16
SLIDE 16

Cassandra Throughput

(Throughout Deployment)

% of Baremetal

70% 80% 90% 100% 110% 120%

Elapsed Time (sec)

50 200 350 500 650 800 950 1100 1250

Proposed KVM (No Background Install)

16

Seamless de-virtualization Eventual bare-metal performance

slide-17
SLIDE 17

Storage Throughput

Throughput (MB/sec)

0.00 30.00 60.00 90.00 120.00

Bare-metal Deploy Devirt KVM/Local

100 115 112 112 101 112 112 117

Read Write

17

Bare-metal performance

slide-18
SLIDE 18

InfiniBand RDMA latency

Latency (usec)

0.00 0.43 0.85 1.28 1.70

Baremetal Deploy Devirt KVM/Pass

1.61 1.30 1.30 1.30

18

Bare-metal performance

slide-19
SLIDE 19

Conclusion

  • Improved agility and elasticity in bare-metal clouds
  • De-virtualizable VMM with streaming deployment
  • Device-interface-level I/O mediation
  • Achieved quick startup of an OS
  • 8.6 times faster than image copy
  • Preserved high performance & OS-transparency
19
slide-20
SLIDE 20

Future work

  • Generating device mediators from specification
  • Reduce development cost of device mediators
  • More advanced features of IaaS clouds
  • Live migration and checkpointing
20
slide-21
SLIDE 21

Thank you

21