micromon a monitoring framework for tackling distributed
play

MicroMon: A Monitoring Framework for Tackling Distributed - PowerPoint PPT Presentation

MicroMon: A Monitoring Framework for Tackling Distributed Heterogeneity Babar Khalid*+, Nolan Rudolph+, Ramakrishnan Durairajan, Sudarsun Kannan* *Rutgers University, University of Oregon (+co-primary authors) Background Modern


  1. MicroMon: A Monitoring Framework for Tackling Distributed Heterogeneity Babar Khalid*+, Nolan Rudolph†+, Ramakrishnan Durairajan†, Sudarsun Kannan* *Rutgers University, †University of Oregon (+co-primary authors)

  2. Background • Modern applications are increasingly becoming geo-distributed - e.g., Cassandra, Apache Spark • Geo-distributed datacenters (DCs) use heterogeneous resources - storage heterogeneity (e.g., SSD, NVMe, Harddisk) - WAN heterogeneity (e.g., fiber optics, InfiniBand) • Hardware heterogeneity in DCs avoids vendor lockout and reduces operational cost (by combining older/cheaper and newer/expensive hardware) • Careful provisioning can provide high performance at lower cost 2

  3. Problem With Current Systems • Current monitoring frameworks for geo-distributed applications are unidimensional - can only monitor hosts, storage devices, networks in isolation • Lack hardware heterogeneity awareness - e.g. no awareness for storage heterogeneity - could impact I/O intensive applications • Coarse-granular monitoring - unaware of host-level micro-metrics in software and hardware - e.g. page cache, node-level I/O traffic, node’s network queue delays 3

  4. Our Solution - MicroMon • MicroMon is a fined grained monitoring, dissemination, and inference framework • Collects fine-grained ( micrometrics ) software and hardware metrics in end-hosts and network - e.g., page cache utilization, disk read/write throughput in end host • Filters micrometrics into anomalies to efficiently disseminate • Enables replica selection for geo-distributed Cassandra • Preliminary study of Micromon integrated with geo-distributed Cassandra shows high throughput gains 4

  5. Outline • Background • Case Study • Design • Evaluation • Conclusion 5

  6. Case Study - Cassandra • Distributed NoSQL database system deployed geographically • Manages large amounts of structured data in commodity servers • Provides highly available service and no single point of failure • Typically focuses on availability and partition tolerance 6

  7. Cassandra – Replication Node 1 Node 5 Node 2 Update (key) Node 4 Node 3 Client Cassandra Cluster 7

  8. Cassandra – Replication Rack Awareness Rack 1 Node 1 Rack 1 Node 5 Node 2 Update (key) Rack 2 Node 4 Node 3 Client 8

  9. Cassandra – Replication DC Awareness Rack 1 Rack 1 Node 1 Node 1 Rack 1 Rack 1 Node 5 Node 5 Node 2 Node 2 Update (key) Rack 2 Rack 2 Node 4 Node 3 Node 4 Node 3 Client DC: US DC: Europe 9

  10. Cassandra’s Snitch Monitoring • Cassandra uses Snitch to monitor network topology and route requests across replicas • Also provides capability to spread replicas across DCs to avoid correlated failures • Snitch monitors (read) latencies to avoid non-responsive replicas • Different types: Gossiping, MultiRegionSnitch - Gossiping uses rack and datacenter information to gossip across nodes and collect latency information • Problem: No hardware heterogeneity awareness 10

  11. Analysis Goal and Methodology Goal: Highlight the lack of heterogeneity awareness • Replica Configuration • - SSD Replica: Sequential storage b/w - 600MB/s, rand b/w: 180 MB/s - HDD replica: Sequential storage b/w - 120MB/s, rand b/w: 10 MB/s Network latency across replicas same (for this analysis) • Workload – YCSB benchmark • - workload A (50% read and writes) - workload B (95% reads) - workload C (100% reads) 11

  12. Impact of Storage Heterogeneity Awareness 50000 HDD-only SSD-only Snitch 40000 OPS/sec 30000 20000 10000 0 A B C YCSB Workloads • Significant performance impact over optimal SSD-only configuration • Snitch: Lack of awareness to storage hardware heterogeneity 12

  13. Outline • Background • Case Study • Design • Evaluation • Conclusion 13

  14. Our Design: MicroMon • Monitoring and inference framework for geo-distributed applications • Performs micro-metrics monitoring at the host and network-level - micro-metrics includes fine-grained software and hardware metrics • Efficiently disseminates collected micro-metrics • Ongoing - Distributed inference engines to guide application requests to the best replica 14

  15. MicroMon Challenges • Selection Problem: What micrometrics to consider? • Dissemination Problem: How to send all micrometrics? • Inference Problem: How to quickly infer from micrometrics? 15

  16. Design - Micrometrics Selection Huge combinations of micrometrics across app, host OS, and network • • Micrometrics could vary for different application-level metrics e.g. micrometrics for latency different than those for throughput • Our approach: Start with storage and network micrometrics • Identify hardware and software micrometrics using resource usage - e.g. high storage usage -> monitor page cache, read/write latency 16

  17. MicroMon High-level Design Enterprise DC A Enterprise DC B Enterprise Backbone Networking stack Storage stack Networking stack micrometrics at switches micrometrics at DC micrometrics at DC ----- Ingress/Egress ----- Page cache (SW) ----- Transport ----- Port File system (SW) Flags (syn, ack, etc.) Server Packet count Block device driver (SW) Window size Byte count Hard disk (HW) Goodput Drop count Switch Bytes transmitted/received Utilization Round-trip time ----- Buffer ----- ----- Application ----- Collected Avg. queue length Throughput micrometrics Queue drop count Congestion status 17

  18. Reducing Dissemination – Anomaly Reports • Problem: Prohibitive cost of dissemination across thousands of nodes - cost increases with hardware and software components - e.g., SSD’s SMART counters contain close to 32 counters • • Observation: OSes already expose anomalies (indirectly) - e.g. high I/O wait time of process -> higher page cache misses - e.g. sustained storage BW against max. hardware BW - e.g. network I/O queue wait time alludes to TCP congestion • Proposed Idea: Instead of sending thousands of micrometrics to decision agent, only report OS perceived anomalies 18

  19. Reducing Dissemination - Network Telemetry • Network telemetry offers aggregated stats about state of the network • Idea: co-design in-band network telemetry (INT) with end host OS - monitor packets at end host with anomaly reports as payload - get network anomaly reports using INT • Pre-established anomaly thresholds reduce total aggregated stats further INT payload INT header Network End-host anomalies anomalies 19

  20. Scalable Inference - Scoring-based Inference • Simple scoring–based inference in Cassandra - replicas sorted and ranked by network latency • Problem: for bandwidth sensitive applications, need higher weights for WAN-based micrometrics compared to host-level micrometrics • Our approach: - we assign equal weights to all software and hardware micrometrics - use collected micrometrics to calculate a replica score - route request to replicas with higher scores - flexibility to assign higher weights for WAN-based micrometrics • Ongoing: Designing a generic, self-adaptive inference engine 20

  21. Outline • Background • Case Study • Design • Evaluation • Conclusion 21

  22. Evaluation Goals Goals: Understand the impact of storage heterogeneity with Micromon • Understand the impact of storage heterogeneity + network latency • Analyze the page cache impact (see paper for details) • 22

  23. Analysis Methodology Multiple DCs from CloudLab Infrastructure • - three nodes located in UTAH, APT, and Emulab DCs Replica Configuration • - UTAH replica: NVMe storage (seq bw: 600MB/s, rand bw: 180 MB/s) - APT replica: HDD (seq bw: 120 MB/s, rand bw: 10 MB/s) - Emulab master node: HDD (same as above) Network Latencies • - 400us between UTAH (NVMe) replica and master node - 600us between APT (HDD) replica and master node Workload – YCSB benchmark • - workload A (50% read and writes) - workload B (95% reads) - workload C (100% reads) 23

  24. MicroMon’s - Storage Heterogeneity 50000 HDD-only SSD-only Snitch MicroMon 40000 Ops/sec 30000 20000 10000 0 32 64 128 32 64 128 32 64 128 clients clients clients clients clients clients clients clients clients Workload A Workload B Workload C • Snitch lacks storage heterogeneity awareness • MicroMon’s storage heterogeneity awareness provides performance close to SSD-only (optimal) configuration Performance improves by up to 49% for large thread configuration • 24

  25. Storage Heterogeneity + Network Latency Introduce network latency for SSD-only node • 9000 Throughput (ops/s) 7000 Snitch MicroMon 5000 3000 1000 -1000 0ms 1ms 2ms 5ms 10ms 15ms 25ms Network Latency For high network latencies (e.g., beyond 10ms) SSD benefits reduce • 25

  26. Conclusion Datacenter systems are becoming more and more heterogeneous • Deploying geo-distributed applications in heterogeneous datacenters • requires redesign of monitoring mechanisms We propose MicroMon, a fine-grained micrometric monitoring, • dissemination, and inference framework Our on-going work will focus on efficient dissemination and self- • adaptive inference mechanisms 26

  27. Thanks! Questions? Contact: sudarsun.kannan@rutgers.edu ram@cs.uoregon.edu 27

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend