CloudDet Motivation Visualization Challenges CloudDet: Interactive - - PowerPoint PPT Presentation

clouddet motivation visualization challenges
SMART_READER_LITE
LIVE PREVIEW

CloudDet Motivation Visualization Challenges CloudDet: Interactive - - PowerPoint PPT Presentation

CloudDet Motivation Visualization Challenges CloudDet: Interactive Visual Analysis of Anomalous Performances in Cloud Monitoring nodes instead of monitoring applications Computing Systems Scale : Trade-off between system scalability


slide-1
SLIDE 1

CloudDet: Interactive Visual Analysis of Anomalous Performances in Cloud Computing Systems

Ke Xu, Yun Wang, Leni Yang, Yifang Wang, Bo Qiao, Si Qin, Yong Xu, Haidong Zhang, Huamin Qu IEEE Transactions on Visualization and Computer Graphics, 2019

Amirhossein Abbasi Nov 2019

  • 1

CloudDet

2

Motivation

  • Monitoring nodes instead of

monitoring applications

  • Too many false positives, scale

problem.

  • Visualization of anomalies:

Intuitiveness, interaction.

  • Research Contribution:

Detection system, Visualization, Evaluation

3

Visualization Challenges

  • Scale: Trade-off between system scalability and level-of-

detail(LoD)

  • Multi-dimensionality: Temporal patterns, Relation between

metrics

  • Boundary normal/abnormal

4

System Overview

5

What is abnormal and what is not? How to detect?

6

Mathematics!

7

Algorithm Flow

8

Algorithm Flow

9

Only utilizes the most recent data Anomalies have patterns

Design Tasks

  • Overview of anomalies for data query
  • Ranking suspicious nodes dynamically
  • Browse data flexibly
  • Facilitate anomaly detection
  • Similarities of nodes

10

T4 T3 T1 T2 T5

Visualization

11

Encoding Protocol

12

Global Categorical Colors: performance metrics (CPU Frequency, Memory Usage,…)

Linear Color Scheme: Anomaly score Diverging Color Scheme: Difference

  • f performance metrics to average

Spatial and Temporal Views

13

Spatial and Temporal Views

14 Overview Ranking Browse data Facilitate detection Comparison T1 T2 T3 T4 T5

Rank and Performance View

15

Horizon Chart

16

Interactions:

  • Brushing
  • Collapsing
  • Stretching
slide-2
SLIDE 2

Line mode

17

  • Each line for one metric
  • More conventional
  • Normalize data to [-1,1]

PCA mode

18

  • Project a multivariate data to a
  • ne-dimensional time-series data
  • Major Trend

Alternative Designs

19

Measurement Unit Visual Clutter Consumes Space Showing Trends Showing Trends Non-linear time scaling Analyzing Overal Changes Scaling in time domain

Rank and Performance View

20 Overview Ranking Browse data Facilitate detection Comparison T1 T2 T3 T4 T5

The Cluster View

21

  • Dimensionality reduction.
  • Another perspective for anomaly diagnosis
  • White contour: most probably anomaly
  • Gray contour:normal.

The Cluster View

22 Overview Ranking Browse data Facilitate detection Comparison T1 T2 T3 T4 T5 23

Official Video (1:05 min)

What-Why-How Summary

Multi-variate time- series quantitative performance data from compute nodes.

24

What Why How

  • Colors and brightness
  • horizontal and line chart
  • Special glyphs
  • Spatial positions of nodes

and charts.

  • Interactivity: Scrolling,

Brushing, and setting parameters

  • Anomaly Ranking
  • Anomaly inspection
  • Anomaly Clustering

What-Why-How Summary

Multi-variate time- series quantitative performance data from compute nodes.

25

What Why How

  • Colors and brightness
  • horizontal and line chart
  • Special glyphs
  • Spatial positions of nodes

and charts.

  • Interactivity: Scrolling,

Brushing, and setting parameters

  • Anomaly Ranking
  • Anomaly inspection
  • Anomaly Clustering

What-Why-How Summary

Multi-variate time- series quantitative performance data from compute nodes.

26

What Why How

  • Colors and brightness
  • horizontal and line chart
  • Special glyphs
  • Spatial positions of nodes

and charts.

  • Interactivity: Scrolling,

Brushing, and Querying

  • Anomaly Ranking
  • Anomaly inspection
  • Anomaly Clustering

Scalability

Multi-variate time- series quantitative performance data from compute nodes.

27

What Why How

  • Colors and brightness
  • horizontal and line chart
  • Special glyph
  • Spatial positions of nodes

and charts.

  • Interactivity: Scrolling,

Brushing, and setting parameters

  • Anomaly Ranking
  • Anomaly inspection
  • Anomaly Clustering

Scale

Very Scalable: scale linearly with time-series input data size

Evaluation

28

Quantitative Evaluation

29

Case Studies

30

Case Study 1: Bitbrains Datacenter Traces Case Study 2: Live Cloud System Data

500 VMs, One month

1,000,000 nodes, Two weeks [100 data centers with 20 data clusters with 500 nodes each]

User Feedback

31

  • Automated Anomaly Detection: Trust in algorithm,
  • System: Useful and User-friendly, Consistent, too

comprehensive and Overwhelming, Need Tutorial

  • Visualization and Interaction: Helpful, new perspective for
  • verall trend, clear comparison, Confess that they use

chaotic line charts before.

Critique

32

slide-3
SLIDE 3

Positive

  • Alternative designs
  • Super-scalable
  • Perfect evaluation
  • Very Accurate
  • Special Glyphs

33

Negative

  • Better to use non-diverging colors for horizon charts.
  • Minor occlusion in the clustering.
  • Make use of global colors in horizontal chart.
  • Bad way for Assigning the ranks to performance.
  • Empty clusters in spatial overview.
  • Limitation: Just consider recent data and one metric.
  • Limitation: Don’t discuss why using those performance metrics for

anomaly.

34

Question?

35