Webinar Series: Triangulate your Storage Architecture with SvSAN - - PowerPoint PPT Presentation

webinar series
SMART_READER_LITE
LIVE PREVIEW

Webinar Series: Triangulate your Storage Architecture with SvSAN - - PowerPoint PPT Presentation

Webinar Series: Triangulate your Storage Architecture with SvSAN Caching Luke Pruen Technical Services Director What can you expect from this webinar? To answer a simple question How can I create the perfect balanced storage architecture


slide-1
SLIDE 1

Triangulate your Storage Architecture with SvSAN Caching Luke Pruen – Technical Services Director

Webinar Series:

slide-2
SLIDE 2

What can you expect from this webinar?

To answer a simple question

How can I create the perfect balanced storage architecture to meet my capacity, footprint and performance needs, now and in the future?

slide-3
SLIDE 3

Introducing StorMagic

What do we do?

StorMagic SvSAN eliminates the need for physical SANs by exposing the storage of an industry standard server as a virtual SAN thereby dramatically reducing CAPEX and OPEX.

How does SvSAN achieve this?

StorMagic’s virtual SAN converts the internal disk, flash and memory of industry standard servers into robust, cost effective and flexible shared-storage.

Where is this most applicable?

SvSAN is deployed for hyperconverged infrastructure for multi-site enterprises and SMEs and server-based storage arrays as an alternative to a traditional physical SAN.

slide-4
SLIDE 4

1 to Thousands

  • f Sites

Large and small deployments from enterprises with 1000s of sites to SMEs with a single site Global Customer Adoption Within 72 countries,

  • rganisations depend on

StorMagic for sever and storage infrastructure Global Partner Network Wherever you are, StorMagic has resellers, integrators, and server partners to meet your needs Across Many Verticals Retail, health, government, industrial, education, finance, pharma and many more

Introducing StorMagic

slide-5
SLIDE 5

What is a Storage Array?

A disk array is a hardware element that contains a large group

  • f hard disk drives (HDDs). It may contain several disk

drive trays and has an architecture which improves speed and increases data protection. The system is run via a storage controller, which coordinates activity within the unit. Wikipedia Definition:

slide-6
SLIDE 6

Storage Array Components

  • SAN Presentation

̶ iSCSI ̶ Fibre Channel

  • SAN Switch

̶ Fibre channel ̶ Ethernet (iSCSI)

  • Physical Storage Controller

̶ CPU ̶ Memory ̶ Dedicated Storage Hardware

  • Enterprise Drives

̶ 10K or 15K SAS ̶ SSD

SSD

Storage Controller

slide-7
SLIDE 7

Utilise industry standard server as a Virtual SAN

SSD

Storage Controller

SSD

Virtual Storage Appliance Hypervisor

slide-8
SLIDE 8

StorMagic SvSAN: Overview

“SvSAN turns the internal disk, SSD and memory of industry standard servers into highly available shared storage”

slide-9
SLIDE 9

Use Cases: Hyperconverged or Server-Based Storage Array

  • Hyperconverged (HCI)

̶ Shared storage and compute platform ̶ Custom-build and avoid appliance over-provisioning ̶ Perfect for remote, branch and private cloud services

  • Server-Based Storage Array (Server SAN)

̶ Custom-build active/active mirrored shared storage ̶ Fraction of the cost of off-the-shelf SAN ̶ Flexibility to scale capacity and performance as needed

slide-10
SLIDE 10

Triangulate your storage requirements: 3 Axis

10

slide-11
SLIDE 11

Triangulate your requirements: Capacity

Object: Capacity

  • Capacity - High priority
  • Footprint - Low priority
  • Performance - Low priority

Solution: High capacity drives

  • SATA 7.2k rpm 65 – 100 IOPS
  • SAS 10k.15k rpm 140 – 210 IOPS
  • Low cost per GB

Compromise: Performance & Footprint

  • Require more drives to achieve performance (spindles)
  • Larger server footprint
  • Lower cost per GB, higher cost per IOPS

11

slide-12
SLIDE 12

Triangulate your requirements: Footprint

12

Object: Footprint

  • Capacity - Low priority
  • Footprint - High priority
  • Performance - Low priority

Solution: Smaller servers

  • Use smaller U servers
  • 2.5” drives

Compromise: Capacity & Performance

  • Limits the number of drive
  • Limits performance when using spinning disks
slide-13
SLIDE 13

Triangulate your requirements: Performance

13

Object: Performance

  • Capacity - Low priority
  • Footprint - Low priority
  • Performance - High priority

Solution: SSD/Flash

  • SSD/Flash 8.6k to 10 millions IOPs
  • High cost per GB compared to magnetic
  • Add more servers to cluster

Compromise: Capacity & Resource

  • High cost per GB
  • Lower capacity servers????
  • Deduplication available but heavy on server resource
slide-14
SLIDE 14

Triangulate your requirements: How to balance all 3?

14

Priority: All high

  • Capacity - Medium/High priority
  • Footprint - Medium/High priority
  • Performance – Medium/High priority

Priority: Mixed

  • Capacity - Medium/Low priority
  • Footprint - High priority
  • Performance - Medium priority

Priority: Two high

  • Capacity - Low priority
  • Footprint - High priority
  • Performance - High priority
slide-15
SLIDE 15

Triangulate your requirements: No compromise

15

Object: Capacity

  • Capacity - High priority
  • Footprint - High priority
  • Performance - High priority

Solution: Caching Compromise: None

  • Leverage high capacity SAS or SATA drive

for capacity

  • Leverage SSD/Flash & memory for

performance

  • SAS 10k.15k rpm 140 – 210 IOPS
slide-16
SLIDE 16

Caching: The importance of caching

16

Virtualized environments suffer from the ‘I/O blender’ effect

  • Multiple Virtual Machines sharing a set of disks
  • Resulting in predominantly random I/O
  • Magnetic drives provide poor random performance
  • SSD & Flash storage ideal for workloads but expensive

Working sets of data

  • Driven by workloads which are ever changing
  • Refers to the amount of data most frequently accessed
  • Always related to a time period
  • Working sets sizes evolve as workloads change

Caching

  • Combat the I/O blender effect without the expense of all Flash or SSD
  • Working sets of data can be identified and elevated to cache
slide-17
SLIDE 17

Optimising Storage: Write Caching

Stage 1

  • All new data written to SSD
  • Provides low latency and high IOPs
  • Data is marked as “dirty” as is has not be committed to the backing store

Stage 2

  • The write operation is acknowledged immediately to the

server/application Stage 3

  • The “dirty” data is reordered and grouped based on disk locality
  • The data is destaged and written out the backing store sequentially as

possible Stage 4

  • Cache is notified when data has successfully been written to backing

store

  • The data in cache is marked as “clean” and remains in cache until the

space is required

slide-18
SLIDE 18

Caching: Predictive read caching

Intelligent read caching algorithm

  • All read I/Os are monitored and analyzed
  • Most frequently used data – “Hot” data
  • Cache tiers are populated based on access frequency

Tiering

  • RAM: Most frequently accessed data
  • SSD/Flash: Next most frequently accessed data
  • HDD: Infrequently accessed data – “Cold” data

Tracker Module

  • Identifies transient data to prevent cache pollution
  • Promotes and demotes data through tiers
  • Intelligently adjusts tracker window to workload

Sizing

  • Assign cache sizes to meet requirements
  • Grow caches as working sets change
  • Use any combination of Memory, SSD/Flash and Disk
slide-19
SLIDE 19

Caching: Read Ahead & Pinning

Read-ahead mode

  • I/O blender effect aware!
  • Identifies sequential interleaved I/O requests
  • Detects sequential read streams to allow read ahead
  • Pre-fetches data into memory

19

Data pinning mode

  • Pin specific data/workloads in memory
  • Delivers most efficient read performance
  • Databases, VDI, frequently repeated operations
  • Manage multiple pin groups
slide-20
SLIDE 20

Caching: Example Platform Numbers

20

3 x 1.2TB 10k SAS RAID 5 3x1.2TB 10k SAS RAID 5 1xSSD 1xSSD Hardware

  • Cisco USC C220 M3
  • 1 x Intel Xeon CPU E5-2699 v4 @ 2.20GHz
  • 22 Cores
  • 44 Threads
  • 128GB RAM per host

Networking

  • 2 x 1Gb NICs VM Network
  • 2 x 10Gb direct connect, SvSAN iSCSI/Mirror

Storage

  • 1 x RAID5 = 3 x 1.2TB 10K SAS disks
  • 1 x 200GB Samsung SSD
slide-21
SLIDE 21

Caching: Customer Workload

Workload

  • 12 Virtual Machines
  • 78 applications
  • Back up service

500 1000 1500 2000 2500 3000 3500 18:39 21:25 00:11 02:57 05:43 08:29 11:15 14:01 16:47 19:33 22:19 01:05 03:51 06:37 09:23 12:09 14:55 17:41 20:27 23:13 01:59 04:45 07:31 10:17 13:03 15:49 18:35 IOPs Time of Day (UTC)

Throughput IOPs

Read Write 0.001 0.01 0.1 1 10 100 1000 10000 21 MB 110 GB 221 GB 332 GB 443 GB 554 GB 665 GB 776 GB 887 GB 998 GB 1.1 TB 1.2 TB 1.3 TB 1.4 TB 1.5 TB 1.6 TB 1.7 TB 1.8 TB 2.0 TB 2.1 TB 2.2 TB 2.3 TB 2.4 TB 2.5 TB 2.6 TB Number of accesses (logarithmic scale) Thousands

Locality of access

Read Write

5,000,000 10,000,000 15,000,000 20,000,000 25,000,000 1 KB 2 KB 4 KB 8 KB 16 KB 32 KB 64 KB 128 KB 256 KB 512 KB 1 MB 2 MB 4 MB Hit Count Block Size

Block Size Distribution

Writes Read

Read Write Read/Write % 77% 23% Sequential % 49% 39% Average Per Day 991 GB 294 GB Average Block Size 58 KB 54 KB Average IOPS 212 138

slide-22
SLIDE 22

Caching: Customer Workload Results

HDD Only

  • 1 x RAID5 = 3 x 1.2TB 10K SAS disks

HDD & Memory

  • 12GB of memory per host for caching
  • 1 x RAID5 = 3 x 1.2TB 10K SAS disks

HDD & SSD

  • 1 x 200GB Samsung SSD
  • 1 x RAID5 = 3 x 1.2TB 10K SAS disks

HDD, SSD & Memory

  • 12GB of memory per host for caching
  • 1 x 200GB Samsung SSD
  • 1 x RAID5 = 3 x 1.2TB 10K SAS disks

2400 12341 6165 13061 2000 4000 6000 8000 10000 12000 14000 HDD Only HDD & Memory HDD & SSD HDD, SSD & Memory

Total IOPS

21.86 4.94 9.37 5.21 5 10 15 20 25 HDD Only HDD & Memory HDD & SSD HDD, SSD & Memory

AVG Latency (ms)

HDD Only HDD & Memory HDD & SSD HDD, SSD & Memory Total IOPS 2400 12341 6165 13061 AVG Latency (ms) 21.86 4.94 9.37 5.21

slide-23
SLIDE 23

Caching: Synthetic Hero Numbers

23

Total IOPS Average Response (ms) Workload HDD Tiered HDD Tiered 4KB 100% Random Read 4263 207117 60.04 1.24 4KB 100% Random Write 2004 38196 127.17 6.70 4KB 70/30 Read Write 80% Random 1832 79738 147.36 3.21 32KB 100% Sequential Read 7729 85803 33.62 2.98 32KB 100% Sequential Write 8728 15651 29.33 16.36

4263 2004 1832 7729 8728 207117 38196 79738 85803 15651 4KB 100% RND RD 4KB 100% RND WR 4KB 70/30 RD/WR 80% RND 32KB 100% SEQ RD 32KB 100% SEQ WR

IOPS

HDD Tiered 60.04 127.17 147.36 33.62 29.33 1.24 6.70 3.21 2.98 16.36 4KB 100% RND RD 4KB 100% RND WR 4KB 70/30 RD/WR 80% RND 32KB 100% SEQ RD 32KB 100% SEQ WR

AVG Latency (ms)

HDD Tiered

HDD Config

  • 1 x RAID5 = 3 x 1.2TB 10K SAS disks

Tiered Config

  • 1 x RAID5 = 3 x 1.2TB 10K SAS disks
  • 1 x 200GB Samsung SSD
  • 32GB of memory per host for caching

Test

  • vSphere 6.5
  • 2 x Windows VMs
  • IOmeter
  • VMDK on VMFS
slide-24
SLIDE 24

Triangulate your storage architecture: Light bulb moment!

  • Significant storage options available to IT pros today.
  • Optimal storage configurations are very hard to achieve.
  • It starts with understanding your capacity and performance requirements.
  • Balanced with achieving the smallest possible footprint.
  • Don’t blindly head down the path of HCI appliances or off-the-shelf SANs.
  • Consider custom building against your exacting requirements.
  • Leverage 3-tiered caching to meet widest set of requirements

Step one - Conduct deep analysis of your IOPs and Latency requirements. Step two - Triangulate these requirements with the most optimum architecture. Step three - Deploy a solution that meets today’s and tomorrow’s needs.

slide-25
SLIDE 25

Q&A and Further Information

SvSAN Product Information

Product Options SvSAN Editions Standard & Advanced SvSAN license capacity 2, 6, 12 and unlimited TBs License entitlement 2 mirrored servers Maintenance and support Platinum - 24x7 / Gold - 9x5

For further information, please contact: sales@stormagic.com Resources: http://stormagic.com/resources-page/ Whitepapers, datasheets, webinars/videos, case studies This webinar and whitepaper will be posted on Monday 7th May Download your free trial of SvSAN

stormagic.com/trial

Performance analytics: Service Engagement

  • StorMagic Solutions Architect
  • Analyses actual workload
  • 100% clarity on IOPs and latency needs
  • Triangulate your storage architecture
  • 50% discount for applications received by 30/5/17