Branch offices and SMBs: choosing the right hyperconverged solution - - PowerPoint PPT Presentation

branch offices and smbs choosing the right hyperconverged
SMART_READER_LITE
LIVE PREVIEW

Branch offices and SMBs: choosing the right hyperconverged solution - - PowerPoint PPT Presentation

Branch offices and SMBs: choosing the right hyperconverged solution Presenters: Howard Marks Chief Scientist, DeepStorage.net Luke Pruen Director of Technical Services, StorMagic Infrastructures for the Remote Office Howard Marks -


slide-1
SLIDE 1

Presenters: Howard Marks – Chief Scientist, DeepStorage.net Luke Pruen – Director of Technical Services, StorMagic

Branch offices and SMBs: choosing the right hyperconverged solution

slide-2
SLIDE 2

Infrastructures for the Remote Office

Howard Marks - @DeepStorageNet

slide-3
SLIDE 3

Your not so Humble Speaker

  • 30 years of consulting and

writing for trade press

– Now at TechTarget storage sites

  • Chief Scientist DeepStorage, LLC.

– Independent test lab and analysts

  • Co-Host Greybeards on Storage

podcast

  • Executive director The Other Other

Operation

Hmarks@DeepStorage.Net @DeepStorageNet

slide-4
SLIDE 4

Remote Office IT Challenges

  • Not just offices:

– Plants – Stores – Restaurants – Bank/brokerage branches

  • Everything’s limited

– IT Staff – Budget – Connectivity – Space

  • Outside the range of 4hr response
slide-5
SLIDE 5

Let’s Send Everything to the Cloud

  • Limited connectivity

– Low bandwidth – Not redundant – Now critical to mission of office

  • Hardware interfaces

– Time clocks – POS terminals – Bar code/inventory

  • Line of business apps not web
slide-6
SLIDE 6

Conventional Infrastructure

  • Typically 4-5 workloads

– Plus some batch/transfer jobs

  • Two virtualization hosts

– Plenty of horsepower for failover

  • Too complex

– External disk array – Fibre Channel switches

  • Too expensive
  • Too Fragile
slide-7
SLIDE 7

Enter Hyperconvergence

  • Software turns host disk into

shared volume

  • Replicate for resilience
  • SSD for acceleration
  • Lower cost
  • Major vendors in the data

center

slide-8
SLIDE 8

Remote Office Requirements

  • Lightweight storage VM

– Run on 1 socket server – Hypervisor, Windows etc. $5000/socket

  • Scale down

– 2 nodes

  • Enhanced resiliency

– 3-4 day MTTR common – Don’t expose data for that long

  • Management options
slide-9
SLIDE 9

Split Brains and Witnesses (go out in the noonday sun)

  • In split-brain both nodes of a 2-Node

cluster remain active

  • Some HCI solutions have required 3

nodes

– Any 2 nodes make a quorum

  • Witness makes quorum but isn’t node

– Witness should not run on a node – Witness at HQ, on PC, other options

slide-10
SLIDE 10

Summary

  • Remote sites have unique needs
  • Conventional infrastructure – too many too’s
  • HCI could be a good solution

– Be careful of Datacenter products

  • Look for:

– Lightweight Software – High resiliency after failures – Witness flexibility – Management flexibility

slide-11
SLIDE 11
slide-12
SLIDE 12

Luke Pruen – Director of Technical Services, StorMagic

SvSAN Overview

slide-13
SLIDE 13

StorMagic SvSAN: Overview

SvSAN turns the internal disk, SSD and memory of industry standard servers into highly available shared storage

slide-14
SLIDE 14

SvSAN - Features

  • SvSAN VSA – Virtual Storage Appliance

̶ Lightweight software defined storage platform

  • Synchronous Mirroring

̶ Synchronously mirror your storage between as little as two hosts for high availability and protection of your storage

  • Stretch Cluster Support

̶ Mirror storage across separate sites to protect against major outages

  • Centralized, Simplified Management

̶ Control all your SvSAN clusters from one place with simplified management tools

  • Remote Shared Witness - NSH

̶ Flexible cluster witness keeps your mirrored storage in sync and highly available

  • Performance Caching Features

̶ Utilized SSD and system memory to boost your performance

  • Scale Flexibly

̶ Scale-Up and Scale-Out

slide-15
SLIDE 15
  • Tie-breaker service for SvSAN mirrors
  • Prevents data inconsistency AKA ‘split brain’
  • This ensures

̶ In the event of a single failure there is no interruption in service ̶ In the event of multiple failures there is no corruption or loss of data

  • Local or Remote

̶ Supported as Windows service, Linux daemon, packaged VM, Raspbian (Raspberry Pi) ̶ Withstands 3000ms latency ̶ Up to 20% packet loss ̶ 9kbs bandwidth required per SvSAN mirror

  • Single NSH instance for 1000s of mirrors across clusters

VSA

Hypervisor

VM

VSA

Hypervisor

SvSAN Synchronous Mirroring

VM VM VM

WAN NSH

SvSAN - Neutral Storage Host (witness)

slide-16
SLIDE 16
  • Tie-breaker service for SvSAN mirrors
  • Prevents data inconsistency AKA ‘split brain’
  • This ensures

̶ In the event of a single failure there is no interruption in service ̶ In the event of multiple failures there is no corruption or loss of data

  • Local or Remote

̶ Supported as Windows service, Linux daemon, packaged VM, Raspbian (Raspberry Pi) ̶ Withstands 3000ms latency ̶ Up to 20% packet loss ̶ 9kbs bandwidth required per SvSAN mirror

  • Single NSH instance for 1000s of mirrors across clusters

SvSAN - Neutral Storage Host (witness)

slide-17
SLIDE 17
  • Tie-breaker service for SvSAN mirrors
  • Prevents data inconsistency AKA ‘split brain’
  • This ensures

̶ In the event of a single failure there is no interruption in service ̶ In the event of multiple failures there is no corruption or loss of data

  • Local or Remote

̶ Supported as Windows service, Linux daemon, packaged VM, Raspbian (Raspberry Pi) ̶ Withstands 3000ms latency ̶ Up to 20% packet loss ̶ 9kbs bandwidth required per SvSAN mirror

  • Single NSH instance for 1000s of mirrors across clusters

SvSAN - Neutral Storage Host (witness)

slide-18
SLIDE 18
  • Tie-breaker service for SvSAN mirrors
  • Prevents data inconsistency AKA ‘split brain’
  • This ensures

̶ In the event of a single failure there is no interruption in service ̶ In the event of multiple failures there is no corruption or loss of data

  • Local or Remote

̶ Supported as Windows service, Linux daemon, packaged VM, Raspbian (Raspberry Pi) ̶ Withstands 3000ms latency ̶ Up to 20% packet loss ̶ 9kbs bandwidth required per SvSAN mirror

  • Single NSH instance for 1000s of mirrors across clusters

SvSAN - Neutral Storage Host (witness)

slide-19
SLIDE 19

Witness - SvSAN vs vSAN 100 site example

StorMagic SvSAN VMware vSAN ROBO Number of remote nodes 2 2 Number of witness nodes required 1:1000 100 sites: 1 witness 1:1 100 sites: 100 witnesses Witness node vCPU 1vCPU per witness 100 sites: 1vCPU required 2vCPU per witness 100 sites: 200 vCPU required Witness node memory 512MB per witness 100 sites: 512MB 8GB per witness 100 sites: 800GB Witness latency allowance <3000ms RTT <500ms RTT Witness Bandwidth 9 Kbps per mirrored datastore 100 sites: 0.9 Mbps Per 10 VMs 0.24 Mbps 100 sites: 24 Mbps Virtual SAN CPU 1 vCPU per host 10% of hosts total CPU Virtual SAN Memory 1GB per host 2GB per host using SSD caching Optional memory for memory caching Depends on the number of disk groups. Each host must contain a minimum of 32GB to support 7 disk groups

Reference: VMware Virtual SAN Bandwidth Sizing Guide https://www.vmware.com/files/pdf/products/vsan/vmware-virtual-san-6.1-stretched-cluster-bandwidth-sizing.pdf

slide-20
SLIDE 20

SvSAN - Intelligent automated read caching & tiering

Intelligent read caching algorithm

  • All read I/Os are monitored and analyzed
  • Most frequently used data – “Hot” data
  • Cache tiers are populated based on access frequency

Tiering

  • RAM: Most frequently accessed data
  • SSD/Flash: Next most frequently accessed data
  • HDD: Infrequently accessed data – “Cold” data

Sizing

  • Assign cache sizes to meet requirements
  • Grow caches as working sets change
  • Use any combination of Memory, SSD/Flash and Disk

Play to the strengths

  • Play to the strengths of all mediums
  • Memory Highest IOPS
  • SSD/Flash Magnetic drives providing lower price per GB
slide-21
SLIDE 21

SvSAN - management & integration

  • Centralized Management & Monitoring from the Datacenter
  • vCenter StorMagic Integration

̶ StorMagic Dashboard ̶ Single/Multi VSA Deploy ̶ NSH Deploy ̶ VSA Restore ̶ Create, expand and migrate storage

  • Hyper-V Integration

̶ StorMagic Deployment Wizard

  • Monitoring

̶ SNMP v2 & v3 ̶ SMTP ̶ System Center Operations Manager

  • Scripting tool box

̶ Powershell module ̶ Deployment, configuration, firmware upgrades ̶ Plugin script generation

slide-22
SLIDE 22

SvSAN - Summary

  • Lightweight Solution

̶ Architected for smallest possible footprint ̶ Minimal resources required for VSAs and Witness ̶ Powerful features to drive higher performance requirements

  • Eliminate Downtime

̶ Synchronously mirrored storage across multiple servers ̶ No single point of failure or maintenance downtime ̶ Upgrade & replace hardware with no impact

  • Flexible witness

̶ Run remotely or locally ̶ Supported on a wide options of platforms ̶ Lowest requirements of any solution

  • Centrally Deploy, Manage & Monitor

̶ Centrally deployed ̶ Automated though scripting ̶ Central management of thousands of locations

slide-23
SLIDE 23

Q&A and Next Steps

SvSAN Product Information

Product Options SvSAN license 2, 6, 12 and unlimited TBs License entitlement 2 mirrored servers Maintenance and support Platinum - 24x7 / Gold - 9x5

For further information, please contact: sales@stormagic.com Further Reading: An overview of SvSAN - http://stormagic.com/svsan/ SvSAN Data Sheet - http://stormagic.com/svsan-data-sheet/ SvSAN White Paper - http://stormagic.com/svsan-6/ Download your free trial of SvSAN

stormagic.com/trial