High Availability with a minimal Cluster 29. October 2009 Thorsten - - PowerPoint PPT Presentation

high availability with a minimal cluster
SMART_READER_LITE
LIVE PREVIEW

High Availability with a minimal Cluster 29. October 2009 Thorsten - - PowerPoint PPT Presentation

High Availability with a minimal Cluster 29. October 2009 Thorsten Frauf Availability Engineering Sun Microsystems GmbH Open HA Cluster Agenda Motivation Open HA Cluster 2009.06 Minimal HA Configuration Weak Membership COMSTAR / iSCSI


slide-1
SLIDE 1

Open HA Cluster

High Availability with a minimal Cluster

  • 29. October 2009

Thorsten Früauf Availability Engineering Sun Microsystems GmbH

slide-2
SLIDE 2

High Availability with a minimal Cluster 2

Open HA Cluster

Agenda

Motivation Open HA Cluster 2009.06 Minimal HA Configuration

Weak Membership COMSTAR / iSCSI / ZFS Crossbow IPS

Live Demo References

slide-3
SLIDE 3

High Availability with a minimal Cluster 3

Open HA Cluster

Why care about High Availability?

Computer systems provide services:

Web Services, Databases, Business Logic, File Systems, etc.

Downtime is costly

Services should be available as close as possible to 100%

  • f the time

Failures are inevitable:

Software Bugs Hardware components People and Processes Natural Disaster Terrorism

slide-4
SLIDE 4

High Availability with a minimal Cluster 4

Open HA Cluster

The Goal of High Availability

HA Cluster automate the recovery process from inevitable failures to minimize downtime and cost.

slide-5
SLIDE 5

High Availability with a minimal Cluster 5

Open HA Cluster

You don't want your users to see this...

slide-6
SLIDE 6

High Availability with a minimal Cluster 6

Open HA Cluster

Methods to implement HA

Redundant hardware

physical nodes, network adapters, network paths, storage, storage paths, etc.

Software monitoring

physical nodes, applications, network paths, storage paths, etc.

Failover to secondary hardware when problems detected

slide-7
SLIDE 7

High Availability with a minimal Cluster 7

Open HA Cluster

Perceptions of HA Clusters

complex complicated heavyweight difficult to install difficult to use requires special hardware expensive Perceptions not completely unfounded...

slide-8
SLIDE 8

High Availability with a minimal Cluster 8

Open HA Cluster

Typical HA Cluster Hardware config

Two or more physical machines Four or more network adapters on each machine Dedicated interconnects between nodes Shared disk storage

multi-homed disks or network-attached storage

Redundant storage paths from each node Quorum arbitration device etc.

slide-9
SLIDE 9

High Availability with a minimal Cluster 9

Open HA Cluster

Typical HA Cluster software components

Heartbeats Membership Distributed configuration repository Service management Cluster-private networking layer Global file system Network load-balancing etc.

slide-10
SLIDE 10

High Availability with a minimal Cluster 10

Open HA Cluster

Solaris Cluster Architecture

Heartbeats Membership Quorum Disk Fencing Global File Service Failover File service Scalable Service

Software Load Balancing

Global Network Service

Provides Global IP address with failure protection

Monitoring

HA Failover Service

  • Application failover
  • Within nodes or container
  • Containers failover
  • Failover IP address
slide-11
SLIDE 11

High Availability with a minimal Cluster 11

Open HA Cluster

Campus / Metro Cluster

Up to 400KM

slide-12
SLIDE 12

High Availability with a minimal Cluster 12

Open HA Cluster

Solaris Cluster Geographic Edition

Replication Sun StorEdge Availability Suite 4.0 EMC SRDF HDS Truecopy Dataguard for Oracle RAC Script based plugin for MySQL Primary Site Backup Site 9 & 10 Oracle RAC support

Optional Storage Network Optional Heartbeat Network

slide-13
SLIDE 13

High Availability with a minimal Cluster 13

Open HA Cluster

Re-evaluate HA Cluster Complexity

Many use cases (incl. SLA) require all the hardware and software in traditional HA Clusters … but not everything... approach: “good enough” is sufficient as well! Configure, install, and use only the hardware and software components you actually need

slide-14
SLIDE 14

High Availability with a minimal Cluster 14

Open HA Cluster

Goals of Project Colorado

Provide a lightweight, modular, cluster that can run on minimized hardware configurations What has been possible before should still be possible to configure

as much as OpenSolaris allows

slide-15
SLIDE 15

High Availability with a minimal Cluster 15

Open HA Cluster

How to Get There

Port Open HA Cluster source to work with OpenSolaris Add hardware minimization features Leverage OpenSolaris Image Packaging System (IPS) for software modularity and extensibility

Analyze all package dependencies

slide-16
SLIDE 16

High Availability with a minimal Cluster 16

Open HA Cluster

Development context

Open HA Cluster Source Code Solaris Cluster 3.2 12/06

  • 27. Juni 2007

SCX 07/07 runs with SXCE build 68 Solaris Cluster 3.2 02/08 (Update 1) current development SCX 10/07 runs with SXCE build 70b SCX 02/08 runs with SXCE build 79b SCX 06/08 runs with SXCE build 86 SCX 09/08 runs with SXCE build 97 SCX 12/08 runs with SXCE build 101b OHAC 2009.06 runs with OpenSolaris build 111b Open HA Cluster Binary Distribution Solaris Cluster 3.2 01/09 (Update 2) SCX = Solaris Cluster Express OHAC = Open HA Cluster SXCE = Solaris Express Community Edition

slide-17
SLIDE 17

High Availability with a minimal Cluster 17

Open HA Cluster

Solaris Express vs. OpenSolaris

ON (= OS/Net) Java X window system Installer (old & some news) Installer (new) IPS (default) SVR4 package (legacy) encumbered code: CDE / JDS Motif binaries from closed source Zones: native (sparse & full root) lx Zones: ipkg (full root) lx KSH93 KSH88 (default) KSH93 (default) webconsole / JATO Crossbow SVR4 package (default) gnome Solaris Express (Nevada) Binary distribution of usr/src und usr/closed not freely redistributable SPARC x86/x64 OpenSolaris 200X.Y Binary distribution on LiveCD freely redistributable packages (pkg.opensolaris.org Repo) not freely redistributable packages (pkg.sun.com Repo) COMSTAR ZFS root UFS root Motif libs

slide-18
SLIDE 18

High Availability with a minimal Cluster 18

Open HA Cluster

Open HA Cluster 2009.06 (Colorado-I)

Runs on OpenSolaris 2009.06 (SPARC & x86/x64) Many features from Solaris Cluster 3.2 available Free to use (without support)

Support subscriptions available

Installation from IPS package repository

https://pkg.sun.com/opensolaris/ha-cluster

Source is open and freely available at

http://www.opensolaris.org/os/community/ha- clusters/

slide-19
SLIDE 19

High Availability with a minimal Cluster 19

Open HA Cluster

Open HA Cluster 2009.06 Agents

Apache Webserver Apache Tomcat MySQL GlassFish NFS Generic Data Service (GDS) DHCP DNS Kerberos Samba HA Containers

ipkg Zones

slide-20
SLIDE 20

High Availability with a minimal Cluster 20

Open HA Cluster

Hardware Minimization

Using local disks as “Poor man's shared storage” with COMSTAR iSCSI and ZFS Using Crossbow VNICs for private cluster traffic over public network “Weak membership” (preview-only feature) Taken together, allow any two-nodes on the same IP subnet to form a functional cluster.

slide-21
SLIDE 21

High Availability with a minimal Cluster 21

Open HA Cluster

Minimale HA Konfiguration

Heartbeats

  • ver VNICs

Weak Membership local Storage exported as iSCSI targets Failover ZFS File system mirroring iSCSI targets

Monitoring

HA Failover Service

  • Application failover
  • IP address failover
  • HA ZFS failover

external ping targets IPS package repositories

slide-22
SLIDE 22

High Availability with a minimal Cluster 22

Open HA Cluster

Technologies useable for Minimization

Weak Membership Software Quorum Quorum Server Optional Fencing HA ZFS COMSTAR / iSCSI IPsec Crossbow IPS VirtualBox

for training and development

slide-23
SLIDE 23

High Availability with a minimal Cluster 23

Open HA Cluster

HA Cluster „Strong Membership“

Use concept of quorum to ensure cluster consistency in the presence of partitions in space and time

Partition in space (network partition) can cause split-brain Partition in time can cause amnesia

Two-node cluster requires third arbitration device in case of partitions

Typically hardware disk or software quorum server

slide-24
SLIDE 24

High Availability with a minimal Cluster 24

Open HA Cluster

Weak Membership (preview feature)

Run a two-node cluster without a quorum device External “ping target” used as “health check” to arbitrate in case of split-brain

Worst-case, both nodes stay up and provide service OpenSolaris Duplicate Address Detection (DAD) can mitigate somewhat

Places importance of availability above data integrity

Can lead to data loss

slide-25
SLIDE 25

High Availability with a minimal Cluster 25

Open HA Cluster

Why use Weak Membership?

Read-only or read-mostly applications Availability is more important than data integrety

the SLA matches (solution is “good enough”)

Test Cluster with limited resources Demos Development Training

slide-26
SLIDE 26

High Availability with a minimal Cluster 26

Open HA Cluster

iSCSI Storage

IP-based storage networking standard Initiators (clients) send SCSI commands to targets (storage devices) over regular IP networks Alternative to NAS, SAN and DAS The OpenSolaris Common Multiprotocol iSCSI Target (COMSTAR) implements the iSCSI protocol

slide-27
SLIDE 27

High Availability with a minimal Cluster 27

Open HA Cluster

COMSTAR iSCSI for OHAC 2009.06

Each node exports directly-attached disk as iSCSI target Nodes access both disks through iSCSI initiators Mirrored zpool built on top of the two disks HAStoragePlus imports zpool on node hosting the services that need it If one node goes down, local half of mirror still available and accessible from other node

slide-28
SLIDE 28

High Availability with a minimal Cluster 28

Open HA Cluster

COMSTAR iSCSI Configuration

Node 1 Node 2 Mirrored Zpool

Local Disk Local Disk iSCSI Initiator iSCSI Initiator iSCSI Target iSCSI Target

clprivnet0

slide-29
SLIDE 29

High Availability with a minimal Cluster 29

Open HA Cluster

iSCSI with MPxIO and OHAC 2009.06

OpenSolaris Storage Multipathing

Multiple redundant paths to storage Operates above transport layer

Configure MPxIO over redundant private interconnects to provide multiple paths to iSCSI targets (needs strong membership) Benefits of MPxIO with iSCSI in OHAC

Supports RDMA (infiniband) Round-robin load balancing for increased throughput

slide-30
SLIDE 30

High Availability with a minimal Cluster 30

Open HA Cluster

iSCSI with MPxIO Configuration

Node 2 iSCSI Initiator iSCSI Target Node 1 Mirrored Zpool iSCSI Initiator iSCSI Target Local Disk Local Disk

slide-31
SLIDE 31

High Availability with a minimal Cluster 31

Open HA Cluster

Crossbow VNICs

Virtual Network Interface Card (VNIC) Pseudo-network interface VNICs configured on a physical interface # dladm create-vnic -l e1000g0 vnic1

slide-32
SLIDE 32

High Availability with a minimal Cluster 32

Open HA Cluster

Crossbow VNICs with OHAC 2009.06

Cluster private interconnect can use VNICs as endpoints instead of physical adapters Work over dedicated physical adapter or public adapter Use IPsec to protect cluster-private traffic

Though DLPI heartbeats not protected

Resource consolidation: Share physical adapters Easier setup: No dedicated private physical adapters and cabling required

slide-33
SLIDE 33

High Availability with a minimal Cluster 33

Open HA Cluster

OHAC 2009.06 Software Modularization

ha-cluster-full group package

Contains core framework, wizards, agents, man pages, l10n, … (everything)

ha-cluster-minimal group package

Only core framework Add agents, wizards, l10n, man pages, telemetry, etc. individually as needed

Install quorum server and agent builder without core framework

slide-34
SLIDE 34

High Availability with a minimal Cluster 34

Open HA Cluster

Why Minimal Installation is Useful

Minimizing resources (you don't pay for what you don't need)

Disk space Network download bandwidth etc.

Security minimization Minimizing administrative overhead

Both initial and ongoing

slide-35
SLIDE 35

High Availability with a minimal Cluster 35

Open HA Cluster

Installing Open HA Cluster 2009.06

Accept terms of use at pkg.sun.com and download key and certificate to /var/pkg/ssl Set ha-cluster publisher (on all nodes):

# pkg set-publisher \

  • k /var/pkg/ssl/Open_HA_Cluster_2009.06.key.pem \
  • c /var/pkg/ssl/Open_HA_Cluster_2009.06.certificate.pem \
  • O https://pkg.sun.com/opensolaris/ha-cluster/ ha-cluster
slide-36
SLIDE 36

High Availability with a minimal Cluster 36

Open HA Cluster

Installing OHAC 2009.06 (cont)

Install the cluster software (on all nodes): # pkg install ha-cluster-full Configure the cluster (on one node): # /usr/cluster/bin/scinstall

slide-37
SLIDE 37

High Availability with a minimal Cluster 37

Open HA Cluster

Live Demo

Toshiba M10

4 GB main memory 160 GB hard disk OpenSolaris 2009.06 Open HA Cluster 2009.06 VirtualBox 3.0.8

slide-38
SLIDE 38

High Availability with a minimal Cluster 38

Open HA Cluster

References (1)

Open HA Cluster 2009.06 Documentation

http://www.opensolaris.com/learn/features/availability/ http://docs.sun.com/app/docs/prod/open.ha.cluster~2509.1#hic

Solaris Cluster Blog

http://blogs.sun.com/SC

White Paper: Running Open HA Cluster on OpenSolaris with VirtualBox

http://opensolaris.org/os/project/colorado/files/Whitepaper- OpenHAClusterOnOpenSolaris-external.pdf

slide-39
SLIDE 39

High Availability with a minimal Cluster 39

Open HA Cluster

References (2)

HA Clusters Community Group

http://opensolaris.org/os/community/ha-clusters/

Project Colorado

http://opensolaris.org/os/project/colorado/

Project Image Packaging System (IPS)

http://opensolaris.org/os/project/pkg/

Project Crossbow (VNICs)

http://opensolaris.org/os/project/crossbow/

Project COMSTAR (iSCSI)

http://opensolaris.org/os/project/comstar/

slide-40
SLIDE 40

Open HA Cluster

Thank You! Questions?

thorsten.frueauf@sun.com http://blogs.sun.com/tf

Thorsten Früauf Availability Engineering Sun Microsystems GmbH

Open HA Cluster