DSS Data & Storage Services Handling Big Data an overview of - - PowerPoint PPT Presentation

dss
SMART_READER_LITE
LIVE PREVIEW

DSS Data & Storage Services Handling Big Data an overview of - - PowerPoint PPT Presentation

DSS Data & Storage Services Handling Big Data an overview of mass storage technologies ukasz Janyst CERN IT Department GridKA School 2013 CH-1211 Genve 23 Switzerland Karlsruhe, 26.08.2013 www.cern.ch/i t Data & What is Big


slide-1
SLIDE 1

Data & Storage Services

CERN IT Department CH-1211 Genève 23 Switzerland

www.cern.ch/it

DSS

Handling Big Data

an overview of mass storage technologies Łukasz Janyst

GridKA School 2013 Karlsruhe, 26.08.2013

slide-2
SLIDE 2

Data & Storage Services

What is Big Data?

A buzzword typically used to describe data sets that are too big to be stored and processed by conventional means.

slide-3
SLIDE 3

Data & Storage Services

What can we do with it?

  • Analyze anonymous GPS records from 100 million

drivers to help home buyers determine optimal property locations

  • Analyze billions of credit card transactions to

protect from fraud

  • Find trends in the stock market moves
  • Decode human genome
slide-4
SLIDE 4

Data & Storage Services

What can we do with it?

Copy, store, and analyze the internet traffic for more or less questionable reasons

source: Wikipedia

The NSA’s Data Center in Utah - where all the PRISM data is supposedly handled

slide-5
SLIDE 5

Data & Storage Services

What can we do with it?

Process data from over 150 million sensors to find the Higgs boson

slide-6
SLIDE 6

Data & Storage Services

How big is it now?

  • CERN alone currently

stores over 100 petabytes

  • f data, with the

experiments producing around 30 PB annually

  • Facebook stores around

300 billion photos

  • Walmart processes 1

million client transactions per hour and has 2.5 PB

  • NSA builds a data

center capable of handling 12 exabytes

  • f data
slide-7
SLIDE 7

Data & Storage Services

How big is it going to be?

Grow by 50% each year 5200 GB/per person in 2020

International Data Corporation forecasts the digital universe to grow up to 40ZB (40 trillion gigabytes) by 2020.

slide-8
SLIDE 8

Data & Storage Services

What are the challenges?

Capture Store Transmit Process

Scope of this presentation

slide-9
SLIDE 9

Data & Storage Services

Multitude of solutions

slide-10
SLIDE 10

Data & Storage Services

Scaling

Storage systems need to be able to grow with the growing amount of data they handle. Scaling up Scaling out

slide-11
SLIDE 11

Data & Storage Services

Ideal properties

  • Consistent

– commits are atomic across the entire system, all clients see the same data at the same time

  • (Highly) Available

– remains operational at all times, requests are always answered (successfully or otherwise)

  • Tolerant to partitions

– network failures don’t cause inconsistencies, the system continues to operate correctly despite part of it being unreachable

Ideally all distributed systems should be:

slide-12
SLIDE 12

Data & Storage Services

Ideal properties - CAP

In reality however:

Pick two

Brewer’s CAP theorem

Consistent

C A

Available

P

Partition tolerant

slide-13
SLIDE 13

Data & Storage Services

Typical components

Object store Metadata system Protocol handlers Clients

Caveat: not necessarily logically separate - may be tightly coupled and interleaved

slide-14
SLIDE 14

Data & Storage Services

Object stores

Distributed Object Store - typically, a collection of uncorrelated flexible-sized data containers (objects) spread across multiple data servers

10c39527b893c798a93e8997772f65a8

(Hashed) key Data Blob

slide-15
SLIDE 15

Data & Storage Services

Object-node mapping

  • Algorithmic

– object location can be computed by the client or server using object name (key) and other inputs (cluster state) – Dynamo, CEPH

  • Manager/Cache

– manager node asks storage nodes for an object and caches the location for future reference (XRootD)

  • Index

– central entity (database) knows all the objects and their locations - most of “traditional” storage systems

slide-16
SLIDE 16

Data & Storage Services

Amazon Dynamo

  • The output space of the hash function is treated like

a ring

  • A node is assigned a random value denoting it’s

position in the ring

  • An object is assigned to a node by hashing the key

and walking the ring clockwise to find a node with a position larger than the key.

  • Replicas are stored to the

subsequent nodes

slide-17
SLIDE 17

Data & Storage Services

CEPH - RADOS

  • Each object is first mapped

to a placement group depending on the key and replication level

  • Placement groups are assigned to nodes and disks

using a stable, pseudo random mapping algorithm depending on cluster map (CRUSH).

  • Cluster map is managed by monitors and replicated

to storage nodes and clients.

slide-18
SLIDE 18

Data & Storage Services

Chunks, stripes, replicas

  • Replicas

– fairly simple, little metadata, performance – space issues: knapsack problem, expensive for archiving

  • Chunks

– solves the knapsack problem, distributes the load – still requires replicating for safety, much more metadata

  • Stripes

– relatively cheap archiving – more metadata, knapsack problem

For performance, space and safety reasons, the data may be distributed in many different ways

slide-19
SLIDE 19

Data & Storage Services

RAIN - Erasure codes

  • RAIN - redundant array of inexpensive nodes

(RAID implementation across nodes instead of disks)

  • Used to increase fault tolerance by adding extra

stripes correlating the info contained in the base stripes. Multiple techniques:

  • Hamming parity
  • Reed-Solomon error correction
  • Low-density parity-check
slide-20
SLIDE 20

Data & Storage Services

System topology

  • Spread replicas/chunks/stripes between failure

domains:

– Different disks, nodes, racks, switches, power supplies,

  • r entire data centers if possible
  • There is even some research on reducing heat

production by appropriately scheduling disk writes. Data placement needs to take into account system topology.

slide-21
SLIDE 21

Data & Storage Services

Data locality

  • Computation is most efficient when executed close

to data it operates on

  • Core concept of Hadoop, where nodes are typically

both storage and computation nodes

  • HDFS exposes interfaces allowing job schedulers

to dispatch jobs close to data: often the same node

  • r rack
slide-22
SLIDE 22

Data & Storage Services

Metadata services

  • POSIX-like trees

– familiar, used since decades – very hard to scale out

Group and organize objects into human-browsable groups, manage quotas, ownership, group attributes...

  • Accounts/Containers/Objects

– trivially scalable – may be hard to adjust legacy software

slide-23
SLIDE 23

Data & Storage Services

CEPH Filesystem

  • Runs on top of RADOS
  • Maps files and directories hierarchies to RADOS
  • bjects
  • Does dynamic tree partitioning
  • Metadata cluster may grow or contract - nodes are

stateless facades for accessing data in RADOS

slide-24
SLIDE 24

Data & Storage Services

Amazon S3 approach

  • Proprietary technology
  • Most likely it’s Dynamo with:

– HTTP interface – accounting system for billing – user authentication/authorization mechanisms

  • User accounts consist of buckets
  • Buckets are sets of files
  • account-bucket-file tuples are likely used as keys of

Dynamo objects

slide-25
SLIDE 25

Data & Storage Services

Backups-Archiving

  • Back up - copy important data to a

different kind of media - cheaper, more resilient to some natural phenomena Some data may need to be moved to cheaper or more reliable media.

  • Archive - move inactive data to a cheaper but safer

and possibly less available system Backups and archives of big data are likely even bigger data!

slide-26
SLIDE 26

Data & Storage Services

HSM and Tiers

  • Hierarchical Storage Manager -

transparently move data files between media types depending on how soon and how often they are accessed

  • Tier Storage - assigning different

categories of data (more/less critical, active/inactive, ...) to different kind of storage technologies, often manually

slide-27
SLIDE 27

Data & Storage Services

Clients

  • APIs

– direct use – integrating into commonly used tools as plug-ins

  • Mount points

– through widespread protocols (NFS, CIFS/Samba, ...) – dedicated drivers (typically FUSE)

  • Commandline and GUIs

– through widespread software (web browsers) – custom tools

slide-28
SLIDE 28

Data & Storage Services

Access requirements

  • User authentication

– is the system exposed to multiple users? – X.509, Kerberos, user/password, etc.

  • Transmission encryption

– are the channels secure or data sensitive – symmetric/asymmetric

  • Access patterns

– Is put/get enough? – Do we need partial reads, vector reads? – What about updates?

  • Filesystem/bucket operations

– list, stat, chown, etc.

slide-29
SLIDE 29

Data & Storage Services

Efficiency considerations

  • Latency

– support for logical streams and priorities – allow for multiple queries at once and provide a way of disambiguating responses

  • Bandwidth

– protocol overhead – compression (both headers and payload)

  • Server-side CPU intensiveness

– Do requests need to be decompressed? – Does it need to parse a ton of text/XML?

slide-30
SLIDE 30

Data & Storage Services

HTTP

  • HTTP is indisputable king of the cloud

communication protocols

– not because it’s particularly efficient, but because clients are built into pretty much every computer

  • There’s problems with it, mainly:

– does not allow out-of-order or interleaved responses

  • reasonable performance only for big, one-shot downloads

– protocol overhead:

  • many headers sent with each request, most of which are

redundant

slide-31
SLIDE 31

Data & Storage Services

HTTP 2.0 (a.k.a SPDY)

  • SPDY kind of a virtual transport protocol for HTTP.

– It does not really change the request or response format, just the way they are transported over the wire

  • Prioritization and multiplexing

– introduce multiple logical streams within one connection – responses may interleave each other

  • Header dictionary

– only changing headers are sent over the wire and they are compressed

An effort to fix the most important issues with HTTP

slide-32
SLIDE 32

Data & Storage Services

Outlook

  • Need for handling bigger and bigger data will likely

push out POSIX completely

– write-once read-many, put/get/remove – the main cost is in moving applications to use the new semantics

  • Metadata services are bottlenecks

– likely replaced by deterministic data placement

Prediction is very difficult, especially about the future.

  • Niels Bohr
slide-33
SLIDE 33

Data & Storage Services

Questions

Thanks for your attention! Questions? Comments?

slide-34
SLIDE 34

Data & Storage Services

Notes

  • Some interesting reading:

– Dynamo: Amazon’s Highly Available Key-value Store – RADOS: A Scalable, Reliable Storage Service for Petabyte-scale Storage Clusters

  • Most of the artwork in this presentation comes from:

Open Icon Library