HPSS Treefrog Introduction HUF 2017 - - PowerPoint PPT Presentation

hpss treefrog introduction
SMART_READER_LITE
LIVE PREVIEW

HPSS Treefrog Introduction HUF 2017 - - PowerPoint PPT Presentation

HPSS Treefrog Introduction HUF 2017 http://www.hpss-collaboration.org Disclaimer Forward looking information including schedules and future software reflect current planning that may change and should not be taken as commitments by IBM or the


slide-1
SLIDE 1

http://www.hpss-collaboration.org

HPSS Treefrog Introduction

HUF 2017

slide-2
SLIDE 2

http://www.hpss-collaboration.org

Disclaimer

Forward looking information including schedules and future software reflect current planning that may change and should not be taken as commitments by IBM or the other members of the HPSS collaboration.

2

slide-3
SLIDE 3

http://www.hpss-collaboration.org

HPSS Treefrog Goals

Manage and share data across the life

  • f your mission’s projects,

procurements, infrastructure, deployment, user access, and staffing cycles. Store, protect, and error correct project data across a wide variety of local and remote classic and cloud storage products and services. Effectively exploit and scale tape and

  • ther high latency storage by using

data containers to group and store files and data objects!

3

slide-4
SLIDE 4

http://www.hpss-collaboration.org

A Single User Namespace

Managed across industry storage devices and solutions called storage endpoints:

§ Cloud § HSMs including HPSS § Optical § Tape § File system § Disk § SSD

Managed across data repositories

§ Storage endpoints provide real storage for data repositories. § Repositories are wholly contained inside a storage endpoint.

4

slide-5
SLIDE 5

http://www.hpss-collaboration.org

Manage Data by Project

§ Projects provide the nexus between data management and data organization. § Administrators manage project policies including

§ storage quotas

§ Storage access § Service limits § Access authorization

§ Users store data within the projects and group data within data containers (called managed data sets)

§ Data are share amongst project members (allowed users)

§ Project members will have different roles:

§ Owner, reader, writer, modify, delete

§ Data will be owned by the project.

§ Insures data will always have an owner. § Allows for easy on and off boarded of users.

5

slide-6
SLIDE 6

http://www.hpss-collaboration.org

Policy Defined Storage Management

§Policies determine how and where data are stored. §Make multiple copies of data:

§ At ingest from the golden copy § After a delay from a managed copy

§Control data recall:

§ Assign primary recall copy § Assign failover copies § Block recall of copies from storage endpoint requiring administrator authorization

6

slide-7
SLIDE 7

http://www.hpss-collaboration.org

Smart Data Storage

§Manage data containers not individual data

  • bjects and files.

§Grouped data will be stored as an immutable collection of files or objects called a managed data set. §As a bonus, grouping data benefits high latency storage.

§ Decreases the number of tape syncs. § Allows for all data to be recalled with a single IO.

§Data will be grouped into date sets using a data retention format. §The Treefrog interface will make grouping data simple.

7

slide-8
SLIDE 8

http://www.hpss-collaboration.org

Parallel Data Transfer

§Managed Data Sets may be broken into smaller fragments.

§ Based on storage policy settings. § Fragments are contiguous sections

  • f Treefrog managed data set that

are distributed across repositories.

§Maximum degree of parallelism will be based on configuration.

8

Transfer Transfer Transfer Huge Object Fragment #1 Huge Object Fragment #2 Large Object Small Small Dataset Fragment #1 Dataset Fragment #2 Dataset Fragment #3 Manifest Huge Object Large Object Small Small Repository 1 Repository 2 Repository 3

slide-9
SLIDE 9

http://www.hpss-collaboration.org

Data Redundancy via Erasure Coding

§Parity fragments will be generated based on storage policy settings.

§ The number of fragments that may be recovered will be based on the number of parity fragments created.

9

Transfer Transfer Transfer Huge Object Fragment #1 Huge Object Fragment #2 Large Object Small Small Dataset Fragment #1 Dataset Fragment #2 Dataset Fragment #3 Dataset Parity Fragment Manifest Huge Object Large Object Small Small Repository 1 Repository 2 Repository 3 Repository 4

slide-10
SLIDE 10

http://www.hpss-collaboration.org

First copy

More About Storage Policies

§A copy of a data set may be:

§ Stored to a single repository § Fragmented to a single repository § Fragmented across multiple repositories

§Changing storage policies

  • nly moves data when

required.

10

Transfer Transfer Transfer Fragment #1 Fragment #2 Large Object Small Small Dataset Fragment #1 Dataset Fragment #2 Dataset Fragment #3 Dataset Parity Fragment Manifest Huge Object Large Object Small Small Repository Repository Repository Repository Second Copy Repository

slide-11
SLIDE 11

http://www.hpss-collaboration.org

Simple Insertion of New Storage Endpoints

§Copy agent based on Apache Jclouds Blobstore.

§ Copy agent interface will be extensible. § AWS, Google Cloud Storage, Azure, and Rackspace already supported. § HPSS interface is planned.

§Adding a storage endpoint will be as simple as adding a new Jclouds interface.

11

slide-12
SLIDE 12

http://www.hpss-collaboration.org

Data and Metadata Verification

§Each fragment will be stored with with a checksum. §Treefrog can verify both the metadata and data of managed data sets.

§ Administrators use storage policies to control the verification settings and subsequent overhead.

§Metadata Verification will verify the location, checksum, and size of each fragment in the repository match the value Treefrog has stored.

§ Metadata Verification will not access the data.

§Data Verification will verify the checksum of each fragment.

§ Data Verification may access the data. § Treefrog will use the built in verification on storage systems that have it. § Treefrog will stage fragments to verify checksum.

12

slide-13
SLIDE 13

http://www.hpss-collaboration.org

All of that in an Extreme Scale Architecture

§Scale-out design allows incremental horizontal growth by adding new servers and devices. §Load Balancing using HAProxy. §Agents may run at the client to take advantage of available processing power and reduce store and forwards.

13

slide-14
SLIDE 14

http://www.hpss-collaboration.org

But wait there’s more!!!

In addition HPSS Treefrog will: §Decrease software development delivery time. §Decrease software deployment time. §Enable user installation. §Increase timely access to trending technology. §Increase use of trending programming language skills and open software. §Avoid impact to on-going HPSS core services development.

14

slide-15
SLIDE 15

http://www.hpss-collaboration.org

Treefrog will be an HPSS Interface

15

Spectrum Scale Interface SwiftOnHPSS FUSE Filesystem Parallel FTP HPSS Client API RHEL Core Server & Mover computers Intel Power Massively scalable global HPSS namespace enabled by DB2 Extreme-scale high-performance automated HSM Disk Tape Block or Filesystem Disk Tiers Hardware Vendor Neutral Spectrum Scale Client API for 3rd party applications Enterprise Ÿ LTO Tape IBM Ÿ Oracle Ÿ Spectra Logic

HPSS Treefrog interface & services

Cloud, Object & File Storage and Services including LTFS

slide-16
SLIDE 16

http://www.hpss-collaboration.org

Treefrog will use Existing Technologies

Existing Products

§ Only configuration changes are required

Extendable Functionality

§ Open Source code or library

Treefrog Specific Code

§ Code specific to the Treefrog application § Requires from-scratch development

16

slide-17
SLIDE 17

http://www.hpss-collaboration.org

Treefrog will use Existing Technologies

17

slide-18
SLIDE 18

http://www.hpss-collaboration.org

Questions?

18