Next-generation Magnetic Recording CSCI 333 April 8, 2019 Last - - PowerPoint PPT Presentation

next generation magnetic recording
SMART_READER_LITE
LIVE PREVIEW

Next-generation Magnetic Recording CSCI 333 April 8, 2019 Last - - PowerPoint PPT Presentation

Next-generation Magnetic Recording CSCI 333 April 8, 2019 Last Class: SSDs Interface: Read from pages As many times as we want Program (write to) pages Once -> then need to erase before rewriting Limited endurance


slide-1
SLIDE 1

Next-generation Magnetic Recording

CSCI 333 April 8, 2019

slide-2
SLIDE 2

Last Class: SSDs

  • Interface:
  • Read from pages
  • As many times as we want
  • Program (write to) pages
  • Once -> then need to erase before rewriting
  • Limited endurance -> need to wear level
  • Erase whole blocks
  • Erasing is slow
  • Need to perform GC -> migrate live data
  • FTLs wear many hats
  • L2P page translation, wear leveling, GC, ECC, …
slide-3
SLIDE 3

This Class: “Spinning Rust”

  • (Abbreviated recap) Hard Disk Drives
  • Basic Design/Geometry
  • Performance characteristics
  • Shingled Magnetic Recording
  • Concepts and interface
  • Position in the storage stack
  • Other SMR Interfaces/Opportunities
  • Skylight
  • IMR
slide-4
SLIDE 4

Next Class

  • Filters. Why shift schedule?
  • Hopefully inspire final project ideas
  • Original DAM model paper is rough… looking for more

interesting/clear presentation of material

  • For next class: read the Bloom filter paper
  • Optionally read the quotient filter paper
  • Optionally read the cuckoo filter paper
  • Goals:
  • Understand/articulate problem(s) that filters solve
  • Describe the high-level design and parameters
slide-5
SLIDE 5

Hard Disk Drives (HDDs)

  • High capacity, low cost
  • Predictable performance
  • “Unwritten contract”: LBAs near each other are

more efficient to access than LBAs that are far away

slide-6
SLIDE 6

HDDs

Disk Head (seeks in/out) Platters (rotate) Sector (unit of transfer) Tracks (concentric circles)

slide-7
SLIDE 7

Performance Observations

  • Setup (placing the disk head) is expensive O(10 ms)
  • seeking to target track
  • Up to a full rotational delay to locate target sector
  • Once the disk head is in place, data transfer is

quite fast O(100s MiB/s)

slide-8
SLIDE 8

213 216 219 222 225 228 2−2 20 22 24 26 Read size (bytes) Effective Bandwidth (MB/sec) HDD

Performance Goal: build a system where data is written sequentially (i.e., no random writes)

slide-9
SLIDE 9

Keeping HDDs Relevant

  • HDDs compete on $/GiB, not performance
  • As capacity goes up, $/GiB down
  • Problem:
  • Capacity gains traditionally result of reduced

track width to increase density

  • Physical limits restrict our ability to shrink

tracks further

  • We’re stuck… unless?
slide-10
SLIDE 10

[https://blog.seagate.com/craftsman-ship/hamr-next-leap-forward-now/]

slide-11
SLIDE 11

Shingled Magnetic Recording (SMR)

  • Increases HDD density by overlapping tracks

Perpendicular Magnetic Recording

slide-12
SLIDE 12

Shingled Magnetic Recording (SMR)

  • Increases HDD density by overlapping tracks

Perpendicular Magnetic Recording Shingled Magnetic Recording

  • Insight: Read head is more precise than write head
  • Technique: Overlap next track, but leave enough
  • f “lower” track visible for safe reading
slide-13
SLIDE 13

SMR Introduces Challenges

  • Writing data becomes harder
  • No random writes
  • No overwrites
  • Must garbage collect to reclaim space
slide-14
SLIDE 14

No Random Writes

If we don’t write to zones append-only, we could lose data

slide-15
SLIDE 15

No Overwrites

Must perform out-of-place updates, or suffer a read-modify-write of entire zone

slide-16
SLIDE 16

Garbage Collection

  • 1. Copy live data from source to destination
  • 2. Reclaim old zone
slide-17
SLIDE 17

Garbage Collection

  • 1. Copy live data from source to destination
  • 2. Reclaim old zone
slide-18
SLIDE 18

Recall HDD Observations

  • Problem: Seeking is slow
  • Solution: perform large sequential I/Os

Takeaway: HDD performance optimizations translate into SMR correctness

slide-19
SLIDE 19

Persistent Storage File System Application

user space OS kernel

Simplified Storage Stack

data = read(LBA), write(data,LBA)

Question: who enforces the SMR write constraints?

SMR

slide-20
SLIDE 20

Drive Managed vs. Host Managed

File System

SMR Zoned Access

SMR Translation Layer (STL)

Software Firmware

+ Easy to Deploy


  • Limited HW resources

+ Flexible
 + Shares host resources

SMR Translation Logic

File System

Read and write LBAs Read and write LBAs Read LBAs, write to zones

slide-21
SLIDE 21

Zoned Block Commands

  • Conventional Zones
  • Random write capabilities of “normal” disks
  • Sequential-write-required zones
  • Query zone status
  • Append blocks to zone’s write pointer
  • Reset zone write pointer (reclaim space)

Conventional zone(s) Sequential write required zones

slide-22
SLIDE 22

SMR Opportunities

  • Other SMR interfaces have been proposed
  • Caveat Scriptor
  • Configurable zone layouts (Flex) [Feldman ’18]
  • Interlaced Magnetic Recording (IMR)
  • Combines HAMR and overlapping tracks
slide-23
SLIDE 23

Caveat Scriptor

Basic Idea:

  • Drive characteristics are exposed to the user
  • User can write anywhere, but data may be lost

[Kadekodi ’15 HotStorage]

slide-24
SLIDE 24

Interlaced Magnetic Recording

[Feldman ’18 ;login:]

r R

Figure 3: Depiction of interlaced track recording

[Hwang ’16 Transactions on Magnetics]

slide-25
SLIDE 25

Magnetic Recording

(a) CMR (b) SMR (c) IMR

Figure 1: Track layout for CMR, SMR, and IMR.

[Wu ’18 HotStorage]

slide-26
SLIDE 26

Open Questions

  • Translation layer design
  • Garbage collection schemes
  • Creating and using new interfaces
  • SMR-aware key-value stores
  • Integrating SMR maintenance with DS work
slide-27
SLIDE 27

Let’s Think About Designs

  • What are our options?
  • Static or dynamic?
  • What do you think is done in practice (“Archive”

DM-SMR drives available at big box stores)?

  • Skylight designed & performed benchmarks to

tease out drive parameters