Disk Management Disk Structure Disk Scheduling RAID Disk Block - - PDF document

disk management
SMART_READER_LITE
LIVE PREVIEW

Disk Management Disk Structure Disk Scheduling RAID Disk Block - - PDF document

CPSC 410 / 611 : Operating Systems Disk Management Disk Management Disk Structure Disk Scheduling RAID Disk Block Management Modern Secondary Storage Technologies Disk Management Disk Structure Disk Scheduling RAID


slide-1
SLIDE 1

CPSC 410 / 611 : Operating Systems Disk Management 1

Disk Management

  • Disk Structure
  • Disk Scheduling
  • RAID
  • Disk Block Management
  • Modern Secondary Storage Technologies

Disk Management

  • Disk Structure
  • Disk Scheduling
  • RAID
  • Disk Block Management
  • Modern Secondary Storage Technologies
slide-2
SLIDE 2

CPSC 410 / 611 : Operating Systems Disk Management 2

Disk Structure

cylinder

track sector Disk speed:

  • seek time :

head moves to correct track

  • rotational delay :

wait until sector is under head

  • transfer time :

transfer data between disk and memory

Disk Performance

Seek Time : Ts n = number of tracks traversed m = “track traversal time” s = startup time Rotational Delay (Latency Time): TR r = # revolutions per time unit Transfer Time: TT b = # bytes to be transferred N = number of bytes on track Disk Access Time: s n m T

S

+ × = r T

R

2 1 = rN b T

T = T R S

T T T T + + =

slide-3
SLIDE 3

CPSC 410 / 611 : Operating Systems Disk Management 3

Disk Management

  • Disk Structure
  • Disk Scheduling
  • RAID
  • Disk Block Management
  • Modern Secondary Storage Technologies

Disk Scheduling

Q: Does it pay off to think about scheduling policy in disk queue?

application application

kernel

device driver file system disk queue

Evaluation: Compare time for service for given request sequence. Distinguish only by cylinder.

slide-4
SLIDE 4

CPSC 410 / 611 : Operating Systems Disk Management 4

24 49 74 99 124 149 174 199

FCFS Scheduling

Request Sequence: 98, 183, 37, 122, 14, 124, 65, 67

total head movement: 640 tracks

FCFS Pros: – simple – fair FCFS Cons: – poor average service time

24 49 74 99 124 149 174 199

Shortest-Seek-Time-First (SSTF)

Request Sequence: 98, 183, 37, 122, 14, 124, 65, 67

total head movement: 236 tracks

Always serve “closest” request. SSTF Pros: – short service times SSTF Cons: – Starvation!

slide-5
SLIDE 5

CPSC 410 / 611 : Operating Systems Disk Management 5

24 49 74 99 124 149 174 199

Elevator Algorithm (SCAN)

Request Sequence: 98, 183, 37, 122, 14, 124, 65, 67

total head movement: 236 tracks

Continuously scan disk from one end to the other. SCAN Pros: – short service times SCAN Cons:

– When scanning, few requests after us, since just past through. – Problem: When we change direction at end, requests there are very new. 24 49 74 99 124 149 174 199

Circular SCAN (C-SCAN)

Request Sequence: 98, 183, 37, 122, 14, 124, 65, 67 Reduce variance in service time by always starting at the beginning of the disk.

slide-6
SLIDE 6

CPSC 410 / 611 : Operating Systems Disk Management 6

24 49 74 99 124 149 174 199

LOOK, C-LOOK

Disk Management

  • Disk Structure
  • Disk Scheduling
  • RAID
  • Disk Block Management
  • Modern Secondary Storage Technologies
slide-7
SLIDE 7

CPSC 410 / 611 : Operating Systems Disk Management 7

RAID

Observation: Traditional secondary storage devices are slow! Improve their performance by using multiple devices in parallel: RAID – Redundant Arrays of Independent Disks – Redundant Arrays of Inexpensive Disks (Berkeley) Common characteristics: – Array of physical disks that are visible as single device to OS. – Data is distributed across physical drives of array. – Redundant disk capacity is used for error detection/correction.

RAID (cont)

Approach: Replace single large-capacity disk with array of smaller- capacity disks. Benefits: – Improved I/O performance – Enables incremental upgrade Problems: – Reliability: more devices increase the probability of failure. Solution: redundancy

  • =

= = > = ) ( ] [ ) ( . . ] [ ) ( dt t R t E MTTF e t R g e t t P t R

F t F λ

slide-8
SLIDE 8

CPSC 410 / 611 : Operating Systems Disk Management 8

Raid (cont 2)

Raid Striping Redundancy + = bit-level block-level

block

bit-level striping

blocks

block-level striping

RAID Level 0

“Block-level Striped Set without Parity”

blocks

block-level striping

slide-9
SLIDE 9

CPSC 410 / 611 : Operating Systems Disk Management 9

  • Problem:

– cost (100% redundancy)

  • Performance

– READs : good (with multithreading and “split reads”) – WRITEs: small performance penalty.

RAID Level 1

“Mirrored Set without Parity”

mirrors

RAID Level 2

  • Head and spindles synchronized
  • Small strips
  • Error correction code calculated over

bits of data disks. (Hamming Code)

  • Appropriate for systems with many

failures.

  • Typically not implemented.

block

“Memory-Style Error-Correcting Parity”

slide-10
SLIDE 10

CPSC 410 / 611 : Operating Systems Disk Management 10

RAID Level 3

  • Heads and spindles synchronized.
  • Small strips.
  • Simple parity bits instead of ECC.

S0

block

S1 S2 S3 S4

parity Disk 1 fails:

1 2 3 4

) ( . . S S S S S S P g e ⊕ ⊕ ⊕ = =

2 3 4 1

S S S S S ⊕ ⊕ ⊕ =

“Bit-Interleaved Parity”

RAID Level 4

  • Same as RAID 3, but with block-

level striping.

  • No synchronization across disks.
  • Large strips.
  • Each strip on parity disk contains

parity information for all corresponding strips.

  • Parity computation upon READ:

block

) ( ' 1 ) ( 1 ) ( 4 ) ( 1 ) ( 1 ) ( ) ( ' 1 ) ( 2 ) ( 3 ) ( ) ( ' 1 ) ( 2 ) ( 3 ) ( ' 4 ) ( ) ( 1 ) ( 2 ) ( 3 ) ( 4 i X i X i X i X i X i X i X i X i X i X i X i X i X i X i X i X i X i X i X ⊕ ⊕ = ⊕ ⊕ ⊕ ⊕ ⊕ = ⊕ ⊕ ⊕ = ⊕ ⊕ ⊕ =

“Block level Parity”

slide-11
SLIDE 11

CPSC 410 / 611 : Operating Systems Disk Management 11

RAID Level 5

  • Same as RAID 4, but parity spread

across all disks.

  • No synchronization across disks.
  • Large strips.

block

“Striped Set with Interleaved Parity”

RAID Level 6

  • Same as RAID 5, but uses 2 bits to

store “parity”.

  • No synchronization across disks.
  • Large strips.
  • Uses ECC instead of parity.
  • Tolerates two failures.
  • In practice, a second drive can fail

during recovery from first drive failure. block

“Striped Set with Dual Interleaved Parity”

slide-12
SLIDE 12

CPSC 410 / 611 : Operating Systems Disk Management 12

Nested Levels: RAID Level 1+0 = RAID 10

Raid 10 = “Mirrored Set in a Striped Set”

blocks

block-level striping Raid 0 Raid 1 Raid 10 Raid 1

Disk Management

  • Disk Structure
  • Disk Scheduling
  • RAID
  • Disk Block Management
  • Modern Secondary Storage Technologies
slide-13
SLIDE 13

CPSC 410 / 611 : Operating Systems Disk Management 13

Disk Formatting

  • Bare disk:
  • Physical formatting:
  • “cut” into sectors
  • identify sectors
  • add space for error detection/correction
  • Logical formatting:
  • add blank directory, FAT, free space list, ...

ecc 1 ecc 2 ecc 3 ecc 4 ecc 5 ecc X 1 2 3 4 5 X FAT DIR DIR F/L ...

Framing

  • Character count
  • Starting and ending flags, with bit

stuffing

  • Physical layer coding violations
  • Starting and ending chars, with

character stuffing

5 1 2 3 4 8 1 2 3 4 5 6 7 1 2

character count

DLE STX a b DLE DLE c DLE ETX

stuffed DLE

framing pattern: 01111110

011011111011111011111010010

stuffed bits

binary Manchester

lack of transition

slide-14
SLIDE 14

CPSC 410 / 611 : Operating Systems Disk Management 14

Bad Block Management

  • One or more blocks become unreadable/unwriteable: bad blocks
  • Off-line management of bad blocks:

– Run bad-block detection program and put bad blocks on bad- block list. (Either remove them from free list or mark entry in FAT.) – May have to run file recovery utility.

  • On-line management:

– Have the device driver map the bad block onto a good block – Block X goes bad. Whenever OS requests block X, the disk transparently accesses a replacement block Y. – Problem: interferes with scheduling!

Disk Management

  • Disk Structure
  • Disk Scheduling
  • RAID
  • Disk Block Management
  • Modern Secondary Storage Technologies
slide-15
SLIDE 15

CPSC 410 / 611 : Operating Systems Disk Management 15

NAND Flash Memory

  • Flash chips are arranged in 8kB blocks.
  • Each block is divided into 512B pages.
  • Flash memory does not support “overwrite” operations.
  • Only supports a limited number of “erase” operations.
  • This is handled in the Flash Translation Layer (FTL)

Flash Translation Layer (FTL)

slide-16
SLIDE 16

CPSC 410 / 611 : Operating Systems Disk Management 16

Performance Comparison: IOPS Performance Comparison: Throughput

slide-17
SLIDE 17

CPSC 410 / 611 : Operating Systems Disk Management 17

Performance Comparison: Latency Intel 3D XPoint (Optane)

slide-18
SLIDE 18

CPSC 410 / 611 : Operating Systems Disk Management 18

Optane (cont)