Mirror File System A Multiple Server File System John Wong CTO - - PowerPoint PPT Presentation

mirror file system
SMART_READER_LITE
LIVE PREVIEW

Mirror File System A Multiple Server File System John Wong CTO - - PowerPoint PPT Presentation

Mirror File System A Multiple Server File System John Wong CTO John.Wong@TwinPeakSoft.com Twin Peaks Software Inc. Page 1 of Multiple Server File System Conventional File System EXT3/UFS and NFS Manage files on a single server


slide-1
SLIDE 1

Page 1 of

John Wong

CTO

John.Wong@TwinPeakSoft.com

Twin Peaks Software Inc.

Mirror File System

A Multiple Server File System

slide-2
SLIDE 2

Page 2 of

Multiple Server File System

  • Conventional File System – EXT3/UFS

and NFS

– Manage files on a single server and its storage

devices

  • Multiple Server File system

– Manage files on multiple servers and their

storage devices

slide-3
SLIDE 3

Page 3 of

Problems

  • Single resource is vulnerable
  • Redundancy provides a safety net

– Disk level

=> RAID

– Storage level

=> Storage Replication

– TCP/IP level

=> SNDR

– File System level

=> CFS, MFS

– System level

=> Clustering system

– Application

=> Database

slide-4
SLIDE 4

Page 4 of

Why MFS?

  • Many advantages over existing

technologies

slide-5
SLIDE 5

Page 5 of

Local File System

Data

EXT3

Application 1 Application 2

Kernel Space User Space

Disk Driver

EXT3 manages file on the local server’s storage devices

slide-6
SLIDE 6

Page 6 of

Application Application

NFS (Client mount)

Application Application Data

EXT3/UFS NFSD

Network File System

Client Server

NFS manages file on remote server’s storage devices

slide-7
SLIDE 7

Page 7 of

rsync, tar

NFS (Client mount) Application Application Data B

EXT3/UFS

NFSD

EXT3 | NFS

EXT3/UFS

Data B

Client Server

Application

Applications can only use either one, not both.

slide-8
SLIDE 8

Page 8 of

  • Combine these two file systems to manage file on

both local and remote servers storage devices

  • - at the same time
  • - in real time

EXT3 + NFS ??

slide-9
SLIDE 9

Page 9 of

Application Application Data EXT3/UFS Application Application Data EXT3/UFS

Passive MFS Server

MFS = EXT3 + NFS

Active MFS Server MFS

NFS

User Space Kernel Space

slide-10
SLIDE 10

Page 10 of

  • MFS is a kernel loadable module
  • loaded on top of EXT3/UFS and NFS
  • Standard VFS interface
  • Provide Complete Transparency
  • to users and applications
  • to underlining file systems

Building Block Approach

slide-11
SLIDE 11

Page 11 of

Q & A

Application Application Data A Application Application Data B

MFS MFS

slide-12
SLIDE 12

Page 12 of

  • Building block approach
  • - Building upon existing EXT3, NFS, NTFS, CIFS

infrastructures

  • No metadata is replicated
  • - Superblock, Cylinder group, file allocation map

are not replicated.

  • Every file write operation is checked by file system
  • - file consistency, integrity
  • Live file, not raw data replication
  • - The primary and backup copy both are live files

Advantages

slide-13
SLIDE 13

Page 13 of

  • Interoperability
  • - Two nodes can be different systems
  • - Storage systems can be different
  • Small granularity
  • - Directory level, not entire file system
  • One to many or many to one replication

Advantages

slide-14
SLIDE 14

Page 14 of

  • Fast replication
  • - Replication in Kernel file system module
  • Immediate failover
  • - No need to fsck and mount operation
  • Geographically dispersed clustering
  • - Two nodes can be separated by hundreds of miles
  • Easy to deploy and manage
  • - Only one copy of MFS running on primary server is

needed for replication

Advantages

slide-15
SLIDE 15

Page 15 of

Why MFS?

  • Better Data Protection
  • Better Disaster Recovery
  • Better RAS
  • Better Scalability
  • Better Performance
  • Better Resources Utilization
slide-16
SLIDE 16

Page 16 of

File System Framework

SOLARIS Internal, Core Kernel Architecture, Jim Mauro. Richard McDougall, PRENTICE HALL

Optical drive Network

File System Operation calls

User Applications

System Call Interface

File Operation System Calls Other System calls

read () write ()

  • pen ()

c l

  • s

e ( ) mkdir () rmdir () link () ioctl () creat () lseek () mount () umount () Statfs() sync () Vnode interfaces VFS interfaces UFS (2) NFS (2) VxFS HSFS QFS UFS (1) NFS (1) PCFS

PCFS

Data Data Data Data

File System Operation calls

slide-17
SLIDE 17

Page 17 of

MFS Framework

User Applications

System Call Interface File Operation System Calls File System Operation calls Other System calls

read () write ()

  • pen ()

c l

  • s

e ( ) mkdir () rmdir () link () ioctl () creat () lseek () mount () umount () Statfs() sync () Vnode interfaces VFS interfaces UFS (2) NFS (2) VxFS HSFS QFS UFS(1) NFS (1) PCFS

PCFS

Network Optical drive

Data Data Data Data

MFS

Vnode VFS interface

slide-18
SLIDE 18

Page 18 of

  • Transparent to users and applications
  • No re-compilation or re-link needed
  • Transparent to existing file structures
  • Same pathname access
  • Transparent to underlying file systems
  • UFS, NFS

Transparency

slide-19
SLIDE 19

Page 19 of

  • Conventional Mount
  • One directory, one file system
  • MFS Mount
  • One directory, two or more file systems

Mount Mechanism

slide-20
SLIDE 20

Page 20 of

# mount –F mfs host:/ndir1/ndir2 /udir1/udir2

  • First mount the NFS on a UFS directory
  • Then mount the MFS on top of UFS and NFS
  • Existing UFS tree structure /udir1/udir2

becomes a local copy of MFS

  • Newly mounted host:/ndir1/ndir2 becomes a

remote copy of MFS

  • Same mount options as NFS except no ‘-o hard’
  • ption

Mount Mechanism

slide-21
SLIDE 21

Page 21 of

# /usr/lib/fs/mfs/mfsck mfs_dir

  • After MFS mount succeeds, the local copy may

not be identical to the remote copy.

  • Use mfsck (the MFS fsck) to synchronize them.
  • The mfs_dir can be any directory under MFS

mount point.

  • Multiple mfsck commands can be invoked at the

same time.

MFS mfsck Command

slide-22
SLIDE 22

Page 22 of

READ/WRITE Vnode Operation

  • All VFS/vnode operations received by MFS
  • READ related operation: read, getattr,….

those operations only need to go to local copy (UFS).

  • WRITE related operation: write, setattr,…..

those operations go to both local (UFS) and remote (NFS) copy simultaneously (using threads)

slide-23
SLIDE 23

Page 23 of

  • Directory Level
  • Mirror any UFS directory instead of entire UFS

file system

  • Directory A mirrored to Server A
  • Directory B mirrored to Server B
  • Block Level Update
  • Only changed block is mirrored

Mirroring Granularity

slide-24
SLIDE 24

Page 24 of

# /usr/lib/fs/mfs/msync mfs_root_dir

  • A daemon that synchronizes MFS pair after a

remote MFS partner fails.

  • Upon a write failure, MFS:
  • Logs name of file to which the write operation failed
  • Starts a heartbeat thread to verify the remote MFS

server is back online

  • Once the remote MFS server is back online, msync

uses the log to sync missing files to remote server.

MFS msync Command

slide-25
SLIDE 25

Page 25 of

Active/Active Configuration

Server Server

Application Application Data A UFS Application Application Data B UFS

Active MFS Server MFS MFS Active MFS Server

NFS NFS

slide-26
SLIDE 26

Page 26 of

MFS uses UFS, NFS file record lock. Locking is required for the active-active configuration. Locking enables write-related vnode operations as atomic operations. Locking is enabled by default. Locking is not necessary in active-passive configuration.

MFS Locking Mechanism

slide-27
SLIDE 27

Page 27 of

  • Real-time
  • - Replicate file in real-time
  • Scheduled
  • - Log file path, offset and size
  • - Replicate only changed portion of a file

Real -Time and Scheduled

slide-28
SLIDE 28

Page 28 of

  • Online File Backup
  • Server File Backup, active  passive
  • Server/NAS Clustering, active  Active

Applications

slide-29
SLIDE 29

Page 29 of

Application Application Data NTFS Application Application Data NTFS

Remote Server

MFS = NTFS + CIFS

Window Desktop/Laptop MFS

CIFS

slide-30
SLIDE 30

Page 30 of

MFS

User Desktop/Laptop

Online File Backup

Real-time or Scheduled time

Folder

ISP Server

MFS

Folder

MFS

Folder

LAN or Wan LAN or Wan

slide-31
SLIDE 31

Page 31 of

Secondary

Server Replication

Mirror File System Mirror File System App Primary Email Mirror File System Mirroring Path : /home : /var/spool/mail Heartbeat

slide-32
SLIDE 32

Page 32 of

Enterprise Clusters

Mirror File System Mirror File System

Mirroring Path

App App App App App Mirror File System Mirror File System Central Mirror File System App App