Reducing Replication Bandwidth for Distributed Document Databases
Lianghong Xu1, Andy Pavlo1, Sudipta Sengupta2 Jin Li2, Greg Ganger1 Carnegie Mellon University1, Microsoft Research2
Reducing Replication Bandwidth for Distributed Document Databases - - PowerPoint PPT Presentation
Reducing Replication Bandwidth for Distributed Document Databases Lianghong Xu 1 , Andy Pavlo 1 , Sudipta Sengupta 2 Jin Li 2 , Greg Ganger 1 Carnegie Mellon University 1 , Microsoft Research 2 Document-oriented Databases { { " _id " :
Lianghong Xu1, Andy Pavlo1, Sudipta Sengupta2 Jin Li2, Greg Ganger1 Carnegie Mellon University1, Microsoft Research2
{ "_id" : "55ca4cf7bad4f75b8eb5c25c", "pageId" : "46780", "revId" : "41173", "timestamp" : "2002-03-30T20:06:22", "sha1" : "6i81h1zt22u1w4sfxoofyzmxd” "text" : “The Peer and the Peri is a comic [[Gilbert and Sullivan]] [[operetta ]] in two acts… just as predicting,…The fairy Queen, however, appears to … all live happily ever after. " } { "_id" : "55ca4cf7bad4f75b8eb5c25d”, "pageId" : "46780", "revId" : "128520", "timestamp" : "2002-03-30T20:11:12", "sha1" : "q08x58kbjmyljj4bow3e903uz” "text" : "The Peer and the Peri is a comic [[Gilbert and Sullivan]] [[operetta ]] in two acts… just as predicted, …The fairy Queen, on the other hand, is ''not'' happy, and appears to … all live happily ever after. " }
Update
2 ¡
Operation logs Operation logs
Secondary Secondary
Primary Database
3 ¡
{ "_id" : "55ca4cf7bad4f75b8eb5c25c", "pageId" : "46780", "revId" : "41173", "timestamp" : "2002-03-30T20:06:22", "sha1" : "6i81h1zt22u1w4sfxoofyzmxd” "text" : “The Peer and the Peri is a comic [[Gilbert and Sullivan]] [[operetta ]] in two acts… just as predicting,…The fairy Queen, however, appears to … all live happily ever after. " } { "_id" : "55ca4cf7bad4f75b8eb5c25d”, "pageId" : "46780", "revId" : "128520", "timestamp" : "2002-03-30T20:11:12", "sha1" : "q08x58kbjmyljj4bow3e903uz” "text" : "The Peer and the Peri is a comic [[Gilbert and Sullivan]] [[operetta ]] in two acts… just as predicted, …The fairy Queen, on the other hand, is ''not'' happy, and appears to … all live happily ever after. " }
Operation logs Operation logs
Secondary Secondary
Primary Database
4 ¡
{ "_id" : "55ca4cf7bad4f75b8eb5c25c", "pageId" : "46780", "revId" : "41173", "timestamp" : "2002-03-30T20:06:22", "sha1" : "6i81h1zt22u1w4sfxoofyzmxd” "text" : “The Peer and the Peri is a comic [[Gilbert and Sullivan]] [[operetta ]] in two acts… just as predicting,…The fairy Queen, however, appears to … all live happily ever after. " } { "_id" : "55ca4cf7bad4f75b8eb5c25d”, "pageId" : "46780", "revId" : "128520", "timestamp" : "2002-03-30T20:11:12", "sha1" : "q08x58kbjmyljj4bow3e903uz” "text" : "The Peer and the Peri is a comic [[Gilbert and Sullivan]] [[operetta ]] in two acts… just as predicted, …The fairy Queen, on the other hand, is ''not'' happy, and appears to … all live happily ever after. " }
5 ¡
Modified Region Duplicate Region Chunk Boundary Deduped Data Incoming Data
6 ¡
Modified Region Duplicate Region Chunk Boundary Incoming Data Deduped Data
7 ¡
Modified Region Duplicate Region Chunk Boundary Incoming Data Deduped Data
8 ¡
Modified Region Duplicate Region Chunk Boundary Incoming Data Dedup’ed Data
Delta!
9 ¡
20GB sampled Wikipedia dataset MongoDB v2.7 / / 4MB Oplog batches
10 ¡
Primary Node Client Secondary Node
Source documents Insertion & Updates Database Oplog sDedup Encoder Unsynchronized
Dedup’ed
Oplog Re-constructed
Replay sDedup Decoder Oplog syncer Database Source documents Source Document Cache
11 ¡
12 ¡
Target Document Consistent Sampling Similarity Sketch Rabin Chunking
32 17 25 41 12 41 32
Feature Index Table
Candida andidate Documents Documents
41 32 32 25 38 41 12 32 17 38 41 12 39 32 22 15
Doc #1 Doc #2 Doc #3
32 25 38 41 12 32 17 38 41 12
Doc #2 Doc #3
Doc #1
Doc #2
Doc #3 Similarity Similarity Sc Scor
13 ¡
Source Document Cache Rank Candidates Score 1 2 1 2 2 1 Doc #1 Doc #2 Doc #3 Initial Ranking Initial Ranking Final Ranking Final Ranking Rank Candidates Cached? Score 1 Yes es 4 1 Yes es 3 2 No 2 Doc #1 Doc #3 Doc #2
Is doc cached? If yes, reward +2
14 ¡
15 ¡
9.9 26.3 38.4 38.9 2.3 4.6 9.1 15.2 10 20 30 40 50
4KB 1KB 256B 64B Compression Ratio Chunk Size sDedup trad-dedup 20GB sampled Wikipedia dataset
16 ¡
34.1 47.9 57.3 61.0 80.2 133.0 272.5 780.5 200 400 600 800
4KB 1KB 256B 64B Memory (MB) Chunk Size sDedup trad-dedup 20GB sampled Wikipedia dataset
17 ¡
18 ¡
– Much greater data reduction than traditional dedup – Up to 38x compression ratio for Wikipedia – Resource-effjcient design with negligible overhead
– More diverse datasets – Dedup for local database storage – Difgerent similarity search schemes (e.g., super-fingerprints)
19 ¡