The Impact of Solid State Drive on Search Engine Cache Management - - PowerPoint PPT Presentation

the impact of solid state drive on search engine cache
SMART_READER_LITE
LIVE PREVIEW

The Impact of Solid State Drive on Search Engine Cache Management - - PowerPoint PPT Presentation

The Impact of Solid State Drive on Search Engine Cache Management Jiancong Tong Ph.D. candidate at Nankai University Visiting student at University of Melbourne lingfenghx@gmail.com with Gang Wang, Xiaoguang Liu (Nankai University) and


slide-1
SLIDE 1

The Impact of Solid State Drive on Search Engine Cache Management

Jiancong Tong

Ph.D. candidate at Nankai University Visiting student at University of Melbourne lingfenghx@gmail.com with Gang Wang, Xiaoguang Liu (Nankai University) and Jianguo Wang, Eric Lo, Man Lung Yiu (Hong Kong Polytechnic University)

Monash University May 12, 2014

slide-2
SLIDE 2

Background and Motivation Research Questions and Answers

Outline 1 Background and Motivation 2 Research Questions and Answers RQ1: What is the impact of SSD on buffer management? RQ2: How could we deal with that?

1 / 30

slide-3
SLIDE 3

Background and Motivation Research Questions and Answers

Hard Disk Drive

Hard Disk Drive (HDD) Magnetic head How does HDD work? [Garcia et al., 2000]

2 / 30

slide-4
SLIDE 4

Background and Motivation Research Questions and Answers

Hard Disk Drive

Hard Disk Drive (HDD) Magnetic head How does HDD work? [Garcia et al., 2000]

Random read latency of HDD

◮ Seek time ◮ Rotational latency ◮ Transfer time

2 / 30

slide-5
SLIDE 5

Background and Motivation Research Questions and Answers

Hard Disk Drive

Hard Disk Drive (HDD) Magnetic head How does HDD work? [Garcia et al., 2000]

Random read latency of HDD

◮ Seek time ◮ Rotational latency ◮ Transfer time

Caching technology is used to reduce the latency.

2 / 30

slide-6
SLIDE 6

Background and Motivation Research Questions and Answers

What is a Cache?

Small, fast memory used to improve average access time to large, slow storage media.

3 / 30

slide-7
SLIDE 7

Background and Motivation Research Questions and Answers

What is a Cache?

Small, fast memory used to improve average access time to large, slow storage media. Exploits locality: both spacial and temporal.

3 / 30

slide-8
SLIDE 8

Background and Motivation Research Questions and Answers

What is a Cache?

Small, fast memory used to improve average access time to large, slow storage media. Exploits locality: both spacial and temporal. Almost everything is a cache in computer architecture...

3 / 30

slide-9
SLIDE 9

Background and Motivation Research Questions and Answers

What is a Cache? (Cont.)

◮ Cache Hit: the requested data is found in the memory ◮ Cache Miss: the requested data is not found in the memory Cache hit Cache miss

4 / 30

slide-10
SLIDE 10

Background and Motivation Research Questions and Answers

What is a Cache? (Cont.)

◮ Cache Hit: the requested data is found in the memory ◮ Cache Miss: the requested data is not found in the memory Cache hit Cache miss

Hit ratio =

#Hits #Memory accesses

4 / 30

slide-11
SLIDE 11

Background and Motivation Research Questions and Answers

What is a Cache? (Cont.)

◮ Cache Hit: the requested data is found in the memory ◮ Cache Miss: the requested data is not found in the memory Cache hit Cache miss

Hit ratio =

#Hits #Memory accesses

Miss ratio =

#Misses #Memory accesses

4 / 30

slide-12
SLIDE 12

Background and Motivation Research Questions and Answers

What is a Cache? (Cont.)

◮ Cache Hit: the requested data is found in the memory ◮ Cache Miss: the requested data is not found in the memory Cache hit Cache miss

Hit ratio =

#Hits #Memory accesses

Miss ratio =

#Misses #Memory accesses

Hit ratio + Miss ratio = 1

4 / 30

slide-13
SLIDE 13

Background and Motivation Research Questions and Answers

What is Solid State Drive?

Hard Disk Drive (HDD) with magnetic moving head Solid State Drive (SSD) based on semiconductor chips ◮ SSD: New Faster (10 ∼ 100x) HDD with compatible interface

5 / 30

slide-14
SLIDE 14

Background and Motivation Research Questions and Answers

What is Solid State Drive?

Hard Disk Drive (HDD) with magnetic moving head Solid State Drive (SSD) based on semiconductor chips ◮ SSD: New Faster (10 ∼ 100x) HDD with compatible interface ◮ Strong technical merits: [Chen et al., 2009] 1 Lower power consumption 2 More compact size 3 Better shock resistance 4 Extraordinarily faster random data access

5 / 30

slide-15
SLIDE 15

Background and Motivation Research Questions and Answers

The Rise of SSD Words of Pioneer (by Jim Gray, 2006) Tape is dead; Disk is tape; Flash is disk; RAM locality is King.

6 / 30

slide-16
SLIDE 16

Background and Motivation Research Questions and Answers

The Rise of SSD Words of Pioneer (by Jim Gray, 2006) Tape is dead; Disk is tape; Flash is disk; RAM locality is King. SSD in Large-Scale System Architectures

◮ Google 2008 (or later) ◮ Baidu 2008 ◮ Facebook 2010 ◮ Myspace 2010 ◮ Oracle 2011 ◮ Microsoft Azure 2012

6 / 30

slide-17
SLIDE 17

Background and Motivation Research Questions and Answers

The Rise of SSD Words of Pioneer (by Jim Gray, 2006) Tape is dead; Disk is tape; Flash is disk; RAM locality is King. SSD in Large-Scale System Architectures

◮ Google 2008 (or later) ◮ Baidu 2008 ◮ Facebook 2010 ◮ Myspace 2010 ◮ Oracle 2011 ◮ Microsoft Azure 2012

Trends Flash memory based SSD is replacing and is going to completely replace HDD as the major storage medium!

6 / 30

slide-18
SLIDE 18

Background and Motivation Research Questions and Answers

Challenges

Existing caching policies were originally designed for HDD

◮ HDD: Very slow random read (compare to sequential read) ◮ Cache design principle: minimize random read

7 / 30

slide-19
SLIDE 19

Background and Motivation Research Questions and Answers

Challenges

Existing caching policies were originally designed for HDD

◮ HDD: Very slow random read (compare to sequential read) ◮ Cache design principle: minimize random read

However...

[Tong et al., 2013]

7 / 30

slide-20
SLIDE 20

Background and Motivation Research Questions and Answers

Challenges

Existing caching policies were originally designed for HDD

◮ HDD: Very slow random read (compare to sequential read) ◮ Cache design principle: minimize random read

However...

[Tong et al., 2013]

What now?

7 / 30

slide-21
SLIDE 21

Background and Motivation Research Questions and Answers

Research Questions RQ1: What is the impact of SSD on buffer management?

8 / 30

slide-22
SLIDE 22

Background and Motivation Research Questions and Answers

Research Questions RQ1: What is the impact of SSD on buffer management?

◮ Are the existing cache techniques designed for HDD-based

search engine still good for SSD-based search engine?

8 / 30

slide-23
SLIDE 23

Background and Motivation Research Questions and Answers

Research Questions RQ1: What is the impact of SSD on buffer management?

◮ Are the existing cache techniques designed for HDD-based

search engine still good for SSD-based search engine?

◮ What measure(s) should be used to define a ‘good’ cache

policy in this case?

8 / 30

slide-24
SLIDE 24

Background and Motivation Research Questions and Answers

Research Questions RQ1: What is the impact of SSD on buffer management?

◮ Are the existing cache techniques designed for HDD-based

search engine still good for SSD-based search engine?

◮ What measure(s) should be used to define a ‘good’ cache

policy in this case?

◮ If the performance of caching is improved or degraded, what

does that mean to the entire system?

8 / 30

slide-25
SLIDE 25

Background and Motivation Research Questions and Answers

Research Questions RQ1: What is the impact of SSD on buffer management?

◮ Are the existing cache techniques designed for HDD-based

search engine still good for SSD-based search engine?

◮ What measure(s) should be used to define a ‘good’ cache

policy in this case?

◮ If the performance of caching is improved or degraded, what

does that mean to the entire system? RQ2: How could we deal with that?

8 / 30

slide-26
SLIDE 26

Background and Motivation Research Questions and Answers

Research Questions RQ1: What is the impact of SSD on buffer management?

◮ Are the existing cache techniques designed for HDD-based

search engine still good for SSD-based search engine?

◮ What measure(s) should be used to define a ‘good’ cache

policy in this case?

◮ If the performance of caching is improved or degraded, what

does that mean to the entire system? RQ2: How could we deal with that?

◮ What to do if the efficiency of the entire system is affected by

such impact?

8 / 30

slide-27
SLIDE 27

Background and Motivation Research Questions and Answers

Research Questions RQ1: What is the impact of SSD on buffer management?

◮ Are the existing cache techniques designed for HDD-based

search engine still good for SSD-based search engine?

◮ What measure(s) should be used to define a ‘good’ cache

policy in this case?

◮ If the performance of caching is improved or degraded, what

does that mean to the entire system? RQ2: How could we deal with that?

◮ What to do if the efficiency of the entire system is affected by

such impact?

◮ Can we propose better cache policies for SSD-based systems?

8 / 30

slide-28
SLIDE 28

Background and Motivation Research Questions and Answers

1 Background and Motivation 2 Research Questions and Answers RQ1: What is the impact of SSD on buffer management? RQ2: How could we deal with that?

9 / 30

slide-29
SLIDE 29

Background and Motivation Research Questions and Answers

Large-scale experimental study

RQ1 can be answered by evaluating the effectiveness of existing caching policies on an SSD-based search engine.

Settings

◮ Datasets

Web documents: 12,000,000 (∼100GB) Queries: 1,000,000

◮ Device

SSD: ADATA 256GB SSD HDD: Seagate 3TB 7200rpm

◮ System: Apache Lucene ◮ Measure: Query time (NOT hit ratio)

10 / 30

slide-30
SLIDE 30

Background and Motivation Research Questions and Answers

Overview of search engine architecture

[Wang et al., 2013]

11 / 30

slide-31
SLIDE 31

Background and Motivation Research Questions and Answers

Overview of search engine architecture

[Wang et al., 2013]

12 / 30

slide-32
SLIDE 32

Background and Motivation Research Questions and Answers

Overview of search engine architecture

[Wang et al., 2013]

12 / 30

slide-33
SLIDE 33

Background and Motivation Research Questions and Answers

Index Server and List (Index) Cache Posting Lists (Inverted Lists)

◮ Clayton (2, 4, 5, 10, 100, 107, ..., 1,000,000) ◮ Monash (1, 2, 3, 23, ...)

13 / 30

slide-34
SLIDE 34

Background and Motivation Research Questions and Answers

Index Server and List (Index) Cache Posting Lists (Inverted Lists)

◮ Clayton (2, 4, 5, 10, 100, 107, ..., 1,000,000) ◮ Monash (1, 2, 3, 23, ...)

Index Server

13 / 30

slide-35
SLIDE 35

Background and Motivation Research Questions and Answers

Which lists should be kept in the cache?

14 / 30

slide-36
SLIDE 36

Background and Motivation Research Questions and Answers

Which lists should be kept in the cache? Metrics to consider

◮ High Frequency First? ◮ High Frequency/Size First?

14 / 30

slide-37
SLIDE 37

Background and Motivation Research Questions and Answers

Previous conclusion [Baeza-Yates et al., 2007]

Frequency/Size is a better metric (for HDD).

15 / 30

slide-38
SLIDE 38

Background and Motivation Research Questions and Answers

Previous conclusion [Baeza-Yates et al., 2007]

Frequency/Size is a better metric (for HDD).

The reason

On HDD

◮ Random read is 130 ∼ 170x slower than sequential read

= ⇒ List access time is (almost) a constant!

15 / 30

slide-39
SLIDE 39

Background and Motivation Research Questions and Answers

Previous conclusion [Baeza-Yates et al., 2007]

Frequency/Size is a better metric (for HDD).

The reason

On HDD

◮ Random read is 130 ∼ 170x slower than sequential read

= ⇒ List access time is (almost) a constant!

15 / 30

slide-40
SLIDE 40

Background and Motivation Research Questions and Answers

Previous conclusion [Baeza-Yates et al., 2007]

Frequency/Size is a better metric (for HDD).

The reason

On HDD

◮ Random read is 130 ∼ 170x slower than sequential read

= ⇒ List access time is (almost) a constant!

◮ It is #access that dominates, not |access data|

= ⇒ Short lists are favored (more lists can be held)

15 / 30

slide-41
SLIDE 41

Background and Motivation Research Questions and Answers

Previous conclusion [Baeza-Yates et al., 2007]

Frequency/Size is a better metric (for HDD).

The reason

On HDD

◮ Random read is 130 ∼ 170x slower than sequential read

= ⇒ List access time is (almost) a constant!

◮ It is #access that dominates, not |access data|

= ⇒ Short lists are favored (more lists can be held)

◮ High benefit gained when size is taken into account

15 / 30

slide-42
SLIDE 42

Background and Motivation Research Questions and Answers

How does SSD change the story? [Wang et al., 2013] Differences

On SSD

◮ Random read is 130 ∼ 170x slower than sequential read

= ⇒ List access time is (almost) a constant!

◮ Random read is only 2 ∼ 10x slower than sequential read

= ⇒ List access time varies a lot!

16 / 30

slide-43
SLIDE 43

Background and Motivation Research Questions and Answers

How does SSD change the story? [Wang et al., 2013] Differences

On SSD

◮ Random read is 130 ∼ 170x slower than sequential read

= ⇒ List access time is (almost) a constant!

◮ Random read is only 2 ∼ 10x slower than sequential read

= ⇒ List access time varies a lot!

16 / 30

slide-44
SLIDE 44

Background and Motivation Research Questions and Answers

How does SSD change the story? [Wang et al., 2013] Differences

On SSD

◮ Random read is 130 ∼ 170x slower than sequential read

= ⇒ List access time is (almost) a constant!

◮ Random read is only 2 ∼ 10x slower than sequential read

= ⇒ List access time varies a lot!

◮ Long lists have slower access time

= ⇒ Admitting too many short lists offers less benefit

16 / 30

slide-45
SLIDE 45

Background and Motivation Research Questions and Answers

How does SSD change the story? (Cont.) Read access latency of posting lists of varying lengths

17 / 30

slide-46
SLIDE 46

Background and Motivation Research Questions and Answers

How does SSD change the story? (Cont.) Read access latency of posting lists of varying lengths Conjecture 1

The best cache policy may have changed

17 / 30

slide-47
SLIDE 47

Background and Motivation Research Questions and Answers

Evaluation on Index Server Static List Cache Policy

  • n HDD
  • n SSD

18 / 30

slide-48
SLIDE 48

Background and Motivation Research Questions and Answers

Evaluation on Index Server Static List Cache Policy

  • n HDD
  • n SSD

Conjecture 1

The best cache policy may have changed – Confirmed

18 / 30

slide-49
SLIDE 49

Background and Motivation Research Questions and Answers

Evaluation on Index Server Static List Cache Policy

  • n HDD
  • n SSD

Conjecture 1

The best cache policy may have changed – Confirmed ⋆Do not rely on your knowledge acquired in HDD era.

18 / 30

slide-50
SLIDE 50

Background and Motivation Research Questions and Answers

Is Hit Ratio Reliable?

19 / 30

slide-51
SLIDE 51

Background and Motivation Research Questions and Answers

Is Hit Ratio Reliable? Previously on HDD

Now on SSD Reliable ? As list access is a constant NO So, can measure the effectiveness of the cache ?

19 / 30

slide-52
SLIDE 52

Background and Motivation Research Questions and Answers

Is Hit Ratio Reliable? Previously on HDD

Now on SSD Reliable ? As list access is a constant NO So, can measure the effectiveness of the cache ?

Conjecture 2

Cache hit ratio might not be reliable.

19 / 30

slide-53
SLIDE 53

Background and Motivation Research Questions and Answers

Is Hit Ratio Reliable? (Cont.) Static List Cache Policy (on SSD)

20 / 30

slide-54
SLIDE 54

Background and Motivation Research Questions and Answers

Is Hit Ratio Reliable? (Cont.) Static List Cache Policy (on SSD) Conjecture 2

Cache hit ratio might not be reliable – Confirmed

20 / 30

slide-55
SLIDE 55

Background and Motivation Research Questions and Answers

Is Hit Ratio Reliable? (Cont.) Static List Cache Policy (on SSD) Conjecture 2

Cache hit ratio might not be reliable – Confirmed ⋆Use query time, not hit ratio

20 / 30

slide-56
SLIDE 56

Background and Motivation Research Questions and Answers

Document Server and Its Cache

Document Server

21 / 30

slide-57
SLIDE 57

Background and Motivation Research Questions and Answers

Document Server and Its Cache (Cont.)

22 / 30

slide-58
SLIDE 58

Background and Motivation Research Questions and Answers

Evaluation on Document Server

23 / 30

slide-59
SLIDE 59

Background and Motivation Research Questions and Answers

Evaluation on Document Server Document retrieval is no longer the bottleneck on document server

23 / 30

slide-60
SLIDE 60

Background and Motivation Research Questions and Answers

Evaluation on Document Server Document retrieval is no longer the bottleneck on document server Conjecture 3

Is it the same case for the whole system?

23 / 30

slide-61
SLIDE 61

Background and Motivation Research Questions and Answers

The Impact on the Entire System Efficiency Time break down for query processing (on SSD) All caches are enabled

24 / 30

slide-62
SLIDE 62

Background and Motivation Research Questions and Answers

The Impact on the Entire System Efficiency Time break down for query processing (on SSD) All caches are enabled Conjecture 3

Is it the same case for the whole system? – Confirmed

24 / 30

slide-63
SLIDE 63

Background and Motivation Research Questions and Answers

The Impact on the Entire System Efficiency Time break down for query processing (on SSD) All caches are enabled Conjecture 3

Is it the same case for the whole system? – Confirmed ⋆Disk accessing is no longer the bottleneck!

24 / 30

slide-64
SLIDE 64

Background and Motivation Research Questions and Answers

1 Background and Motivation 2 Research Questions and Answers RQ1: What is the impact of SSD on buffer management? RQ2: How could we deal with that?

25 / 30

slide-65
SLIDE 65

Background and Motivation Research Questions and Answers

The Impact on the Entire System Efficiency (Cont.) Time break down for query processing (on SSD) All caches are enabled

Problems The time spent on intersection and ranking should be reduced Solution Adopting full-term-ranking-cache (FTRC) [Alting¨

  • vde et al., 2011]

and two-term-intersection-cache (TTIC) [Long and Suel, 2005]

26 / 30

slide-66
SLIDE 66

Background and Motivation Research Questions and Answers

Incorporating PLC+FTRC+TTIC

27 / 30

slide-67
SLIDE 67

Background and Motivation Research Questions and Answers

New Static List Caching Policy

28 / 30

slide-68
SLIDE 68

Background and Motivation Research Questions and Answers

New Static List Caching Policy

Can we convert the latencies of random seek and sequential read into a uniform measure?

28 / 30

slide-69
SLIDE 69

Background and Motivation Research Questions and Answers

New Static List Caching Policy (Cont.) Latency-Aware Caching (Static): BLOCK [Tong et al., 2013]

◮ For term t, B(t) represents the I/O cost that can be saved during the whole query evaluating process if ℓ(t) is kept in the cache ◮ Estimating model and existing static caching policies:

29 / 30

slide-70
SLIDE 70

Background and Motivation Research Questions and Answers

New Static List Caching Policy (Cont.) Latency-Aware Caching (Static): BLOCK [Tong et al., 2013]

◮ For term t, B(t) represents the I/O cost that can be saved during the whole query evaluating process if ℓ(t) is kept in the cache ◮ Estimating model and existing static caching policies:

29 / 30

slide-71
SLIDE 71

Background and Motivation Research Questions and Answers

New Static List Caching Policy (Cont.) Latency-Aware Caching (Static): BLOCK [Tong et al., 2013]

◮ For term t, B(t) represents the I/O cost that can be saved during the whole query evaluating process if ℓ(t) is kept in the cache ◮ Estimating model and existing static caching policies: ◮ New method: BLOCK

29 / 30

slide-72
SLIDE 72

Background and Motivation Research Questions and Answers

New Static List Caching Policy (Cont.) Latency-Aware Caching (Static): BLOCK [Tong et al., 2013]

◮ For term t, B(t) represents the I/O cost that can be saved during the whole query evaluating process if ℓ(t) is kept in the cache ◮ Estimating model and existing static caching policies: ◮ New method: BLOCK 1 access = 1 random seek + several sequential reads = 1 random seek + a few equivalent random seeks

29 / 30

slide-73
SLIDE 73

Background and Motivation Research Questions and Answers

New Static List Caching Policy (Cont.) Experimental Results

  • n SSD
  • n HDD

2 4 6 8 10 12 14 16 512 1024 1536 2048 2560 3072 3584

Average disk access latency (ms) Cache size (MB)

QTFDF QTF/MECH BLOCK 0.9 1.0 1.1 1.2 1.3 1.4 1.5 2560 3072 3584 5 10 15 20 25 512 1024 1536 2048 2560 3072 3584

Average disk access latency (ms) Cache size (MB)

QTFDF QTF MECH BLOCK 1.0 1.5 2.0 2.5 3.0 3.5 2560 3072 3584

30 / 30

slide-74
SLIDE 74

Background and Motivation Research Questions and Answers

New Static List Caching Policy (Cont.) Experimental Results

  • n SSD
  • n HDD

2 4 6 8 10 12 14 16 512 1024 1536 2048 2560 3072 3584

Average disk access latency (ms) Cache size (MB)

QTFDF QTF/MECH BLOCK 0.9 1.0 1.1 1.2 1.3 1.4 1.5 2560 3072 3584 5 10 15 20 25 512 1024 1536 2048 2560 3072 3584

Average disk access latency (ms) Cache size (MB)

QTFDF QTF MECH BLOCK 1.0 1.5 2.0 2.5 3.0 3.5 2560 3072 3584

◮ BLOCK outperforms other methods on SSD

30 / 30

slide-75
SLIDE 75

Background and Motivation Research Questions and Answers

New Static List Caching Policy (Cont.) Experimental Results

  • n SSD
  • n HDD

2 4 6 8 10 12 14 16 512 1024 1536 2048 2560 3072 3584

Average disk access latency (ms) Cache size (MB)

QTFDF QTF/MECH BLOCK 0.9 1.0 1.1 1.2 1.3 1.4 1.5 2560 3072 3584 5 10 15 20 25 512 1024 1536 2048 2560 3072 3584

Average disk access latency (ms) Cache size (MB)

QTFDF QTF MECH BLOCK 1.0 1.5 2.0 2.5 3.0 3.5 2560 3072 3584

◮ BLOCK outperforms other methods on SSD 1 BLOCK beats QTFDF (by far the best policy on HDD) by 20%-60%

30 / 30

slide-76
SLIDE 76

Background and Motivation Research Questions and Answers

New Static List Caching Policy (Cont.) Experimental Results

  • n SSD
  • n HDD

2 4 6 8 10 12 14 16 512 1024 1536 2048 2560 3072 3584

Average disk access latency (ms) Cache size (MB)

QTFDF QTF/MECH BLOCK 0.9 1.0 1.1 1.2 1.3 1.4 1.5 2560 3072 3584 5 10 15 20 25 512 1024 1536 2048 2560 3072 3584

Average disk access latency (ms) Cache size (MB)

QTFDF QTF MECH BLOCK 1.0 1.5 2.0 2.5 3.0 3.5 2560 3072 3584

◮ BLOCK outperforms other methods on SSD 1 BLOCK beats QTFDF (by far the best policy on HDD) by 20%-60% 2 BLOCK outperforms QTF/MECH (identical on SSD) too

30 / 30

slide-77
SLIDE 77

Background and Motivation Research Questions and Answers

New Static List Caching Policy (Cont.) Experimental Results

  • n SSD
  • n HDD

2 4 6 8 10 12 14 16 512 1024 1536 2048 2560 3072 3584

Average disk access latency (ms) Cache size (MB)

QTFDF QTF/MECH BLOCK 0.9 1.0 1.1 1.2 1.3 1.4 1.5 2560 3072 3584 5 10 15 20 25 512 1024 1536 2048 2560 3072 3584

Average disk access latency (ms) Cache size (MB)

QTFDF QTF MECH BLOCK 1.0 1.5 2.0 2.5 3.0 3.5 2560 3072 3584

◮ BLOCK outperforms other methods on SSD 1 BLOCK beats QTFDF (by far the best policy on HDD) by 20%-60% 2 BLOCK outperforms QTF/MECH (identical on SSD) too ◮ BLOCK is better even on HDD (though not that significantly)

30 / 30

slide-78
SLIDE 78

References

Alting¨

  • vde, I. S., Ozcan, R., Cambazoglu, B. B., and Ulusoy, ¨
  • O. (2011).

Second chance: A hybrid approach for dynamic result caching in search engines. In ECIR, pages 510–516. Baeza-Yates, R. A., Gionis, A., Junqueira, F., Murdock, V., Plachouras, V., and Silvestri, F. (2007). The impact of caching on search engines. In SIGIR, pages 183–190. Chen, F., Koufaty, D. A., and Zhang, X. (2009). Understanding intrinsic characteristics and system implications of flash memory based solid state drives. In SIGMETRICS/Performance, pages 181–192. Garcia, J. M., Frohock, M., Reding, E., Reding, J. A., Garcia, M. F., DeLuca, S. A., and Whalen, E. (2000). Microsoft SQL Server(tm) 2000 Administrator’s Companion. Microsoft Press. Long, X. and Suel, T. (2005). Three-level caching for efficient query processing in large web search engines. In WWW, pages 257–266. Tong, J., Wang, G., and Liu, X. (2013). Latency-aware strategy for static list caching in flash-based web search engines. In CIKM, pages 1209–1212. Wang, J., Lo, E., Yiu, M. L., Tong, J., Wang, G., and Liu, X. (2013). The impact of solid state drive on search engine cache management. In SIGIR, pages 693–702.

slide-79
SLIDE 79

Thanks for your time!

Questions & Comments?

For more details of RQ1, please refer to: ◮

  • J. Wang, E. Lo, M, Yiu, J. Tong, G. Wang, X. Liu,

The impact of Solid State Disk on Search Engine Cache Management, In SIGIR, 2013, pp. 693-702.

For more details of RQ2, please refer to: ◮

  • J. Tong, G. Wang, X. Liu,

Latency-Aware Strategy for Static List Caching in Flash-based Web Search Engines, In CIKM, 2013, pp. 1209-1212.