an imitation learning approach for cache replacement
play

An Imitation Learning Approach for Cache Replacement Evan Z. Liu, - PowerPoint PPT Presentation

An Imitation Learning Approach for Cache Replacement Evan Z. Liu, Milad Hashemi, Kevin Swersky, Paruhasarathy Ranganathan, Junwhan Ahn The Need for Faster Compute Small cache improvements can make large difgerences! (Beckman, 2019) E.g., 1%


  1. An Imitation Learning Approach for Cache Replacement Evan Z. Liu, Milad Hashemi, Kevin Swersky, Paruhasarathy Ranganathan, Junwhan Ahn

  2. The Need for Faster Compute Small cache improvements can make large difgerences! (Beckman, 2019) E.g., 1% cache hit rate improvement → 35% ● decrease in latency (Cidon, et. al., 2016) Caches are everywhere: CPU chips ● Operating Systems ● Databases ● Web applications ● Our goal : Faster applications via betuer cache replacement policies (htups://openai.com/blog/ai-and-compute/)

  3. TL;DR: I. We approximate the optimal cache replacement policy by (implicitly) predicting the future II. Caching is an aturactive benchmark for the general reinforcement learning / imitation learning communities

  4. Cache Replacement Evict Cache A B C A B D A B D Miss Hit (100x faster) Miss Accesses D A C Goal: Evict the cache lines to maximize cache hits

  5. Cache Replacement Evict Mistake Cache A B C A B D A B D Miss Hit Miss Accesses D A C

  6. Cache Replacement Optimal decision Cache A B C A B D A B D Miss Hit Miss Accesses D A C

  7. Cache Replacement Reuse distance d t (line): number of accesses from access t until the line is reused d 0 (A) = 1, d 0 (B) > 2, d 0 (C) = 2 Cache A B C A B D A B D Miss Hit Miss Accesses D A C Optimal Policy (Belady’s): Evict the line with the greatest reuse distance (Belady, 1966)

  8. Belady’s Requires Future Information Reuse distance d t (line): number of accesses from access t until the line is reused Problem: Computing reuse distance requires knowing the future So in practice, we use heuristics , e.g.: Least-recently used (LRU) ● Most-recently used (MRU) ● … but these pergorm poorly on complex access patuerns

  9. Leveraging Belady’s Idea: approximate Belady’s from past accesses Training Predicted decision Optimal decision Learned Belady’s Model . . . . . . Past accesses Current Future access accesses

  10. Prior Work Evict line X Trained on Current line cache Belady’s friendly or averse? Traditional Hawkeye / Algorithm Glider Past Current Current accesses access cache state Current state-of-the-aru (Shi et. al., ‘19, Jain et. al., ‘18)

  11. Prior Work Evict line X + binary classifjcation is relatively Trained on Current line cache easy to learn Belady’s friendly or averse? - traditional algorithm can’t Traditional Hawkeye / express optimal policy Algorithm Glider Past Current Current accesses access cache state Current state-of-the-aru (Shi et. al., ‘19, Jain et. al., ‘18)

  12. Our contribution: Our Approach Directly approximate Belady’s via imitation learning Evict line X Trained on Current line cache Trained on Evict line X Belady’s friendly or averse? Belady’s Traditional Model Hawkeye / Algorithm Glider . . . Past Current Current Past Current Current accesses access cache state accesses access cache state Current state-of-the-aru Our proposal (Shi et. al., ‘19, Jain et. al., ‘18)

  13. Cache Replacement Markov Decision Process Evict Cache A B C A B D A B D Miss Hit Miss Accesses D A C Similar to Wang, et. al., 2019

  14. Cache Replacement Markov Decision Process Current cache Evict contents Cache A B C A B D A B D Miss Hit Miss Past accesses Current access Accesses D A C Similar to Wang, et. al., 2019

  15. Cache Replacement Markov Decision Process Cache A B C A B D A B D Miss Hit Miss Accesses D A C Similar to Wang, et. al., 2019

  16. Cache Replacement Markov Decision Process Evict Cache A B C A B D A B D Miss Hit Miss Accesses D A C Similar to Wang, et. al., 2019

  17. Leveraging the Optimal Policy Typical imitation learning setuing Observation: Not all errors are equally bad (Pomerlau, 1991, Ross, et. al., 2011, Kim, et. al., 2013) Learning from optimal policy yields ● greater training signal optimal action state state Learned policy Approximate optimal policy Learned policy optimize, e.g., Concretely: minimize a ranking loss

  18. Reuse Distance as an Auxiliary Task Observation: predicting reuse distance is correlated with cache replacement Cast this as an auxiliary task (Jaderberg, et. al., 2016) ● Loss Policy Reuse distance State embedding State s t

  19. Results Optimal cache-hit rate LRU cache-hit rate ~19% cache-hit rate increase over Glider (Shi, et. al., 2019) on memory-intensive SPEC2006 applications (Jaleel, et. al., 2009) ~64% cache-hit rate increase over LRU on Google Web Search

  20. A Note on Practicality address embedding Linear Layer This work : Establish a proof-of-concept Per-byte address embedding Reduce embedding size from 100MB to <10KB ● ● ~6% cache-hit rate increase on SPEC2006 vs. Glider ● ~59% cache-hit rate increase on Google Web Address: 0x 12 C5 A1 ... Search vs. LRU Byte 1 Byte 2 Byte 3

  21. A Note on Practicality address embedding Linear Layer This work : Establish a proof-of-concept Per-byte address embedding Reduce embedding size from 100MB to <10KB ● ● ~6% cache-hit rate increase on SPEC2006 vs. Glider ● ~59% cache-hit rate increase on Google Web Address: 0x 12 C5 A1 ... Search vs. LRU Byte 1 Byte 2 Byte 3 Future work: Production ready learned policies ● Smaller models via distillation (Hinton, et. al., 2015) , pruning (Janowsky, 1989, Han, et. al., 2015, Sze, et. al., 2017), or quantization ● Target domains with longer latency and larger caches (e.g., sofuware caches)

  22. A New Imitation / Reinforcement Learning Benchmark Bellemare, et. al., 2012, Levine, et. al., 2016, Lillicrap, et. al., 2015 Silver, et. al., 2017, OpenAI, 2019, Evict Vinyals, et. al., 2019 A B C Miss D + plentiful data - limited / expensive data + plentiful data - delayed real-world utility + immediate real-world impact + immediate real-world impact Open-source cache replacement Gym environment coming soon!

  23. Takeaways A new state-of-the-aru approach for cache replacement by imitating the ● oracle policy Future work: making this production ready ○ A new benchmark for imitation learning / reinforcement learning research ●

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend