CS4617 Computer Architecture
Lecture 5: Memory Hierarchy 3 Dr J Vaughan September 22, 2014
1/37
CS4617 Computer Architecture Lecture 5: Memory Hierarchy 3 Dr J - - PowerPoint PPT Presentation
CS4617 Computer Architecture Lecture 5: Memory Hierarchy 3 Dr J Vaughan September 22, 2014 1/37 Six basic cache optimisations Average memory access time = Hit time + Miss rate Miss penalty Thus, cache optimisations can be divided into 3
1/37
2/37
3/37
4/37
5/37
6/37
7/37
8/37
9/37
10/37
11/37
◮ A direct mapped cache of size N has about the same miss rate
12/37
◮ Clock cycle time2−way = 1.36 × Clock cycle time1−way ◮ Clock cycle time4−way = 1.44 × Clock cycle time1−way ◮ Clock cycle time8−way = 1.52 × Clock cycle time1−way
13/37
14/37
15/37
16/37
17/37
◮ Make cache faster? ◮ Make cache larger? ◮ Do both by adding another level of cache
18/37
19/37
20/37
21/37
22/37
23/37
24/37
25/37
◮ Everything in L1 is likely to be in L2 also =
◮ If |L2| just a little bigger than |L1|, the local miss rate,
◮ Does set associativity make sense for L2 caches? 26/37
27/37
28/37
◮ Multilevel inclusion ◮ L1 data are always in the L2 cache
◮ L2 cache must invalidate all L1 blocks that map onto the L2
◮ This causes a higher L1 miss rate
29/37
30/37
31/37
32/37
33/37
34/37
35/37
36/37
◮ Index the cache ◮ Compare addresses
◮ Page-level protection is part of virtual to physical address
◮ Cache flushing on process switch because virtual addresses are
◮ Aliasing due to different processes using different virtual
37/37