large pages may be harmful on numa systems
play

Large Pages May Be Harmful on NUMA Systems Fabien Gaud - PowerPoint PPT Presentation

Large Pages May Be Harmful on NUMA Systems Fabien Gaud Bap?ste Lepers Jeremie Decouchant Simon Fraser University CNRS Grenoble University Jus?n


  1. Large ¡Pages ¡May ¡Be ¡Harmful ¡on ¡ NUMA ¡Systems ¡ Fabien ¡Gaud ¡ Bap?ste ¡Lepers ¡ Jeremie ¡Decouchant ¡ Simon ¡Fraser ¡University ¡ CNRS ¡ Grenoble ¡University ¡ Jus?n ¡Funston ¡ Alexandra ¡Fedorova ¡ Vivien ¡Quéma ¡ Simon ¡Fraser ¡University ¡ Simon ¡Fraser ¡University ¡ Grenoble ¡INP ¡

  2. Virtual-­‑to-­‑physical ¡transla?on ¡is ¡done ¡ by ¡the ¡TLB ¡and ¡page ¡table ¡ ¡ ¡ TLB hit Virtual address TLB Physical address TLB miss Page table Typical TLB size: 1024 entries (AMD Bulldozer), 512 entries (Intel i7). 2

  3. Virtual-­‑to-­‑physical ¡transla?on ¡is ¡done ¡ by ¡the ¡TLB ¡and ¡page ¡table ¡ ¡ ¡ TLB hit Virtual address TLB Physical address TLB miss 43 cycles Page table Typical TLB size: 1024 entries (AMD Bulldozer), 512 entries (Intel i7). 3

  4. To ¡reduce ¡the ¡number ¡of ¡TLB ¡misses, ¡ developers ¡can ¡use ¡“large ¡pages” ¡ ¡ Page size 512 entries coverage 1024 entries coverage 4KB (default) 2MB 4MB 2MB 1GB 2GB 1GB 512GB 1024GB In Linux: - Manually: mmap( … , flags | MAP_HUGETLB) - Automatically: using Transparent Huge Pages (THP). THP uses 2MB pages for anonymous memory and clusters groups of 4K pages periodically. 4

  5. Large ¡pages ¡known ¡advantages ¡& ¡ downsides ¡ Known advantages: • Fewer TLB misses • Fewer page allocations (reduces contention in the kernel memory manager) Known downsides: • Increased memory footprint • Memory fragmentation 5

  6. New ¡observa?on: ¡large ¡pages ¡may ¡hurt ¡ performance ¡on ¡NUMA ¡machines ¡ Machine A, 24 cores Perf. improvement relative 30 to default Linux (%) 20 10 0 -10 -20 THP -30 B C D E F I L M S U U W W K M p w S S S T U c T G C P P A A m r S P G a . C R a m . . D . . . . . t C E . . C e B C B . B B C r D A D e i C a A x m n M . j b 2 s u 0 b l t i p l y Machine B, 64 cores Perf. improvement relative 109 70 51 30 to default Linux (%) 20 10 0 -10 -20 THP -43 -30 B C D E F I L M S U U W W K M p w S S S U c T P T P m S P G C G A A a r . C R a m . . D . . . . . t C E . . C e B C B . B A B C r D D e i C a A x m n M . j b 2 s b u 0 l 6 t i p l y

  7. Machines ¡are ¡NUMA ¡ Remote memory accesses hurt performance Memory Memory 8GB/s 160 cycles 3GB/s 300 cycles Node 1 CPU0 CPU1 CPU2 CPU3 Node 2 Node 3 Memory Memory 7

  8. Machines ¡are ¡NUMA ¡ Contention hurts performance even more. Memory Memory 1200 cycles ! Node 1 CPU0 CPU1 CPU2 CPU3 Node 2 Node 3 Memory Memory 8

  9. Large ¡pages ¡on ¡NUMA ¡machines ¡(1/2) ¡ void *a = malloc(2MB); Node 0 Node 1 Node 2 Node 3 With 4K pages, load is balanced. 9

  10. Large ¡pages ¡on ¡NUMA ¡machines ¡(1/2) ¡ void *a = malloc(2MB); Node 0 Node 1 Node 2 Node 3 With 2M pages, data are allocated on 1 node => contention. 10

  11. Large ¡pages ¡on ¡NUMA ¡machines ¡(1/2) ¡ HOT PAGE void *a = malloc(2MB); Node 0 Node 1 Node 2 Node 3 With 2M pages, data are allocated on 1 node => contention. 11

  12. Performance ¡example ¡(1/2) ¡ App. Perf. % of time % of time Imbalance Imbalance increase spent in spent in 4K (%) 2M (%) THP/4K TLB miss TLB miss (%) 4K 2M CG.D -43 0 0 1 59 SSCA.20 17 15 2 8 52 SpecJBB -6 7 0 16 39 Using large pages, 1 node is overloaded in CG, SSCA and SpecJBB. Only SSCA benefits from the reduction of TLB misses. 12

  13. Large ¡pages ¡on ¡NUMA ¡machines ¡(2/2) ¡ PAGE-LEVEL void *a = malloc(1.5MB); // node 0 FALSE SHARING void *b = malloc(1.5MB); // node 1 Node 0 Node 1 Node 2 Node 3 Page-level false sharing reduces the maximum achievable locality. 13

  14. Performance ¡example ¡(2/2) ¡ App. Perf. Local Local increase Access Access THP/4K Ratio 4K Ratio 2M (%) (%) (%) UA.C -15 88 66 The locality decreases when using large pages. 14

  15. Can ¡exis?ng ¡memory ¡management ¡ algorithms ¡solve ¡the ¡problem? ¡ 15

  16. Exis?ng ¡memory ¡management ¡ algorithms ¡do ¡not ¡solve ¡the ¡problem ¡ We run the application with Carrefour[1], the state-of-the-art memory management algorithm. Carrefour monitors memory accesses and places pages to minimize imbalance and maximize locality. Perf. improvement relative Carrefour solves imbalance / locality issues on some applications 30 to default Linux (%) 20 10 0 -10 THP -20 Carrefour-2M -30 C L U U M w S S U G A A r S P a m . C E . . t . B B C r D e i A C x m M . j 2 b u 0 b l t i p l y But does not improve performance on some other applications (hot pages or page-level false sharing) [1] DASHTI M., FEDOROVA A., FUNSTON J., GAUD F.,LACHAIZE R., LEPERS B., QUEMA V., AND ROTH M. Traffic management: A holistic approach to memory placement on NUMA systems. ASPLOS 2013. 16

  17. We ¡need ¡a ¡new ¡memory ¡management ¡ algorithm ¡ 17

  18. Our ¡solu?on ¡– ¡ Carrefour-­‑LP ¡ • Built on top of Carrefour. • By default, 2M pages are activated. • Two components that run every second: Reactive component Conservative component Splits 2M pages Promotes 4K pages Detects and removes “hot When the time spent pages” and page-level handling TLB misses is “false sharing”. high. Deactivate 2M page allocation Forces 2M page allocation In case of contention in the page fault handler. • We show in the paper that the two components are required. 18

  19. Implementa?on ¡ Reactive component (splits 2M pages) Sample memory accesses using IBS A page represents more YES than 5% of all Split and interleave the hot page accesses and is accessed from multiple nodes? 19

  20. Implementa?on ¡ Reactive component (splits 2M pages) Sample memory accesses using IBS • Compute observed local access ratio (LAR 1 ) • Compute the LAR that would have been obtained if each page was placed on the node that accessed it the most. LAR1 can be YES Run carrefour significantly improved? NO • Compute the LAR that would have been obtained if each page was split and then placed on the node that accessed it the most. LAR1 can be YES significantly Split all 2M pages and run carrefour improved? 20

  21. Implementa?on ¡challenges ¡ Reactive component (splits 2M pages) Sample memory accesses using IBS COSTLY • Compute observed local access ratio (LAR 1 ) • Compute the LAR that would have been obtained if each page was placed on the node that accessed it the most (without splitting). LAR1 can be YES Run carrefour significantly improved? IMPRECISE NO • Compute the LAR that would have been obtained if each page was split and then placed on the node that accessed it the most. LAR1 can be YES COSTLY significantly Split all 2M pages and run carrefour improved? 21

  22. Implementa?on ¡challenges ¡ Reactive component (splits 2M pages) • We only have few IBS samples. • The LAR with “2M pages split into 4K pages” can be wrong. • We try to be conservative by running Carrefour first and only splitting pages when necessary (splitting pages is expensive). • Predicting that splitting a 2M page will increase TLB miss rate is hard. This is why the conservative component is required. 22

  23. Implementa?on ¡ Conservative component Monitor time spent in TLB miss (hardware counters) YES Cluster 4K pages and force 2M pages allocation > 5% Monitor time spent in page fault handler (kernel statistics) YES Force 2M pages allocation > 5% 23

  24. Evalua?on ¡ The reactive and conservative components work together. Perf. improvement relative Machine A, 24 cores 30 to default Linux (%) 20 10 0 Carrefour-2M -10 Conservative Reactive -20 Carrefour-LP -30 C L U U M w S S U S P G A A r a m . . . t C E . B B C r D i e C A x m M . j b 2 b u 0 l t i p l y Perf. improvement relative Machine B, 64 cores 32 46 46 45 30 to default Linux (%) 20 10 0 Carrefour-2M -10 Conservative Reactive -20 Carrefour-LP -40 -30 C L U U M w S S U S P G A A r a m . . . t C E . B B C r D e i A C x m M . j 2 b b u 0 l t 24 i p l y

  25. Evalua?on ¡ • On the selected set of applications, our solution performs up to: • 46% better than Linux • 50% better than THP. (The full set of applications is available in the paper.) • Overhead: • Less than 3% CPU overhead. 25

  26. Conclusion ¡ • Large pages can hurt performance on NUMA systems. • We identified two new issues when using large pages on NUMA systems: “hot pages” and “page-level false sharing”. • We designed a new algorithm, Carrefour-LP, that: • Splits large pages when they hurt performance. • Promotes 4K pages and uses 2M page allocation when beneficial. • Carrefour-LP restores the performance when it was lost due to large pages and makes their benefits accessible to applications. 26

  27. Ques?ons? ¡

  28. 28

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend