SLIDE 11 CUG 2010
Early experiments that overlap communica+on with communica+on are promising for future HPC systems
!" #!" $!" %!" &!" '!!" '#!" '$!" '%!" '&!" #!!" '!#$" ('#" #(%" )*+,"-.,/0" 123"245/,..,.6"7#6$6&8"59,:12";<4,=>."9,4"123"945/,.."
)=.?*:@"" A4*@*:=B"
!" #!" $!!" $#!" %!!" %#!" &!!" $!%'" #$%" %#("
)*+,"-.,/0"
123"245/,..,.6"7%6'689"5:,;12"<=4,>?.":,4"123":45/,.."
)>.@*;A"" B4*A*;>C"
- Overlapping MPI communica+on with other consecu+ve, data independent MPI
Communica+on
- Here: itera+ve execu+on of two consecu+ve MPI_Allreduce with small and larger
messages on Hopper Cray XT5
- GTS shicer or pusher rou+nes have such consecu+ve MPI communica+on
- Overlapping MPI_Allreduce with larger messages (~1K bytes) pays off when ra+o of
threads/sockets per node is reasonable
- Future HPC systems are expected to have many communica+on channels per node