future trends in hardw are and softw are for use in
play

Future Trends in Hardw are and Softw are for use in Simulation - PowerPoint PPT Presentation

Future Trends in Hardw are and Softw are for use in Simulation Steve Feldman VP/IT, CD-adapco April, 2009 H igh P erformance C omputing Building Blocks CPU I/O Interconnect Software General CPU Maximum clock speeds have


  1. Future Trends in Hardw are and Softw are for use in Simulation Steve Feldman VP/IT, CD-adapco April, 2009

  2. H igh P erformance C omputing Building Blocks • CPU • I/O • Interconnect • Software

  3. General CPU Maximum clock speeds have remained relatively constant. • Higher speeds require too much power and cooling. Cpu’s simply are not getting faster. Manufacturers have and will be able to put more cores on • each chip (dual core, quad-core, many-core) Memory bandwidth will likely be the constraint for HPC to • efficiently employ many-core architectures. CD-adapco products treat each core as a single cpu. No • special tricks required to run in parallel. Each core requires an hpc (or Power session) license just like any other cpu whether it resides on the same node or a different one.

  4. GPGPU Single precision floating point speed is exceptional but • double precision is far less interesting. Programming is difficult. No (current) standard for uniform • programming. OpenCL has just been approved. Questionable memory access speed from main memory. •

  5. I/O Performance Many cpu’s all writing at once can overwhelm a cheap I/O • system. Transients in particular write lots of data. • RAID systems allow for the possibility of disk failures • and/or disk striping for better performance. SATA in RAID seems to have good performance at reasonable prices. I/O hardware can be hung off of Infiniband, GigE, fibre • channel or directly connected to a dedicated I/O server. There are a number of parallel file systems and associated • hardware that can handle intensive I/O demands of a large HPC cluster.

  6. Interconnect Allows cpu’s to exchange information • Important characteristics • – Bandwidth – How quickly can I stream large arrays? – Latency – How long for a node to signal back that the data (of any size) was received. CD-adapco solvers have some loops that are bandwidth • bound and others that are latency bound. Transients are more interconnect sensitive than steady- • state analyses. Small problems (fewer cells/cpu) are more interconnect • sensitive than large ones. Higher node counts require better interconnects.

  7. Interconnect Characteristics Interconnect Bandwidth(Mbytes/s) Latency(µs) GigE 100 40-60 10GigE 1000 40-60 Myrinet “D” (IP) 50 30 Myrinet “D” (native) 162 10 Infiniband (TCP/IP) 100 50 Infiniband (native-SDR) <1000 2-5 Infiniband (native-DDR) <2000 1.8-5 Infiniband (native-QDR) <4000 ? Cray Rapid Array 1000 2 Shared Mem (ibm pwr4) 2012 3

  8. Software Components OS • OS Management (Cluster Software) • Job Management (Queueing) • IO Management •

  9. OS CD-adapco supports all the major Unix flavors • – AIX, HP-UX, Solaris (shorter list every year) CD-adapco supports major Linux flavors • – Red Hat Enterprise – Suse Enterprise – Others may or may not work but are not supported Microsoft Windows • – Windows Server 2008 HPC support complete for STAR-CD and STAR-CCM+ – WinXP64 and Vista are supported but not for multi-node clusters

  10. Cluster Management Software Propagates the OS, upgrades, changes to all nodes • Provides views of all nodes from one location • A cluster is never as easy to maintain as one single • instance of an OS but neither should it be N times harder for N individual nodes.

  11. Queuing Software Submits jobs in an orderly fashion • Resource managers – applies open cpu’s to queued jobs •

  12. I/O Management May be as simple as NFS (Unix/Linx) or standard Windows • shared drives May be much more complicated dealing with multiple I/O • nodes, parallel file servers (on I/O nodes) and clients (on compute nodes).

  13. Questions for Panel (I) CPU - Intel and AMD (and IBM Power) have all solved the • problem of memory bandwidth to quad cores. Will this approach scale to higher core counts? How will higher core counts affect interconnect needs? Will • Infiniband be able to keep up with increasing numbers of cores connected to a single path? Will something else emerge? How will higher core counts affect I/O streams? What will • high performance I/O systems look like in the future? Will pNFS work or will we continue to rely on proprietary parallel file systems?

  14. Questions for Panel (II) GPU’s have weak double precision performance, memory • throughput is questionable and there is no current single environment to program all makes. Will they ever become part of the CAE/HPC world? How? Clusters are not easy to manage. Is there anything to look • forward to in Software that will make them look more like a single machine to the IT admins? What will a supercomputer look like 5 or more years from • now? Is there a “breakthrough” architecture that will allow faster – clock rates on the horizon? – Are there any other disruptive technologies to look for?

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend