hpc infrastructures hpc infrastructures
play

HPC Infrastructures HPC Infrastructures Moreno Baricevic CNR-INFM - PowerPoint PPT Presentation

Distributed Applications, Web Services, Tools and GRID Infrastructures for Bioinformatics HPC Infrastructures HPC Infrastructures Moreno Baricevic CNR-INFM DEMOCRITOS, Trieste NETTAB 2006 - Santa Margherita di Pula (CA) - July 10-13, 2006


  1. Distributed Applications, Web Services, Tools and GRID Infrastructures for Bioinformatics HPC Infrastructures HPC Infrastructures Moreno Baricevic CNR-INFM DEMOCRITOS, Trieste NETTAB 2006 - Santa Margherita di Pula (CA) - July 10-13, 2006

  2. SOFTWARE INFRASTRUCTURE SOFTWARE INFRASTRUCTURE Overview Overview Users' Parallel Applications Users' Serial Applications Parallel Environment: MPI/PVM Software Tools for Applications GRID-enabling software (compilers, scientific libraries) Management Software (installation, administration, monitoring, resource management) O.S. Network Storage + (fast interconnection (shared and parallel services among nodes) file systems) 2

  3. SOFTWARE INFRASTRUCTURE SOFTWARE INFRASTRUCTURE Overview (Michelangelo @ CILEA) Overview (Michelangelo @ CILEA) Fortran, C/C++ codes Fortran, C/C++ codes MVAPICH INTEL, PGI, GNU compilers BLAS, LAPACK, ScaLAPACK, ATLAS, ACML, FFTW libraries LCG-2 / gLite (EGEE II) C3Tools, SSH, blade , ad-hoc scripts Ganglia, Nagios PBS/TORQUE batch system + MAUI scheduler InfiniBand, NFS LINUX CentOS Gigabit Ethernet SAN + GFS 3

  4. COMPATIBILITY ISSUES COMPATIBILITY ISSUES Kernel vs new hardware Kernel vs new hardware Latest vanilla kernel: 2.6.16 .9 CentOS kernel: 2.6.9 -22 UnionFS v1.1.4 ↔ kernel 2.6.9 ÷ 2.6.14 InfiniBand IBGD-1.8.2 ↔ kernel ≤ 2.6.11 GFS cluster 1.01 ↔ kernel ≤ 2.6.14 GFS cluster 1.02 ↔ kernel 2.6.15 patched by FC5 Qlogic qla2xxx (severe bug fixed) ↔ kernel ≥ 2.6.15 AMD CPU Dual Core 275 ↔ kernel ≥ 2.6.12 [Up to May 2006] 4

  5. COMPATIBILITY ISSUES COMPATIBILITY ISSUES Kernel vs new hardware Kernel vs new hardware kernel version 2.6. ... 9 10 11 12 13 14 15 16 ... - Latest vanilla kernel 2.6.16 - CentOS kernel 2.6.9 - UnionFS v1.1.4 2.6.9 ÷ 2.6.14 - InfiniBand IBGD-1.8.2 ≤ 2.6.11 - GFS cluster 1.01 ≤ 2.6.14 2.6.15 - GFS cluster 1.02 (patched by FC5) - Qlogic qla2xxx (severe bug fixes) ≥ 2.6.15 - AMD CPU Dual Core 275 ≥ 2.6.12 [Up to May 2006] 5

  6. COMPATIBILITY ISSUES COMPATIBILITY ISSUES Kernel vs new hardware Kernel vs new hardware Roll up your own kernel and patch as needed! vanilla kernel 2.6.16.16 UnionFS 1.1.4 (patched) IBGD 1.8.2 (patched) GFS cluster 1.02 (patched) Qlogic qla2xxx (bug fixed) AMD CPU Dual Core 275 (supported) 6

  7. CLUSTER SERVICES CLUSTER SERVICES PRIVATE NETWORK NTP NTP CLUSTER-WIDE TIME SYNC DNS DNS DYNAMIC HOSTNAMES RESOLUTION SERVER / MASTERNODE DHCP INSTALLATION / CONFIGURATION LAN (+switches backup and configuration) TFTP NFS SHARED FILESYSTEM REMOTE ACCESS SSH SSH FILE TRANSFER PARALLEL COMPUTATION (MPI) LDAP/NIS/... LDAP/NIS/... AUTHENTICATION ... 7

  8. CLUSTER MANAGEMENT CLUSTER MANAGEMENT Installation Installation Installation can be performed: - interactively - non-interactively Interactive installations: - finer control Non-interactive installations: - minimize human intervention and let you save a lot of time - are less error prone - are performed using programs (such as RedHat Kickstart) which: - “simulate” the interactive answering - can perform some post-installation procedures for customization 8

  9. CLUSTER MANAGEMENT CLUSTER MANAGEMENT Installation Installation MASTERNODE Ad-hoc installation once forever (hopefully), usually interactive: - local devices (CD-ROM, DVD-ROM, Floppy, ...) - network based (PXE+DHCP+TFTP+NFS) CLUSTER NODES One installation reiterated for each node, usually non-interactive. Nodes can be: 1) disk-based 2) disk-less (not to be really installed) 9

  10. CLUSTER MANAGEMENT CLUSTER MANAGEMENT Cluster Nodes Installation Cluster Nodes Installation 1) Disk-based nodes - CD-ROM, DVD-ROM, Floppy, ... Time expensive and tedious operation - HD cloning: mirrored raid, dd and the like A “template” hard-disk needs to be swapped or a disk image needs to be available for cloning, configuration needs to be changed either way - Distributed installation: PXE+DHCP+TFTP+NFS More efforts to make the first installation work properly (especially for heterogeneous clusters), (mostly) straightforward for the next ones 2) Disk-less nodes - Live CD/DVD/Floppy - NFS - NFS + UnionFS - initrd (RAM disk) 10

  11. CLUSTER MANAGEMENT CLUSTER MANAGEMENT Existent toolkits Existent toolkits Are generally made of an ensemble of already available software packages thought for specific tasks, but configured to operate together, plus some add-ons. Sometimes limited by rigid and not customizable configurations, often bounded to some specific LINUX distribution and version. May depend on vendors' hardware. Free and Open - OSCAR (Open Source Cluster Application Resources) - NPACI Rocks - xCAT (eXtreme Cluster Administration Toolkit) - OpenSCE (Open Scalable Cluster Environment) - Warewulf Commercial - IBM CSM (Cluster Systems Management) - Scyld Beowulf - HP, SUN and other vendors' Management Software... 11

  12. CLUSTER MANAGEMENT CLUSTER MANAGEMENT Network-based Distributed Installation Network-based Distributed Installation PXE DHCP TFTP INITRD INSTALLATION ROOTFS OVER NFS Kickstart/Anaconda NFS + UnionFS Customization Customization through through Post-installation UnionFS layers 12

  13. CLUSTER MANAGEMENT CLUSTER MANAGEMENT Network-based Distributed Installation Network-based Distributed Installation DHCPDISCOVER PXE DHCP DHCPOFFER IP Address / Subnet Mask / Gateway / ... Network Bootstrap Program (pxelinux.0) CLIENT / COMPUTING NODE SERVER / MASTERNODE DHCPREQUEST PXE DHCP DHCPACK PXE tftp get pxelinux.0 DHCP PXE TFTP TFTP INITRD tftp get pxelinux.cfg/HEXIP PXE+NBP TFTP tftp get kernel foobar PXE+NBP TFTP tftp get initrd foobar.img kernel foobar TFTP 13

  14. CLUSTER MANAGEMENT CLUSTER MANAGEMENT Network-based Distributed Installation Network-based Distributed Installation get NFS:kickstart.cfg kernel + initrd NFS get RPMs anaconda+kickstart NFS CLIENT / COMPUTING NODE SERVER / MASTERNODE tftp get tasklist kickstart: %post TFTP Installation tftp get task#1 kickstart: %post TFTP tftp get task#N kickstart: %post TFTP tftp get pxelinux.cfg/default kickstart: %post TFTP tftp put pxelinux.cfg/HEXIP kickstart: %post TFTP 14

  15. CLUSTER MANAGEMENT CLUSTER MANAGEMENT Network-based Distributed Installation Network-based Distributed Installation mount /hopeless/roots/root kernel + initrd NFS+UnionFS ROOTFS over NFS+UnionFS CLIENT / COMPUTING NODE SERVER / MASTERNODE mount /hopeless/roots/overlay kernel + initrd NFS+UnioNFS mount /hopeless/roots/gfs kernel + initrd NFS+UnionFS mount /hopeless/clients/IP kernel + initrd NFS+UnionFS RW /hopeless/roots/192.168.10.1 /hopeless/roots/gfs RO RO /hopeless/roots/overlay RO /hopeless/roots/root Resultant file system RW! 15 DELETED FILEs NEW FILEs

  16. CLUSTER MANAGEMENT CLUSTER MANAGEMENT Administration Tools Administration Tools Requirements: ✔ cluster-wide command execution ✔ cluster-wide file distribution and gathering ✔ must be simple, efficient, easy to use for CLI addicted C3 tools - The Cluster Command and Control tool suite allows configurable clusters and subsets of machines concurrently execution of commands supplies many utilities cexec (parallel execution of standard commands on all cluster nodes) cexecs (as the above but serial execution, useful for troubleshooting and debugging) cpush (distribute files or directories to all cluster nodes) cget (retrieves files or directory from all cluster nodes) crm (cluster-wide remove) ... and many more http://www.csm.ornl.gov/torc/C3/ DSH - Distributed Shell http://www.netfort.gr.jp/~dancer/software/dsh.html.en 16

  17. CLUSTER MANAGEMENT CLUSTER MANAGEMENT Monitoring Tools Monitoring Tools Ad-hoc scripts (BASH, PERL, ...) + cron Ganglia excellent graphic tool XML data representation web-based interface for visualization http://ganglia.sourceforge.net/ Nagios complex but can interact with other software configurable alarms, SNMP, E-mail, SMS, ... optional web interface http://www.nagios.org/ 17

  18. CLUSTER MANAGEMENT CLUSTER MANAGEMENT Ganglia at work /1 Ganglia at work /1 18

  19. CLUSTER MANAGEMENT CLUSTER MANAGEMENT Ganglia at work /2 Ganglia at work /2 19

  20. STORAGE STORAGE Shared and Parallel File Systems Shared and Parallel File Systems A shared file system to ease management and supply a centralized repository: Performance is not an issue! NFS – Network File System A file system to deal with intensive I/O operations both serial and parallel (parallel file system). Available choices: Performance IS an issue! GFS – Global File System GPFS – Global Parallel File System PVFS – Parallel Virtual File System Lustre 20

  21. STORAGE STORAGE Shared File System: NFS Shared File System: NFS Central repository for: packages (installation/updates) cluster-wide configurations libraries non-critical executables (not needed at boot-up) sporadic non I/O intensive operations ... Can supply the root file system (and/or UnionFS layers) for disk- less nodes and can export the /home file system as well. 21

  22. STORAGE STORAGE Parallel File System: GFS - - Features Parallel File System: GFS Features designed from scratch as a cluster-based distributed file system works in a SAN/LAN environment single system image style view of the file system (consistency) fully 64bit journaled works with LVM volume managers scalable 22

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend