An Introduction to Princeton’s New Computing Resources: IBM Blue Gene,

advertisement
An Introduction to Princeton’s New
Computing Resources: IBM Blue Gene,
SGI Altix, and Dell Beowulf Cluster
PICASso Mini-Course
October 18, 2006
Curt Hillegas
Introduction
•
•
•
•
•
SGI Altix - Hecate
IBM Blue Gene/L – Orangena
Dell Beowulf Cluster – Della
Storage
Other resources
TIGRESS High Performance
Computing Center
Terascale
Infrastructure for
Groundbreaking
Research in
Engineering and
Science
Partnerships
• Princeton Institute for Computational
Science and Engineering (PICSciE)
• Office of Information Technology (OIT)
• School of Engineering and Applied Science
(SEAS)
• Lewis-Sigler Institute for Integrative
Genomics
• Astrophysical Sciences
• Princeton Plasma Physics Laboratory
(PPPL)
SGI Altix - Hecate
•
•
•
•
•
64 1.5 GHz Itanium2 processors
256 GB RAM (4 GB per processor)
NUMAlink interconnect
5 TB local disk
360 GFlops
SGI Altix – Itanium 2
• 1.5 GHz
• 4 MB L3 Cache
– 256 KB L2 Cache
– 32 KB L1 Cache
SGI Altix - NUMAlink
•
•
•
•
•
NUMAlink 4
3.2 GB/s per direction
Physical latency – 28 ns
MPI latency – 1 ms
Up to 256 processors
SGI Altix - Software
• SLES 9 with SGI ProPack
•
•
•
•
•
•
•
•
•
•
•
– 2.6.5-7.252-sn2 kernel
Intel Fortran compilers v8.1
Intel C/C++ compilers v8.1
Intel Math Kernel Libraries v7
Intel vtune
Torque/Maui
OpenMP
MPT (SGI mpich libraries)
fftw-2.1.5, fftw-3.1.2
hdf4, hdf5
ncarg
petsc
IBM Blue Gene/L - Orangena
•
•
•
•
•
•
2048 700 MHz Power4 processors
1024 nodes
512 MB RAM (256 MB per processor)
5 Interconnects including a 3D torus
8 TB local disk
4.713 TFlops
IBM Blue Gene/L – Full system
architecture
• 1024 nodes
–
–
–
–
–
2 PowerPC 440 cpus
512 MB RAM
1 rack
35 kVA
100 kBTU/hr
–
–
–
–
–
Service node
Front end node
8 storage nodes
8 TB GPFS storage
1 Cisco switch
• 2 racks of supporting servers and disks
IBM Blue Gene/L
IBM Blue Gene/L - networks
•
•
•
•
•
3D Torus network
Collective (tree) network
Barrier network
Functional network
Service network
IBM Blue Gene/L - Software
• LoadLeveler (coming soon)
• mpich
• XL Fortran Advanced Edition V9.1
– mpxlf, mpf90, mpf95
• XL C/C++ Advanced Edition V7.0
•
•
•
•
– Mpcc, mpxlc, mpCC
fftw-2.1.5 and fftw-3.0.1
hdf5-1.6.2
netcdf-3.6.0
BLAS, LAPACK, ScaLAPACK
IBM Blue Gene/L – More…
• http://orangena.Princeton.EDU
• http://orangena-sn.Princeton.EDU
Dell Beowulf Cluster - Della
•
•
•
•
•
•
•
512 3.2 GHz Xeon processors
256 nodes
2 TB RAM (4 GB per processor)
Gigabit Ethernet
64 nodes connected to Infiniband
3 TB local disk
1.922 TFlops
Dell Beowulf Cluster –
Interconnects
• All nodes connected with Gigabit
Ethernet
– 1 Gb/s
– MPI latency ~ 30 ms
• 64 nodes connected with Infiniband
– 10 Gb/s
– MPI latency ~5 ms
Dell Beowulf Cluster - Software
• Elders RHEL 4 based image
– 2.6.9-42.0.3.ELsmp kernel
•
•
•
•
•
•
Intel compilers
Torque/Maui
OpenMPI-1.1
fftw-2.1.5, fftw-3.1.2
R-2.1.3
MatlabR2006a
Dell Beowulf Cluster – More…
• https://della.Princeton.EDU
• https://della.Princeton.EDU/ganglia
Storage
•
•
•
•
•
38 TB delivered
GPFS filesystem
At least 200 MB/s
Installation at the end of this month
Fees to recover half the cost
Getting Access
• 1 – 3 page proposal
• Scientific background and merit
• Resource requirements
–
–
–
–
# concurrent cpus
Total cpu hours
Memory per process/total memory
Disk space
• A few references
• curt@Princeton.EDU
Other resources
• adrOIT
• Condor
• Programming help
Questions
Download