NEC and CSCS A First Ivy Bridge Deployment

advertisement
NEC and CSCS
A First Ivy Bridge Deployment
Jason Temple, Luc Corbeil, HPC Solutions, CSCS
What is CSCS
Office building
•  5 floors
•  2’600 m2
•  Standard Minergie
•  Offices for 55 people
•  Two conference rooms
© CSCS 2013
2
CSCS in a nutshell
•  An autonomous unit of the Swiss Federal Institute of Technology in
Zurich (ETH Zurich)
•  founded in 1991 in Manno
•  relocated to Lugano in 2012
•  Develops and promotes technical and scientific services
•  for the Swiss research community in the field of high-performance
computing
•  Enables world-class scientific research
•  by pioneering, operating and supporting leading-edge supercomputing
technologies
© CSCS 2013
3
Management of Computing Resources
•  Two groups of highly skilled HPC system administrators
•  National Systems
–  Systems that are operated for the benefit of the Swiss scientific
community
–  Time allowed based on a call for proposal with peer-review
•  HPC Solutions
–  Hosting and management of HPC resources for Swiss organizations
–  Current partners:
–  MeteoSwiss (24/7 operational weather forecast)
–  EPFL/BlueBrain
–  CHiPP (data processing from LHC)
New project for HPC Solutions
•  Community of ETHZ professors needed HPC resources
–  Theoretical physics
–  Seismology
–  Polymer Physics
–  Astrophysics
–  Molecular dynamics
•  CSCS acts as a broker
–  Pooling of resources
–  Reduced overhead for each professor
–  Provider of a turnkey solution
Requirements
•  Common analysis of the requirements
–  Scientific
–  Subset of applications
–  Gromacs
–  VASP
–  AMR
–  SSE
–  Technical
–  Facilities constraints
–  Cluster design
–  Reliability, Availability, Serviceability (RAS)
–  Best solution within budget
Solution
•  NEC’s offer was selected
–  Full competitive WTO process
•  First Intel Ivy Bridge cluster in Switzerland
–  Maximum power consumption was lower than expected
–  Throughput per node improved
System Hardware Specifications
340 Standard Compute Nodes with two Sockets and 20 Cores
–  Two Intel Ivy Bridge EP E5-2660v2 @ 2.2 GHz
–  312 + 4 nodes with 32 GB (DDR3 1600 MHz), 24 with 64GB
© CSCS 2013
8 of 21
System Hardware Specifications
•  FDR fat tree, fully non-blocking, Mellanox SX6036
•  >300TB NEC LXFS v3 (based on Lustre 2.3), >15 GB/s
–  Built with NEC SNA260-FS building blocks
•  <100 kW under typical load
9
System Management, Compilers and Tools
•  LX C3 (LX Cluster Command and Control)
•  Node provisioning
•  Cluster monitoring
•  Slurm scheduler (installed and managed by CSCS)
•  Intel Cluster Studio
•  Intel C, C++, Fortran Compiler
•  OpenMP Version 3
•  Intel MPI (MPI 2.2 for C/C++ and Fortran)
•  Mathematical Libraries
•  Intel VTune XE for Linux
© CSCS 2013
10
•  Video
NEC: solid HPC partners
•  Very professional approach
–  Detailed installation planning
–  Planned followed with very minor variations
–  No bad surprises
•  Great benchmarking expertise
–  Prudent, conservative commitments
–  Performance commitments were all exceeded
•  Efficient support
–  Personalized attention
–  Pro-active replacement of suspicious hardware
Conclusion: Success Story!
Thank you for your attention.
Questions?
13
Download