Purdue RP Update, September 2010

advertisement
Purdue RP Highlights
TeraGrid Round Table
September 23, 2010
Carol Song
Purdue TeraGrid RP PI
Rosen Center for Advanced Computing
Purdue University
Infrastructure udpates
• Steele moved to a new data center
• New addition to community clusters – the Rossmann
cluster
– Installed, up and running in early Sept.
– Currently 8800+ cores
– HP ProLiant DL165 G7 nodes with dual 12-core AMD Opteron
6172 processors (24 cores per node)
– 48 GB or 96 GB RAM, and 250 GB of local disk on each node
– 10Gbit Ethernet
– 150 TB Lustre filesystem
• Purdue Condor resource now have ~ 42,300 cores
TeraGrid Round Table, 5/20/2010
Storage
• DC-WAN mounted and used at Purdue
– Working on Lustre lnet routers to reach compute nodes
by the end of October.
– Installing a system to act as a Lustre WAN router now,
and expect to have necessary configuration in place
during IU's next DC-WAN maintenance window.
– Testing and deployment to follow
• Distributed Replication Service
– Sharing spinning disk to DRS today
– Investigating integration with Hadoop Filesystem (HDFS)
TeraGrid Round Table, 5/20/2010
New Developments in Condor Pool
• Virtual Machine “Universe”
Running on student Windows labs today – with VMWare
Integrating now: KVM and libVirt on cluster (Steele) nodes
TeraGrid Round Table, 5/20/2010
VM Controller
• Summer project
• Undergrad student
summer intern (E.
Albersmeyer from
CIT Technology)
• Skills Used
• C# Programming
• VMware Knowledge
• UNIX Admin
Knowledge
• Windows 7
• Control VM state (on, off depending on user activity)
• Provide usage information to system owner – important in a
community resource environment
• Study performance trade-offs
TeraGrid Round Table, 5/20/2010
Cloud Computing: Wispy
• Purdue staff operating experimental cloud resource
– Built with Nimbus from UC
– Current Specs
• 32 nodes (128 cores):
– 16 GB RAM
– 4 cores per node
– Public IP space for VM guests
• Available for allocation in POPS now
• Use case
TeraGrid Round Table, 5/20/2010
Wispy – Use Cases
• Used in Virtual Clusters
– Publications using Purdue’s Wispy cited below
• NEES project exploring using Wispy to provision ondemand clusters for quick turn-around of wide parallel
jobs
• With OSG team, using Wispy (and Steele) to run VMs
for STAR project
• Working with faculty at Marquette Univ. to use Wispy
in Fall 2010 course to teach cloud computing concepts
•“CloudBLAST: Combining MapReduce and Virtualization on Distributed Resources for
Bioinformatics Applications” by A. Matsunaga, M. Tsugawa and J. Fortes. eScience 2008.
•“Sky Computing”, by K. Keahey, A. Matsunaga, M. Tsugawa, J. Fortes, to appear in IEEE
Internet Computing, September 2009
TeraGrid Round Table, 5/20/2010
Clouded Computational Sciences
• Craig A. Struble
Department of Mathematics, Statistics, and Computer Science
Marquette University
• Fall 2010 semester
• A course covering topics related to science clouds and cloud
architectures
• Students should be able to
–
–
–
–
Describe different cloud architectures
Identify prevalent cloud middleware
Address problems in computational sciences using a cloud
Understand security and ethical concerns surrounding science clouds
• Required to do
–
–
–
–
–
Instantiate
Instantiate
Instantiate
Instantiate
…
a single VM serving a web page, etc.
4 node cluster of VMs, running Condor jobs.
4,8,16 nodes of VM clusters, running parallel, MPI based apps
an 8-node Hadoop cluster to process a large text DB
TeraGrid Round Table, 9/23/2010
Purdue – NCAR Collaboration
• Develop Earth Science Gateways
• Leverage NCAR’s work on ESG and ESG-curator (may have
new names now) and Purdue’s Climate Model gateway
(CCSM portal)
• Build around CCSM4, the new version
• Architectural changes – service oriented
– CCSM4 web services
• Purdue migrating its CCSM3 gateway to also use the new
web services
• Publish climate model data to ESG
• NCAR workflow application will use CCSM4 web service to
run on TG
TeraGrid Round Table, 9/23/2010
ExTENCI
• Joint project between OSG and TeraGrid
• Kickoff meeting: August 19 at Fermilab
• Distributed File System (Lustre-WAN)
– CMS/ATLAS HEP
– Ralph Roskies
• Workflow & Client Tools
– SCEC & Protein Folding
– Daniel S. Katz / Mike Wilde
• Job Submission Paradigms
– Cactus Application
– Shantenu Jha / Miron Livny
• Virtual Machines (Carol Song/Sebastien Goasguen)
– STAR & CMS
– Demonstrated: STAR cloud VM running at Purdue, joining their simulation cloud
– discussed using Wispy's SOAP interface with a CMS group to let their Glidein-WMS
infrastructure start resources on the cloud to join their Glidein pool. Plan to
demonstrate this soon.
TeraGrid Round Table, 9/23/2010
Download