The DOSAR VO Statement of Work

advertisement
The DOSAR VO Statement of Work
DOSAR (Distributed Organization for Scientific and Academic Research) is a 'grassroots' grid organization which evolved from the Southern Analysis Region consortium of
D0 institutions as that group expanded beyond the realms of both D0 and the southern
region of the US.
DOSAR is involved in many aspects of 'community' grid activities, such as deploying
desktop computer clusters and generic computing resources for grid computing. The
following ongoing grid computing activities and development at DOSAR institutions will
be conducted:
Integration Testbed and Production at Louisiana Tech University
At Louisiana Tech University, both Integration Testbed and Production work are being
conducted. The Integration work is spearheaded by Box Leangsuksun, a well-know
expert in the area of High Available grid computing. Leangsuksun, assisted by Physicist
Dick Greenwood, and their students, is applying techniques successfully developed for
the HA-OSCAR software to the OSG software stack. In order to eliminate single-pointfailures, the following features will be incorporated in the OSG software: self-healing;
failure detection; recovery; automatic fail-over; and automatic fail-back. The production
work, directed by Dick Greenwood, presently includes SAMGrid-OSG MC Production at
LTU and at the Center for Computation and Technology (CCT) at Louisiana State
University (LSU) via the Louisiana Optical Network Initiative (LONI). Starting in
January, 2007, D0 data reprocessing will begin on 250 new Dell processors on LONI grid
using SAMGrid-OSG, and later in the Spring, 2007, ATLAS production will begin on the
LONI grid using the OSG stack.
Tier 2 Production and D0 Reprocessing at Sao Paulo
Another one of the D0 Reprocessing sites employing SAMGrid-OSG is SPRACE, the
grid computing center at Sao Paulo, Brazil. SPRACE is an active OSG CMS Tier2
associated with FNAL Tier1.
Production and Cluster Development at Iowa State
At Iowa State University (Jim Cochran and Charles Zaruba), a cluster is being developed
to do ATLAS production. Also, there is an ongoing effort to install MS Virtual Server
(similar to VMware) on about 50 departmental Windows machines to allow Linux batch
operation under Condor when these units are otherwise idle.
Tier 2 Production and Monitoring Software Development at UT-Arlington
At UT Arlington, Kaushik De is co-leading the production software development for
ATLAS and Jae Yu is directing an effort to produce MonALISA based ATLAS
distributed analysis monitoring. UT Arlington and the University of Oklahoma (OU) are
co-hosts of the ATLAS Southwest Tier 2 computing center.
Tier 2 Production, D0 Production, and Campus/community Grid, and Cyber
infrastructure (CI) Activities at the University of Oklahoma
At OU, a Condor pool has been created that currently consists of about 200 student lab
PCs, but is scheduled to increase to 750 by the Spring. It has the OSG middleware stack
installed, and is already being used by OSG VOs (including DOSAR and D0) and local
OU users. Non High Energy Physics science being conducted on that Condor pool
includes materials science (nano technology) and data storage research (data encoding for
storage media). The project recently got additional funding by a CI-TEAM grant, which
will enable us to more efficiently incorporate the already existing campus Condor pool
into both research and education. The OUHEP group has also created a Condor pool
using desktop computers dedicated to High Energy Physics, which is being used for both
ATLAS and D0 production and is currently also serving as a data storage cache for the
SAMGrid-OSG interface for the entire D0 collaboration.
Download