The DOSAR VO Statement of Work DOSAR (Distributed Organization for Scientific and Academic Research) is a 'grassroots' grid organization which evolved from the Southern Analysis Region consortium of D0 institutions as that group expanded beyond the realms of both D0 and the southern region of the US. DOSAR is involved in many aspects of 'community' grid activities, such as deploying desktop computer clusters and generic computing resources for grid computing. The following ongoing grid computing activities and development at DOSAR institutions will be conducted: Integration Testbed and Production at Louisiana Tech University At Louisiana Tech University, both Integration Testbed and Production work are being conducted. The Integration work is spearheaded by Box Leangsuksun, a well-know expert in the area of High Available grid computing. Leangsuksun, assisted by Physicist Dick Greenwood, and their students, is applying techniques successfully developed for the HA-OSCAR software to the OSG software stack. In order to eliminate single-pointfailures, the following features will be incorporated in the OSG software: self-healing; failure detection; recovery; automatic fail-over; and automatic fail-back. The production work, directed by Dick Greenwood, presently includes SAMGrid-OSG MC Production at LTU and at the Center for Computation and Technology (CCT) at Louisiana State University (LSU) via the Louisiana Optical Network Initiative (LONI). Starting in January, 2007, D0 data reprocessing will begin on 250 new Dell processors on LONI grid using SAMGrid-OSG, and later in the Spring, 2007, ATLAS production will begin on the LONI grid using the OSG stack. Tier 2 Production and D0 Reprocessing at Sao Paulo Another one of the D0 Reprocessing sites employing SAMGrid-OSG is SPRACE, the grid computing center at Sao Paulo, Brazil. SPRACE is an active OSG CMS Tier2 associated with FNAL Tier1. Production and Cluster Development at Iowa State At Iowa State University (Jim Cochran and Charles Zaruba), a cluster is being developed to do ATLAS production. Also, there is an ongoing effort to install MS Virtual Server (similar to VMware) on about 50 departmental Windows machines to allow Linux batch operation under Condor when these units are otherwise idle. Tier 2 Production and Monitoring Software Development at UT-Arlington At UT Arlington, Kaushik De is co-leading the production software development for ATLAS and Jae Yu is directing an effort to produce MonALISA based ATLAS distributed analysis monitoring. UT Arlington and the University of Oklahoma (OU) are co-hosts of the ATLAS Southwest Tier 2 computing center. Tier 2 Production, D0 Production, and Campus/community Grid, and Cyber infrastructure (CI) Activities at the University of Oklahoma At OU, a Condor pool has been created that currently consists of about 200 student lab PCs, but is scheduled to increase to 750 by the Spring. It has the OSG middleware stack installed, and is already being used by OSG VOs (including DOSAR and D0) and local OU users. Non High Energy Physics science being conducted on that Condor pool includes materials science (nano technology) and data storage research (data encoding for storage media). The project recently got additional funding by a CI-TEAM grant, which will enable us to more efficiently incorporate the already existing campus Condor pool into both research and education. The OUHEP group has also created a Condor pool using desktop computers dedicated to High Energy Physics, which is being used for both ATLAS and D0 production and is currently also serving as a data storage cache for the SAMGrid-OSG interface for the entire D0 collaboration.