Open Science Grid James “Kent” Blackburn OSG Resources Manager Senior Scientist LIGO Laboratory California Institute of Technology OSG’s Coverage of the CI “Bubble” Diagram Training Instrumentation Control Help Desk Researcher Security Viewing Data Generation Program Security Collab Tools Control Policy and Funding Management Security and Access Access Control Human Support Funding Agencies Authentication Education And Outreach Publishing Network Resource Providers Campuses Authorization Security Security Input Analysis Archive Simulation Computation Program OSG Display Tools Retrieval Data Sets Storage Search Schema Data Metadata Directories Ontologies Data . Input 3D Imaging Security Display and Visualization OSG Consortium The Open Science Grid OSG is a Cyberinfrastructure for Research The OSG is a framework for large scale distributed resource sharing, addressing the technology, policy and social requirements of sharing • • The Open Science Grid’s mission is to help satisfy the ever-growing computing and data management requirements of researchers by enabling them to share a greater percentage of available computer cycles and software with less effort. The OSG is a distributed, common cyberinfrastructure spanning campus, regional, national and international boundaries. At over 50 provider sites, independently-owned and managed resources make up the distributed facility; agreements between members provided the glue; their requirements drive the evolution; their effort helps make it happen. • The facility is dedicated to high throughput computing and is open to researchers from all domains. OSG Consortium Partners Academia Sinica Argonne National Laboratory (ANL) Boston University Brookhaven National Laboratory (BNL) California Institute of Technology Center for Advanced Computing Research Center for Computation & Technology at Louisiana State University Center for Computational Research, The State University of New York at Buffalo Center for High Performance Computing at the University of New Mexico Columbia University Computation Institute at the University of Chicago Cornell University DZero Collaboration Dartmouth College Fermi National Accelerator Laboratory (FNAL) Florida International University Georgetown University Hampton University Indiana University Indiana University-Purdue University, Indianapolis International Virtual Data Grid Laboratory (iVDGL) Kyungpook National University Laser Interferometer Gravitational Wave Observatory (LIGO) Lawrence Berkeley National Laboratory (LBL) Lehigh University Massachusetts Institute of Technology National Energy Research Scientific Computing Center (NERSC) National Taiwan University New York University Northwest Indiana Computational Grid Notre Dame University Pennsylvania State University Purdue University Rice University Rochester Institute of Technology Sloan Digital Sky Survey (SDSS) Southern Methodist University Stanford Linear Accelerator Center (SLAC) State University of New York at Albany State University of New York at Binghamton State University of New York at Buffalo Syracuse University T2 HEPGrid Brazil Texas Advanced Computing Center Texas Tech University Thomas Jefferson National Accelerator Facility University of Arkansas Universidade de São Paulo Universideade do Estado do Rio de Janerio University of Birmingham University of California, San Diego University of Chicago University of Florida University of Illinois at Chicago University of Iowa University of Michigan University of Nebraska - Lincoln University of New Mexico University of North Carolina/Renaissance Computing Institute University of Northern Iowa University of Oklahoma University of South Florida University of Texas at Arlington University of Virginia University of Wisconsin-Madison University of Wisconsin-Milwaukee Center for Gravitation and Cosmology Vanderbilt University Wayne State University What The OSG Offers • Low-threshold access to many distributed computing and storage resources • A combination of dedicated, scheduled, and opportunistic computing • The Virtual Data Toolkit software packaging and distributions • Grid Operations, including facility-wide monitoring, validation, information services and system integration testing • Operational security • Troubleshooting of end-to-end problems • Education and Training The OSG as a Community Alliance • The OSG is a grass-roots endeavor bringing together research institutions throughout the U.S. and the World. The OSG Consortium brings together the stakeholders. The OSG Facility brings together resources and users. • The OSG’s growing alliance of universities, national laboratories, scientific collaborations and software developers, contribute to the OSG, share ideas and technologies reap the benefits of the integrated resources through both agreements with fellow members and opportunistic use. • • An active engagement effort adds new domains and resource providers to the OSG Consortium. Training is offered at semi-annual OSG Consortium meetings and through educational activities organized in collaboration with TeraGrid. One to three day hands-on training sessions are offered around the U.S and abroad for users, administrators and developers. OSG Community Structure Virtual Organizations (VOs) • The OSG community shares/trades in groups (VOs) not individuals VO Management Service • VO management services allow registration, administration and control of members within VOs • Facilities trust and authorize VOs • Compute and storage services prioritize according to VO group membership Campus Grid Image courtesy: UNM OSG and WAN VO Management & Applications Set of Available Resources Experimental Project Grid Image courtesy: UNM Campus Grids • They are a fundamental building block of the OSG The multi-institutional, multi-disciplinary nature of the OSG is a macrocosm of many campus IT cyberinfrastructure coordination issues. • Currently OSG has three operational campus grids on board: Fermilab, Purdue, Wisconsin Working to add Clemson, Harvard, Lehigh • Elevation of jobs from Campus CI to OSG is transparent • Campus scale brings value through Richness of common software stack with common interfaces Higher common denominator makes sharing easier Greater collective buying power with venders Synergy through common goals and achievements Current OSG Resources • OSG has more than 50 participating institutions, including self-operated research VOs, campus grids, regional grids and OSG-operated VOs • Provides about 10,000 CPU-days per day in processing • Provides 10 Terabytes per day in data transport • CPU usage averages about 75% • OSG is starting to offer support for MPI Weekly OSG Process Hours Facts and Figures from First Year of Operations • OSG contributed an average of over one thousand CPU-days per day for two months to the D0 physics experiment • OSG provided the LHC collaboration more than 30% of their processing cycles worldwide, in which up to 100 Terabytes per day were transferred across more than 7 storage sites • LIGO has been running workflows of more than 10,000 jobs across more than 20 different OSG sites. • A climate modeling application has accumulated more than 10,000 CPU days of processing on the OSG. • The Kuhlman Lab completed structure predictions for ten proteins, consuming more than 10,000 CPU-days on the OSG. Facing the CI Challenge Together • OSG is looking for a few partners to help deploy campus wide grid infrastructure that integrates with local enterprise infrastructure and the national CI • OSG’s Engagement Team is available to help scientists get their applications running on OSG Low impact starting point Help your researchers gain significant compute cycles while exploring OSG as a framework for your own campus CI • Send your inquires to osg@renci.org • Learn more about the OSG at http:www.opensciencegrid.org