Rule-Based Distributed Data Management iRODS 1.0 - Jan 23, 2008

advertisement
Rule-Based Distributed Data
Management
iRODS 1.0 - Jan 23, 2008
http://irods.sdsc.edu
Reagan W. Moore
Mike Wan
Arcot Rajasekar
Wayne Schroeder
San Diego Supercomputer Center
{moore, mwan, sekar, schroede}@sdsc.edu
http://irods.sdsc.edu
http://www.sdsc.edu/srb/
Data Management Goals
• Support for data life cycle
• Shared collections -> data publication -> reference
collections
• Support for socialization of collections
• Process that governs life cycle transitions
• Consensus building for collection properties
• Generic infrastructure
• Common underlying distributed data management
technology
• iRODS - integrated Rule-Oriented Data System
NSF Software Development for
CyberInfrastructure Data
Improvement Project
• #0721400: Data Grids for Community Driven
Applications
• Three major components:
• Maintain highly successful Storage Resource Broker
(SRB) data grid technology for use by the NSF
research community
• Create an open source production version of the
integrated Rule-Oriented Data System (iRODS)
• Support migration of collections from the current SRB
data grid to the iRODS data grid
Why Data Grids (SRB)?
• Organize distributed data into shared collections
• Improve the ability for researchers to collaborate on
national and international scales
• Provide generic distributed data management mechanisms
•
•
•
•
•
•
•
•
•
Logical name spaces (files, users, storage systems)
Collection metadata
Replicas, versions, backups
Optimized data transport
Authentication and Authorization across domains
Support for community specific clients
Support for vendor specific storage protocols
Support for remote processing on data, aggregation in containers
Management of all phases of the data life cycle
Using a SRB Data Grid - Details
DB
SRB Server
Metadata Catalog
SRB Server
•User asks for data
•Data request goes to SRB Server
•Server looks up information in catalog
•Catalog tells which SRB server has data
•1st server asks 2nd for data
•The 2nd SRB server supplies the data
Extremely Successful
• Storage Resource Broker (SRB) manages 2 PBs of data
in internationally shared collections
• Data collections for NSF, NARA, NASA, DOE, DOD, NIH,
LC, NHPRC, IMLS: APAC, UK e-Science, IN2P3, WUNgrid
•
•
•
•
•
•
•
•
•
•
•
•
Astronomy
Bio-informatics
Earth Sciences
Ecology
Education
Engineering
Environmental science
High energy physics
Humanities
Medical community
Oceanography
Seismology
Data grid
Digital library
Data grid
Collection
Persistent archive
Digital library
Data grid
Data grid
Data Grid
Digital library
Real time sensor data, persistent archive
Digital library, real-time sensor data
• Goal has been generic infrastructure for distributed data
Date
Project
Data Grid
NSF / NVO
NSF / NPACI
Hayden
Pzone
NSF / LDAS-SALK
NSF / SLAC-JCSG
NSF / TeraGrid
NIH / BIRN
NCAR
LCA
Digital Library
NSF / LTER
NSF / Portal
NIH / AfCS
NSF / SIO Explorer
NSF / SCEC
LLNL
CHRON
Persistent Archive
NARA
NSF / NSDL
UCSD Libraries
NHPRC / PAT
RoadNet
UCTV
LOC
Earth Sci
TOTAL
5/17/02
6/30/04
11/29/07
GBs of
1000Хs of GBs of
1000Хs of Users with GBs of
1000Хs of Users with
data stored
files
data stored
files
ACLs data stored
files
ACLs
17,800
1,972
6,800
438
239
514
5,139
1,083
41
31
1
77
51,380
17,578
7,201
812
4,562
4,317
80,354
5,416
8,690
4,694
113
47
16
563
685
3,366
80
380
178
49
66
47
2,962
148
88,216
39,697
8,013
28,799
207,018
23,854
282,536
20,400
70,334
3,787
14,550
7,590
161
17,640
169
2,493
7,257
40,747
325
77
100
380
227
68
67
55
3,267
445
2
2
158
33
27
19
3
5
4
1
233
1,745
462
1,734
15,246
6
48
49
601
1,737
35
384
21
27
52
260
2,620
733
2,750
168,931
18,934
12,863
42
53
94
1,202
3,545
2,338
6,443
36
460
21
27
73
5
5
7
2
63
2,785
127
81
20,054
202
58
119
29
28 TB
6 mil
194 TB
40 mil
4,635
5,023
7,499
5,205
2,576
3,557
7,140
6,644
6,136
1,023 TB
6,430
84,984
1,328
966
1,569
2
192
652
200 mil
58
136
29
28
30
5
8
5
5,539
Generic Infrastructure
• Data grids manage data distributed
across multiple types of storage systems
• File systems, tape archives, object ring buffers
• Data grids manage collection attributes
• Provenance, descriptive, system metadata
• Data grids manage technology evolution
• At the point in time when new technology is
available, both the old and new systems can be
integrated
SRB Maintenance
• Release of SRB 3.5.0 was made on
December 3, 2007
• Eliminated a security vulnerability through use
of bind variables for interacting with databases
• Implemented additional community
requirements
• Provided bulk replication and backup
• Automated reconnection on network read-write
failures
• Incorporated bug fixes
Why iRODS?
• Need to verify assertions about the purpose
of a collection
• Socialization of data collections, map from creator
assertions to community expectations
• Need to manage exponential growth in
collection size
• Improve support for all phases of data life cycle from
shared data within a project, to published data in a
digital library, to reference collections within an archive
• Data life cycle is a way to prune collections, and
identify what is valuable
• Need to minimize labor by automating
enforcement of management policies
Starting Requirements
• Base capabilities upon features required by
scientific research communities
• Started with features in SRB data grid, but needed to
understand impact of management policies and
procedures
• Incorporate trustworthiness assessment
criteria from the preservation community
• Other criteria include human subject approval, patient
confidentiality, time-dependent access controls
• Promote international support for iRODS
development to ensure international use
Data Management Applications
• Data grids
• Share data - organize distributed data as a collection
• Digital libraries
• Publish data - support browsing and discovery
• Persistent archives
• Preserve data - manage technology evolution
• Real-time sensor systems
• Federate sensor data - integrate across sensor streams
• Workflow systems
• Analyze data - integrate client- & server-side workflows
Approach
• To meet the diverse requirements, the
architecture must:
•
•
•
•
•
•
•
Be highly modular
Be highly extensible
Provide infrastructure independence
Enforce management policies
Provide scalability mechanisms
Manipulate structured information
Enable community standards
Observations of Production
Data Grids
• Each community implements different
management polices
• Community specific preservation objectives
• Community specific assertions about
properties of the shared collection
• Community specific management policies
• Need a mechanism to support the
socialization of shared collections
• Map from assertions made by collection
creators to expectations of the users
Tension between Common and
Unique Components
• Synergism - common infrastructure
• Distributed data
• Sources, users, performance, reliability, analysis
• Technology management
• Incorporate new technology
• Unique components - extensibility
• Information management
• Semantics, formats, services
• Management policies
• Integrity, authenticity, availability, authorization
Data Grid Evolution
• Implement essential components needed for
synergism
• Storage Resource Broker - SRB
• Infrastructure independence
• Data and trust virtualization
• Implement components needed for specific
management policies and processes
•
•
•
•
integrated Rule Oriented Data System - iRODS
Policy management virtualization
Map processes to standard micro-services
Structured information management and transmission
Initial iRODS Design
Next-generation data grid technology
• Open source software - BSD license
• Unique capability - Virtualization of
management policies
• Map management policies to rules
• Enforce rules at each remote storage location
• Highly extensible modular design
• Management procedures are mapped to microservices that encapsulate operations performed at the
remote storage location
• Can add rules, micro-services, and state information
• Layered architecture
• Separation of client protocols from storage protocols
Using an iRODS Data Grid - Details
DB
iRODS Server
Rule Engine
Metadata Catalog
Rule Base
iRODS Server
Rule Engine
•User asks for data
•Data request goes to iRODS Server
•Server looks up information in catalog
•Catalog tells which iRODS server has data
•1st server asks 2nd for data
•The 2nd iRODS server applies rules
Data Virtualization
Access Interface
Storage Protocol
Storage System
Traditional
approach:
Client talks
directly to
storage system
using Unix I/O:
Microsoft Word
Data Virtualization (Digital Library)
Access Interface
Digital Library
Storage Protocol
Storage System
Client talks to
the Digital
Library which
then interacts
with the
storage system
using Unix I/O
Data Virtualization (iRODS)
•Map from the
Access Interface
Standard Micro-services
Data Grid
Standard Operations
Storage Protocol
Storage System
actions requested by
the access method to
a standard set of
micro-services.
•The standard microservices use standard
operations.
•Separate protocol
drivers are written
for each storage
system.
iRODS Release 1.0
• Open source software available at wiki:
• http://irods.sdsc.edu
• Since January 23, 2008, more than 590
downloads by projects in 18 countries:
• Australia, Austria, Belgium, Brazil, China,
France, Germany, Hungary, India, Italy,
Norway, Poland, Portugal, Russia, Spain,
Taiwan, UK, and the US
Core Components
• Framework
• Infrastructure that ties together the layered environment
• Drivers
• Infrastructure that interacts with commercial protocols (database,
storage, information resource)
• Clients
• Community specific access protocols
• Rules
• Management policies specific to a community
• Micro-services
• Management procedures specific to a community
• Quality assurance
• Testing routines for code validation
• Maintenance
• Bug fixes, help desk, chat, bugzilla, wiki
Rule Specification
• Rule - Event : Condition : Action set :
Recovery Procedure
•
•
•
•
Event - atomic, deferred, periodic
Condition - test on any state information attribute
Action set - chained micro-services and rules
Recovery procedure - ensure transaction
semantics in a distributed world
• Rule types
• System level, administrative level, user level
Distributed Management System
Data
Transport
Rule
Engine
Metadata
Catalog
Persistent
State
information
Policy
Management
Virtualization
Server
Side
Workflow
Execution
Engine
Execution
Control
Scheduling
Messaging
System
integrated Rule-Oriented Data System
Client Interface
Admin Interface
Rule Invoker
Rule
Modifier
Module
Current
State
Rule
Rule
Base
Config
Modifier
Module
Consistency
Check
Module
Metadata
Modifier
Module
Service
Manager
Consistency
Check
Module
Consistency
Check
Module
Confs
Resources
Resource-based
Services
Micro
Service
Modules
Metadata-based
Services
Micro
Service
Modules
Metadata
Persistent
Repository
iRODS Data Grid Capabilities
• Remote procedures
• Atomic / deferred / periodic
• Procedure execution / chaining
• Structured information
• Structured information
•
•
•
•
•
Metadata catalog interactions / 205 queries
Information transmission
Template parsing
Memory structures
Report generation / audit trail parsing
iRODS Data Grid Capabilities
• Rules
•
•
•
•
User / administrative / internal
Remote web service invocation
Rule & micro-service creation
Standards / XAM, SNIA
• Installation
• CVS / modules
• System dependencies
• Automation
iRODS Data Grid
• Administration
•
•
•
•
User creation
Resource creation
Token management
Listing
• Collaborations
• Development plans
• International collaborators
• Federations
Three Major Innovations
1. Management virtualization
• Expression of management policies as rules
• Expression of management procedures as
remote micro-services
• Expression of assertions as queries on
persistent state information
•
Required addition of three more logical
name spaces for rules, micro-services, and
state information
Second Major Innovation
• Recognition of the need to support structured
information
• Manage exchange of structured information between
micro-services
• Argument passing
• Memory white board
• Manage transmission of structured information
between servers and clients
• C-based protocol for efficiency
• XML-based protocol to simplify client porting (Java)
• High performance message system
Third Major Innovation
• Development of the Mounted Collection
interface
• Standard set of operations (20) for extracting
information from a remote information resource
• Allows data grid to interact with autonomous resources
which manage information independently of iRODS
• Structured information drivers implement the
information exchange protocol used by a particular
information repository
• Examples
• Mounted Unix directory
• Tar file
SDCI Project at SDSC
• Implement using spiral development. Iterate
across development phases:
• Requirements - driven by application communities
• Prototypes - community specific implementation of a
new feature / capability
• Design - creation of generic mechanism that is
suitable for all communities
• Implementation - robust, reliable, high performing code
• Maintenance - documentation, quality assurance, bug
fixes, help desk, testing environment
• We find communities eager to participate in
all phases of spiral development
Example External Projects
• International Virtual Observatory Alliance
(federation of astronomy researchers)
• Observatoire de Strasbourg ported the IVOA VOSpace
interface on top of iRODS
• This means the astronomy community can use their
web service access interface to retrieve data from
iRODS data grids.
• Parrot grid interface
• Ported by Douglas Thain (University of Notre Dame)
on top of iRODS. Builds user level file system across
GridFTP, iRODS, http, and other protocols
Collaborators - Partial List
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
Aerospace Corporation
Academia Sinica, Taiwan
UK security project, King's College London
BaBar High Energy Physics
Biomedical Informatics Research Network
California State Archive
Chinese Academy of Sciences
CASPAR - Cultural, Artistic, and Scientific knowledge for Preservation Access and Retrieval
CineGrid - Media cyberinfrastructure
DARIAH - Infrastructure for arts and humanities in Europe
D-Grid - TextGrid project, Germany
DSpace Foundation digital library
Fedora Commons digital library
Institut national de physique nucleaire et de physique des particules
IVOA - International Virtual Observatory Alliance
James Cook University
KEK - High Energy Accelerator Research Organization, Japan
LOCKSS - Lots of Copies Keep Stuff Safe
Lstore - REDDnet Research and Education Data Depot network
NASA Planetary Data System
National Optical Astronomy Observatory
Ocean Observatory Initiative
SHAMAN - Sustaining Heritage through Multivalent Archiving
SNIA - Storage networking Industry Association
Temporal Dynamics of Learning Center
Project Coordination
• Define international collaborators
• Technology developers for a specific development
phase for a specific component.
• Collaborators span:
•
•
•
•
•
Scientific disciplines
Communities of practice (digital library, archive, grid)
Technology developers
Resource providers
Institutions and user communities
• Federations within each community are
essential for managing scientific data life
cycle
Scientific Data Life Cycle
• Shared collection
• Used by a project to promote collaboration between
distributed researchers
• Project members agree on semantics, data formats,
and manipulation services
• Data publication
• Requires defining context for the data
• Provenance, conformance to community format
standards
• Reference collections
• Community standard against which future research
results are compared
Scientific Data Life Cycle
• Each phase of the life cycle requires
consensus by a broader community
• Need mechanisms for expressing the
new purpose for the data collection
• Need mechanisms that verify
•
•
•
•
Authoritative source
Completeness
Integrity
Authenticity
Why iRODS?
• Collections are assembled for a purpose
•
•
•
•
Map purpose to assessment criteria
Use management policies to meet assertions
Use management procedures to enforce policies
Track persistent state information generated by
every procedure
• Validate criteria by queries on state information
and on audit trails
Data Management
iRODS - integrated Rule-Oriented Data System
Data Management
Environment
Management
Functions
Data Management
Infrastructure
Physical
Infrastructure
Conserved
Control
Remote
Properties
Mechanisms
Operations
Assessment Management Capabilities
Criteria
Policies
Data grid Π Management virtualization
Persistent
Rules
Micro-services
State
Data grid Π Data and trust virtualization
Database
Rule Engine
Storage
System
Why iRODS?
•
Can create a theory of data management
•
•
Prove compliance of data management system with specified
assertions
Three components
1. Define the purpose for the collection, expressed as assessment
criteria, management policies, and management procedures
2. Analyze completeness of the system
•
•
•
•
For each criteria, persistent state is generated that can be audited
Persistent state attributes are generated by specific procedure
versions
For each procedure version there are specific management policy
versions
For each criteria, there are governing policies
3. Audit properties of the system
• Periodic rules validate assessment criteria
Major iRODS Research Question
• Do we federate data grids as was done
in the SRB, by explicitly crossregistering information?
• Or do we take advantage of the
Mounted Collection interface and
access each data grid as an
autonomous information resource?
• Or do we use a rule-based database
access interface for interactions
between iCAT catalogs?
Federation Between IRODS Data Grids
Data Access Methods (Web Browser, DSpace, OAI-PMH)
Data Collection A
Data Grid
Data Collection B
Data Grid
• Logical resource name space
• Logical resource name space
• Logical user name space
• Logical user name space
• Logical file name space
• Logical file name space
• Logical rule name space
• Logical rule name space
• Logical micro-service name
• Logical micro-service name
• Logical persistent state
• Logical persistent state
Mounted Collections
• Minimizes dependencies between the
autonomous systems
• Supports retrieval from the remote information
resource, but not pushing of information
• Pull environment
• Can be controlled by rules that automate
interactions
• Chained data grids
• Central archive (archive pulls from other data grids)
• Master-slave data grids (slaves pull from master)
Rule-based Database Access
Interface
• Support interactions by querying the
remote iCAT catalog’s database
• Expect to support publication of schemata
• Ontology-based reasoning on semantics
• Can be used for both deposition and retrieval
of information
• Simplifies exchange of rules and possibly of
micro-services
iRODS Development
• NSF - SDCI grant “Adaptive Middleware for
Community Shared Collections”
• iRODS development, SRB maintenance
• NARA - Transcontinental Persistent Archive
Prototype
• Trusted repository assessment criteria
• NSF - Ocean Research Interactive
Observatory Network (ORION)
• Real-time sensor data stream management
• NSF - Temporal Dynamics of Learning Center
data grid
• Management of IRB approval
iRODS Development Status
• Production release is version 1.0
• January 24, 2008
• International collaborations
• SHAMAN - University of Liverpool
• Sustaining Heritage Access through Multivalent
ArchiviNg
•
•
•
•
UK e-Science data grid
IN2P3 in Lyon, France
DSpace policy management
Shibboleth - collaboration with ASPIS
Planned Development
•
•
•
•
•
•
•
•
•
•
•
•
•
•
GSI support (1)
Time-limited sessions via a one-way hash authentication
Python Client library
GUI Browser (AJAX in development)
Driver for HPSS (in development)
Driver for SAM-QFS
Porting to additional versions of Unix/Linux
Porting to Windows
Support for MySQL as the metadata catalog
API support packages based on existing mounted collection driver
MCAT to ICAT migration tools (2)
Extensible Metadata including Databases Access Interface (6)
Zones/Federation (4)
Auditing - mechanisms to record and track iRODS metadata changes
For More Information
Reagan W. Moore
San Diego Supercomputer Center
moore@sdsc.edu
http://www.sdsc.edu/srb/
http://irods.sdsc.edu/
Collaborations
Requirements
Capability
Prototypes
Design
Generic Mainversion tenance
Technology
Lead
Framework
Core server
TPAP, OOI, TDLC, BIRN, Й
SDCI
SDCI
SDCI
SDCI
Wan (SDSC)
Rule Engine
TPAP, BIRN, TDLC
SDCI
SDCI
SDCI
SDCI
Rajasekar (SDSC)
Rule Base
TPAP
SDCI
SDCI
SDCI
SDCI
Rajasekar (SDSC)
iCAT catalog
TPAP, TDLC, BIRN, OOI
TPAP, SDCI
SDCI
SDCI
SDCI
Schroeder (SDSC)
Extensible iCAT
TPAP
TPAP, SDCI
SDCI
SDCI
SDCI
Schroeder (SDSC)
SDCI
SDCI
SDCI
SDCI
Wan (SDSC)
TPAP, OOI
SDCI
SDCI
SDCI
Teragrid, TPAP, OOI
ExecServer
(Scheduler
Xmessaging Server OOI, TPAP, LSST, Kepler
Transport Layer
DLI2, OOI, Cinegrid, LSST, TPAP, PDS
SDCI
SDCI
SDCI
SDCI
Wan, Rajasekar
(SDSC)
Wan (SDSC)
Transport Layer
(XML)
Monitoring System
TPAP, DLI2, OOI
TPAP, SDCI
SDCI
SDCI
SDCI
Wan (SDSC)
IN2P3, BIRN
IN2P3
IN2P3
IN2P3
IN2P3
Nief(IN2P3)
TPAP, SDCI
SDCI
SDCI
SDCI
Wan (SDSC)
Wan, Rajasekar,
Schroeder (SDSC)
Schroeder, Wan
(SDSC)
Schroeder (SDSC)
Mounted Collection TPAP, OOI, Teragrid, NSDL
Federation
TPAP, OOI, TDLC, NSDL, BIRN
TPAP, SDCI
SDCI
SDCI
SDCI
Master-slave
catalogs
Authentication
BIRN
BIRN, SDCI
SDCI
SDCI
SDCI
BIRN, TPAP, Teragrid, OSG, NeSC
BIRN, SDCI
SDCI
SDCI
SDCI
Collaborations
Capability
Requirements
Prototypes
Design
Generic
version
Main-tenance Technology Lead
Drivers
iCAT database
drivers
Database access
drivers
Storage drivers
TPAP, SPAWAR, NSDL
SDCI, TPAP
SDCI
SDCI
SDCI
Schroeder (SDSC)
TDLC, TPAP, ROADnet, OOI
ROADnet, TPAP
SDCI
SDCI
SDCI
Schroeder (SDSC)
TPAP, Teragrid, LSST, SNIA, ZIH
SDCI
SDCI
SDCI
SDCI
Structured
Teragrid, TPAP
Information Drivers
Teragrid, TPAP,
SDCI, NeSC
SDCI
SDCI
SDCI
SRB-iRODS
migration
ZIH, JCU, BIRN, TPAP
TPAP, SDCI
SDCI
SDCI
SDCI
Wan (SDSC),
Pasquinelli (Sun)
Wan (SDSC), Hasan
(SHAMAN), Bady
(RMBC)
Schroeder, Rajasekar,
Wan (SDSC)
GSI authentication
SPAWAR, OSG, TeraGrid, NeSC, AS, BMBF
SPAWAR
SDCI
SDCI
SDCI
Schroeder (SDSC)
Shibboleth
authentication
ASPIS, SHAMAN
NeSC
ASPIS
ASPIS
ASPIS
Hasan (SHAMAN),
Beitz(MU)
Collaborations
Capability
Requirements
Prototypes
Design
Generic
version
Maintenance
Technology
Lead
Clients
Jargon - Java I/O
library
Web client
BIRN, TPAP, OOI, Teragrid, TDLC
BIRN, SDCI, TPAP
BIRN, SDCI
BIRN, SDCI
BIRN, SDCI
Gilbert (SDSC)
BIRN, Teragrid, OSG, TPAP, LUScid
SDCI
SDCI
SDCI
Lu, Antoine (SDSC)
inQ - Window s
brow ser
iCommands - shell
commands
Parrot
TPAP, DLI2, OOI, Teragrid, NSDL
SDCI, TPAP,
LUScid
SDCI, TPAP, NSDL
SDCI
SDCI
SDCI
Teragrid, OOI, TPAP
SDCI
SDCI
SDCI
SDCI
PetaShare, Condor
ND
ND
ND
ND
Cow art, Zhu
(SDSC)
Schroeder, Wan,
Rajasekar (SDSC)
Thain (ND)
PHP client API
TPAP, TDLC
TPAP
SDCI
SDCI
SDCI
PERL load library
TPAP, JCU
JCU
JCU
JCU
JCU
Python load library
TPAP, NCRIS
TPAP
SDCI
SDCI
SDCI
Fedora
NSDL, KCL FNL, RMBC, D-Grid, DARIAH
NSDL, KCL
NSDL
NSDL
NSDL
DSpace
MIT, TPAP, AS
MIT
TPAP
MIT
MIT
LOCKSS
LOCKSS
LOCKSS
LOCKSS
LOCKSS
LOCKSS
Lstore
REDDnet
REDDnet
REDDnet
REDDnet
REDDnet
Rosenthal
(Stanf
ord)
Sheldon
(ACCRE)
Archivist Toolkit
UCSD Libraries
UCSD Libraries
UCSD Libraries
UCSD Libraries
UCSD Libraries
Westbrook (UCSD)
Internet Archive
NDIIPP, NSDL, TPAP, CDL
NDIIPP
CDL
CDL
CDL
Zhu (SDSC)
SRM
AS, OGF, EGEE, DEISA
AS
AS
AS
AS
(AS)
GridFTP
Archer, DEISA
Archer
Archer
Archer
Archer
Sim (JCU)
VOSpace
IVOA
NVO
NVO
NVO
NVO
Sensor Stream
Interface
OOI
OOI
OOI
OOI
OOI
Video Streaming
Interface
Cinegrid
Cinegrid
Cine-grid
Cinegrid
Cinegrid
Wagner (UCSD), Lu
(SDSC)
Rajasekar, Lu
(SDSC), Arrott
(UCSD)
Wan (SDSC)
Lu (SDSC),
Atkinson (JCU)
Atkinson (JCU)
DeTordy (SDSC),
Atkinson (JCU)
Zhu (SDSC),
Hodges (KCL),
Aschenbrenner
(UG)
Smith (MIT)
Collaborations
Capability
Requirements
Prototypes
Design
Generic
version
Maintenance
Technology
Lead
Rules
Administrative
Teragrid, BIRN, TPAP, NOAO, BaBar, NeSC Teragrid, TPAP
SDCI
SDCI
SDCI
Trustw orth-iness
TPAP, SHAMAN
TPAP, SHAMAN
TPAP
TPAP
TPAP
IRB
TDLC
TDLC
TDLC
TDLC
TDLC
Rajasekar,
Schroeder, Chen
(SDSC)
DeTorcy, Marciano
(SDSC), Giaretta
(CASPAR),
Bady(RMBC),
Thaller (Planets)
Hou(SDSC)
HIPAA
BIRN
BIRN
BIRN
BIRN
BIRN
Rajasekar (SDSC)
Collaborations
Capability
Requirements
Prototypes
Design
Generic
version
Maintenance
Technology
Lead
Micro-services
Administrative
Teragrid, BIRN, TPAP, NOAO, BaBar, NeSC Teragrid, TPAP
SDCI
SDCI
SDCI
Audit trails
TPAP, RMBC
TPAP
SDCI
SDCI
SDCI
ERA
TPAP, MU
TPAP
TPAP, SDCI
TPAP, SDCI
TPAP, SDCI
Trustw orth-iness
TPAP, SHAMAN
TPAP, SHAMAN
TPAP
TPAP
TPAP
IRB
TDLC
TDLC
TDLC
TDLC
TDLC
HIPAA
BIRN
BIRN
BIRN
BIRN
BIRN
Rajasekar,
Schroeder (SDSC)
DeTorcy, Schroeder
(SDSC), Bady
(UMBC)
Moore, DeTorcy
(SDSC), Treloar
(MU)
DeTorcy, Marciano,
Kremenek, Moore
(SDSC)
Hou, DeTorcy,
Moore (SDSC)
Rajasekar (SDSC)
Collaborations
Capability
Requirements
Prototypes
Design
Generic
version
Maintenance
Technology
Lead
Quality Assurance
Unit testing
TPAP, BIRN, Teragrid
SDCI
SDCI
SDCI
SDCI
Function testing
TPAP, OOI
SDCI
SDCI
SDCI
SDCI
Load testing
TPAP, OOI, KEK, IN2P3
SDCI, NARA
SDCI
SDCI
SDCI
Performance testing TPAP, IN2P3
SDCI, NARA
SDCI
SDCI
SDCI
Security testing
SDCI
SDCI
SDCI
SDCI
Rajasekar,
Schroeder, Wan,
DeTorcy (SDSC)
Lindsey, Tooby
(SDSC), Nief
(IN2P3)
Lindsey (SDSC),
Smorul (U Md)
Lindsey (SDSC),
Smorul (U Md), Iida
(KEK)
Lindsey (SDSC)
Recovery procedureTPAP, OOI, SHAMAN
testing
System failure
TPAP, OOI, IN2P3
testing
Recoverability
TPAP, OOI, NARA, SHAMAN, IN2P3
testing
Integration testing TPAP, OOI, SHAMAN
SDCI
SDCI
SDCI
SDCI
Lindsey (SDSC)
SDCI
SDCI
SDCI
SDCI
Lindsey (SDSC)
SDCI
SDCI
SDCI
SDCI
SDCI
SDCI
SDCI
SDCI
Parallel execution
testing
Tuning
OOI
SDCI
SDCI
SDCI
SDCI
TPAP, OOI, BIRN, IN2P3, SHAMAN
SDCI
SDCI
SDCI
SDCI
Lindsey (SDSC),
Smorul (U Md)
Zhu, Lindsey
(SDSC)
Wan, Lindsey
(SDSC)
Rajasekar, Wan,
Schroeder (SDSC)
TPAP, TDLC, BIRN
Collaborations
Capability
Requirements
Prototypes
Design
Generic
version
Maintenance
Technology
Lead
Documentation
Micro-service
creation
Micro-services
SHAMAN, IN2P3, TPAP
SHAMAN, IN2P3,
TPAP
SHAMAN, IN2P3,
TPAP
SDCI, TPAP
SDCI, TPAP
SDCI, TPAP
SDCI, TPAP
SDCI, TPAP
SDCI, TPAP
SHAMAN, IN2P3,
TPAP
SHAMAN, IN2P3,
TPAP
SHAMAN, IN2P3,
TPAP
SHAMAN, IN2P3,
TPAP
SAA demonstration TPAP
TPAP
TPAP, SDCI
TPAP, SDCI
TPAP, SDCI
Tutorials
AS, NeSC, Teragrid, TPAP
TPAP, SDCI
TPAP, SDCI
TPAP, SDCI
TPAP, SDCI
Concepts
TPAP, OGF
TPAP, SDCI
SDCI, TPAP
SDCI, TPAP
SDCI, TPAP
Rajasekar, Tooby
(SDSC)
Rajasekar, Tooby,
Schroeder (SDSC),
Hasan (SHAMAN),
Nief (IN2P3)
Rajasekar, Tooby,
Schroeder (SDSC),
Hasan (SHAMAN),
Nief (IN2P3)
Marciano, DeTorcy,
Moore (SDSC)
Moore, Schroeder,
Rajasekar,
Wan(SDSC), Berry
(NeSC), Hasan
(SHAMAN), Li
(CAS)
Moore (SDSC)
Clients
SHAMAN, IN2P3, TPAP, AS
Community
Sustainability
Standards
DSpace, Fedora, IA
SDCI
SDCI
SDCI
SDCI
Tooby (SDSC)
OGF, SNIA, Sun, SHAMAN
SDCI, TPAP
SDCI, TPAP
SDCI, TPAP
SDCI, TPAP
Moore (SDSC), Lee
(AC)
SHAMAN, IN2P3, TPAP, AS
Download