The iPlant Collaborative Cyberinfrastructure aka Development of Public Cyberinfrastructure to Support Plant Science " Nirav Merchant! University of Arizona! PowerPoint Does Rocket Science--and Better Techniques for Technical Reports Essay by Edward Tufte What is iPlant?" • iPlantʼs mission is to build the CI to support plant biologyʼs Grand Challenge solutions" • Grand Challenges were not defined in advance, but identified through engagement with the community" • A virtual organization with Grand Challenge teams relying on national cyberinfrastructure " • Long term focus on sustainable food supply, climate change, biofuels, ecological stability, etc" • Hundreds of participants globally… Working group members at >50 US institutions, USDA, DOE, etc." Brief History" • Formally approved by National Science Board – 12/2007" • Funding by NSF – February 1st, 2008 " • iPlant Kickoff Conference at CSHL – April 2008" o ~200 participants" Grand Challenge Workshops – Sept-Dec 2008" CI workshop – Jan 2009" Grand Challenge White Paper Review – March 2009" Project Recommendations – March 2009" Project Kickoffs – May 2009 & August 2009" First Release of Discovery Environments – April 2010" The paradigm shi- Classic paradigm: You produce data, analyze, interpret (end to end) Conven=onal paradigm: Consor=um/centers produce data and you consume it New Paradigm: Consor=um/centers have produced data and crea=ng “cyber infrastructure” to tackle the “grand challenge” GC Projects Recommended by the iPlant Board of Directors March 2009 Initial Projects: Plant Tree of Life – iPToL – May ʻ09 "+Taxonomic Intelligence "+ APWeb2 "+ Social Networking Website Genotype to Phenotype – iPG2P – Aug ʻ09 "+ Image Analysis Platform" Trait Evolution, Brian Omeara" – Post-tree analysis and mapping of ancestral traits" Tree Reconciliation, Todd Vision" – Large-scale reconciliation of gene trees, co-evolving parasites, etc., with species trees" Big Trees, Alexandros Stamatakis" – HPC Phylogenetic inference with 500K taxa" Tree Visualization Michael Sanderson; Karen Cranston" – Cross cutting group for the viz needs of all! Data Integration, Val Tannen, Bill Piel" – Cross cutting group for the data integration needs of all" Data Assembly, Doug Soltis, Pam Soltis, Michael Donoghue" – Community and network building, data assembly" iPlant Genotype to Phenotype Working Groups" • NextGen Sequencing" – Establishing an informatics pipeline that will allow the plant community to process NextGen sequence data" • Statistical Inference" – Developing a platform using advanced computational approaches to statistically link genotype to phenotype" • Modeling Tools" – Developing a framework to support tools for the construction, simulation and analysis of computational models of plant function at various scales of resolution and fidelity" • Visual Analytics" – Generating, adapting, and integrating visualization tools capable of displaying diverse types of data from laboratory, field, in silico analyses and simulations" • Data Integration" – Investigating and applying methods for describing and unifying data sets into virtual systems that support iPG2P activities" What is Cyberinfrastructure? (Originally about TeraGrid)" WWW.TERAGRID.ORG It was six men of Indostan, To learning much inclined, Who went to see the elephant, (Though all of them were blind), That each by observation Might satisfy his mind. It’s a Grid! It’s a Network! They are HPC Centers! It’s a Common Software Environ! And More!: - Viz - Facilities - Data collections It’s Apps and Support! It’s Storage! … The iPlant Cyberinfrastructure" User Grand Challenge Workflows, iPlant Interfaces Third Party Tools, iPlant-­‐built Tools, Community Contributed Tools and Data! iPlant Discovery Environments Job Submission Workflow Management Service/Data APIs iRODS, Grid Technologies, Condor, RESTful Services iPlant Middleware Compute Storage Persistent Virtual Machines TeraGrid Open Science Grid UA/ASU/TACC Physical Infrastructure Build a CI that’s robust, leverages national infrastructure, and can grow through community contribution! Open Source Philosophy, Commercial Quality Process" • iPlant is open in every sense of the word:" – Open access to source" – Open API to build a community of contributors" – Open standards adopted wherever possible" – Open access to data (where users so choose). " • iPlant code design, implementation, and quality control will be based in best industrial practice" Portfolio of Activities" • Maintaining a balance of “past, present, future” strategies" – “Past”: make services, systems, and support available to existing bioinformatics projects, either to enhance them or simply make critical tools more widely available. " – “Present” build the best bioinformatics software tools that todayʼs technologies can provide. " – “Future” track emerging technologies, and where appropriate stimulate research into the creation and use of those technologies. " Portfolio of Activities" • In a nutshell:" – 12 Working groups in the two grand challenges, each of which is defining requirements for DE development. " Each group not only has discussions that leads to final projects, but they also spawn prototyping efforts, tech eval projects, tool support projects, etc. " – Services group: provide cycles, storage, hosting, etc. to users. " – A comprehensive technology evaluation program to find, borrow, or build relevant technologies, headlined by the semantic web effort." – A number of ancillary projects related to grand challenges, i.e. APWEB, high throughput image analysis " – The Core development/integration effort. " Systems and Services" • Provide access for problems like these on large scale systems" • Provide the storage infrastructure for biological data (again, in support of existing projects)" • Provide cloud style VM infrastructure for service hosting. " iPlant : Connecting Users, Ideas and Resources" The core foundation component comprises of :" Data layer" Registry and Integration layer" Compute and Analysis layer" Interaction and Collaboration layer" iPlant: Using proven technologies" • Data layer: providing access to raw and ingested data sets including high throughput data transfers! • • • • • • • iRODS" GridFTP , Aspera" Dspace (DuraSpace), OpenArchive initiative" Content Distribution Networks (CDN)" High performance storage @ TACC (Lustre)" MySQL and Postgres database clusters" Connection to established data sources (NCBI, TAIR, Gramene)" • Connection to DataOne, DataNet initiatives" • Cloud style storage (similar to Amazon S3 and Walrus)" iPlant: Using proven technologies" • Registry and Integration Layer: Connecting services, data and meta data elements with semantic understanding ! • Meta data catalog management " • Provenance tracking (W7 model)" • Integrated Registry and Service discovery servers" • Data Client and Data Provider Ontology development Kit" • Semantic Architecture (OWL based SSWAP)" iPlant: Using proven technologies" • Compute and Analysis Layer: Connecting tasks with scalable platforms and algorithms! • Virtualization (Xen clusters)" • High Performance Computing at TACC and TeraGrid" • • • • • Grid (Condor, BOINC, Gearman)" Cloud (Eucalyptus, Nimbus)" Reconfigurable Hardware (GP GPU, FPGA)" Checkpoint & Restart (DMTCP)" Scaling and parallelizing code (MPI)" • Workflow engines (DAGman, Pegasus, Kepler)" iPlant: Using proven technologies" • Interaction and Collaboration layer: Providing end user access to unified services and data, from API to large scale visualization! • • • • • • • • • • Google Web Toolkit (GWT driven front end)" Messaging bus (Java Mule, RabbitMQ, XMPP/Jabber)" RESTful web services (web API access)" Single sign-on/identity management (Shibboleth. Oauth ?)" Transparent HPC integration (TeraGrid science gateway and TACC resources" Integration with desktop applications (via web services)" Collaboration platforms (openmeeting, webex wiki, mailman)" Shared analysis (shared workflows, desktop view)" Sharing data (DOI, persistent URL, CDN, social networks)" Large scale visualization (Large Tree, Paraview, SAGE)" Storage Services" • We have also begun offering storage to a number of projects connected to the grand challenges in some way, as well as iPlant internal." – IRODS interface" – Corral at TACC, a local storage array at UA" • Data arriving now for 1KP project, Gates C3/C4 project. " Cloud Services" • iPlant is now offering “cloud” style hosting services." • Dynamically launch virtual servers hosted by iPlant." • Still in prototype " Arrival of “As a Service” models Cyberinfrastructure Is “Research as a Service” SaaS: So;ware as a Service (e.g. Clustering/Assembly is a service) PaaS: PlaIorm as a Service IaaS plus core so;ware capabiliEes on which you build SaaS (e.g. Hadoop/MapReduce is a PlaIorm) IaaS: Infrastructure as a Service (get computer Eme with a credit card and with a Web interface like EC2) 22 What do working groups want ?" • • • • • • • • Wiki" Shared storage" WebEX" CMS" Google apps" Machine for prototyping/development" Change management s/w (git/svn)" Access to compute grid/cluster" What iPlant wants" • Ability to integrate single sign on (sso) with all services we offer (api, cloud, grid, irods etc)" • Leverage credentials from users home institutions" • Lower the barrier to access while still being secure" • Emphasis on ease of access to “research as a service”" Phases of a project" • Enthusiasm " • Disillusionment " • Panic " • Search for the guilty " • Punishment of the innocent " • Praise and honor for the non-participants " 25