Overview of UK e-Science and its enabling Grid Middleware Dr Steven Newhouse Technical Director, London e-Science Centre, Imperial College, London (http://www.lesc.ic.ac.uk/) s.newhouse@doc.ic.ac.uk • • • • Overview What is the Grid? Current Grid Infrastructures UK e-Science Programme The future… • Content Acknowledgements: – Ian Foster (Globus co-PI) – Miron Livney (Condor PI) 2 1 Grid Computing 3 The Grid: A Brief History • Early 90s – Gigabit testbeds, metacomputing • Mid to late 90s – Early experiments (e.g., I-WAY), academic software projects (e.g., Globus, Legion), application experiments • 2002 – Dozens of application communities & projects – Major infrastructure deployments – Significant technology base (esp. Globus ToolkitTM) – Growing industrial interest – Global Grid Forum: ~500 people, 20+ countries 4 2 The Grid – a definition “Secure resource sharing & coordinated problem solving in dynamic, multi-institutional virtual organisations” – On-demand, ubiquitous access to computing, data, and services – New capabilities constructed dynamically and transparently from distributed services 5 What is an e-Scientist? • Applied Scientists are becoming e-scientists: – computational & data services • HPC, genetic databases, electronic journals – remote sensors • Large Hadron Collider, SOHO – personal devices • workstations, mobile phones, PDA’s – supporting collaborative working • ‘collaboratories’: global scientific collaborative communities e-science enabled through Grid middleware 6 3 CERN’s Large Hadron Collider 1800 Physicists, 150 Institutes, 32 Countries 100 PB of data by 2010; 50,000 CPUs? 7 Data Grids for High Energy Physics |}@~@v12 h0S s W XY kkV `*W xYj U1w dve%f* ii%i !" 8@8LNM ,-O-*4 kU_1^@j %f0U ]W c Y w U1^ j X 0 ' 1 / 2 / ) $ 1 3 * 4 % 3 6 5 7 . S T*U1V UW XY[Z \%]^ _T_V `X X W ^ a6b U2c U1V def0^X U1_ X g ! 89#*$& 8 89LM ,*%-O*-*4 S T*U1V UY V UGh i%i(Z j V W aa U1V X@b kU*V2X U1_` ^l !" m Y _Tnj V W aa U1V U1lGUc U1^ jW Xo1hpqRd j UW ^X W r U :;6<=31.6>*?@ GL QR -O-4 Tier 0 3GA G5 DI G!R{P > *47* { Tier 1 '(*) +&,*-. E .7 ,0; %D2 3 7) 5 7 4*0;2D 3 72) #%72) ,6; %D 32 7) Tier 2 !P @LNQR - O-4 $I ,*- 4-{R7*764*7*4I #.6>* 7) !" #%$& # - ?* ! 8GF H #%$#& - ? # - ?* # - ?* !" LNM ,-O-*4 $I ,*- 4 -J3K - 7*@ 3 %R% Tier 4 %¡¢£%¢¤¦¥ ;6A%B !C #$& !P @LNQR - O*-%4 &413 E { v !" #%$& !" #$& !" #%$& !" #%$& s T1dX W _W Xj Xt@`V uG`^vY ^Y w dX W X[Z _T*Y ^^ U1w Xb g m Y _T6W ^ Xj W j ] j UGtW w w1T*Y@c UGo1h i0kT*dX W _ W XRj Xt`V u*W ^ a9`^v`^ U`Vx`V U _T*Y ^^ U1w X y*lYj Yz `1V0j T*U*X U_T*Y ^^ U1w XX T1`]w lG\U_Y _T*Ul6\dj T1U W ^ X@j W j ]@j UvX U1VRc U*V §%¤¦¨¢©9¡«ª ¥Gª «R¡£%¬ 8 4 The Globus Toolkit in One Slide • Grid protocols (GSI, GRAM, …) enable resource sharing within virtual orgs; toolkit provides reference implementation ( = Globus Toolkit services) O ² Ñ »N®»¿»%¯ Ù RÆÅ ©GT S Ó6¯Nà ¾ ¿+Ô[à ¯ É Ù ¤¦¥6ª R R¤¦Ê ¥6¬ Å ¤¦VURÊ ¤ Í ° ¯àN¯+¾ ®½:»Y ° º[¿¾»v°¾¸ ²+½XW ÄÆÅÈÇ ­®¯+° ° ¯+Õ¦²»¯ ¾ ½vÖ0²³¿»v¾ ²+½ Î ¿»¯+Ï2¯¯ ± ¯+° Ó6¯ ± ²+° »%¯+° ÎNMMOQPnÎ ° ¾ Â É Ä %¡ Ð ° ¯àN¾ ®»° ¸âá M ½ Ñ ²+° Õ¦¿»v¾ ²+½ Ð Ñ ¹ º » ¼ + ¯ v ½ » ¾ ³ ¿ % » Á ¯ À ¿ ³ % » + ² ° ( ¸ Ò Å ¤¦Ê ¨¢%¥ ÂN¾ ®³²[Ö6¯+° ¸+Ò /1023#45687:9 M ½v¯· O ¯+° Ö0¯+° ³[° ¯¿»%¯ ± ° ²·¸ Ð ÂN¾ ®³²[Ö6¯+° ¸+Ò Ç «Ëª Ì ¥0¨¢Ê © ³[° ¯Â¯+½v»v¾ ¿ à ä(åæçè%æÛé åêëæììîí æïð ìè%æå ;#< 023#=>0?A@ ; 03#B C 4 # 3 F D G E 0 3 G H # 3 . 4 8 I A ? @ 3 ¥G¨¢¤¦Í ­®¯+° ­®¯+° ± ° ²³¯®®´µ 4J3#K < 3LHJ0H ± ° ²³¯®®£´ã ' »¼v¯+°®¯+° Ö+¾ ³¯ ¶° ²·¸ ¶° ²·¸â´ã ÄÆ×ØÚÙ Ð ¯)( à*( Î °@¾ Â,+.-¶Ò É Ä R¡ × ¤ÛÌN ¨¢ÊN¤ ØÚÜRÜ ÊNª ¥R ÞÝ Ù ª ¢ª ¤¦ßÞ¤¦¢¥GÍ ñóòÛô=õö[õ÷øõÚù úüû ýÁþ ÿ ò =úÁþ£ýÚù õ ö Áô ö[õõ ù úüýÁþ£ÿ úÁôR÷Áú õ ô Áô= ú ÿ RúÛ÷ÛõÁô ÿÈÁ ý ö[ý ö ! õÈô#" ù õ$&% 9 Condor np#8d a*c]q] ^#h y .u p |! ~ np#8d a*c]q] ^#hni*[h,q 8_*\>qG^#[]`npi#[jh*q rts!uvxwzys!{}|!~ kml hL[]`onpiL[h,q Z[]^Lq[N[ja,_*q] ^#hnpiL[jh*q F |!v u| , |! ~ b]^La,cdNZ[]\>^L_*`>a,[fegch*cji*[j` Z[]\>^L_,`>a*[ 10 5 Condor within the Grid ~ vx| xz u ~s | v y¡y ¢vxs!u 11 Mathematicians Solve NUG30 • Looking for the solution to the NUG30 quadratic assignment problem • An informal collaboration of mathematicians and computer scientists • Condor-G delivered 3.46E8 CPU seconds in 7 days (peak 1009 processors) in U.S. and Italy (8 sites) £¤}¥*¦¥§}¨¥§}¤¥,£¥©¥,£ª}¥*£¦¥ £«}¥¬}¥§S£¥§¥¤¥§}¬¥§¦¥§§¥ £©}¥§}ª¥,£­¥©«}¥ª}¥§«¥,£¬¥ ¨¥,£¨}¥­}¥§}­¥,£§¥,££¥§© Ù ¤¦¥Gª®}¯N° ű¢Ø [ ¢«¤³² Ç ¬ª²´® ¥6«¤¦Ì ¥0¤¦²µ RÌ Ê «ÌNR 12 6 GriPhyN: CMS Data Reconstruction Master Condor job running at Caltech 2) Launch secondary job on WI pool; input files via Globus GASS Secondary Condor job on WI pool 5) Secondary reports complete to master 1) Caltech workstation 6) Master starts reconstruction jobs via Globus jobmanager on cluster 3) 100 Monte Carlo jobs on Wisconsin Condor pool 9) Reconstruction job reports complete to master 4) 100 data files transferred via GridFTP, ~ 1 GB each 7) GridFTP fetches data from UniTree NCSA Linux cluster 8) Processed objectivity database stored to UniTree Å ÊN ¥G¥ ¶ ª ¢¡ª ² Ù R · U ²£ ¥6¢¤ Ì NCSA UniTree - GridFTPenabled FTP server 13 Grid Middleware Requirements • Security – X.509 PKI with certificate authority – Mapping of Grid Identity to local UNIX account • Connections – Defined ports or port range through firewalls – GridFTP (modified FTP protocols) • Data – Very large data transport requirements – Prescribed time frames (QoS) 14 7 UK e-Science Funding 2000-03 ÇÈ$É ¾XÊJ¾XËÂC¼»ÌzÄQÅÀÍÀ¿À¼»½ Î Ê ¸ ¹»º]¼»½ ¾¿À¼»¾ ³ º]ÁJ¾¾ÂJ½ ¿ÀÃ!ÄQÅÀÆNÆ1½ ÁJÁC¾¾ Ç ½ ÂJ¾¼»ÁCÅÀ ÏÐ ÑCÒÓ»ÔCÕÀÑJÖ × Ø´Ù»ÚÀÙ»Û ÒÜ8Ò Ú Ô8Ý)ÕÀÞ Ò ÏÐ ÑCÒÓ»ÔJÕÀÑCÖ × ßà Ù ÑCÒ Ú ÒX×C× ÙáÚÀâ´ã ÕÀäåÕÀÑ â Ð ÚÀÙ ÔJÐ Õ Ú Ý,ÕÀÞ Ò È ¾¿À¾XÂC½ ¼}ÄQÌGËXÎ Î ¾¿ÀÃ¾Ê ø ¼»Ëô¾XÆN½ ¼ ³ ø ùù Î ½ ¼»ËÁJ½ ÅÀ¿º]Í ùù ÀÅ Â:Á æ)çÀè Ý ãxéëê1ìîí ÜNïðÏjñ#ò éëê1ìîí Ü1ï ç ÑCÕ Û Ñ Ù ÜNÜNÒ Ý)Ò×JÒ Ù ÑJÓ»ú ã ÕÀû Ú Ó»Ð Þ × éë Ü ï ê ü>ö Ü8ïëðGÏñ#ò é êí N çÀç ß Ý ãxéëê1ý>þ ÜNï ÿ,ÿ è Ý ãxéë ê ÜNï Ø Ý ãxéë ê Ü1ï æ Ý ãxéëê1ü Ü1ï æ,è Ý ãxéë ê ÜNï £80m Collaborative projects æ,çÀè Ý ãxéëêìîü ÜNï ã Ý ãxéëêí ÜNï ó ¿ÀôGÍÀÊJÁCÂJ½ ËXÎjÄQÅÀÎ Î Ëõ>ÅÀÂCËXÁC½ ÅÀ¿ éëêNöQ÷ ÜNï e-Science Director : Professor Tony Hey 15 Regional e-Science Centres _d !#"%$'&&(*),+.-%/10L2!3 è&è Ð ÔCÒ× ]Þ Ù × Û Õ à æ â Ð Ú >ûÀÑ Û ú é1Ù ÔJÐ Õ ÀÚ Ù Þ ã Ò Ú ÔCÑJÒï Ò à Ó Ù ×áÔCÞ Ò ÿ)ÒXÞ Ù ×JÔ ã>Ù Ñ â Ð 4 ã æ Ý Ø´ÙáÚ Ó»úÒ×JÔCÒÑ Ï Ïû>Þ Ð Ú æ ÕÀÑ â Ý ß è ÕÀûÀÔCú Ù ÜÔCÕ Ú ã>Ù Ü>ÑCÐ âÛ Ò òÜÒXÑCÐ Ù Þ ã ÀÕ Þ Þ Ò Û Ò é Õ ÚÀâ Õ Ú ï 16 8 Grid Support Centre • First point of contact for ‘getting on the Grid’ • Grid Starter Kit (GSK): Globus, Condor & SRB • Provides: – user support for deployment & debugging – software distribution and maintenance of GSK – central point of contact with Globus team • Virtual Centre drawing on expertise at: – – – – Rutherford Appleton Laboratory (Oxford) Daresbury Labs (Manchester) Edinburgh Parallel Computing Centre (NeSC) University of Manchester (Manchester Computing & ESNW) • http://www.grid-support.ac.uk 17 Grid Engineering Task Force • Experts from the Regional & Grid Support Centres • Working Groups – – – – – – – Globus 2: Deployment on contributed Grid resources Directory Services: National GIIS: ginfo.grid-support.ac.uk Security: X.509 PKI through National CA, local RA’s Firewalls: At a university & centre level Grid User Management: Account creation & accounting Integration: Verify the functionality of the ‘Grid’ Access Grid: Collaborative workspace through multicast • Basic UK grid infrastructure demonstrated at NeSC opening 18 9 Access Grid • Collaborative work among large groups • ~50 sites worldwide • Use Grid services for discovery, security • www.scglobal.org Presenter mic Presenter camera Ambient mic (tabletop) 57686:9<;8;>=%?@BADC157?EFHGG9<IJFKML9<?; U #V WTX#VTW Y[a Z]\ #^ _'` W ^#` ` W#bdc ` Audience camera NONPNQSRH6T689;T;8E3?@BADQSFH?19E Applied User Base • Regional Centre Projects • EPSRC Pilot Projects – – – – Geodise: Engineering Optimisation Reality Grid: Distributed mesoscale materials modelling Discovery Net: Data mining on distributed high throughput data sources Comb-e-Chem: Synthesis of new chemical compounds by combinatorial methods – MyGrid: Bioinformatics workbench infrastructure – DAME: Distributed analysis of jet engine sensor data • PPARC Projects – GridPP: UK activities within EU-DataGrid – AstroGrid: Federation & datamining of astronomical data – Grid1D: Development of distributed peer-to-peer workflow framework • Other research councils: BBSRC, NERC, MRC 20 10 U #V WTX#VTW Y[a Z]\ #^ _'` W ^#` ` W#bdc ` Demanding Applications Visualisation NOW Computation Data • Networking – – – – Latency Bandwidth Multicast QoS U #V WTX#VTW Y[a Z]\ #^ _'` W ^#` ` W#bdc ` • Resources 3 Years – – – – HPC Throughput Storage Data 21 Networking Challenges • UK Backbone – 2001: 2.5Gbps – 2002: 2x10Gbps • Links to regional networks (MANs) – 2.5 Gbps • Local MANs to HEI’s – 100Mbps to 1Gbps • Within an HEI – 10Mbs to 1Gpbs 22 11 U #V WTX#VTW Y[a Z]\ #^ _'` W ^#` ` W#bdc ` Computational Resources • Provided through Regional Centres – Hardware: commodity clusters to HPC machines – Software: Linux, Solaris, Tru64, AIX… – NB: No hardware provision within UK e-Science • Accessible to UK e-Science community – Centre’s have contributed resources to the Grid – No centralised accounting – No centralised management of user accounts 23 U #V WTX#VTW Y[a Z]\ #^ _'` W ^#` ` W#bdc ` Example Collective Service: Community Authorization 1. CAS request, with resource names and operations 2. CAS reply, with capability and resource CA info User 3. Resource request, authenticated with capability user/group membership CAS Does the collective policy authorize this request for this user? resource/collective membership collective policy information Resource Is this request authorized by the capability? local policy information 4. Resource reply Is this request authorized for the CAS? efRHg?RihH9<RH?jBklRHGInmKo9qp9>r2g9<6:sH9<ItuFHGwvw9<jB68LInFHKoL9<?; 24 12 U #V WTX#VTW Y[a Z]\ #^ _'` W ^#` ` W#bdc ` The Grid World: Current Status • Dozens of major Grid projects in scientific & technical computing/research & education • Considerable consensus on key concepts and technologies – Open source Globus Toolkit™ an evolving de-facto standard for major protocols & services – Condor: long development history & robust infrastructure • Industrial interest emerging rapidly in Globus toolkit – IBM: Adopting & deploying Globus – Platform: Providing Globus software & professional services • Opportunity: convergence of eScience and eBusiness requirements & technologies 25 U #V WTX#VTW Y[a Z]\ #^ _'` W ^#` ` W#bdc ` The Grid Opportunity: eScience and eBusiness • Physicists worldwide pool resources for peta-op analyses of petabytes of data • Civil engineers collaborate to design, execute, & analyze shake table experiments • An insurance company mines data from partner hospitals for fraud detection • An application service provider offloads excess load to a compute cycle provider • An enterprise configures internal & external resources to support eBusiness workload 26 13 U #V WTX#VTW Y[a Z]\ #^ _'` W ^#` ` W#bdc ` “Web Services” • Increasingly popular standards-based framework for accessing network applications – W3C standardization; Microsoft, IBM, Sun, others • WSDL: Web Services Description Language – Interface Definition Language for Web services • SOAP: Simple Object Access Protocol – XML-based RPC protocol; common WSDL target • WS-Inspection – Conventions for locating service descriptions • UDDI: Universal Description, Discovery, & Integration – Directory for Web services 27 U #V WTX#VTW Y[a Z]\ #^ _'` W ^#` ` W#bdc ` Open Grid Services Architecture • Service orientation to virtualize resources • Define fundamental Grid service behaviors – Core set required, others optional x A unifying framework for interoperability & establishment of total system properties • Integration with Web services and hosting environment technologies x Leverage tremendous commercial base Standard IDL accelerates community code x • Delivery via open source Globus Toolkit 3.0 x Leverage GT experience, code, mindshare 28 14 U #V WTX#VTW Y[a Z]\ #^ _'` W ^#` ` W#bdc ` Summary • Grid Middleware is here & being widely deployed – Condor: very mature & robust – Globus: high quality research software • UK e-Science programme – International recognition – Diverse application community • Production Grids are emerging – UK e-Science Grid – European Data Grid – NASA’s Information Power Grid 29 U #V WTX#VTW Y[a Z]\ #^ _'` W ^#` ` W#bdc ` Further Information • UK Grid Support Centre – http://www.grid-support.ac.uk • Globus Toolkit – http://www.globus.org • Condor – http://www.cs.wisc.edu/condor • Contact: – s.newhouse@doc.ic.ac.uk 30 15