Document 13346895

advertisement
Overview of UK e-Science and
its enabling Grid Middleware
Dr Steven Newhouse
Technical Director,
London e-Science Centre,
Imperial College, London
(http://www.lesc.ic.ac.uk/)
s.newhouse@doc.ic.ac.uk
•
•
•
•
Overview
What is the Grid?
Current Grid Infrastructures
UK e-Science Programme
The future…
• Content Acknowledgements:
– Ian Foster (Globus co-PI)
– Miron Livney (Condor PI)
2
1
Grid Computing
3
The Grid: A Brief History
• Early 90s
– Gigabit testbeds, metacomputing
• Mid to late 90s
– Early experiments (e.g., I-WAY), academic software projects
(e.g., Globus, Legion), application experiments
• 2002
– Dozens of application communities & projects
– Major infrastructure deployments
– Significant technology base (esp. Globus ToolkitTM)
– Growing industrial interest
– Global Grid Forum: ~500 people, 20+ countries
4
2
The Grid – a definition
“Secure resource sharing & coordinated problem solving
in dynamic, multi-institutional virtual organisations”
– On-demand, ubiquitous access to computing, data, and
services
– New capabilities constructed dynamically and
transparently from distributed services
5
What is an e-Scientist?
• Applied Scientists are becoming e-scientists:
– computational & data services
• HPC, genetic databases, electronic journals
– remote sensors
• Large Hadron Collider, SOHO
– personal devices
• workstations, mobile phones, PDA’s
– supporting collaborative working
• ‘collaboratories’: global scientific collaborative communities
e-science enabled through Grid middleware
6
3
CERN’s Large Hadron Collider
1800 Physicists, 150 Institutes, 32 Countries
100 PB of data by 2010; 50,000 CPUs?
7
Data Grids for
High Energy Physics
|}@~@€v1‚ƒ‚2„
h0S ‰ sŠ W XY kkV `‹*W xYj U1w dve%f*Œ ii%i
!" 8@8LNM ,-O-*4
Š kU_‰1^@j %f0UŽ ]W c Y w U1^
j X
0
'
1
/
2
/
)
$
1
3
*
4
%
3
6
5
7
.
S T*U1V UW XY[Z \%]^ _T_V `X X W ^ a6b U2c U1V def0^X U1_ X g
! 89#*$& 8
89LM ,*%-O*-*4
S T*U1V UY V UGh i%i(Z j V W aa
U1V X@b kU*V2X U1_` ^l
!"
m Y _Tnj V W aa
U1V U1lGUc U1^
jW Xo1hpqRd
j UW ^X W r U
:;6<=31.6>*?@ GL…QR -O-4 Tier 0
3GA G5 DI G†!R{P > *47* {‡
Tier 1
'(*) +&,*-.
E .7 ,0;
%D2 3 7)
5 7 4*0;2D 3 72)
#%72) ,6;
%D 32 7)
Tier 2
!P @LNQR - O-4
$I ,*-
4-{R7*764*7*4I #.6>* 7)
!" #%$&
# -
?*
! 8GF H #%$#& -
? # -
?* # -
?*
!" LNM ,-O-*4
$I ,*-
4 -J3K - 7*@ 3 ‘“’%”–•—R˜–™›šœž’%Ÿ •”
Tier 4
‘“’%˜–˜–¡¢”£’%œ¢¤¦¥
;6A%B !C #$&
!P @LNQR - O*-%4
&413 E { ˆ v !" #%$& !" #$& !" #%$& !" #%$&
s T1dX W _W Xj Xt@`V uG`^vY ^Y w dX W X[Z _T*Y ^^ U1w Xb g
m Y _T6W ^ Xj W j ]
j UGtW w w1T*Y@c UGo1h i0kT*dX W _ W XRj Xt`V u*W ^ a9`^v`^ U`Vx`V U
_T*Y ^^ U1w X y*lYj Yz `1V0j T*U*X U_T*Y ^^ U1w XX T1`]w lG\U_Y _T*Ul6\dj T1U
W ^ X@j W j ]@j UvX U1VRc U*V
‘§’%¤¦¨¢©9¡«ª ¥Gª ”«•—R¡£’%Ÿ¬•” 8
4
The Globus Toolkit in One Slide
• Grid protocols (GSI, GRAM, …) enable resource
sharing within virtual orgs; toolkit provides reference
implementation ( = Globus Toolkit services)
O ² Ñ »N®»¿»%¯
Ù RÆÅ ©GT
S
Ó6¯NÃ ¾ ¿+Ô[Ã ¯
É Ù ¤¦¥6ª R —R•¤¦Ê…¥6Ÿ¬•š Å ¤¦•VU—Rʅ¤ Í ° ¯àN¯+¾ ®½:»Y…° º[¿¾»v°¾¸ ²+½XW
ÄÆÅÈÇ ­®¯+°
° ¯+Õ¦²»¯
¾ ½vÖ0²³¿»v¾ ²+½ Î ¿»¯+Ï2¯¯ ± ¯+° Ó6¯ ± ²+° »%¯+°
ÎNMMOQPnÎ °
¾ Â
É Ä •—%¡
Ð ° ¯àN¾ ®»° ¸âá
M
½ Ñ ²+°
Õ¦¿»v¾ ²+½
Ð
Ñ
›
¹
º
»
¼
+
¯
v
½
»
¾
³
¿
%
»
Á
¯
À
¿
³
%
»
+
²
°
(
¸
Ò
Å ¤¦Ê…¨¢•—%¥
š
ÂN¾ ®³²[Ö6¯+° ¸+Ò
/1023#45687:9
M ½vˆ¯· O ¯+° Ö0¯+°
³[° ¯¿»%¯ ± ° ²·¸
Ð ÂN¾ ®³²[Ö6¯+° ¸+Ò
Ç œ«Ë•ª ̅¥0•¨¢Ê © ³[° ¯Âˆ¯+½v»v¾ ¿…Ã
ä(åæçè%æÛé…åêëæììîí æïˆð ìè%æå
;#< 023#=>0?A@ ; 03#B
C
4
#
3
F
D
G
E
0
3
G
H
#
3
.
4
8
I
A
?
@
3
¥G¨¢•¤¦Í
­®¯+°
­®¯+°
± ° ²³¯®®ž´›µ
4J3#K < 3LHJ0H
± ° ²³¯®®£´›ã
' »¼v¯+°®¯+° Ö+¾ ³¯
¶° ²·¸
¶° ²·¸â´›ã
ÄÆזØÚÙ
Ð ¯)( à*( Î °@¾ Â,+.-¶Ò
É Ä •—R¡ × ¤ÛÌNŸ ¨¢•ÊN¤ ØÚÜRÜ Ÿ ÊNª ¥—RŸ œÞÝ Ù ª œ¢ª ”–¤¦ßÞ¤¦œ¢¥GÍ
ñóòÛô=õö[õ÷øõÚù úüû…ýÁþ ÿ ò =úÁþ£ýÚù 
õ ö
Áô ö[õõ ù úüýÁþ£ÿ úÁôR÷Áú
õ
ô Áô=
ú ÿ RúÛ÷ÛõÁô ÿÈÁ
ý ö[ý ö ! õÈô#" ù õ$&%
9
Condor
npƒ#ƒ8d…„ a*c]q]„ ^#h
y ‘.u
Ž‡ ’
“ Žp” |!–
~ •
npƒ#ƒ8d…„ a*c]q]„ ^#h†n‡i*[h,q
8_*\>qG^#€[]`‚npi#[jh*q
rts!uvxwzys!{}|!~
kml hL[]`onpiL[h,q
Z[]€^Lq[‰ˆNŠ‹[ja,_*q]„ ^#hŒnpiL[jh*q
—F˜ |!v Ž u|
“,™›šœ |!–
~ •
b]^La,cdNZ[]\>^L_*`>a,[fegch*cji*[j`
Z[]\>^L_,`>a*[
10
5
Condor within the Grid
ž ~ ™ vx| ”” ‘ œxŸz ” u ™ ~–s Ÿ | v ™ y¡y Ž œ ‘¢vxs!u‡‘ ™›œ
11
Mathematicians Solve NUG30
• Looking for the solution to the
NUG30 quadratic assignment
problem
• An informal collaboration of
mathematicians and computer
scientists
• Condor-G delivered 3.46E8 CPU
seconds in 7 days (peak 1009
processors) in U.S. and Italy (8
sites)
£¤}¥*¦¥§}¨¥§}¤¥,£¥©¥,£ª}¥*£›¦¥
£«}¥¬}¥§S£¥§¥¤¥§}¬¥§¦¥§›§¥
£©}¥§}ª¥,£­¥©«}¥ª}¥§›«¥,£¬¥
¨¥,£¨}¥­}¥§}­¥,£§¥,££¥§›©
Ù ¤¦¥Gª‹®}¯N° Ł±¢Ø •[”–Ÿ œ¢œ«¤³² Ç Ÿ¬ª‹²´® Ÿ •¥6™«¤¦Ì…¥0¤¦•œ²µ —R̅ʟ œ«ÌN—Rœ
12
6
GriPhyN: CMS Data
Reconstruction
Master Condor
job running at
Caltech
2) Launch secondary job on WI pool;
input files via Globus GASS
Secondary
Condor job on WI
pool
5) Secondary
reports complete
to master
1) Caltech
workstation
6) Master starts
reconstruction jobs
via Globus
jobmanager on
cluster
3) 100 Monte
Carlo jobs on
Wisconsin Condor
pool
9) Reconstruction
job reports
complete to master
4) 100 data files
transferred via
GridFTP, ~ 1 GB
each
7) GridFTP fetches
data from UniTree
NCSA Linux cluster
8) Processed
objectivity
database stored
to UniTree
Å ÊNŸ ¥G¥ ¶ Ÿ •ª œ¢¡–ª ² Ù —R•Ÿ œ‰·›— Uœ›š ²£Ÿ ¥6™¢¤ •Ì
NCSA UniTree
- GridFTPenabled FTP
server
13
Grid Middleware Requirements
• Security
– X.509 PKI with certificate authority
– Mapping of Grid Identity to local UNIX account
• Connections
– Defined ports or port range through firewalls
– GridFTP (modified FTP protocols)
• Data
– Very large data transport requirements
– Prescribed time frames (QoS)
14
7
UK e-Science Funding 2000-03
ÇÈ$É ¾XÊJ¾XËÂC¼»ÌzÄQÅÀÍÀ¿À¼»½ Î Ê
¸ ¹»º]¼»½ ¾¿À¼»¾
³
º]ÁJ¾¾ÂJ½ ¿ÀÃ!ÄQÅÀÆNÆ1½ ÁJÁC¾¾
Ç ½ ÂJ¾¼»ÁCÅÀÂ
ÏÐ ÑCÒÓ»ÔCÕÀÑJÖ ×
شٻÚÀÙ»Û ÒÜ8Ò Ú Ô8Ý)ÕÀÞ Ò
ÏÐ ÑCÒÓ»ÔJÕÀÑCÖ ×
ßà Ù ÑCÒ Ú ÒX×C× ÙáÚÀâ´ã ÕÀäåÕÀÑ â Ð ÚÀÙ ÔJÐ Õ Ú Ý,ÕÀÞ Ò
È ¾¿À¾XÂC½ ¼}ÄQÌGËXÎ Î ¾¿ÀþÊ
ø ¼»Ëô¾XÆN½ ¼ ³
ø ùù Î ½ ¼»ËÁJ½ ÅÀ¿º]Í ùù ÀÅ Â:Á
æ)çÀè Ý ãxéëê1ìîí ÜNïðÏjñ#ò éëê1ìîí Ü1ï
ç ÑCÕ Û Ñ Ù ÜNÜNÒ
Ý)Ò×JÒ Ù ÑJÓ»ú ã ÕÀû Ú Ó»Ð Þ × éë
Ü ï
ê ü>ö Ü8ïëðGÏñ#ò é êí N
çÀç ß Ý ãxéëê1ý>þ ÜNï
ÿ,ÿ è Ý ãxéë
ê ÜNï
Ø Ý ãxéë
ê Ü1ï
æ Ý ãxéëê1ü Ü1ï
æ,è Ý ãxéë
ê ÜNï
£80m Collaborative projects
æ,çÀè Ý ãxéëêìîü ÜNï
ã Ý ãxéëêí ÜNï
ó ¿ÀôGÍÀÊJÁCÂJ½ ËXÎjÄQÅÀÎ Î Ëõ>ÅÀÂCËXÁC½ ÅÀ¿ éëêNöQ÷ ÜNï
e-Science Director : Professor Tony Hey
15
Regional e-Science Centres
_‹d !#"%$'&&(*),+.-%/10L2!3
è&è Ð ÔCÒ×
]Þ Ù × Û Õ à
æ â Ð Ú >ûÀÑ Û ú
é1Ù ÔJÐ Õ ÀÚ Ù Þ ã Ò Ú ÔCÑJÒï
Ò à Ó Ù ×áÔCÞ Ò
ÿ)ÒXÞ Ù ×JÔ
ã>Ù Ñ â Ð 4
ã æ Ý Ø´ÙáÚ Ó»úÒ×JÔCÒÑ
Ï Ïû>Þ Ð Ú
æ ÕÀÑ â
Ý ß è ÕÀûÀÔCú Ù ÜÔCÕ Ú
ã>Ù Ü>ÑCÐ âÛ Ò
òÜÒXÑCÐ Ù Þ ã ÀÕ Þ Þ Ò Û Ò
é Õ ÚÀâ Õ Ú ï
16
8
Grid Support Centre
• First point of contact for ‘getting on the Grid’
• Grid Starter Kit (GSK): Globus, Condor & SRB
• Provides:
– user support for deployment & debugging
– software distribution and maintenance of GSK
– central point of contact with Globus team
• Virtual Centre drawing on expertise at:
–
–
–
–
Rutherford Appleton Laboratory (Oxford)
Daresbury Labs (Manchester)
Edinburgh Parallel Computing Centre (NeSC)
University of Manchester (Manchester Computing & ESNW)
• http://www.grid-support.ac.uk
17
Grid Engineering Task Force
• Experts from the Regional & Grid Support Centres
• Working Groups
–
–
–
–
–
–
–
Globus 2: Deployment on contributed Grid resources
Directory Services: National GIIS: ginfo.grid-support.ac.uk
Security: X.509 PKI through National CA, local RA’s
Firewalls: At a university & centre level
Grid User Management: Account creation & accounting
Integration: Verify the functionality of the ‘Grid’
Access Grid: Collaborative workspace through multicast
• Basic UK grid infrastructure demonstrated at NeSC
opening
18
9
Access Grid
• Collaborative work
among large groups
• ~50 sites worldwide
• Use Grid services for
discovery, security
• www.scglobal.org
Presenter
mic
Presenter
camera
Ambient mic
(tabletop)
57686:9<;8;>=%?@BADC157?EFHGG9<IJFKML9<?;
U #V WTX#VTW
Y[a Z]\ #^ _'` W ^#`
` W#bdc `
Audience camera
NONPNQSRH6T689;T;8E3?@BADQSFH?19E
Applied User Base
• Regional Centre Projects
• EPSRC Pilot Projects
–
–
–
–
Geodise: Engineering Optimisation
Reality Grid: Distributed mesoscale materials modelling
Discovery Net: Data mining on distributed high throughput data sources
Comb-e-Chem: Synthesis of new chemical compounds by combinatorial
methods
– MyGrid: Bioinformatics workbench infrastructure
– DAME: Distributed analysis of jet engine sensor data
• PPARC Projects
– GridPP: UK activities within EU-DataGrid
– AstroGrid: Federation & datamining of astronomical data
– Grid1D: Development of distributed peer-to-peer workflow framework
• Other research councils: BBSRC, NERC, MRC
20
10
U #V WTX#VTW
Y[a Z]\ #^ _'` W ^#`
` W#bdc `
Demanding Applications
Visualisation
NOW
Computation
Data
• Networking
–
–
–
–
Latency
Bandwidth
Multicast
QoS
U #V WTX#VTW
Y[a Z]\ #^ _'` W ^#`
` W#bdc `
• Resources
3 Years
–
–
–
–
HPC
Throughput
Storage
Data
21
Networking Challenges
• UK Backbone
– 2001: 2.5Gbps
– 2002: 2x10Gbps
• Links to regional
networks (MANs)
– 2.5 Gbps
• Local MANs to HEI’s
– 100Mbps to 1Gbps
• Within an HEI
– 10Mbs to 1Gpbs
22
11
U #V WTX#VTW
Y[a Z]\ #^ _'` W ^#`
` W#bdc `
Computational Resources
• Provided through Regional Centres
– Hardware: commodity clusters to HPC machines
– Software: Linux, Solaris, Tru64, AIX…
– NB: No hardware provision within UK e-Science
• Accessible to UK e-Science community
– Centre’s have contributed resources to the Grid
– No centralised accounting
– No centralised management of user accounts
23
U #V WTX#VTW
Y[a Z]\ #^ _'` W ^#`
` W#bdc `
Example Collective Service:
Community Authorization
1. CAS request, with
resource names
and operations
2. CAS reply, with
capability
and resource CA info
User
3. Resource request,
authenticated with
capability
user/group
membership
CAS
Does the
collective policy
authorize this
request for this
user?
resource/collective
membership
collective policy
information
Resource
Is this request
authorized by
the
capability?
local policy
information
4. Resource reply
Is this request
authorized for
the CAS?
efRHg?RihH9<RH?jBklRHGInmKo9qp9>r2g9<6:sH9<ItuFHGwvw9<jB68LInFHKoL9<?;
24
12
U #V WTX#VTW
Y[a Z]\ #^ _'` W ^#`
` W#bdc `
The Grid World: Current Status
• Dozens of major Grid projects in scientific & technical
computing/research & education
• Considerable consensus on key concepts and
technologies
– Open source Globus Toolkit™ an evolving de-facto standard
for major protocols & services
– Condor: long development history & robust infrastructure
• Industrial interest emerging rapidly in Globus toolkit
– IBM: Adopting & deploying Globus
– Platform: Providing Globus software & professional services
• Opportunity: convergence of eScience and eBusiness
requirements & technologies
25
U #V WTX#VTW
Y[a Z]\ #^ _'` W ^#`
` W#bdc `
The Grid Opportunity:
eScience and eBusiness
• Physicists worldwide pool resources for peta-op
analyses of petabytes of data
• Civil engineers collaborate to design, execute, &
analyze shake table experiments
• An insurance company mines data from partner
hospitals for fraud detection
• An application service provider offloads excess load
to a compute cycle provider
• An enterprise configures internal & external
resources to support eBusiness workload
26
13
U #V WTX#VTW
Y[a Z]\ #^ _'` W ^#`
` W#bdc `
“Web Services”
• Increasingly popular standards-based framework for
accessing network applications
– W3C standardization; Microsoft, IBM, Sun, others
• WSDL: Web Services Description Language
– Interface Definition Language for Web services
• SOAP: Simple Object Access Protocol
– XML-based RPC protocol; common WSDL target
• WS-Inspection
– Conventions for locating service descriptions
• UDDI: Universal Description, Discovery, & Integration
– Directory for Web services
27
U #V WTX#VTW
Y[a Z]\ #^ _'` W ^#`
` W#bdc `
Open Grid Services Architecture
• Service orientation to virtualize resources
• Define fundamental Grid service behaviors
– Core set required, others optional
x A unifying framework for interoperability &
establishment of total system properties
• Integration with Web services and hosting
environment technologies
x
Leverage tremendous commercial base
Standard IDL accelerates community code
x
• Delivery via open source Globus Toolkit 3.0
x
Leverage GT experience, code, mindshare
28
14
U #V WTX#VTW
Y[a Z]\ #^ _'` W ^#`
` W#bdc `
Summary
• Grid Middleware is here & being widely deployed
– Condor: very mature & robust
– Globus: high quality research software
• UK e-Science programme
– International recognition
– Diverse application community
• Production Grids are emerging
– UK e-Science Grid
– European Data Grid
– NASA’s Information Power Grid
29
U #V WTX#VTW
Y[a Z]\ #^ _'` W ^#`
` W#bdc `
Further Information
• UK Grid Support Centre
– http://www.grid-support.ac.uk
• Globus Toolkit
– http://www.globus.org
• Condor
– http://www.cs.wisc.edu/condor
• Contact:
– s.newhouse@doc.ic.ac.uk
30
15
Download