The UK National Grid Service http://www.ngs.ac.uk Andrew Richards – CCLRC, RAL History Pre NGS • L2G – Level 2 Grid deployed Easter 2003 – • contains members from each of the ten UK e-Science Centres and CCLRC plus Centres of Excellence Requirement for dedicated resources Rollout of NGS • NGS Evaluation and Installation May 2003 -> December 2003 • Pre-Production use Started April 2004 (‘Early Bird Users’) • Full Production Started – September 2004 (All Hands Meeting 2004 – Nottingham) Funding – JISC funded £1.06M Hardware (2003-2006) • Oxford, White Rose Grid (Leeds) - Compute Cluster • Manchester – Data Cluster • E-Science Centre funded 420k for additional Data Cluster – JISC funded £900k staff effort including 2.5SY at CCLRC e-Science Centre • The e-Science Centre leads and coordinates the project for the JISC funded clusters Objectives • Build on recent advances in grid technology to offer a stable highly available production quality grid service • Interoperability with other grids such as those deployed by the Enabling Grids for eScience in Europe (EGEE) project and the LHC Computing Grid (LCG) • To provide the broad user community with access to large scale resources through the grid interface • To meet users’ evolving requirements for a production grid service by working with the user community, through the User Group established by the Core Programme • To cooperate effectively with the Engineering Task Force and other teams NGS Today NGS “Tomorrow” GOSC Timeline NGS WS Service NGS Expansion (Bristol, Cardiff…) NGS Production Service NGS WS Service 2 OGSA-DAI NGS Expansion WS2 plan WS plan Q2 Q3 Q4 Q1 2004 Q2 Q3 2005 Q4 Q1 Q2 Q3 2006 OMII release gLite release 1 EGEE gLite alpha release EGEE gLite release OMII Release Grid Operation Support Centre Web Services based National Grid Infrastructure UK e-Infrastructure NGS Hardware Compute Cluster Data Cluster •64 dual CPU Intel 3.06 GHz (1MB cache) nodes •2GB memory per node •2x 120GB IDE disks (1 boot, 1 data) •Gigabit network •Myrinet M3F-PCIXD-2 •Front end (as node) •Disk server (as node) with 2x Infortrend 2.1TB U16U SCSI Arrays (UltraStar 146Z10 disks) •20 dual CPU Intel 3.06 GHz nodes •4GB memory per node •2x120GB IDE disks (1 boot, 1 data) •Gigabit network •Myrinet M3F-PCIXD-2 •Front end (as node) •18TB Fibre SAN ( Infortrend F16F 4.1TB Fibre Arrays (UltraStar 146Z10 disks) •PGI compilers •Intel Compilers, MKL •PBSPro •TotalView Debugger •RedHat ES 3.0 •PGI compilers •Intel Compilers, MKL •PBSPro •TotalView Debugger •Oracle 9i RAC •Oracle Application server •RedHat ES 3.0 NGS Software NGS Database Service • Database Service on CCLRC – RAL and Manchester Sites • • • Oracle 9i RAC (hold license for 10g if/when required) Using OCFS Oracle Application Server • • • 5 node Clustered Database Single node Database for SRB Use of Transparent Application Failover (TAF) NGS SRB • Single instance Oracle 9i for MCAT at CCLRC – RAL, duplicated at Manchester • SRB Version 3.2.1, using GSI authentication mechanisms • SRB vaults at each of the NGS core sites • Single NGS zone for core nodes – Option to use Federation for other zones. Monitoring BDII Nagios Ganglia GITS Testing User Support http://www.grid-support.ac.uk support@grid-support.ac.uk NGS Portal User Registration ‘Early Bird’ Users April 2004 Example Users Projects • e-Minerals • e-Materials • Orbital Dynamics of Galaxies • Bioinformatics (using BLAST) • GEODISE project • UKQCD Singlet meson project • Census data analysis • MIAKT project • e-HTPX project. • RealityGrid (chemistry) Users • Leeds • Oxford • UCL • Cardiff • Southampton • Imperial • Liverpool • Sheffield • Cambridge • Edinburgh • QUB • BBSRC • CCLRC User Registration (Process) User Applies Accepted (Via Website) (User Notified) Application QC User added to NGS VO (includes SRB account) Application Submitted to Peer Review Panel Approved ? Rejected (User Notified) User added NGS-USER and NGS-ANNOUNCE Mailing lists User Example GODIVA Diagnostics Study of Oceanography Data EVE Excitations and Visualisation Project Integrative Biology Simulation of Strain in Soft Tissue under Gravity Production Status • • • • • • • • • • TODAY 4 new core nodes operational 125 Users registered (25 since 1 September ’04) Grid enabled – Globus v2 (VDT distribution v1.2) at present BDII information service (GLUE + MDS Schemas) Data Services – Oracle and SRB Growing base of user applications MyProxy and CA services provided by UK Grid Support Centre VO Management Software – LCG-VO User support: Helpdesk • • • • • Next… Other Middleware [gLite/OMII etc…] NGS Portal Resource Broker SRB production service Contact Information NGS Grid Coordinator Andrew Richards: a.j.richards@rl.ac.uk NGS Website http://www.ngs.ac.uk Helpdesk (via Grid Support Centre) Email: support@grid-support.ac.uk