Polish Infrastructure for Supporting Computational Science in the European Research Space Recent Advancements of National Grid Infrastructure in Poland (PL-Grid) Jacek Kitowski ACC Cyfronet AGH, Krakow, Poland www.plgrid.pl/en Cracow Grid Workshop – CGW’10 11-13.10.2010, Krakow, Poland Acknowledgements ACC Cyfronet AGH Kazimierz Wiatr Marian Bubak Łukasz Dutka Karol Krawentek Andrzej Oziębło Maciej Malawski Zofia Mosurska Robert Pająk Marcin Radecki Renata Słota Mariusz Sterzel Tomasz Szepieniec Agnieszka Szymańska Teresa Ozga ICM Piotr Bała Maciej Filocha PCSS Norbert Meyer Bartek Palak Krzysztof Kurowski Tomasz Piontek Dawid Szejnfeld Paweł Wolniewicz WCSS Jerzy Janyszek Bartłomiej Balcerek TASK Rafał Tylman Mścislaw Nakonieczny 2 Outline General Introduction to PL-Grid Project Motivation, consortium and objectives PL-Grid and EGI Organization of the project Current Status and Advancements Status of hardware infrastructure Operational Centre – tasks and services e-Infrastructure software, software packages and user’s tools Training and user support Security aspects Conclusions – Future work 3 PL-Grid Consortium Consortium Creation – agreement signed January 2007 a response to requirements of Polish scientists and due to ongoing Grid activities in Europe within the framework of EGI_DS. Consortium members – make up of 5 Polish supercomputing and networking centres: Academic Computer Centre CYFRONET AGH, Krakow (coordinator) Interdisciplinary Centre for Mathematical and Computational Modelling, Warsaw University Poznan Supercomputing and Networking Centre Academic Computer Centre, Gdansk Wroclaw Centre for Networking and Supercomputing PL-Grid Project proposal got funded by the European Regional Development Fund as part of the Innovative Economy Program on March 2, 2009. Result: First working NGI in Europe in the framework of EGI.eu 4 Rationales behind PL-Grid Consortium The Consortium consists of five High Performance Computing Polish Centres representing Communities, coordinated by ACC Cyfronet Participation in international and national projects 35+ international projects FP5, FP6, FP7 on Grids (~50% common) (~45 currently, including EDA) 15+ Polish projects (~50% common) Polish scientific communities ~75% highly rated Polish publications in 5 Communities Computational resources Top500 list European/Worldwide integration Activities EGEE I-III, EGI_DS, EGI, e-IRG, PRACE, DEISA, OMII, EU Unit F3 „Research Infrastructure” Experts, Representatives in EGI Council, Members of EGI Executive Board National Network Infrastructure available Rank 161 181 444 Pionier National Project TOP500 – June 2010 Site System Cores Rmax (TFlops) Rpeak (TFlops) Cyfronet AGH Krakow Cluster Platform 3000 BL2x220, L54xx 2.5 Ghz, Infiniband / 2010 Hewlett-Packard 6144 39.93 55.54 ACTION Cluster Xeon E5345 Infiniband / 2008, Action 5336 38.17 49.73 Cluster Platform 3000 BL 2x220, E5530 2.4 GHz, Infiniband GDR / 2010, Hewlett-Packard 3456 26.22 31.24 Gdansk University of Technology, CI Task PCSS Poznan 5 PL-Grid Project - Basic Data The Project is co-funded by the European Regional Development Fund as part of the Innovative Economy Program. Total budget: 83 M PLN (~ 21 M EUR) • Personel cost 27 M PLN (~7 M EUR) • Equipment cost 33 M PLN (~8 M EUR) • Other cost 23 M PLN (~6 M EUR) Funding from the EC: 68 M PLN (~ 17 M EUR) Project duration: 01 January 2009 – 31 December 2011 Beneficiary: Academic Computer Centre Cyfronet AGH, Krakow, Poland Contract number: POIG.02.03.00-00-007/08 Project website: www.plgrid.pl/en 6 Main Objectives of PL-Grid Polish Grid is developing a common base infrastructure – compatible and interoperable with European and Worldwide Grids Capacity to construct specialized, domain Grid systems – including services and tools focused on specific types of applications This approach should enable efficient use of available financial resources Plans for HPC and Scalability Computing enabled with more focus on domain specific Grids Advanced Service Platforms Application Application Domain Grid Application Computing Power 215 Tflops Storage 2500 TBytes Support from PL-Grid staff on using advanced Grid tools Support on porting legacy codes to Grid environment Support on designing applications for PL-Grid environment Application Offer for the Users Domain Grid Domain Grid Domain Grid Grid infrastructure (Grid services) PL-Grid Clusters High Performance Computers National Computer Network PIONIER Data repositories 7 PL-Grid Building Blocks PL-Grid software comprises: user tools (portals, systems for applications management and monitoring, result visualization and other purposes, compatible with the lower-layer software used in PL-Grid) software libraries virtual organization systems: certificates, accounting, security, dynamic data management systems: metadata catalogues, replica management, file transfer resource management systems: job management, applications, grid services and infrastructure monitoring, license management, local resource management, monitoring Users Grid Application Programming Interface Grid portals, development tools Virtual organizations and security systems Grid services LCG/gLite (EGEE) UNICORE (DEISA) Other Grids systems Basic Grid services Grid resources Distributed data repositories National computer network Distributed computational resources Three Grid structures are maintained: production research development / testing 8 PL-Grid – first operational NGI in Europe EGI.eu – European organization being developed to coordinate the European Grid Infrastructure, based on the federation of individual National Grid Initiatives (NGI), to support a multi-disciplinary user community. PL-Grid tasks in EGI (planned/fixed) Grid operation and oversight of the e-Infrastructure Coordination of resource allocation and of brokering support for VOs from NGIs Computational Chemistry – organization and management of Computational Chemistry and Material Science and Technology Specialized Support Centre and EGI liaisons Development of unified middleware via European Middleware Initiative PL-Grid Integration Activity in the framework of European Grid Infrastructure On 30 March 2010, Poland - as the first country in Europe - has initiated functioning of the National Grid Initiative (NGI) Scientific application porting, especially concerning UNICORE architecture, within Application Porting SSC 9 Organization of the PL-Grid project P1 PROJECT MANAGEMENT Coordination P2 PLANNING AND DEVELOPMENT OF INFRASTRUCTURE Strategic Planning Dissemination P6 P3 OPERATIONS CENTER P5 P4 GRID SOFTWARE AND USERS TOOLS DEVELOPMENT SECURITY CENTER EGEE DEISA …. SUPPORT FOR VARIOUS DOMAIN GRIDS Training 10 Current Status and Advancements 12 Status of Hardware Infrastructure Current status (October 2010): Computational power CYFRONET: 29,3 Tflops ICM: 13,1 Tflops PCSS: 23,2 Tflops Disk storage CYFRONET: 313 TBytes ICM: 433 TBytes PCSS: 300 TBytes Total: 65,6 Tflops Total: 1046 TBytes Plans until the end of 2010: 185 Tflops 1900 TBytes Plans until the end of the Project (end of 2011): 215 Tflops 2500 TBytes 13 Operational Center – just in operation Tasks Coordination of operations Management and accounting Collaboration with EGI and PRACE/DEISA Users’ requirements analysis for operational issues Running infrastructure for: Production Developers Research Future consideration: Computational Cloud Data Cloud Internal and External Clouds Virtualization aspects 14 Services of the Operational Center for Users Operational Center aims at facilitating access to the infrastructure by simplifying the procedures and deployment of useful tools: System of registration of account management of the PL-Grid user • available at https://konto.plgrid.pl/ • required entry in the Polish database of „People of Science” or confirmation of the scientific tutor • grid access to PL-Grid resources 5 centers – gLite 1 center – UNICORE • local access to the queue system „zeus” cluster in ACC CYFRONET AGH • ability of application for a grid certificate on-line • application for access to computational services in other centers Helpdesk system in PL-Grid • enables reporting and tracking issues • available at https://helpdesk.plgrid.pl (for registered users) • access also by e-mail: helpdesk@plgrid.pl • manual: https://wiki.plgrid.pl/doku.php?id=pakiet5:publiczne:podrecznik_uzytkownika_pl-grid → System Pomocy Helpdesk 15 Services of the Operational Center for Users (cont’d) Operational Center cares for a proper functioning of the infrastructure for PL-Grid users by pro-active monitoring of the following infrastructure elements: availability of the infrastructure services software packages supported by PL-Grid Provision of the conformity of the PL-Grid and European (EGI) infrastructures software operational procedures security procedures Advanced work on the „PL-Grid grants” idea Integration of the data presentation concerning resources usage for a user Work on provision of the integrated user portal 16 Development of the e-Infrastructure Software and the User’s Tools Close cooperation of 8 programming and testing groups, about 20 people Installation and provision for testing purposes gLite, Unicore and QosCosGrid About 30 various configurations of virtual machines with installed software used for development and testing of the tools for users – the choice of the technology made Functional, conformity and efficiency tests of selected packages of the research software made in order to perform the deployment and support of the new tools and services on the production level 17 Development of the e-Infrastructure Software and the User’s Tools (cont’d) Direct contact with new users on the basis of a survey, available at: www.plgrid.pl/ankieta Requirements of the Polish users (results ~190 surveys) considered in the new applications, tools and services developed and tested in the framework of the Package 4 Large group of users cooperating with the software and tools team Department of Chemistry of the Jagiellonian University Department of Bioinformatics and Telemedicine of the Collegium Medicum of the Jagiellonian University University of Adam Mickiewicz Poznan Technical University Wrocław Technical University Administrators of the computing centers … 18 Development of the e-Infrastructure Software and the User’s Tools (cont’d) Test versions of tools for users and systems administrators: Grid Resource Bazaar, mobile access to the infrastructure, new security modules Extending of the GridSpace2 platform with a set of new functions, support for new scripting languages and integration with new grid services Integration of the Migrating Desktop and gEclipse tools with various PL-Grid middleware services Extension and deployment of FiVO – a new tool for VO management and monitoring Performance and functional tests of middleware service QosCosGrid and integration with gLite and Unicore infrastructure at the queue systems level Implementation and provision of the advanced graphical interfaces, visualization and tasks and data management for selected user applications via the Vine Toolkit Integration of the selected tools and web applications with Liferay portal framework and Nagios monitoring system 19 Grid Resource Bazaar Providing resources to users with required qualities of services Required = specified in Service Level Agreement RU – Resource Users RP – Resource Providers 20 Grid Resource Bazaar (cont’d) Resource Allocation-related Operation Model 21 Grid Resource Bazaar (cont’d) http://grid.cyfronet.pl/bazaar 22 GridSpace2 Experiment Workbench in the PL-Grid Virtual Laboratory Use of distributed computational resources and data repositories High-level tools offered for the user for in-silico experiments 23 GridSpace2 - working with the Experiment Workbench The PL-Grid Virtual Laboratory provides an Experiment Workbench – a Web 2.0 interface supporting the development and execution of in silico experiments. Working with the Experiment Workbench: Open the workbench in your browser (https://wl.plgrid.pl) Log into one of the available servers with your PLGrid account Start your usual work. Your files are already there. Your code snippets may play the role of scripts (in Python, Perl, Ruby etc.), bash commands or input to external applications (such as Gaussian or Gnuplot); All the files you generate may be viewed with visualization tools (such as Jmol) Save your work – the experiment is now visible among your other files Share the experiment with other members of your research team 24 GridSpace2 - sample experiment in ViroLab environment Patient’s data Medical examination HIV genetic sequence put into database http://gs2.cyfronet.pl/ http://www.virolab.org Experiment in-silico Collect HIV genetic sequences from database Perform sequence matching Calculate virus resistance 25 Migrating Desktop Platform – the Framework for Grid Applications Description: The Migrating Desktop Platform is a powerful and flexible user interface to Grid resources that gives a transparent user work environment and easy access to resources and network file systems independently of the system version and hardware It allows the user to run applications and tools, manage data files, and store personal settings independently of the location or the terminal Main product features: Migrating Desktop main window Simple integration with grid applications Easy job defining, submission, monitoring and visualization of results Support for batch and interactive jobs Handling of sequential and parallel applications Intuitive management of grid data Easy extendable framework 26 Migrating Desktop Platform – the Framework for Grid Applications (cont’d) Application support: The key feature: the possibility of easy adding various tools, applications and supporting different visualization formats The Migrating Desktop offers a framework that can be easily extended on the basis of a set of well-defined plug-ins used for: accessing data, defining job parameters, pre-processing job parameters, and visualization of job results Open architecture of the Migrating Desktop speeds up the application integration process and makes that product significantly more flexible than specialized tools (e.g. portals) designed only for a specific application Job defining and visualization of results (examples of Migrating Desktop plug-ins) 27 g-Eclipse – to Access the Power of the Grid Features: The g-Eclipse framework provides tools to customize Grid users’ applications, to manage remote resources and to support the development cycle of new Grid applications The framework consists of general Grid workbench tools that can be extended for many different Grid and Cloud middleware’s (such as gLite, UNICORE, Globus Toolkit, Amazon Cloud) Benefits from using g-Eclipse: Grid application users will benefit from the desktop-like access to Grid resources Grid operators and resource providers will be able to reduce the time-to-service by using the Grid management and Grid site configuration tools Grid application developers will reduce the timeto-market for new Grid applications by accelerating the development and deployment cycle 28 g-Eclipse – to Access the Power of the Grid (cont’d) g-Eclipse elements (functionalities): Grid Virtual Organisation management allows the dynamic creation and management of Virtual Organisations and their resources Grid job management supports Grid users and developers in creating and managing Grid jobs independent of the middleware Grid file management provides access and management of local and foreign resources which are seamlessly integrated into the Eclipse Workbench Grid application deployment supports Grid application developers and Grid application users with the deployment of their applications on the Grid Grid visualisation tools offer functionalities that allow the visualisation of scientific and numerical calculations Grid workflow builder supports the creation and management of middleware independent workflows by use of a provided graphical editor Grid command console is a command line interface to access the Grid g-Eclipse provides API and User Interface modules for all above functionalities. 29 FiVO/QStorMan toolkit – the goals Provisioning PL-Grid users an appropriate quality of access to storage systems with the following means: Definition of the non-functional requirements for the storage systems, e.g. expected transfer rate of the read/write operations or parameters related to expected availability level of the resource or security. Monitoring of the storage systems parameters, e.g. a distributed file system. Data localization management based on the provided requirements and the current state of an environment. QStorMan is a part of the Framework for Inteligent Virtual Organization (FiVO) 30 FiVO/QStorMan toolkit - architecture and components Components of the toolkit: Portal – a graphical user interface for defining requirements for storage systems GOM (Grid Organizational Memory) – a knowledge base for storing semantic description of the storage systems along with defined requirements SMED – a knowledge-supported monitoring system SES (Storage Element Selection) library – a programming library for selecting storage resource based on provided requirements and current state of the environment SE – Storage Element 31 QCG – QosCosGrid Middleware Description: The QosCosGrid (QCG) middleware is an integrated e-infrastructure offering advanced job and resource management capabilities to deliver to end-users supercomputer-like performance and structure By connecting many computing clusters together, QosCosGrid offers easy-touse mapping, execution and monitoring capabilities for variety of applications, such as parameter sweep, workflows, MPI or hybrid MPIOpenMP QCG architecture 32 QCG – QosCosGrid Middleware (cont’d) Benefits: for End Users – efficient and secure access to dynamically discovered computational resources located in the PL-Grid environment for Cluster and Grid Administrators – a great opportunity to share and use their clusters more efficiently by scheduling local and Grid end users’ applications according to defined policies and constraints Several portal based and mobile applications have been developed and integrated with QCG infrastructure to give the users intuitive - experiment and domain oriented - access to resources. QCG mobile client 33 Vine Toolkit – Distributed Resources at Your Fingertips for Developers and End Users General description: Vine Toolkit was designed as an environment to facilitate the development and integration of web-based applications with HPC resources, Grid services and various existing large-scale computing infrastructures managed by Grid middleware, such as gLite, Unicore, Globus, QosCosGrid and GRIA. It is possible to create a computational web-based Science Gateway using a modular structure and existing Vine components. Main features: Single Sign-On – users log in once and use all resources Support for different grid infrastructures Advanced web applications (integration with Adobe Flex / BazeDS technology) Extensible, module based open architecture based on plugins Many applications included „out-of-the-box” such as: Job Manager, File Manager, Resource Manager, Credential Manager, Certificate Manager, GSI-SSHTerm applet Support for scientific applications: ABINIT, Quantum Espresso (and others…) Integration with portal environments Liferay and GridSphere installed on Tomcat server 34 Vine Toolkit – Distributed Resources at Your Fingertips for Developers and End Users (cont’d) For whom: web applications developers – a base for building advanced computational scientific portal gateways and applications with many ready to use components, integration layer for portal frameworks and advanced installation engine administrators of grid infrastructures – possibility of deployment of a web portal “out-of-the-box” to access existing grid infrastructure end users – many ready to use web applications for grid computations and data management in grid infrastructure scientists – dedicated web applications for scientific applications for different scientific domains Example of the advanced web interface in the Nano-Science Gateway based on Vine Toolkit 35 Scientific software packages Access to software packages is provided to users through: gLite UNICORE Examples of available packages in various fields: biology: AutoDock, BLAST, ClustalW2, CPMD, Gromacs, NAMD quantum chemistry: ADF, CFOUR, Dalton, GAMESS, Gaussian, Molcas, Molpro, MOPAC, NWChem, OpenBabel, Siesta, TURBOMOLE physics: ANSYS FLUENT, Meep numerical computations and simulation: Mathematica, MATLAB other: Blender, POV-Ray Users may report us their expectations through a survey available at: http://www.plgrid.pl/ankieta The system of testing the software packages in the grid environment has been prepared and deployed The correctness of functioning of the packages is monitored automatically in each of the centers 36 Training and Users’ Support Basic training on access to the PL-Grid infrastructure through gLite and UNICORE conducted in all centers participating in the project – in Gdańsk / Kraków / Poznań / Warszawa / Wrocław More advanced training started Similar (free) training may be conducted in other centers, if necessary eLearning courses offered through the Blackboard system ( available for registered users of the PL-Grid infrastructure) Helpdesk system implemented it’s a novel support system for people using the Project resources it involves the technical support and organization of the current users’ support by the experts (maintenance of trouble tickets) tickets in Helpdesk may be created by sending an email to: helpdesk@plgrid.pl or through the online system, available at: https://helpdesk.plgrid.pl 37 Security in PL-Grid Provision of two CAs – PKI certification centers – for grid users Project and implementation of the SimpleCA system, facilitating the users obtaining PKI certificates and their usage Project and implementation of the secure configuration of the infrastructure, in conformity with the most actual security standards Project of the system monitoring the conformance of the configuration deployed in the centers with the security policy Creation of the group of experts from the field of security, in order to continuously monitor the environment, immediate react on incidents and support users and administrators Prototype version of the system of correlation of information about the attacks on the infrastructure (ACARM-ng) Audits of applications crucial for grid security 38 Conclusions Good and promising partial results of the PL-Grid Project Justified expectations to Project completion according to its goals and requirements Generic services and generic applications in development Futher developement needed, as identified currently, mainly on Domain Specific Grids Request from the users’ communities Capacity for organization of future development according to Expertise and experience Strong scientific potential of the users’ communities being represented by PL-Grid Partners Wide international cooperation concerning the Consortium and individual Partners, good recognition worldwide Good managerial capacity 39 http://www.plgrid.pl/en 40