La Investigación generadora de riqueza WINCO`05 México D.F., México, 13 de Abril de 2005 Prof. Mateo Valero, UPC, Barcelona Outline High Performance Research Group at UPC Centers of Supercomputing: CEPBA CIRI BSC Future context in Europe Networks of Excellence Large Scale Facilities 2 Erkki Liikanen Brussels, 21 April 2004 3 Erkki Liikanen Brussels, 21 April 2004 4 Erkki Liikanen Brussels, 21 April 2004 5 Basic concepts about research Fundamental/basic research versus applied research Good versus bad research Good research produces always wealth Products of good research: Short/Medium/Long Term Research Papers and patents Educated people Good education is a key component in this picture Cross-Pollination between Research groups and companies are the other part of the movie To promote good research is the only way Europe has to be competitive in a short/long future 6 Historia Tesis: 1974-1980 FIB Mucha dificultad Crear departamento: asignaturas, contratar,.. Empezar a investigar Situación española… no hay $, no existe nada,… CICYT Decisiones Estratégicas: Arquitectura de Computadores Supercomputadores 7 Computer Architecture Computer Architecture is a rapidly changing subject Computer Architects must deal with both technology and applications Technology changes very fast New applications emerge continuously CMOS Technology is coming to an end A new group of applications is appearing There is a great opportunity for high performance architectures for these applications 8 Supercomputers Faster computers in the world Used to simulate Mainly fabricated by USA companies No experience in Spain Europe uses and produces software 9 Entrada de España en la EU Internacionalización de la Investigación Nuevas oportunidades Antes y despues de 1986 Proyectos industriales Usamos estos proyectos para crecer en investigación básica 10 CICYT: Spanish Projects HLL-oriented Architectures Parallel Architectures for Symbolic Computation TIC89-300 TIC91-1036 Architecture, Tools and Operating Systems for Multiprocessors Microkernel/applications Cooperation in Multiprocessor Systems TIC89-392 TIC94-439 Parallelism Exploitation in High Speed Architectures Architectures and Compilers for Supercomputers High Performance Computing High Performance Computing II TIC89-299 TIC92-880 TIC95-429 TIC98-511-C02-01 UdZ, URV and ULPGC High Performance Computing III High Performance Computing IV TIC2001-995-C02-01 TIC2004-7739-C02- UVall 1990 1991 1992 1993 1994 1995 1996 1997 1998 1999 2000 2001 2002 2003 2004 11 People: large research group Full 6 19 Associate PhD 25 Associate no PhD Assist PhD 79 researchers 34 PhD 45 doing PhD Assist no PhD 21 3 5 PhD student Now 12 Research topics and target platforms Task T1: Computer architecture T2: Compilers, execution environments and tools T3: Algorithms and applications Subtask Target architecture T1.1: Processor microarchitecture SP T1.2: Memory hierarchy SP and small MP T1.3: Code generation and optimization SP T2.1: Sequential code optimization SP T2.2: OpenMP All MP T2.3: Extensible execution environments All MP and GRID T2.4: Tools All MP and GRID T3.1: Numerical applications All of them T3.2: Non-numerical applications SP and small MP T3.3: Distributed applications All MP and GRID 13 Computer architecture (uniprocessor) Dynamically scheduled (superscalar) Front-end engines: instruction fetch mechanisms and branch predictors Speculative execution: data and address prediction Organization of resources: • Register file, functional units, Cache organization, prefetching … Kilo-instruction Processors Not only performance: area and power consumption 14 Computer architecture (uniprocessor) Statically scheduled (VLIW) Organization of resources (functional units and registers) Not only performance: area and power consumption Advanced vector and multimedia architectures Vector units for superscalar and VLIW architectures Memory organization Data and instruction level parallelism 15 Computer architecture (multiprocessor) Multithreaded (hyperthreaded) architectures Front-end engines and trace caches Speculation at different levels Dynamic management of thread priorities Shared-memory multiprocessors Memory support for speculative parallelization (hardware and software) 16 System software (multiprocessor) OpenMP compilation Proposal of language extensions to the standard Compiler technology for OpenMP OpenMP runtime systems Parallel library for OpenMP (SGI Origin, IBM SP2, Intel hyperthreaded, …) Software DSM (Distributed Shared Memory) Intelligent runtimes: load balancing, data movement, … 17 System software (multiprocessor) Scalability of MPI IBM BG/L with 64K processors Prediction of messages Programming models for the GRID Grid superscalar 18 Algorithms and applications Solvers for linear equations systems Out-of-core kernels and applications STORM Metacomputing tool for performing stochastic simulations Data bases Sorting, communication in join operations Query optimization Memory management in OLTP 19 PhD programs Research topics are part of our PhD programs Computer Architecture and Technology at UPC Quality award: MCD2003-126 42% total number of credits, 2003-04 30 PhD in the last 5 years (66% of whole program) 20 International collaboration Industry University and research laboratories Intel, IBM Watson, Toronto, Haifa and Germany Labs (*), Hewlett-Packard, STMicroelectronics, SGI, Cray, Compaq Univ. of Illinois at Urbana-Champaign, WisconsinMadison, California at Irvine, William and Mary, Delft, KTH, … (more than 60) Major research laboratories in USA: NASA Ames, San Diego (SDSC), Lawrence Livermore (LLNL), … Standardization committees OpenMP Futures in the Architecture Review Board (ARB) … with joint publications and developments 21 (*) Part of the CEPBA-IBM Research Institute research agreement International collaboration Pre- and post doctoral short/medium/long stays Industry: Intel, SUN, IBM, Hewlett-Packard, ... University: Univ. of Illinois at Urbana-Champaign, Univ. of California at Irvine, Univ. of Michigan, ... Visiting professors and researchers More than 70 talks in our weekly seminar (last two years, external researchers) http://research.ac.upc.es/HPCseminar/ PhD courses: • 2001-02: 5 courses (135 hours) • 2002-03: 4 courses (110 hours) • 2003-04: 6 courses (165 hours) 22 Some of the seminar guests Krste Asanovic (MIT) Venkata Krishnan (Compaq-DEC) Trevor Mudge (U. Michigan) Jim E. Smith (U. Wisconsin) Luiz A. Barroso (WRL) Josh Fisher (HP Labs) Michel Dubois (USC) Ronny Ronnen (Intel, Haifa) Josep Torrellas (UIUC) Per Stenstrom (U. Gothenburg) Wen-mei Hwu (UIUC) Jim Dehnert (Transmeta) Fred Pollack (Intel) Sanjay Patel (UIUC) Daniel Tabak (George Mason U.) Walid Najjar (Riverside) Paolo Faboroschi (HP Labs) Eduardo Sánchez (EPFL) Guri Sohi (U. Michigan) Jean-Loup Baer (Washington Uni.) Miron Livny (U. Wisconsin) Tomas Sterling (NASA JPL) Maurice V. Wilkes (AT&T Labs) Theo Ungerer (Karlsruhe) Mario Nemirovsky (Xstreamlogic) Gordon Bell (Microsoft) Timothy Pinkston (U.S.C.) Walid Najjar (Riverside) Roberto Moreno (ULPGC) Kazuki Joe (Nara Women U.) Alex Veidenbaum (Irvine) G.R. Gao (U. Delaware) Ricardo Baeza (U.de Chile,Santiago) Gabby M. Silberman (CAS-IBM) Sally A. McKee (U. Utah) Evelyn Duesterwald (HP-Labs) Yale Patt (Austin) Burton Smith (Tera) Doug Carmean (Intel, Oregon) David Baker (BOPS) 23 International collaboration Mobility programs (*) Access: Transnational Access for Researchers (2000-03) Access-2: Transnational Access for Researchers (200204) Networks of Excellence HIPEAC: High-Performance Embedded Architectures and Compilers, in evaluation (www.hipeac.org) CoreGRID (*): middleware for GRID, in evaluation 24 (*) Projects of the European Center for Parallelism of Barcelona Industrial technology transfer European projects (IST and FET) Attract international companies to establish branches or laboratories in Barcelona (*) INTONE: Innovative OpenMP Tools for Non-experts (2000-03) DAMIEN: Distributed Applications and Middleware for Industrial use of European Networks (2001-03) POP: Performance Portability of OpenMP (2001-04) Antitesys: A Networked Training Initiative for Embedded Systems Design (2002-04) EASi Engineering: S. Girona (*) Intel Labs: R. Espasa (*) and T. Juan (*), A. Gonzalez (*), Hewlett-Packard Labs Professors of the Computer Architecture Department (full or part time dedication) 25 Industrial Relationships Compaq HP • • • Roger Espasa (VSSAD) Toni Juan (VSSAD) Marta Jimenez (VSSAD) • • Jesus Corbal (VSSAD) Alex Ramirez (WRL) • BSSAD IBM Sabbaticals Interns Partnerships Intel Sabbaticals • Josep Llosa (Cambridge) • • Daniel Ortega Javier Zalamea • • Software Prefetching Two-Level Register File Interns Parnerships Sun Microsystems Interns • • • • Pedro Marcuello Ramon Canal Esther Salami Manel Fernande Microsoft Interns • Xavi Serrano (CAS) • Daniel Jimenez (CAS) • 3 more people in 2001 Parnerships • Supercomputing (CIRI) • Low Power • Databases • Binary Translation Faculty Awards Interns • Adrian Cristal (Haifa) • Alex Ramirez (MRL) • Pedro Marcuello (MRL) Parnerships • Semantic Gap • Smart Registers • Memory Architecture for Multithreaded Processors • Speculative Vector Processors Labs in Barcelona MRL and BSSAD Advisory Board of MRL Xstream, Flowstorm, Kambaya Advising committee ST- Microelectronics Analog Devices 26 Conclusions Large (90) team with experience in many topics Good production Computer architecture System software Algorithms and applications >100 PhD thesis Publications in top conferences (>400) and journals (>150) Prototypes (3) used in research laboratories 25 professionals in industry Long track of international collaborations Academic Industrial 27 Outline High Performance Research Group at UPC Centers of Supercomputing: Future context in Europe CEPBA CIRI BSC Networks of Excellence Large Scale Facilities Conclusions 28 Centro Europeo de Paralelismo de Barcelona CEPBA CEPBA Depts. CER CEPBA DAC • HPC experience • October 1991 • R+D on parallelism • Training RME, LSI, FEN, FA • Computing needs •Technology transfer • European context 30 CEPBA Activities Technological expert • Service Developments • Training • Technology transfer – T.T. Management –R&D –24 proyectos 31 Service • Access • Users support IBM 64 Power3 Net. SMP 32GB Mem 400 GB Disk COMPAQ Parsytec 16 Pentium II 2CPUs nodes 3 Networks Fast Ethernet Myrinet HS link 1GB Mem 30 GB Disk 12 alpha 21164 SMP 2GB Mem 32 GB Disk SGI 64 R10000 CC-NUMA 8GB Mem 360 GB Disk 32 European Mobility Programs Joint CEPBA - CESCA projects Stays and access to resources Proyect Period Funding Visitors UPC HCM 1993-1997 950 112 49 PECO 1995-1996 160 17 15 TMR 1996-2000 935 133 46 IHP 2000-2003 700 28 8 2745 290 118 Total 33 Technology Transfer 1986... 1991... 1994 1995 1996 1997 1998 1999 2000... 23 Projects R+D projects CEPBA Technology Transfer Management 3 cluster projects 28 Subprojects Technical management & Dissemination Technological partner & developments 34 R&D Projects 92 93 94 95 Supernode II 96 97 98 99 00 01 Intone Dimemas & Paraver Sep-tools Permpar Parmat Tools Damien Asra Parallelization Sloegat Identify Hipsid Metacomputing Promenvir Promenvir+ ST-ORM BMW Phase DDT Bonanova Apparc Nanos System 35 T.T. Management 1994 1995 1996 1997 1998 1999 2000 PCI-PACOS 35 Proposals 28 projects PACOS 2 PCI-II CEPBA-TTN Promote proposals to EC Technical management of projects Dissemination 36 PCI-PACOS Ayto. Barcelona Uitesa UPC-EIO Hesperia Neosystems UPC-EIO AMES CIMNE TGI UPM-DATSI Iberdrola Uitesa UPV Metodos Cuantitativos Gonfiesa CESCA CESGA AZTI UPC-LIM Tecnatom UMA 37 PCI-II Ospedali Galliera Le Molinette Parsytec PAC EDS Italeco Geospace Intecs Univ. Leiden Ferrari Genias P3C ENEL EDF CSR4 Reiter Kenijoki CANDEMAT CIMNE CEPBA-UPC Volkswagen Ricardo PAC Intera SP Intera UK UPC-DIT CEPBA-UPC Inisel Espacio Infocarto UPC-TSC CEPBA-UPC Cari Verona AIS PAC Univ. Cat. Milan Cristaleria Española UNICAN CEPBA-UPC 38 HPCN TTN network 39 CEPBA-TTN INDO CEPBA-UPC Soler y Palau CIMNE CEPBA-UPC Iberdrola SAGE CEPBA-UPC Torres Software Greenhouse CEPBA-UPC CEPBA CESCA UMA UNICAN UPM BCN COSIVER Mides UPC-EIO CASA Envision GTD Intespace RUS ST Mecanica DERBI AUSA CEPBA-UPC CEBAL-ENTEC NEOSYSTEMS SENER CIC UNICAN 40 References: Technology promotion European Comission AMES AUSA INDO Ayto. BCN BCN COSIVER CEBAL-ENTEC DERBI HESPERIA Métodos Cuantitativos Mides NEOSYSTEMS QuantiSci SL Software Greenhouse Soler y Palau ST Mecánica Torres AIS CariVerona EDF-LNH EDS Italy ENEL-CRIS Ferrari Auto Genias Geospace Intecs Sistemi Intespace Italeco Kemijoki Le Molinette Ospedali Galliera Parsytec QuantiSci LTD Reiter Ricardo Volswagen CESCA CESGA CIMNE CRS4 P3C PAC RUS Catholic Univ. Milan Politecnico di Milano UNICAN UPM UMA Univ. of Leiden UPC-DIT UPC-EIO UPC-LIM UPC-OE UPC-RME UPC-TSC IPM-DATSI UPV AZTI Candemat CASA CIC Cristaleria Española Envison Gonfiesa GTD Iberdrola Indra Espacio Infocarto SAGE SENER Tecnatom TGI Uitesa 41 Budget (PACOS,PCI-II,TTN) Managed ESPRIT Funding: 11.8 Mecus (total) Distribución Gestión CEPBA Técnico CEPBA Resto UPC Resto CAT Resto SP Resto EU 42 CEPBA-IBM Research Institute 43 CIRI’s mission CEPBA-IBM Research Institute (CIRI) was a research and development partnership between UPC and IBM. •Established in October 2000 with an initial commitment of four years. •Its mission was to contribute to the community through R&D in Information Technology. •Its objectives are: Research & Development External R&D Support Technology Transfer Education •CEPBA (European Center for Parallelism Barcelona) was a deep computing research center, at the Technical University of Catalonia (UPC), which was created in 1991. 44 Organization Management and Technical Boards evaluate project performance and achievements and recommend future directions. 70 people were collaborating with the Institute: Board of Directors (4) Institute’s Professors (10) Associate Professors (9) Researchers (5) PhD Students (21) Graduate Students (3) Undergraduate Students (18) 45 Introduction: CIRI areas of interest Deep Computing Computer Architecture Performance Tools: Numerical Codes Web Application Servers Parallel Programming Grid Code Optimization Vector Processors Network Processors Data Bases Performance Optimization DB2 Development & Testing 46 CIRI R&D Philosophy Technology Web Metacomputing •ST-ORM Applications •Steel stamping •Structural analysis •MGPOM •MPIRE Computer Architecture OpenMP •OMPTrace •activity •hw counters •Nested Parallelism •Precedences •Indirect access MPI •UTE2paraver •OMPITrace Performance Visualization Paraver Dimemas •Collective, mapping •Scheduling System scheduling RunTime Scheduling •Dynamic load balancing •Page migration •Self analysis •Performance Driven Proc. Alloc. •Process and memory control 47 Barcelona Supercomputing Center Centro Nacional de Supercomputación Professor Mateo Valero Director High Performance Computing Aircraft, Automobile Design Experiment Climate and Weather Modeling Theory Fusion Reactor, Accelerator Design, Material Science, Astrophysics Computing & Simulation What drives HPC ? “The Need for Speed...” Computational Needs of Technical, Scientific, Digital Media and Business Applications Approach or Exceed the Petaflops/s Range CFD Wing Simulation 512x64x256 Grid (8.3 x10e6 mesh points) 5000 FLOPS per mesh point, 5000 time steps/cycles 2.15x10e14 FLOPS Source: A. Jameson, et al Materials Science Magnetic Material: Current: 2000 atoms; 2.64 TF/s, 512 GB Future: HDD Simulation - 30 TF/s, 2 TB Electronic Structures: Current: 300 atoms; 0.5 TF/s, 100 GB Future: 3000 atoms; 50 TF/s, 2TB CFD Full Plane Simulation 512x64x256 Grid (3.5 x10e17 mesh points) 5000 FLOPS per mesh point, 5000 time steps/cycles 8.7x10e24 FLOPS Source: D. Balley, NERSC Spare Parts Inventory Planning Digital Movies and Special Effects ~1e14 FLOPs per frame 50 frames/sec 90 minute movie - 2.7e19 FLOPs ~ 150 days on 2000 1GFLOP/s CPUs Source: Pixar Source: B. Dietrich, IBM Modelling the optimized deployment of 10000 part numbers across 100 part depots and requries: - 2x10e14 FLOP/s (12 hours on 10, 650 MHz CPUs) - 2.4 PetaFlop/s sust. performance (1 hour turn-around time) Industry trend to rapid, frequent modeling for timely business decision support driver higher sustained performance Applications for Supercomputers Aircraft/car simulations Life Science (Proteins, Human Organs,…) Atmosphere Stars Nanomaterials Drugs Regional/Global Climate/Weather/Pollution High Energy Physics Combustion Image Processing Suitable applications for massively parallel systems Source: Rick Stevens, Argonne National Lab and The University of Chicago Motivation Significant contribution to advancement of Science in Spain, enabling supercomputing capacity, scientific-technical synergies, and cost rationalization thanks to economies of scale Powerful tool to assist research and development centers, public and private, generating impulses for a new technological environment Mission ”Investigate, develop and manage technology to facilitate the advancement of science” Objectives Research in Supercomputing and Computer Architecture Collaborate in R&D e-Science projects with prestigious scientific teams Manage BSC supercomputers to accelerate relevant contributions to research areas where intensive computing is an enabling technology Organización del BSC Presidencia CONSEJO RECTOR PATRONATO Comisión Ejecutiva Comisión Científica Asesora Director Director Asociado Comité de Acceso Dirección de Tecnologías de la Información Dirección de Biomedicina y Ciencias de la Vida Dirección de Ciencias de la Tierra Dirección de Astronomia y Espacio Dirección de Química y Ciencia de los Materiales Dirección de Física e Ingeniería Dirección de Operaciones Dirección de Desarrollo de Negocio I+D en I.T. e-Ciencia Gestión 57 Dirección Administración y Finanzas IT research and development projects Continuation of CEPBA (European Center for Parallelism in Barcelona) research lines Deep Computing Performance Tools Parallel programing Grid Code Optimization Computer Architecture Vector processors Network processors e-Science projects R&D collaborations Computational Biology Computational Chemistry Computational Physics Information based Medicine Management projects Supercomputer Management System Administration Users support Business Development External Relations Promotion Technology Transfer Education Administration Accounting and Finances Human Resources MareNostrum PowerPC 970 FX processors (dual processors) 4GB ECC 333 DDR memory per node 3 networks Myrinet Gigabit 10/00 Ethernet Diskless network support Linux cluster MareNostrum: System description 27 Compute Racks (RC01-RC27) • 162 BC chassis w/OPM and gigabit ether switch • 2268 JS20+ nodes w/myrinet daughter card 4 Myrinet Racks (RM01-RM04) • 10 clos256+256 myrinet switches • 2 Myrinet spines 1280s 7 Storage Server Racks (RS01-RS07) • 40 p615 storage servers 6/rack • 20 FastT 100 3/rack • 20 EXP100 3/rack 1 Gigabit Network Racks • 1 Force10 E600 for Gb network • 4 Cisco 3550 48-port for 10/100 network 1 Operations Rack (RH01) • 7316-TF3 display • 2 p615 mgmt nodes • 2 HMC model 7315-CR2 • 3 Remote Async Nodes • 3 Cisco 3550 • 1 BC chassis (BCIO) Processor: PowerPC 970FX Blades, blade center and racks JS20 Processor Blade • 2-way 2.2 GHz Power PC 970 SMP • 4GB memory (512KB L2 cache) • Local IDE drive (40 GB) • 2x1Gb Ethernet on board 6 chassis in a rack (42U) • Myrinet daughter card • 168 processors • 336GB memory Blade Center • 14 blades per chassis (7U) • 28 processors • 56GB memory • Gigabit ethernet switch MareNostrum: Floor plan with cabling 27 racks + 1 BC chassis • 4564 processors • 9TB memory Blade centers Storage servers Gigabit switch Myrinet racks Operations rack 10/100 switches Blade centers Storage servers Gigabit switch Myrinet racks Operations rack 10/100 switches 256 blades connected to 1 clos 256+256 Myrinet 1280 blades connected to 5 clos 256+256 Myrinet and 1 spine 1280 2282 blades connected to 10 clos 256+256 Myrinet and 2 spine 1280 20 x 7TB storage nodes Management rack, Force 10 Gigabit, 10/100 Cisco switches Blade center racks 6 Blade Centers per rack 27 racks + 1 Blade Center Cabling per rack 84 fiber cables to myrinet switch 6 Gb to Force10 E600 6 10/100 cat5 to Cisco Blade Center Blade Center Blade Center Blade Center Blade Center Blade Center Myrinet racks 10 Clos 256x256 switches 2 Spine 1280 Interconnect up to 256 Blades Connect to Spine (64 ports) Interconnect up to 10 Clos 256x256 switches Monitoring using 10/100 connection Myrinet racks 64 to Spine Clos 256x256 Spine 1280 256 Blades or Storage servers 320 from Clos Gb Subsystem: Force 10 E600 Interconnection of Blade Centers Used for system boot of every blade center 212 internal network cables 170 for blades 42 for p615 76 connection available to external connection Storage nodes Total of 20 storage nodes, 20 x 7 TBytes Each storage node 2xP615 FastT100 EXP100 P615 P615 FastT100 EXP100 Cabling per node 2 2 2 1 Myrinet Gb to Force10 E600 10/100 cat5 to Cisco Serial P615 P615 FastT100 EXP100 P615 P615 FastT100 EXP100 Management rack Contents BCIO Display 2 HMC 2 x p615 3 x Cisco 3 x 16-port Remote Async Nodes Mare Nostrum Supercluster Barcino ePOWER User Community System Gateway Hispania Data Server IBM IBM Gigabit Ethernet Interconnect Myrinet or Infiniband interconnect Tarraco Visualization Cluster Ebro Rhodanus Tiberis Nilus Three Rivers contract boundary 74 MareNostrum Outline High Performance Research Group at UPC Centers of Supercomputing: CEPBA CIRI BSC Future context in Europe Networks of Excellence Large Scale Facilities 81 Erkki Liikanen Brussels, 21 April 2004 82 Future Networks of Excellence HiPEAC InSyT European FP VII Technology Platforms Large Scale Facilities DEISA 83 HiPEAC Objectives to help companies identify and select the best architecture solutions for scaling up high-performance embedded processors in the coming years to unify and focus academic research efforts through a processor architecture and compiler research roadmap to address the increasingly slow progression of sustained processor performance by jointly developing processor architecture and compiler optimizations to explore novel approaches for achieving regular and smooth scaling up of processor performance with technology, and to explore the impact of a wide range of post-Moore's law technologies on processor architecture and programming paradigms. 84 Partners Leading partner per country Chalmers University, Sweden CNRS, France Delft University, The Netherlands Edinburgh University, UK Ghent University, Belgium INRIA, France University of Augsburg, Germany University of Patras, Greece University of Pisa, Italy UPC Barcelona Industrial STMicro, Switserland Infineon, Germany Ericsson, Sweden Virtutech, Sweden IBM Haïfa, Israel Kayser Italia, Italy Philips Research, The Netherlands 85 HIPEAC Topics Compiler optimizations Common compiler platform Processor architecture Processor performance Power-aware design Real-time systems Special purpose architectures Multithreading and multiprocessors Dynamic optimization Common simulation platform New processor paradigms 86 FET Programme FET is the IST programme nursery of novel and emerging scientific ideas. Its mission is to promote research that is of a long-term nature or involves particularly high risks, compensated by the potential of a significant societal or industrial impact. As such, FET is not constrained by the IST programme priorities but rather aims to open new possibilities and set new trends for future research programmes in Information Society Technologies. FET goals will be achieved in two ways: Via the proactive scheme: a 'top down' approach which sets the agenda for a small number of strategic areas holding particular promise for the future, and Via the open scheme: a 'roots up' approach available to a wider range of ideas 87 FET: Advanced Computing Architectures Critical mass Education High quality European PhD program Cross-pollination industry-academia Coordinate research by leading European groups Fuel industry with European students Feedback academia with real problems Avoid emigration of specialists European students recruited by European companies Attracts IT companies to Europe 88 Future Emerging Technologies Make Europe leader in microprocessor design Support advanced research in Computer architectures Advanced compilers Micro-kernel operating systems Targets 10+ year horizon 10x to 100x performance increase 10x power reduction 89 10+ year Horizon Future CMOS technology will continue to dominate the market for the next 25 years However, we must be ready for CMOS alternatives • Quantum computing, Molecular computing, … Europe is well positioned in embedded processors, applications and technology Tomorrow’s embedded processor are today’s highperformance processors Europe need to remain the leader in the future embedded domain 90 FET Topics Emerging architectures Supporting technology System on a Chip architectures (SoC) Chip Multiprocessors (CMP) Reconfigurable logic Compiler optimizations Operating system level integration Key elements Reduce design and verification effort Develop infrastructure for system and application development Reuse of system architectures for wide range of targets 91 Future Networks of Excellence HiPEAC InSyst IP projects, SCALA European FP VII Technology Platforms Large Scale Facilities DEISA 92 Muchas Gracias