Daniel Schall, Volker Höfner, Prof. Dr. Theo Härder TU Kaiserslautern Outline Energy efficiency in database sytems Multi-Core vs. Cluster WattDB Recent Current Future Work 2 Motivation More and more data Bigger servers In-memory technology Electricity Cost 3 Power Breakdown Load between 0 – 50 % Energy Consumption: 50 – 90%! ‘‘Analyzing the Energy Efficiency of a Database Server“, D. Tsirogiannis, S. Harizopoulos, and M. A. Shah SIGMOD 2010 ‘‘Distributed Computing at Multi-dimensional Scale“, Alfred Z. Spector Keynote on MIDDLEWARE 2008 4 Growth of Main Memory makes it worse Power (Watt) % power@utilization 100 80 60 40 energyproportional behavior 20 0 20 40 60 80 100 % System utilization In-memory data management assumes continuous peak loads! Energy consumption of memory linearly grows with size and dominates all other components across all levels of system utilization Mission: Energy-Efficiency! Energy cost > HW and SW cost Energy Efficiency = Work Energy Consumption ‚‚Green IT‘‘ 7 Average Server Utilization Google Servers: load at about 30 % SPH AG: load between 5 and 30 % 8 Energy Efficiency - Related Work Software Delaying queries Optimize external storage access patterns Force sleep states „Intelligent“ data placement Hardware Sleep states Optimize energy consumption when idle Select energy-efficient hardware Dynamic Voltage Scaling Narrow approaches Only small improvements 9 Goal: Energy-Proportionality Power (Watt) % power@utilization 100 80 2 60 1 40 energyproportional behavior 20 0 20 40 60 80 100 % System utilization 1) reduce idle power consumption 2) eliminate disproportional energy consumption From Multi-Core to Multi-Node Power Core CPU Core CPU Core CPU Core CPU L1Cache Cache L1Cache Cache L1Cache Cache L1Cache Cache Main L2 Cache memory Main L2 Cache memory power@utilization 100 80 60 L3 Cacheswitch 1Gb ethernet Main L2 Cache memory (Watt) % Main L2 Cache memory L1Cache Cache L1Cache Cache L1Cache Cache L1Cache Cache Core CPU Core CPU Core CPU Core CPU 40 20 0 11 20 40 60 80 100 % System utilization A dynamic cluster of wimpy nodes energy-proportional DBMS Load Time 12 Cluster Overview Light-weighted nodes, low-power hardware Each node Intel Atom D510 CPU 2 GB DRAM 80plus Gold power supply 1Gbit Ethernet interconnect 23 W (idle) - 26 W (100% CPU) 41 W (100% CPU + disks) Considered Amdahl-balanced Scale down the CPUs to the disks and network! 13 … 14 Shared Disk AND Shared Nothing Physical hardware layout: Shared Disk every node can access every page local vs. remote latency Logical implementation: Shared Nothing: data is mapped to node n:1 exclusive access transfer of control Combine the benefits of both worlds! 15 Recent Work SIGMOD 2010 Programming Contest First prototype distributed DBMS BTW 2011 Demo Track Master node powering cluster up/down acc. to load SIGMOD 2011 Demo Track Energy-proportional query processing 16 Current Work Incorporate GPU-Operators improved energy-efficiency? more tuples/Watt? Monitoring & Load Forecasting For management decisions act instead of react Energy-Proportional Storage storage needs vs. processing needs 17 Future Work Policies for powering up / down nodes Load distribution and balancing among nodes Which use cases fit for the proposed architecture, which don‘t? Alternative hardware configurations Heterogeneous HW environment SSDs, other CPUs Energy-efficient self-tuning 18 Current Work Table Partition Partition Partition Node1 Node2 Node3 19 Future Work Table Partition Partition Partition Node1 Node2 Node3 20 Conclusion Energy consumption matters! Current HW is not energy-proportional Systems most of the time at 20% - 50% utilization WattDB as a prototype for an energy-proportional DBMS Several challenges ahead 21 Thank You! Energy Proportionality on a Cluster Scale 22