S316103 : Life with the Sun Oracle Database Machine v2 TURKCELL’s EXADATA Journey TURKCELL DBAs Metin YILMAZ & Ferhat ŞENGÖNÜL ORACLE RACK Pack Team Sanjay SINGH September 2010 Turkcell Bio • • • • • Leading Telco in Turkey 33.9 million subscribers in Turkey as of June 30, 2010 Third biggest GSM operator in Europe in terms of subscribers Turkcell started its operations in February 1994 Turkcell co-operates in 8 other countries and has over 60 million subscribers total • First and only Turkish company listed on NYSE • Over 500 Oracle Databases, 150+ Production • DB Machine hosts our biggest Database from DWH Domain Overview • Turkcell’s largest 100 TB (~250 TB uncompressed) DB was migrated to DBM v2, now only 25 TB with the help of HCC on Full SAS Rack • There are over 50K reports run every month on this DB, Performance improvement up-to 400x, average 10x • Ability to scale linearly: add one more Rack and forget Business Drivers - Why Exadata ? EMC DMX-4 70 TB HITACHI USP-V 50 TB Server Model Cpu Type Number Of Cpu Threads Total Main Memory Total Storage Capacity Storage Connection Technology Storage Maximum IO Throughput Capacity Server + Storage units Total Power Server + Storage units Total Form Factor Approximate Data Backup Duration Number of Backup Tape Cartridges Per Backup Sun M9000 Sparc 7 176 Threads ORACLE Exadata V2 OLD SYSTEM Sun M9000 Sun Sparc 7 2.52 GHz 176 512 GB 120 TB Fiber Channel (32 x 4 Gtbit/s) 5 GB/s 57 KVA NEW SYSTEM Oracle Exadata V2 Xeon® E5540 Processors 2.53 GHz 128 576 GB 30 TB Infiniband ( 8 x 40 Gbit 21 GB/s 20 KVA 11 Rack 1 Rack 44 Hours 159 14 Hours 57 Business Drivers - Why Exadata ? Simplified Architecture • • Full Rack replaces Sun M9000 & ten(10) Storage Cabinets Single Vendor Strategy • It took Turkcell a few years to perfect current environment – Exadata up & running in 5 days • We need to address/deal with one party if a problem occurs Effortless Scalability: one to Multi-Rack • • • Data volume explodes – data size doubles every year (45TB to 100TB) Old storage environment was maxed out – no chance to scale out Old env 600+ (EMC DMX4 & Hitatchi USP-V) disks with max I/O throughput of 5 GB/sec. vs 168 SAS disks with a max I/O throughput of 21 GB/sec. 0 0-5 dk 5-10 dk 10-30 dk 30-45 dk 45-60 dk 1-1.5 saat 1.5-2 saat 2-2.5 saat 2.5-3 saat 3-3.5 saat 3.5- 4 saat 4 saat… 0-5 dk 5-10 dk 10-30 dk 30-45 dk 45-60 dk 1-1.5 saat 1.5-2 saat 2-2.5 saat 2.5-3 saat 3-3.5 saat 3.5- 4 saat 4 saat… 0-5 dk 5-10 dk 10-30 dk 30-45 dk 45-60 dk 1-1.5 saat 1.5-2 saat 2-2.5 saat 2.5-3 saat 3-3.5 saat 3.5- 4 saat 4 saat… 0-5 dk 5-10 dk 10-30 dk 30-45 dk 45-60 dk 1-1.5 saat 1.5-2 saat 2-2.5 saat 2.5-3 saat 3-3.5 saat 3.5- 4 saat 4 saat… 0-5 dk 5-10 dk 10-30 dk 30-45 dk 45-60 dk 1-1.5 saat 1.5-2 saat 2-2.5 saat 2.5-3 saat 3-3.5 saat 0-5 dk 5-10 dk 10-30 dk 30-45 dk 45-60 dk 1-1.5 saat 1.5-2 saat 2-2.5 saat 2.5-3 saat 3.5- 4 saat Business Drivers - Why Exadata ? 6000 Avg Time 3000 2486 2000 1000 • • • • Exadata Migration 5000 Weekly Report Count & Run-time 4000 2944 4827 4750 3284 2227 25 min 27 min 7 min 3 min 3 min 1403 25 min 1161 1023 665 489 648 541 361 192 57 454 376 346 314 271 260 2523519 199342214 3 2 1 3 419 193 179 107 34282588 177 1394049332987 1125245351775 883328 9 5 2 4 0 6 4 2 1 1 1 Performance Needs There over 50K reports run every month on this DB Average report run time is reduced from 27 minutes to 3 minutes !!! Reports completed less than 5 mins went from %45 to %90 Reports running more than 4 hours went from 87 to 1 The Project Overview Planned as simple migration • No Application version or interface changes • Migrate to new Exadata V2 machine • Upgrade to new Oracle Version (from 10.2.0.4 to 11.2.0.1) • Move from Single Instance to RAC 4 Months (Apr-July) testing of all the components • Crash Tests (working closely with support) • Network Transfer Speed Tests (10G vs IB) • Backup/Restore Tests • Performance Tests (RAT, Report run times, data load times) Turkcell’s BI Environment Exadata Source DBs ETL AB Initio Oracle ODI Amount Of Data • • • • • Reporting MSTR Other DWH DBs 1.5 Billion CDR per Day 600 -1000 GB raw data extracted from 20+ source database 5 TB Data on file system processed 2-3 TB loaded into databases, mostly into Exadata Approximately 500 TB Customer data stored in multiple Databases • 250 TB (25 TB compressed) on Exadata, 250 TB (100 TB compressed) on non-Exadata Turkcell’s BI Environment – Application Footprint ETL Environment • AB Initio – runs on Sun E25K and 8 Node Linux Cluster • Data loaded daily between 19:00 - 08:00 (SLA) • Sql*Loader used to load tables with 16 parallel threads Reporting Environment • • • • MSTR (Microstrategy) is used mostly 300+ Users 3000 distinct reports with 50k run per month Runs between 08:00 – 20:00 (SLA) Project Challenges • Will we fit into 30 TB ? (~100 TB 10g compressed) • How to move that much data in 2 days ? • 100TB 10g compressed, how much of it can be moved before/after • How much data needs to be moved during mig window • What kind of Network infrastructure is needed to support such xfr rate • 8-node RAC ? (earlier attempts to use RAC did not go through) • Training needs for RAC and exadata storage architecture • Rollback plan: Parallel run (load data to both DB) Environment Migration Facts • Insert/append over DB Links • Platform and version change forced us to use insert over db-link • None of the other methods like TTS, ASM Rebalance was applicable • Used in-house pl/sql utility to perform migration • Metadata (Tablespaces were re-organized) • Insert/Append with 128 parallel sessions partition-wise • 40TB data was transferred during the migration window • • • • Transfer rate 300mB/sec, around 1 TB per hour approximately Completed in 36 hours. High CPU load on comp nodes because of HCC (expected) SQL*Net compressing the data 2-3x (this was a nice finding) • 20 TB readonly data was transferred before, ~30TB after • Network Infrastructure • 10Gb-to-IB (Voltaire 4036E) gateway used to connect old system Migration Facts • After the test migration we ended-up staying on Exadata • End users were so happy, they did not let us go back • 3 Weeks before the go-live date we were live • Parallel runs continued for few weeks until we completely feel comfortable on Exadata • Stability of the system under real load was proved for various load patterns • Backup/Restore tests were completed • Questions on Migration ? Compression in Action Old System 10gR2 Compression • ~2-3 times ~250TB raw data to 100TB Exadata V2 with EHCC • Raw Data 250TB to 25TB (Data)+ 5TB (Temp) = 30TB • EHCC - Compress ratio ~7-10x • Archive compression is efficient but high CPU consumption SORT COMPRESS SIZE (GB) RATIO NOSORT NOCOMP 137,59 1 NOSORT Q_HIGH 21,21 6,48 SORT_A Q_HIGH 12,18 11,29 SORT_B Q_HIGH 15,37 8,95 SORT_A_B Q_HIGH 11,64 11,80 http://ferhatsengonul.wordpress.com/2010/08/09/getting-the-most-from-hybrid-columnar-compression/ Performance Gains Report Name Old System Exadata by…X CRC Control Report 0:15:48.73 0:05:06.07 X2 prepaid bireysel toplam harcanan kontor tl 8:02:10.59 1:51:33.20 X4.3 eom_equipment_utilization report 0:38:17.77 0:00:23.34 X163 Eom equipment arpu report 0:09:46.25 0:02:08.00 X4.5 Son 10 gun icinde yapilan gonderimler 0:17:57.95 0:00:37.61 X45.7 Rapor 01. Satış Kanalı Kırılımında Blackberry 0:03:22.24 0:00:00.66 X487 İnternet Paketleri Postpaid paketiçi ve paket 0:05:41.34 0:01:00.34 X4.3 Connectcard Fatura tutarları 0:31:32.38 0:00:46.51 X66.3 Connectcard Aktivasyon 0:25:21.00 0:00:54.88 X44.9 • There over 50K reports run every month • Performance improvement is up-to 400x for some reports and on average it is 10x Life with Exadata No need to be scared about RAC • • • • We use scan-IP, rest is assured: loads are distributed over the nodes We use heavily PQ; so far no problems running slaves over multiple nodes. Tons of log files could be a little overwhelming for troubleshooting Temporary Tablespace should be at least twice as more DB Machine is indeed redundant • We lost entire cell or individual hard disk or flash disk but database stayed up, no interruptions for the service Enterprise Manager is must • Grid Control • SQL Monitor Implementation Challenges Hardware Failure Sun: issues with failed IB switch - all node hung. • Faulty switch replaced • Hanging due to configuration setup of switch Disk Failure – no down time , everything continuing working • Disk replaced online – just replaced all taken care automaticaly Issues work in progress • • Cannot explain why cell encountered “root filesystem readonly” • Bug: 9752748 Not all the Instance’s are registered with the scan-vip after re-cycle listeners • Registers after a few hours. • No impact when partly registered. Implementation Challenges SQL Loader session hangs on enq ST contention • Bug9772225 – “ST enqueue acquisition issue” - Complex fix in review • Bug9326041 - “Buffer pre-warm issue” W/A: _db_cache_pre_warm=FALSE Exadata problem ORA-07445: core dump [kdzdcol_skip_slots()] • Bug 9655552 – EHCC optimization issues – W/A : _arch_comp_dbg_scan"=16 RS-7445 [Serv MS absent] • Bug 9252694 MS FREQUENTLY REPORTS LOW VIRTUAL HEAP MEMORY AND SHUTS DOWN - Fixed 11.2.1.3.X • bug 9217088 : using memcpy (on cell nodes) • Disable chained row optimization in Exadata • bug.9338087 is fixed in 11.2.1.2.4 User Feedbacks • We heard before that infrastructure changes would give us performance gains but this time we were surprised with it, it was well over our expectations. Now we can take faster actions in this competitive environment. Director of Marketing Department • XDDS is fantastic in a single word, none of the reports take more than 10 minutes, It was taking 3-4 hours before now it completes in 3 minutes. It sounds like un-real but it is real. Power end-user from Finance Department • It was a never ending race to match the business' performance and capacity needs. With the Database Machine V2, we have outperformed our user's expectations and we are prepared for the future growth. Veteran System Admin • “You started to scare me MSTR” updated her status on facebook. End-user from Marketing Department What is waiting for the future • Adding second Rack • We’ll see if it will scale :) • Partition nodes for test/dev purposes • Several more DWH Databases could be consolidated into another full Rack • Evaluate Exadata for OLTP Environments • Real-Time DWH I would like to Thank here to • Turkcell’s DB and OS Admins for completing migration before it’s planned date and keeping it running successfully • Oracle RAC Pack Team: Sanjay, Su, Martina • Oracle DWH Global Leaders Team: Cemil, Reiner • Oracle Turkey’s local Sales Support and Support Organization Thank YOU Questions Ferhat ŞENGÖNÜL ferhat.sengonul@turkcell.com.tr www.turkcell.com.tr