CS 252 Graduate Computer Architecture Lecture 2 - Metrics and Pipelining Krste Asanovic Electrical Engineering and Computer Sciences University of California at Berkeley http://www.eecs.berkeley.edu/~krste http://inst.eecs.berkeley.edu/~cs252 Review from Last Time • Computer Architecture >> instruction sets • Computer Architecture skill sets are different – – – – 5 Quantitative principles of design Quantitative approach to design Solid interfaces that really work Technology tracking and anticipation • CS 252 to learn new skills, transition to research • Computer Science at the crossroads from sequential to parallel computing – Salvation requires innovation in many fields, including computer architecture • Opportunity for interesting and timely CS 252 projects exploring CS at the crossroads – RAMP as experimental platform 8/30 CS252-Fall’07 2 Review (continued) • Other fields often borrow ideas from architecture • Quantitative Principles of Design 1. Take Advantage of Parallelism 2. Principle of Locality 3. Focus on the Common Case 4. Amdahl’s Law 5. The Processor Performance Equation • Careful, quantitative comparisons – – – – Define, quantity, and summarize relative performance Define and quantity relative cost Define and quantity dependability Define and quantity power • Culture of anticipating and exploiting advances in technology • Culture of well-defined interfaces that are carefully implemented and thoroughly checked 8/30 CS252-Fall’07 3 Metrics used to Compare Designs • Cost – Die cost and system cost • Execution Time – average and worst-case – Latency vs. Throughput • Energy and Power – Also peak power and peak switching current • Reliability – Resiliency to electrical noise, part failure – Robustness to bad software, operator error • Maintainability – System administration costs • Compatibility – Software costs dominate 8/30 CS252-Fall’07 4 Cost of Processor • Design cost (Non-recurring Engineering Costs, NRE) – dominated by engineer-years (~$200K per engineer year) – also mask costs (exceeding $1M per spin) • Cost of die – – – – die area die yield (maturity of manufacturing process, redundancy features) cost/size of wafers die cost ~= f(die area^4) with no redundancy • Cost of packaging – number of pins (signal + power/ground pins) – power dissipation • Cost of testing – built-in test features? – logical complexity of design – choice of circuits (minimum clock rates, leakage currents, I/O drivers) Architect affects all of these 8/30 CS252-Fall’07 5 System-Level Cost Impacts • • • • Power supply and cooling Support chipset Off-chip SRAM/DRAM/ROM Off-chip peripherals 8/30 CS252-Fall’07 6 What is Performance? • Latency (or response time or execution time) – time to complete one task • Bandwidth (or throughput) – tasks completed per unit time 8/30 CS252-Fall’07 7 Definition: Performance • Performance is in units of things per sec – bigger is better • If we are primarily concerned with response time performance(x) = 1 execution_time(x) " X is n times faster than Y" means Performance(X) n = = Performance(Y) 8/30 Execution_time(Y) CS252-Fall’07 Execution_time(X) 8 Performance Guarantees Inputs C B A Execution Rate Average Rate: A > B > C Worst-case Rate: A < B < C 8/30 CS252-Fall’07 9 Types of Benchmark • Synthetic Benchmarks – Designed to have same mix of operations as real workloads, e.g., Dhrystone, Whetstone • Toy Programs – Small, easy to port. Output often known before program is run, e.g., Nqueens, Bubblesort, Towers of Hanoi • Kernels – Common subroutines in real programs, e.g., matrix multiply, FFT, sorting, Livermore Loops, Linpack • Simplified Applications – Extract main computational skeleton of real application to simplify porting, e.g., NAS parallel benchmarks, TPC • Real Applications – Things people actually use their computers for, e.g., car crash simulations, relational databases, Photoshop, Quake 8/30 CS252-Fall’07 10 Performance: What to measure • Usually rely on benchmarks vs. real workloads • To increase predictability, collections of benchmark applications-- benchmark suites -- are popular • SPECCPU: popular desktop benchmark suite – – – – CPU only, split between integer and floating point programs SPECint2000 has 12 integer, SPECfp2000 has 14 integer pgms SPECCPU2006 to be announced Spring 2006 SPECSFS (NFS file server) and SPECWeb (WebServer) added as server benchmarks • Transaction Processing Council measures server performance and cost-performance for databases – – – – 8/30 TPC-C Complex query for Online Transaction Processing TPC-H models ad hoc decision support TPC-W a transactional web benchmark TPC-App application server and web services benchmark CS252-Fall’07 11 Summarizing Performance System Rate (Task 1) Rate (Task 2) A 10 20 B 20 10 Which system is faster? 8/30 CS252-Fall’07 12 … depends who’s selling System Rate (Task 1) Rate (Task 2) Average A 10 20 15 B 20 10 15 Average throughput System Rate (Task 1) Rate (Task 2) Average A 0.50 2.00 1.25 B 1.00 1.00 1.00 Throughput relative to B System Rate (Task 1) Rate (Task 2) Average A 1.00 1.00 1.00 B 2.00 0.50 1.25 Throughput relative to A 8/30 CS252-Fall’07 13 Summarizing Performance over Set of Benchmark Programs Arithmetic mean of execution times ti (in seconds) 1/n i ti Harmonic mean of execution rates ri (MIPS/MFLOPS) n/ [i (1/ri)] • Both equivalent to workload where each program is run the same number of times • Can add weighting factors to model other workload distributions 8/30 CS252-Fall’07 14 Normalized Execution Time and Geometric Mean • Measure speedup up relative to reference machine ratio = tRef/tA • Average time ratios using geometric mean n( I ratioi ) • Insensitive to machine chosen as reference • Insensitive to run time of individual benchmarks • Used by SPEC89, SPEC92, SPEC95, …, SPEC2006 ….. But beware that choice of reference machine can suggest what is “normal” performance profile: 8/30 CS252-Fall’07 15 Vector/Superscalar Speedup • 100 MHz Cray J90 vector machine versus 300MHz Alpha 21164 • [LANL Computational Physics Codes, Wasserman, ICS’96] • Vector machine peaks on a few codes???? 8/30 CS252-Fall’07 16 Superscalar/Vector Speedup • • • 100 MHz Cray J90 vector machine versus 300MHz Alpha 21164 [LANL Computational Physics Codes, Wasserman, ICS’96] Scalar machine peaks on one code??? 8/30 CS252-Fall’07 17 How to Mislead with Performance Reports • • • • • • • • • • • • • • • Select pieces of workload that work well on your design, ignore others Use unrealistic data set sizes for application (too big or too small) Report throughput numbers for a latency benchmark Report latency numbers for a throughput benchmark Report performance on a kernel and claim it represents an entire application Use 16-bit fixed-point arithmetic (because it’s fastest on your system) even though application requires 64-bit floating-point arithmetic Use a less efficient algorithm on the competing machine Report speedup for an inefficient algorithm (bubblesort) Compare hand-optimized assembly code with unoptimized C code Compare your design using next year’s technology against competitor’s year old design (1% performance improvement per week) Ignore the relative cost of the systems being compared Report averages and not individual results Report speedup over unspecified base system, not absolute times Report efficiency not absolute times Report MFLOPS not absolute times (use inefficient algorithm) [ David Bailey “Twelve ways to fool the masses when giving performance results for parallel supercomputers” ] 8/30 CS252-Fall’07 18 Benchmarking for Future Machines • Variance in performance for parallel architectures is going to be much worse than for serial processors – SPECcpu means only really work across very similar machine configurations • What is a good benchmarking methodology? • Possible CS252 project – Berkeley View Techreport has “Dwarves” as major types of code that must run well (http://view.eecs.berkeley.edu) – Can you construct a parallel benchmark methodology from Dwarves? 8/30 CS252-Fall’07 19 Power and Energy • Energy to complete operation (Joules) – Corresponds approximately to battery life – (Battery energy capacity actually depends on rate of discharge) • Peak power dissipation (Watts = Joules/second) – Affects packaging (power and ground pins, thermal design) • di/dt, peak change in supply current (Amps/second) – Affects power supply noise (power and ground pins, decoupling capacitors) 8/30 CS252-Fall’07 20 Peak Power versus Lower Energy Peak A Peak B Power Integrate power curve to get energy Time • System A has higher peak power, but lower total energy • System B has lower peak power, but higher total energy 8/30 CS252-Fall’07 21 CS252 Administrivia Instructor: Prof. Krste Asanovic Office: 645 Soda Hall, krste@eecs Office Hours: M 1-3PM, 645 Soda Hall T. A.: Rose Liu, rfl@eecs Office Hours: W 4-5PM, 751 Soda Hall Lectures: Tu/Th, 9:30-11:00AM 203 McLaughlin Text: Computer Architecture: A Quantitative Approach, 4th Edition (Oct, 2006) Web page: http://inst.eecs.berkeley.edu/~cs252 Lectures available online <6:00 AM day of lecture 8/30 CS252-Fall’07 22 CS252 Updates • Prereq quiz will cover: – – – – Finite state machines ISA designs & MIPS assembly code programming (today’s lecture) Simple pipelining (single-issue, in-order, today’s lecture) Simple caches (Tuesday’s lecture) • Will try using Bspace this term to manage course private material (readings, assignments) – https://bspace.berkeley.edu/portal • Book order late at book store, will try and get you temporary copies of text 8/30 CS252-Fall’07 23 A "Typical" RISC ISA • • • • 32-bit fixed format instruction (3 formats) 32 32-bit GPR (R0 contains zero, DP take pair) 3-address, reg-reg arithmetic instruction Single address mode for load/store: base + displacement – no indirection • Simple branch conditions • Delayed branch see: SPARC, MIPS, HP PA-Risc, DEC Alpha, IBM PowerPC, CDC 6600, CDC 7600, Cray-1, Cray-2, Cray-3 8/30 CS252-Fall’07 24 Example: MIPS ( MIPS) Register-Register 31 26 25 Op 21 20 Rs1 16 15 Rs2 11 10 6 5 Rd 0 Opx Register-Immediate 31 26 25 Op 21 20 Rs1 16 15 Rd immediate 0 Branch 31 26 25 Op Rs1 21 20 16 15 Rs2/Opx immediate 0 Jump / Call 31 26 25 Op 8/30 target CS252-Fall’07 0 25 Datapath vs Control Datapath Controller signals Control Points • Datapath: Storage, FU, interconnect sufficient to perform the desired functions – Inputs are Control Points – Outputs are signals • Controller: State machine to orchestrate operation on the data path – Based on desired function and signals 8/30 CS252-Fall’07 26 Approaching an ISA • Instruction Set Architecture – Defines set of operations, instruction format, hardware supported data types, named storage, addressing modes, sequencing • Meaning of each instruction is described by RTL on architected registers and memory • Given technology constraints assemble adequate datapath – – – – Architected storage mapped to actual storage Function units to do all the required operations Possible additional storage (eg. MAR, MBR, …) Interconnect to move information among regs and FUs • Map each instruction to sequence of RTLs • Collate sequences into symbolic controller state transition diagram (STD) • Lower symbolic STD to control points • Implement controller 8/30 CS252-Fall’07 27 5 Steps of MIPS Datapath Figure A.2, Page A-8 Instruction Fetch Instr. Decode Reg. Fetch Execute Addr. Calc Next SEQ PC Adder 4 Zero? RS1 L M D MUX Data Memory ALU Imm MUX MUX RD Reg File Inst Memory Address IR <= mem[PC]; RS2 Write Back MUX Next PC Memory Access Sign Extend PC <= PC + 4 Reg[IRrd] <= Reg[IRrs] opIRop Reg[IRrt] 8/30 CS252-Fall’07 WB Data 28 5 Steps of MIPS Datapath Figure A.3, Page A-9 Execute Addr. Calc Instr. Decode Reg. Fetch Next SEQ PC Next SEQ PC Adder 4 Zero? RS1 MUX MEM/WB Data Memory EX/MEM ALU A <= Reg[IRrs]; Imm MUX MUX PC <= PC + 4 ID/EX IR <= mem[PC]; Reg File IF/ID Memory Address RS2 Write Back MUX Next PC Memory Access WB Data Instruction Fetch Sign Extend RD RD RD B <= Reg[IRrt] rslt <= A opIRop B WB <= rslt Reg[IR 8/30 rd] <= WB CS252-Fall’07 29 Inst. Set Processor Controller IR <= mem[PC]; PC <= PC + 4 JSR A <= Reg[IRrs]; JR ST RI RR if bop(A,b) opFetch-DCD B <= Reg[IRrt] jmp br Ifetch PC <= IRjaddr r <= A opIRop B LD r <= A opIRop IRim r <= A + IRim WB <= r WB <= Mem[r] PC <= PC+IRim WB <= r Reg[IRrd] <= WB 8/30 CS252-Fall’07 Reg[IRrd] <= WB Reg[IRrd] <= WB 30 5 Steps of MIPS Datapath Figure A.3, Page A-9 Execute Addr. Calc Instr. Decode Reg. Fetch Next SEQ PC Next SEQ PC Adder 4 Zero? RS1 MUX MEM/WB Data Memory EX/MEM ALU MUX MUX ID/EX Imm Reg File IF/ID Memory Address RS2 Write Back MUX Next PC Memory Access WB Data Instruction Fetch Sign Extend RD RD RD • Data stationary control – local decode for each instruction phase / pipeline stage 8/30 CS252-Fall’07 31 Visualizing Pipelining Figure A.2, Page A-8 Time (clock cycles) 8/30 Reg DMem Ifetch Reg DMem Reg ALU DMem Reg ALU O r d e r Ifetch ALU I n s t r. ALU Cycle 1 Cycle 2 Cycle 3 Cycle 4 Cycle 5 Cycle 6 Cycle 7 Ifetch Ifetch CS252-Fall’07 Reg Reg Reg DMem Reg 32 Pipelining is not quite that easy! • Limits to pipelining: Hazards prevent next instruction from executing during its designated clock cycle – Structural hazards: HW cannot support this combination of instructions (single person to fold and put clothes away) – Data hazards: Instruction depends on result of prior instruction still in the pipeline (missing sock) – Control hazards: Caused by delay between the fetching of instructions and decisions about changes in control flow (branches and jumps). 8/30 CS252-Fall’07 33 One Memory Port/Structural Hazards Figure A.4, Page A-14 Time (clock cycles) 8/30 Reg DMem Reg DMem Reg DMem Reg ALU Instr 4 Ifetch ALU Instr 3 DMem ALU O r d e r Instr 2 Reg ALU I Load Ifetch n s Instr 1 t r. ALU Cycle 1 Cycle 2 Cycle 3 Cycle 4 Cycle 5 Cycle 6 Cycle 7 Ifetch Reg Ifetch Ifetch CS252-Fall’07 Reg Reg Reg Reg DMem 34 One Memory Port/Structural Hazards (Similar to Figure A.5, Page A-15) Time (clock cycles) Stall DMem Ifetch Reg DMem Reg ALU Ifetch Bubble Instr 3 Reg Reg DMem Bubble Bubble Ifetch Reg Reg Bubble ALU O r d e r Instr 2 Reg ALU I Load Ifetch n s Instr 1 t r. ALU Cycle 1 Cycle 2 Cycle 3 Cycle 4 Cycle 5 Cycle 6 Cycle 7 Bubble Reg DMem How do you “bubble” the pipe? 8/30 CS252-Fall’07 35 Speed Up Equation for Pipelining CPIpipelined Ideal CPI Average Stall cycles per Inst Cycle Timeunpipelined Ideal CPI Pipeline depth Speedup Ideal CPI Pipeline stall CPI Cycle Timepipelined For simple RISC pipeline, CPI = 1: Cycle Timeunpipelined Pipeline depth Speedup 1 Pipeline stall CPI Cycle Timepipelined 8/30 CS252-Fall’07 36 Example: Dual-port vs. Single-port • Machine A: Dual ported memory (“Harvard Architecture”) • Machine B: Single ported memory, but its pipelined implementation has a 1.05 times faster clock rate • Ideal CPI = 1 for both • Loads are 40% of instructions executed SpeedUpA = Pipeline Depth/(1 + 0) x (clockunpipe/clockpipe) = Pipeline Depth SpeedUpB = Pipeline Depth/(1 + 0.4 x 1) x (clockunpipe/(clockunpipe / 1.05) = (Pipeline Depth/1.4) x 1.05 = 0.75 x Pipeline Depth SpeedUpA / SpeedUpB = Pipeline Depth/(0.75 x Pipeline Depth) = 1.33 • Machine A is 1.33 times faster 8/30 CS252-Fall’07 37 Data Hazard on R1 Figure A.6, Page A-17 Time (clock cycles) and r6,r1,r7 or Ifetch DMem Reg DMem Ifetch Reg DMem Ifetch Reg DMem Ifetch Reg ALU sub r4,r1,r3 Reg ALU Ifetch ALU O r d e r add r1,r2,r3 r8,r1,r9 xor r10,r1,r11 8/30 WB ALU I n s t r. MEM ALU IF ID/RF EX CS252-Fall’07 Reg Reg Reg Reg DMem 38 Reg Three Generic Data Hazards • Read After Write (RAW) InstrJ tries to read operand before InstrI writes it I: add r1,r2,r3 J: sub r4,r1,r3 • Caused by a “Dependence” (in compiler nomenclature). This hazard results from an actual need for communication. 8/30 CS252-Fall’07 39 Three Generic Data Hazards • Write After Read (WAR) InstrJ writes operand before InstrI reads it I: sub r4,r1,r3 J: add r1,r2,r3 K: mul r6,r1,r7 • Called an “anti-dependence” by compiler writers. This results from reuse of the name “r1”. • Can’t happen in MIPS 5 stage pipeline because: – All instructions take 5 stages, and – Reads are always in stage 2, and – Writes are always in stage 5 8/30 CS252-Fall’07 40 Three Generic Data Hazards • Write After Write (WAW) InstrJ writes operand before InstrI writes it. I: sub r1,r4,r3 J: add r1,r2,r3 K: mul r6,r1,r7 • Called an “output dependence” by compiler writers This also results from the reuse of name “r1”. • Can’t happen in MIPS 5 stage pipeline because: – All instructions take 5 stages, and – Writes are always in stage 5 • Will see WAR and WAW in more complicated pipes 8/30 CS252-Fall’07 41 Forwarding to Avoid Data Hazard Figure A.7, Page A-19 or r8,r1,r9 Reg DMem Ifetch Reg DMem Ifetch Reg DMem Ifetch Reg ALU and r6,r1,r7 Ifetch DMem ALU sub r4,r1,r3 Reg ALU O r d e r add r1,r2,r3 Ifetch ALU I n s t r. ALU Time (clock cycles) xor r10,r1,r11 8/30 CS252-Fall’07 Reg Reg Reg Reg DMem 42 Reg HW Change for Forwarding Figure A.23, Page A-37 NextPC mux MEM/WR EX/MEM ALU mux ID/EX Registers Data Memory mux Immediate What circuit detects and resolves this hazard? 8/30 CS252-Fall’07 43 Forwarding to Avoid LW-SW Data Hazard Figure A.8, Page A-20 or r8,r6,r9 Reg DMem Ifetch Reg DMem Ifetch Reg DMem Ifetch Reg ALU sw r4,12(r1) Ifetch DMem ALU lw r4, 0(r1) Reg ALU O r d e r add r1,r2,r3 Ifetch ALU I n s t r. ALU Time (clock cycles) xor r10,r9,r11 8/30 CS252-Fall’07 Reg Reg Reg Reg DMem 44 Reg Data Hazard Even with Forwarding Figure A.9, Page A-21 and r6,r1,r7 or 8/30 DMem Ifetch Reg DMem Reg Ifetch Ifetch r8,r1,r9 CS252-Fall’07 Reg Reg Reg DMem ALU O r d e r sub r4,r1,r6 Reg ALU lw r1, 0(r2) Ifetch ALU I n s t r. ALU Time (clock cycles) Reg DMem 45 Reg Data Hazard Even with Forwarding (Similar to Figure A.10, Page A-21) Reg DMem Ifetch Reg Bubble Ifetch Bubble Reg Bubble Ifetch and r6,r1,r7 or r8,r1,r9 Reg DMem Reg Reg Reg DMem ALU sub r4,r1,r6 Ifetch ALU O r d e r lw r1, 0(r2) ALU I n s t r. ALU Time (clock cycles) DMem How is this detected? 8/30 CS252-Fall’07 46 Software Scheduling to Avoid Load Hazards Try producing fast code for a = b + c; d = e – f; assuming a, b, c, d ,e, and f in memory. Slow code: LW LW ADD SW LW LW SUB SW Rb,b Rc,c Ra,Rb,Rc a,Ra Re,e Rf,f Rd,Re,Rf d,Rd Fast code: LW LW LW ADD LW SW SUB SW Rb,b Rc,c Re,e Ra,Rb,Rc Rf,f a,Ra Rd,Re,Rf d,Rd Compiler optimizes for performance. Hardware checks for safety. 8/30 CS252-Fall’07 47 r6,r1,r7 Reg DMem Ifetch Reg DMem Ifetch Reg DMem Ifetch Reg ALU Ifetch DMem ALU 18: or Reg ALU 14: and r2,r3,r5 Ifetch ALU 10: beq r1,r3,36 ALU Control Hazard on Branches Three Stage Stall 22: add r8,r1,r9 36: xor r10,r1,r11 Reg Reg Reg What do you do with the 3 instructions in between? How do you do it? Where is the “commit”? 8/30 CS252-Fall’07 48 Reg DMem Branch Stall Impact • If CPI = 1, 30% branch, Stall 3 cycles => new CPI = 1.9! • Two part solution: – Determine branch taken or not sooner, AND – Compute taken branch address earlier • MIPS branch tests if register = 0 or 0 • MIPS Solution: – Move Zero test to ID/RF stage – Adder to calculate new PC in ID/RF stage – 1 clock cycle penalty for branch versus 3 8/30 CS252-Fall’07 49 Pipelined MIPS Datapath Figure A.24, page A-38 Instruction Fetch Memory Access Write Back Adder Adder MUX Next SEQ PC Next PC Zero? RS1 MUX MEM/WB Data Memory EX/MEM ALU MUX ID/EX Imm Reg File IF/ID Memory Address RS2 WB Data 4 Execute Addr. Calc Instr. Decode Reg. Fetch Sign Extend RD RD RD • Interplay of instruction set design and cycle time. 8/30 CS252-Fall’07 50 Four Branch Hazard Alternatives #1: Stall until branch direction is clear #2: Predict Branch Not Taken – – – – – Execute successor instructions in sequence “Squash” instructions in pipeline if branch actually taken Advantage of late pipeline state update 47% MIPS branches not taken on average PC+4 already calculated, so use it to get next instruction #3: Predict Branch Taken – 53% MIPS branches taken on average – But haven’t calculated branch target address in MIPS » MIPS still incurs 1 cycle branch penalty » Other machines: branch target known before outcome 8/30 CS252-Fall’07 51 Four Branch Hazard Alternatives #4: Delayed Branch – Define branch to take place AFTER a following instruction branch instruction sequential successor1 sequential successor2 ........ sequential successorn branch target if taken Branch delay of length n – 1 slot delay allows proper decision and branch target address in 5 stage pipeline – MIPS uses this 8/30 CS252-Fall’07 52 Scheduling Branch Delay Slots (Fig A.14) A. From before branch add $1,$2,$3 if $2=0 then delay slot becomes B. From branch target sub $4,$5,$6 add $1,$2,$3 if $1=0 then delay slot becomes if $2=0 then add $1,$2,$3 add $1,$2,$3 if $1=0 then sub $4,$5,$6 C. From fall through add $1,$2,$3 if $1=0 then delay slot sub $4,$5,$6 becomes add $1,$2,$3 if $1=0 then sub $4,$5,$6 • A is the best choice, fills delay slot & reduces instruction count (IC) • In B, the sub instruction may need to be copied, increasing IC • In B and C, must be okay to execute sub when branch fails 8/30 CS252-Fall’07 53 Delayed Branch • Compiler effectiveness for single branch delay slot: – Fills about 60% of branch delay slots – About 80% of instructions executed in branch delay slots useful in computation – About 50% (60% x 80%) of slots usefully filled • Delayed Branch downside: As processor go to deeper pipelines and multiple issue, the branch delay grows and need more than one delay slot – Delayed branching has lost popularity compared to more expensive but more flexible dynamic approaches – Growth in available transistors has made dynamic approaches relatively cheaper 8/30 CS252-Fall’07 54 Evaluating Branch Alternatives Pipeline speedup = Pipeline depth 1 +Branch frequency Branch penalty Assume 4% unconditional branch, 6% conditional branch- untaken, 10% conditional branch-taken Scheduling Branch CPI speedup v. speedup v. scheme penalty unpipelined stall Stall pipeline 3 1.60 3.1 1.0 Predict taken 1 1.20 4.2 1.33 Predict not taken 1 1.14 4.4 1.40 Delayed branch 0.5 1.10 4.5 1.45 8/30 CS252-Fall’07 55 Problems with Pipelining • Exception: An unusual event happens to an instruction during its execution – Examples: divide by zero, undefined opcode • Interrupt: Hardware signal to switch the processor to a new instruction stream – Example: a sound card interrupts when it needs more audio output samples (an audio “click” happens if it is left waiting) • Problem: It must appear that the exception or interrupt must appear between 2 instructions (Ii and Ii+1) – The effect of all instructions up to and including Ii is totalling complete – No effect of any instruction after Ii can take place • The interrupt (exception) handler either aborts program or restarts at instruction Ii+1 8/30 CS252-Fall’07 56 Precise Exceptions In-Order Pipelines Key observation: architected state only change in memory and register write stages. Summary: Metrics and Pipelining • Machines compared over many metrics – Cost, performance, power, reliability, compatibility, … • • • • Difficult to compare widely differing machines on benchmark suite Control VIA State Machines and Microprogramming Just overlap tasks; easy if tasks are independent Speed Up Pipeline Depth; if ideal CPI is 1, then: Cycle Timeunpipelined Pipeline depth Speedup 1 Pipeline stall CPI Cycle Timepipelined • Hazards limit performance on computers: – Structural: need more HW resources – Data (RAW,WAR,WAW): need forwarding, compiler scheduling – Control: delayed branch, prediction • Exceptions, Interrupts add complexity • Next time: Read Appendix C! 8/30 CS252-Fall’07 58