CS654 Advanced Computer Architecture Lec 5 – Performance + Pipeline Review Peter Kemper
Adapted from the slides of EECS 252 by Prof. David Patterson Electrical Engineering and Computer Sciences University of California, Berkeley
CS654 Advanced Computer Architecture Lec 5 Performance + Pipeline - - PowerPoint PPT Presentation
CS654 Advanced Computer Architecture Lec 5 Performance + Pipeline Review Peter Kemper Adapted from the slides of EECS 252 by Prof. David Patterson Electrical Engineering and Computer Sciences University of California, Berkeley Review
Adapted from the slides of EECS 252 by Prof. David Patterson Electrical Engineering and Computer Sciences University of California, Berkeley
2/2/09 CS 654 W&M 2
architect’s responsibility
processors to improve by at least as much as the square of the improvement in Latency
– IC ≈ f(Area) + Learning curve, volume, commodity, margins
– Capacitance x Voltage2 x frequency, Energy vs. power
– Reliability (MTTF vs. FIT), Availability (MTTF/(MTTF+MTTR)
– Performance (1/execTime), SpecRatio
2/2/09 CS 654 W&M 3
– Ratios, Geometric Mean, Multiplicative Standard Deviation
danger
2/2/09 CS 654 W&M 4
How Summarize Suite Performance (1/5)
– But they vary by 4X in speed, so some would be more important than others in arithmetic average
weight?
– Different companies want different weights for their products
computer, yielding a ratio proportional to performance = time on reference computer time on computer being rated
2/2/09 CS 654 W&M 5
How Summarize Suite Performance (2/5)
times bigger than Computer B, then
B A A B B reference A reference B A
execution times on the reference computer drop
2/2/09 CS 654 W&M 6
How Summarize Suite Performance (3/5)
(SPECRatio unitless, so arithmetic mean meaningless)
n n i i
1
to summarize performance:
ratio of the geometric means
= Geometric mean of performance ratios ⇒ choice of reference computer is irrelevant!
2/2/09 CS 654 W&M 7
How Summarize Suite Performance (4/5)
programs in benchmark suite?
variability of distribution using standard deviation
multiplicative rather than arithmetic
the standard mean and standard deviation, and then take the exponent to convert back:
i n i i
SPECRatio StDev tDev GeometricS SPECRatio n ean GeometricM ln exp ln 1 exp
1
=
2/2/09 CS 654 W&M 8
How Summarize Suite Performance (5/5)
distribution has a standard form
– bell-shaped normal distribution, whose data are symmetric around mean – lognormal distribution, where logarithms of data--not data itself--are normally distributed (symmetric) on a logarithmic scale
68% of samples fall in range 95% of samples fall in range
STDEV() that make calculating geometric mean and multiplicative standard deviation easy
gstdev mean gstdev mean
/
2 2,
/ gstdev mean gstdev mean
2/2/09 CS 654 W&M 9
2000 4000 6000 8000 10000 12000 14000
wupwise swim mgrid applu mesa galgel art equake facerec ammp lucas fma3d sixtrack apsi
SPECfpRatio
1372 5362 2712 GM = 2712 GStDev = 1.98
Outside 1 StDev
2/2/09 CS 654 W&M 10
2000 4000 6000 8000 10000 12000 14000
wupwise swim mgrid applu mesa galgel art equake facerec ammp lucas fma3d sixtrack apsi
SPECfpRatio
1494 2911 2086 GM = 2086 GStDev = 1.40
Outside 1 StDev
2/2/09 CS 654 W&M 11
higher-- vs. 1.40--so results will differ more widely from the mean, and therefore are likely less predictable
– 10 of 14 benchmarks (71%) for Itanium 2 – 11 of 14 benchmarks (78%) for Athlon
lognormal distribution (expect 68% for 1 StDev)
2/2/09 CS 654 W&M 12
– When discussing a fallacy, we try to give a counterexample.
– Often generalizations of principles true in limited context – Show Fallacies and Pitfalls to help you avoid these errors
– Once a benchmark becomes popular, tremendous pressure to improve performance by targeted
rules for running the benchmark: “benchmarksmanship.” – 70 benchmarks from the 5 SPEC releases. 70% were dropped from the next release since no longer useful
– Rule of thumb for fault tolerant systems: make sure that every component was redundant so that no single component failure could bring down the whole system (e.g, power supply)
2/2/09 CS 654 W&M 13
≈ 140 years, so disks practically never fail
years; on average, 28 replacements wouldn't fail
= 1000*(5*365*24)*833 /109 = 36,485,000 / 106 = 37 = 3.7% (37/1000) fail over 5 yr lifetime (1.2M hr MTTF)
– little vibration, narrow temperature range ⇒ no power failures
– 3400 - 6800 FIT or 150,000 - 300,000 hour MTTF [Gray & van Ingen 05]
– 3400 - 8000 FIT or 125,000 - 300,000 hour MTTF [Gray & van Ingen 05]
2/2/09 CS 654 W&M 14
– Ratios, Geometric Mean, Multiplicative Standard Deviation
danger
2/2/09 CS 654 W&M 15
– General purpose register architectures, – 80x86: register-memory ISA, MIPS: load-store ISA
– Byte addressing (usually), alignment (some)
– Register, constants/immediate, displacement at least
– 8bit (ASCII), 16 bit (Unicode, halfword), 32 bit (int, word), 64 bit – IEEE 754 floating point 32 bit single, 64 bit double precision
– Data transfer, arithmetic logical, control, floating point
– Jumps, cond. branches, procedure calls, returns, PC-relat. addressing
– Fixed length vs variable length encoding
2/2/09 CS 654 W&M 16
base + displacement
– no indirection
see: SPARC, MIPS, HP PA-Risc, DEC Alpha, IBM PowerPC, CDC 6600, CDC 7600, Cray-1, Cray-2, Cray-3
2/2/09 CS 654 W&M 17
Op
31 26 15 16 20 21 25
Rs1 Rd immediate Op
31 26 25
Op
31 26 15 16 20 21 25
Rs1 Rs2 target Rd Opx Register-Register
5 6 10 11
Register-Immediate Op
31 26 15 16 20 21 25
Rs1 Rs2/Opx immediate Branch Jump / Call
2/2/09 CS 654 W&M 18
desired functions
– Inputs are Control Points – Outputs are signals
path
– Based on desired function and signals
Datapath Controller Control Points signals
2/2/09 CS 654 W&M 19
– Defines set of operations, instruction format, hardware supported data types, named storage, addressing modes, sequencing
architected registers and memory
– Architected storage mapped to actual storage – Function units to do all the required operations – Possible additional storage (eg. MAR, MBR, …) – Interconnect to move information among regs and FUs
diagram (STD)
STD: state transition diagram RTL: register transfer language FU: function unit MAR: memory address register MBR: memory buffer register
2/2/09 CS 654 W&M 20
Memory Access Write Back Instruction Fetch
Execute
L M D ALU
MUX
Memory Reg File
MUX MUX
Data Memory
MUX
Sign Extend
4
Adder
Zero?
Next SEQ PC
Address
Next PC WB Data
Ins t
RD RS1 RS2 Imm
IR <= mem[PC]; PC <= PC + 4 Reg[IRrd] <= Reg[IRrs] opIRop Reg[IRrt]
2/2/09 CS 654 W&M 21
with pipeline registers
Memory Access Write Back Instruction Fetch
Execute
ALU Memory Reg File
MUX MUX
Data Memory
MUX
Sign Extend
Zero?
IF/ID ID/EX MEM/WB EX/MEM
4
Adder
Next SEQ PC Next SEQ PC
RD RD RD
WB Data Next PC
Address
RS1 RS2 Imm
MUX IR <= mem[PC]; PC <= PC + 4 A <= Reg[IRrs]; B <= Reg[IRrt] rslt <= A opIRop B Reg[IRrd] <= WB WB <= rslt
2/2/09 CS 654 W&M 22
IR <= mem[PC]; PC <= PC + 4 A <= Reg[IRrs]; B <= Reg[IRrt] r <= A opIRop B Reg[IRrd] <= WB WB <= r Ifetch
PC <= IRjaddr if bop(A,b) PC <= PC+IRim
br jmp RR
r <= A opIRop IRim Reg[IRrd] <= WB WB <= r
RI
r <= A + IRim WB <= Mem[r] Reg[IRrd] <= WB
LD ST JSR JR
2/2/09 CS 654 W&M 23
Memory Access Write Back Instruction Fetch
Execute
ALU Memory Reg File
MUX MUX
Data Memory
MUX
Sign Extend
Zero?
IF/ID ID/EX MEM/WB EX/MEM
4
Adder
Next SEQ PC Next SEQ PC
RD RD RD
WB Data
– local decode for each instruction phase / pipeline stage
Next PC
Address
RS1 RS2 Imm
MUX
2/2/09 CS 654 W&M 24
I n s t r. O r d e r Time (clock cycles)
Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg
Cycle 1 Cycle 2 Cycle 3 Cycle 4 Cycle 6 Cycle 7 Cycle 5
2/2/09 CS 654 W&M 25
from executing during its designated clock cycle
– Structural hazards: HW cannot support this combination of instructions (single person to fold and put clothes away) – Data hazards: Instruction depends on result of prior instruction still in the pipeline (missing sock) – Control hazards: Caused by delay between the fetching of instructions and decisions about changes in control flow (branches and jumps).
2/2/09 CS 654 W&M 26
Figure A.4, Page A-14
I n s t r. O r d e r Time (clock cycles)
Load Instr 1 Instr 2 Instr 3 Instr 4
Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg
Cycle 1 Cycle 2 Cycle 3 Cycle 4 Cycle 6 Cycle 7 Cycle 5
Reg ALU DMem Ifetch Reg
2/2/09 CS 654 W&M 27
(Similar to Figure A.5, Page A-15)
I n s t r. O r d e r Time (clock cycles)
Load Instr 1 Instr 2 Stall Instr 3
Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg
Cycle 1 Cycle 2 Cycle 3 Cycle 4 Cycle 6 Cycle 7 Cycle 5
Reg ALU DMem Ifetch Reg
Bubble Bubble Bubble Bubble Bubble
How do you “bubble” the pipe?
2/2/09 CS 654 W&M 28
Speedup = Pipeline depth Ideal CPI + Pipeline stall CPI Cycle Timeunpipelined Cycle Timepipelined
pipelined d unpipeline
Time Cycle Time Cycle CPI stall Pipeline 1 depth Pipeline Speedup
= Inst per cycles Stall Average CPI Ideal CPIpipelined + =
For simple RISC pipeline, CPI = 1:
Speedup = AvgInstTimeunpipelined AvgInstTimepipelined = CPIunpipelined CPIpipelined Cycle Timeunpipelined Cycle Timepipelined
If all inst take same number of cycles: CPIunpipelined=Pipeline depth
2/2/09 CS 654 W&M 29
implementation has a 1.05 times faster clock rate
SpeedUpA = Pipeline Depth/(1 + 0) x (clockunpipe/clockpipe) = Pipeline Depth SpeedUpB = Pipeline Depth/(1 + 0.4 x 1) x (clockunpipe/(clockunpipe / 1.05) = (Pipeline Depth/1.4) x 1.05 = 0.75 x Pipeline Depth SpeedUpA / SpeedUpB = Pipeline Depth/(0.75 x Pipeline Depth) = 1.33
2/2/09 CS 654 W&M 30
I n s t r. O r d e r
add r1,r2,r3 sub r4,r1,r3 and r6,r1,r7
xor r10,r1,r11
Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg
Figure A.6, Page A-17
Time (clock cycles)
IF ID/RF EX MEM WB
2/2/09 CS 654 W&M 31
InstrJ tries to read operand before InstrI writes it
nomenclature). This hazard results from an actual need for communication.
I: add r1,r2,r3 J: sub r4,r1,r3
2/2/09 CS 654 W&M 32
InstrJ writes operand before InstrI reads it
This results from reuse of the name “r1”.
– All instructions take 5 stages, and – Reads are always in stage 2, and – Writes are always in stage 5 I: sub r4,r1,r3 J: add r1,r2,r3 K: mul r6,r1,r7
2/2/09 CS 654 W&M 33
InstrJ writes operand before InstrI writes it.
This also results from the reuse of name “r1”.
– All instructions take 5 stages, and – Writes are always in stage 5
I: sub r1,r4,r3 J: add r1,r2,r3 K: mul r6,r1,r7
2/2/09 CS 654 W&M 34
Time (clock cycles)
Figure A.7, Page A-19 I n s t r. O r d e r
add r1,r2,r3 sub r4,r1,r3 and r6,r1,r7
xor r10,r1,r11
Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg
2/2/09 CS 654 W&M 35
Figure A.23, Page A-37 MEM/WR ID/EX EX/MEM Data Memory
ALU
mux mux Registers
NextPC Immediate
mux
What circuit detects and resolves this hazard?
2/2/09 CS 654 W&M 36
Time (clock cycles)
Forwarding to Avoid LW-SW Data Hazard
Figure A.8, Page A-20
I n s t r. O r d e r
add r1,r2,r3 lw r4, 0(r1) sw r4,12(r1)
xor r10,r9,r11
Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg
2/2/09 CS 654 W&M 37
Time (clock cycles) I n s t r. O r d e r
lw r1, 0(r2) sub r4,r1,r6 and r6,r1,r7
Figure A.9, Page A-21
Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg
2/2/09 CS 654 W&M 38
Data Hazard Even with Forwarding
(Similar to Figure A.10, Page A-21)
Time (clock cycles)
I n s t r. O r d e r
lw r1, 0(r2) sub r4,r1,r6 and r6,r1,r7
Reg ALU DMem Ifetch Reg Reg Ifetch ALU DMem Reg
Bubble
Ifetch ALU DMem Reg
Bubble
Reg Ifetch ALU DMem
Bubble
Reg
How is this detected?
2/2/09 CS 654 W&M 39
Try producing fast code for a = b + c; d = e – f; assuming a, b, c, d ,e, and f in memory.
Slow code: LW Rb,b LW Rc,c ADD Ra,Rb,Rc SW a,Ra LW Re,e LW Rf,f SUB Rd,Re,Rf SW d,Rd
Fast code: LW Rb,b LW Rc,c LW Re,e ADD Ra,Rb,Rc LW Rf,f SW a,Ra SUB Rd,Re,Rf SW d,Rd
Compiler optimizes for performance. Hardware checks for safety.
2/2/09 CS 654 W&M 40
– Ratios, Geometric Mean, Multiplicative Standard Deviation
danger
2/2/09 CS 654 W&M 41
10: beq r1,r3,36 14: and r2,r3,r5 18: or r6,r1,r7 22: add r8,r1,r9 36: xor r10,r1,r11
Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch Reg Reg ALU DMem Ifetch
What do you do with the 3 instructions in between? How do you do it? Where is the “commit”?
2/2/09 CS 654 W&M 42
Stall 3 cycles => new CPI = 1.9!
– Determine branch taken or not sooner, AND – Compute taken branch address earlier
– Move Zero test to ID/RF stage – Adder to calculate new PC in ID/RF stage – 1 clock cycle penalty for branch versus 3
2/2/09 CS 654 W&M 43
Adder
IF/ID
Figure A.24, page A-38 Memory Access Write Back Instruction Fetch
Execute
ALU Memory Reg File
MUX
Data Memory
MUX
Sign Extend
Zero?
MEM/WB EX/MEM
4
Adder
Next SEQ PC
RD RD RD
WB Data
Next PC
Address
RS1 RS2 Imm
MUX
ID/EX
2/2/09 CS 654 W&M 44
#1: Stall until branch direction is clear #2: Predict Branch Not Taken
– Execute successor instructions in sequence – “Squash” instructions in pipeline if branch actually taken – Advantage of late pipeline state update – 47% MIPS branches not taken on average – PC+4 already calculated, so use it to get next instruction
#3: Predict Branch Taken
– 53% MIPS branches taken on average – But haven’t calculated branch target address in MIPS » MIPS still incurs 1 cycle branch penalty » Other machines: branch target known before outcome
2/2/09 CS 654 W&M 45
#4: Delayed Branch
– Define branch to take place AFTER a following instruction branch instruction sequential successor1 sequential successor2 ........ sequential successorn branch target if taken – 1 slot delay allows proper decision and branch target address in 5 stage pipeline – MIPS uses this Branch delay of length n
2/2/09 CS 654 W&M 46
Scheduling Branch Delay Slots (Fig A.14)
add $1,$2,$3 if $2=0 then delay slot
add $1,$2,$3 if $1=0 then delay slot add $1,$2,$3 if $1=0 then delay slot sub $4,$5,$6 sub $4,$5,$6 becomes becomes becomes if $2=0 then add $1,$2,$3 add $1,$2,$3 if $1=0 then sub $4,$5,$6 add $1,$2,$3 if $1=0 then sub $4,$5,$6
2/2/09 CS 654 W&M 47
– Fills about 60% of branch delay slots – About 80% of instructions executed in branch delay slots useful in computation – About 50% (≈ 60% x 80%) of slots usefully filled
deeper pipelines and multiple issue, the branch delay grows and need more than one delay slot
– Delayed branching has lost popularity compared to more expensive but more flexible dynamic approaches – Growth in available transistors has made dynamic approaches relatively cheaper
2/2/09 CS 654 W&M 48
Assume 4% unconditional branch, 6% conditional branch- untaken, 10% conditional branch-taken Scheduling Branch CPI speedup v. speedup v. scheme penalty unpipelined stall Stall pipeline 3 1.60 3.1 1.0 Predict taken 1 1.20 4.2 1.33 Predict not taken 1 1.14 4.4 1.40 Delayed branch 0.5 1.10 4.5 1.45
Pipeline speedup = Pipeline depth 1 +Branch frequencyBranch penalty
2/2/09 CS 654 W&M 49
instruction during its execution
– Examples: divide by zero, undefined opcode
processor to a new instruction stream
– Example: a sound card interrupts when it needs more audio
interrupt must appear between 2 instructions (Ii and Ii+1)
– The effect of all instructions up to and including Ii is totalling complete – No effect of any instruction after Ii can take place
program or restarts at instruction Ii+1
Key observation: architected state only change in memory and register write stages.
2/2/09 CS 654 W&M 51
– Ratios, Geometric Mean, Multiplicative Standard Deviation
– Structural: need more HW resources – Data (RAW,WAR,WAW): need forwarding, compiler scheduling – Control: delayed branch, prediction
pipelined d unpipeline
Time Cycle Time Cycle CPI stall Pipeline 1 depth Pipeline Speedup
=