Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
CENG 450 Computer Systems & Architecture Lecture 2 Amirali Baniasadi [email protected] Outline Power & Cost Performance Performance measurement Amdahl's Law Benchmarks History 1. “Big Iron” Computers: Used vacuum tubes, electric relays and bulk magnetic storage devices. No microprocessors. No memory. Example: ENIAC (1945), IBM Mark 1 (1944) History Von Newmann: Invented EDSAC (1949). First Stored Program Computer. Uses Memory. Importance: We are still using the same basic design. Computer Components Output Processor Control (CPU) Memory Input Printer Screen Disk ... keyboard Mouse Disk ... Computer Components Datapath of a von Newman machine OP1 + OP2 ... Op1 Op2 Bus Op1 General-purpose Registers Op2 ALU i/p registers ALU OP1 + OP2 ALU o/p register Computer Components Processor(CPU): Active part of the motherboard Performs calculations & activates devices Gets instruction & data from memory Components are connected via Buses Bus: Collection of parallel wires Transmits data, instructions, or control signals Motherboard Physical chips for I/O connections, memory, & CPU Computer Components CPU consists of Datapath (ALU+ Registers): Performs arithmetic & logical operations Control (CU): Controls the data path, memory, & I/O devices Sends signals that determine operations of datapath, memory, input & output Technology Change Technology changes rapidly HW Vacuum tubes: Electron emitting devices Transistors: On-off switches controlled by electricity Integrated Circuits( IC/ Chips): Combines thousands of transistors Very Large-Scale Integration( VLSI): Combines millions of transistors What next? SW Machine language: Zeros and ones Assembly language: Mnemonics High-Level Languages: English-like Artificial Intelligence languages: Functions & logic predicates Object-Oriented Programming: Objects & operations on objects Moore’s Prediction Moore’s Law: A new generation of memory chips is introduced every 3 years Each new generation has 4 times as much memory as its predecessor Computer technology doubles every 1.5 years: Example: DRAM capacity 100,000 64M 16M K b it c a p a cit y 10,000 4M 1M 1000 256K 100 64K 16K 10 1976 1978 1980 1982 1984 1986 1988 Year o f introduction 1990 1992 1994 1996 Technology => dramatic change Processor logic capacity: about 30% per year clock rate: about 20% per year Memory DRAM capacity: about 60% per year (4x every 3 years) Memory speed: about 10% per year Cost per bit: improves about 25% per year Disk capacity: about 60% per year Question: Does every thing look OK? Software Evolution. Machine language Assembly language High-level languages Subroutine libraries There is a large gap between what is convenient for computers & what is convenient for humans Translation/Interpretation is needed between both Language Evolution swap (int v[], int k) { int temp temp = v[k]; v[k] = v[k+1]; v[k+1] = temp; } swap: muli $2, $5, 4 add $2, $4, $2 lw $15, 0($2) lw $18, 4($2) sw $18, 0($2) sw $15, 4($2) jr $31 0 00 0 0 0 0 0 1 0 1 0 0 0 0 1 0 0 0 00 00 00 00 1 1 0 0 0 0 00 0 0 0 0 0 1 0 0 0 1 1 1 0 0 0 0 1 1 0 0 0 0 0 1 00 0 0 1 1 00 0 1 1 0 0 0 1 1 0 0 0 1 0 0 0 0 00 0 0 0 0 00 00 0 0 0 1 00 0 1 1 0 0 1 1 1 1 0 0 1 0 0 0 0 00 0 0 0 0 00 00 1 0 0 1 0 1 0 1 1 0 0 1 1 1 1 0 0 1 0 0 0 0 00 0 0 0 0 00 00 0 0 0 1 0 1 0 1 1 0 0 0 1 1 0 0 0 1 0 0 0 0 00 0 0 0 0 00 00 1 0 0 0 00 0 0 0 1 1 1 1 1 0 0 0 0 0 0 0 0 00 0 0 0 0 0 0 0 1 0 0 0 High-level language program (in C) Assembly language program (for MIPS) Binary machine language program (for MIPS) HW - SW Components Hardware Memory components Registers Register file memory Disks Functional components Adder, multiplier, dividers, . . . Comparators Control signals Software Data Simple • Characters • Integers • Floating-point • Pointers Structured • Arrays • Structures ( records) Instructions Data transfer Arithmetic Shift Control flow Comparison . . . Things You Will Learn Assembly language introduction/Review How to analyze program performance How to design processor components How to enhance processors performance (caches, pipelines, parallel processors, multiprocessors) The Processor Chip Processor Chip Major Blocks • Example: Intel Pentium • Area: 91 mm2 • ~ 3.3 million transistors ( 1 million for cache memory) Control Data cache Instruction cache Bus Integer datapath Branch Floatingpoint datapath Memory Categories Volatile memory Loses information when power is switched-off Non-volatile memory Keeps information when power is switched-off Types Cache: Volatile Fast but expensive Smaller capacity Placed closer to the processor Main memory Volatile Less expensive More capacity Secondary memory Nonvolatile Low cost Very slow Unlimited capacity Input-Output (I/O) I/O devices have the hardest organization Wide range of speeds Graphics vs. keyboard Wide range of requirements Speed Standard Cost . . . Least amount of research done in this area Our Primary Focus The processor (datapath and control) Implemented using millions of transistors Impossible to understand by looking at each transistor We need abstraction Hides lower-level details to offer simple model at higher level Advantages • Intensive & thorough research into the depths • Reveals more information • Omits unneeded details • Helps us cope with complexity Examples of abstraction: • Language hierarchy • Instruction set architecture (ISA) Instruction Set Architecture (ISA) Instruction set: Complete set of instructions used by a machine ISA: Abstract interface between the HW and lowest-level SW. It encompasses information needed to write machine-language programs including Instructions Memory size Registers used . . . Instruction Set Architecture (ISA) ISA is considered part of the SW Several implementations for the same ISA can exist Modern ISA’s: 80x86/Pentium/K6, PowerPC, DEC Alpha, MIPS, SPARC, HP We are going to study MIPS Advantages: Different implementations of the same architecture Easier to change than HW Standardizes instructions, machine language bit patterns, etc. Disadvantage: Sometimes prevents using new innovations Instruction Set Architecture (ISA) •Instruction Execution Cycle Fetch Instruction From Memory Decode Instruction determine its size & action Fetch Operand data Execute instruction & compute results or status Store Result in memory Determine Next Instruction’s address What Should we Learn? A specific ISA (MIPS) Performance issues - vocabulary and motivation Instruction-Level Parallelism How to Use Pipelining to improve performance Exploiting Instruction-Level Parallelism w/ Software Approach Memory: caches and virtual memory I/O What is Expected From You? • • • • Read textbook & readings! Be up-to-date! Come back with your input & questions for discussion! Appreciate and participate in teamwork! Power? Everything is done by tiny switches Their charge represents logic values Changing charge energy Power energy over time Devices are non-ideal power heat Excess heat Circuits breakdown Need to keep power within acceptable limits N iu m III II iu m uc IV le ar R ea ct or Pe nt Pe nt Pr o iu m iu m iu m Pe nt Pe nt Pe nt i4 86 i3 86 W/cm2 POWER in the real world 1000 100 10 1 Integrated Circuits Costs Die cost = Wafer cost Dies per Wafer * Die yield Percentage of good dies on a wafer Dies per wafer = Wafer Area Die Area * Wafer diameter (2 * die area)1/2 Die yield = Wafer yield * ( 1+ (Defects per unit area * Die area )/ ) - = 4.0 Integrated Circuits Costs-example Find the die yield for a die with a defect density of 0.6 for dies with areas 1.0 and .49. For the larger die: Die yield = (1+ (0.6*1)/4) -4 = for the smaller die: Die yield = (1+ (0.6*0.49)/4) Why? 0.57 -4 = 0.75 Other Costs IC cost = Die cost + Testing cost + Packaging cost Final test yield Packaging Cost: depends on pins, heat dissipation, ... Chip 386DX 486DX2 PowerPC 601 HP PA 7100 DEC Alpha SuperSPARC Pentium Die Packaging $4 $1 $12 $11 $53 $3 $73 $35 $149 $30 $272 $20 $417 $19 Testing $4 $12 $21 $16 $23 $34 $37 Total $9 $35 $77 $124 $202 $326 $473 System Cost: Workstation System Subsystem % of total cost Cabinet Sheet metal, plastic Power supply, fans Cables, nuts, bolts (Subtotal) 1% 2% 1% (4%) Processor DRAM (64MB) Video system I/O system Printed Circuit board (Subtotal) 6% 36% 14% 3% 1% (60%) Keyboard, mouse Monitor Hard disk (1 GB) Tape drive (DAT) (Subtotal) 1% 22% 7% 6% (36%) Motherboard I/O Devices COST v. PRICE Q: What % of company income on Research and Development (R&D)? +50–80% Average Discount (33–45%) gross margin (33–14%) direct costs direct costs (8–10%) component cost component cost (25–31%) avg. selling price +25–100% Gross Margin +33% Direct Costs Component Cost Input: chips, displays, ... component cost Making it: labor, scrap, returns, ... (WS–PC) list price Overhead: R&D, rent, marketing, profits, ... Commission: channel profit, volume discounts, Performance Purchasing perspective given a collection of machines, which has the best performance ? least cost ? best performance / cost ? Design perspective faced with design options, which has the best performance improvement ? least cost ? best performance / cost ? Both require basis for comparison metric for evaluation Our goal is to understand cost & performance implications of architectural choices Two notions of “performance” Plane DC to Paris Speed Passengers Throughput Boeing 747 6.5 hours 610 mph 470 286,700 BAD/Sud Concorde 3 hours 1350 mph 132 178,200 Which has higher performance? ° Time to do the task (Execution Time) – execution time, response time, latency ° Tasks per day, hour, week, sec, ns. .. – throughput, bandwidth Response time and throughput often are in opposition Example • Time of Concorde vs. Boeing 747? • Concord is 1350 mph / 610 mph = 2.2 times faster = 6.5 hours / 3 hours • Throughput of Concorde vs. Boeing 747 ? • Concord is 178,200 pmph / 286,700 pmph = 0.62 “times faster” • Boeing is 286,700 pmph / 178,200 pmph = 1.6 “times faster” • Boeing is 1.6 times (“60%”)faster in terms of throughput • Concord is 2.2 times (“120%”) faster in terms of flying time We will focus primarily on execution time for a single job Definitions Performance is in units of things-per-second bigger is better If we are primarily concerned with response time performance(x) = 1 execution_time(x) " X is n times faster than Y" means Performance(X) n = ---------------------Performance(Y) Performance measurement How about collection of programs? Example: Three machines: A, B and C. Two Programs: P1 and P2. A B C P1 1 10 20 P2 1000 100 20 Arithmetic mean: Weight i * Time i W(1) 500.5 55 20 W(2) 91.9 18 20 W(3) 2 10 20 W(1) W(2) W(3) .5 .9 .99 .5 .1 .01 Performance measurement Other option: Geometric Means (Self study pages 37-39 text book) Metrics of performance Answers per month Operations per second Application Programming Language Compiler ISA (millions) of Instructions per second – MIPS (millions) of (F.P.) operations per second – MFLOP/s Datapath Control Megabytes per second Function Units Transistors Wires Pins Cycles per second (clock rate) Relating Processor Metrics CPU execution time = CPU clock cycles X clock cycle time or CPU execution time = CPU clock cycles ÷ clock rate CPU clock cycles= Instructions X avg. clock cycles per instr. or CPI = CPU clock cycles÷ Instructions CPI tells us something about the Instruction Set Architecture, the Implementation of that architecture, and the program measured Aspects of CPU Performance CPU time = Seconds Program instr. count Program Compiler Instr. Set Arch. Organization Technology = Instructions x Cycles Program Instruction CPI clock rate x Seconds Cycle Aspects of CPU Performance CPU time = Seconds Program instr count = Instructions x Cycles Program Instruction CPI clock rate Program X Compiler X (x) Instr. Set. X X Organization Technology x Seconds X X X Cycle Organizational Trade-offs Application Programming Language Compiler ISA Datapath Control Instruction Mix CPI Function Units Transistors Wires Pins Cycle Time CPI “Average cycles per instruction” CPI = (CPU Time * Clock Rate) / Instruction Count = Clock Cycles / Instruction Count n CPU time = ClockCycleTime * SUM i =1 CPI i * Ii "instruction frequency" n CPI = SUM CPI i * i =1 F i where Fi Invest Resources where time is Spent! = Ii Instruction Count Example (RISC processor) Base Machine (Reg / Reg) Op Freq Cycles CPI(i) ALU 50% 1 .5 Load 20% 5 1.0 Store 10% 3 .3 Branch 20% 2 .4 2.2 % Time 23% 45% 14% 18% Typical Mix How much faster would the machine be if a better data cache reduced the average load time to 2 cycles? How does this compare with using branch prediction to shave a cycle off the branch time? What if two ALU instructions could be executed at once? Example (RISC processor) Base Machine (Reg / Reg) Op Freq Cycles CPI(i) ALU 50% 1 .5 Load 20% 5 1.0 Store 10% 3 .3 Branch 20% 2 .4 2.2 % Time 23% 45% 14% 18% Typical Mix How much faster would the machine be if: A) Loads took “0” cycles? B) Stores took “0” cycles? C) ALU ops took “0” cycles? D)Branches took “0” cycles? MAKE THE COMMON CASE FAST Amdahl's Law Speedup due to enhancement E: ExTime w/o E Speedup(E) = -------------------ExTime w/ E Performance w/ E = --------------------Performance w/o E Suppose that enhancement E accelerates a fraction F of the task by a factor S and the remainder of the task is unaffected then, ExTime(with E) = ((1-F) + F/S) X ExTime(without E) Speedup(with E) = ExTime(without E) ÷ ((1-F) + F/S) X ExTime(without E) Speedup(with E) =1/ ((1-F) + F/S) Amdahl's Law-example A new CPU makes Web serving 10 times faster. The old CPU spent 40% of the time on computation and 60% on waiting for I/O. What is the overall enhancement? Fraction enhanced= 0.4 Speedup enhanced = 10 Speedup overall = 1 0.6 +0.4/10 = 1.56 Example from Quiz 1-2004 a)A program consists of 80% initialization code and of 20% code being the main iteration loop, which is run 1000 times. The total runtime of the program is 100 seconds. Calculate the fraction of the total run time needed for the initialization and the iteration. Which part would you optimize? B) The program should have a total run time of 60 seconds. How can this be achieved? (15 points) Marketing Metrics = Instruction Count / Time * 10^6 MIPS = Clock Rate / CPI * 10^6 •machines with different instruction sets ? •programs with different instruction mixes ? • dynamic frequency of instructions • uncorrelated with performance GFLOPS = FP Operations / Time * 10^9 •machine dependent •often not where time is spent playstation: 6.4 GFLOPS Why Do Benchmarks? How we evaluate differences Different systems Changes to a single system Provide a target Benchmarks should represent large class of important programs Improving benchmark performance should help many programs For better or worse, benchmarks shape a field Good ones accelerate progress good target for development Bad benchmarks hurt progress help real programs v. sell machines/papers? Inventions that help real programs don’t help benchmark Basis of Evaluation Cons • representative Actual Target Workload • portable • widely used • improvements useful in reality • easy to run, early in design cycle • identify peak capability and potential bottlenecks • very specific • non-portable • difficult to run, or measure • hard to identify cause •less representative Full Application Benchmarks Small “Kernel” Benchmarks Microbenchmarks • easy to “fool” • “peak” may be a long way from application performance Successful Benchmark: SPEC 1987 RISC industry mired in “bench marketing”: (“That is 8 MIPS machine, but they claim 10 MIPS!”) EE Times + 5 companies band together to perform Systems Performance Evaluation Committee (SPEC) in 1988: Sun, MIPS, HP, Apollo, DEC Create standard list of programs, inputs, reporting: some real programs, includes OS calls, some I/O SPEC first round First round 1989; 10 programs, single number to summarize performance One program: 99% of time in single line of code New front-end compiler could improve dramatically 800 700 500 400 300 200 100 Benchmark tomcatv fpppp matrix300 eqntott li nasa7 doduc spice epresso 0 gcc SPEC Perf 600 SPEC95 Eighteen application benchmarks (with inputs) reflecting a technical computing workload Eight integer go, m88ksim, gcc, compress, li, ijpeg, perl, vortex Ten floating-point intensive tomcatv, swim, su2cor, hydro2d, mgrid, applu, turb3d, apsi, fppp, wave5 Must run with standard compiler flags eliminate special undocumented incantations that may not even generate working code for real programs Summary CPU time = Seconds Program = Instructions x Cycles Program Instruction x Seconds Cycle Time is the measure of computer performance! Good products created when have: Good benchmarks Good ways to summarize performance If not good benchmarks and summary, then choice between improving product for real programs vs. improving product to get more sales=> sales almost always wins Remember Amdahl’s Law: Speedup is limited by unimproved part of program Readings & More… Reminder: READ: TEXTBOOK: Chapter 1 pages 1 to 47 Moore paper (posted on course web site).