Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
How Moore's Law Works by Jonathan Strickland Gordon Moore at Intel's headquarters AP Photo/Paul Sakuma Introduction to How Moore's Law Works There's a joke about personal computers that has been around almost as long as the devices have been on the market: You buy a new computer, take it home and just as you finish unpacking it you see an advertisement for a new computer that makes yours obsolete. If you're the kind of person who demands to have the fastest, most powerful machines, it seems like you're destined for frustration and a lot of trips to the computer store. While the joke is obviously an exaggeration, it's not that far off the mark. Even one of today's modest personal computers has more processing power and storage space than the famous Cray-1 supercomputer. In 1976, the Cray-1 was state-of-the-art: it could process 160 million floating-point operations per second (flops) and had 8 megabytes (MB) Floating-point numbers of memory. have decimal points in Today, many personal computers can perform more than 10 times that number of them. Integers are simpler floating-point operations in a second and have 100 times the amount of memory. to work with and require less computing power. Meanwhile on the supercomputer front, the Cray XT5 Jaguar at the Oak Ridge National Laborat ory performed a sustained 1.4 petaflops in 2008 [source: Cray]. The prefix peta means 10 to the 15th power -- in other words, one quadrillion. That means the Cray XT5 can process 8.75 million times more flops than the Cray-1. It only took a little over three decades to reach that milestone. Floating-point operation example: 3.99233 x 45.210009 = Integer operation 10 x 13 = 130 If you were to chart the evolution of the computer in terms of processing power, you would see that progress has been exponential. The man who first made this famous observation is Gordon Moore, a co-founder of the microprocessor company Intel. Computer scientists, electrical engineers, manufacturers and journalists extrapolated Moore's Law from his original observation. In general, most people interpret Moore's Law to mean the number of transistors on a 1-inch (2.5 centimeter) diameter of silicon doubles every x number of months. The number of months shifts as conditions in the microprocessor market change. Some people say it takes 18 months and others say 24. Some interpret the law to be about the doubling of processing power, not the number of transistors. And the law sometimes seems to be more of a self-fulfilling prophecy than an actual law, principle or observation. To understand why, it's best to go back to the beginning. A replica of the first transistor AP Photo/Paul Sakuma Semiconductors, Transistors and Integrated Circuits The discovery of semiconductors, the invention of transistors and the creation of the integrated circuit are what make Moore's Law -- and by extension modern electronics -possible. Before the invention of the transistor, the most widely-used element in electronics was the vacuum tube. Electrical engineers used vacuum tubes to amplify electrical signals. But vacuum tubes had a tendency to break down and they generated a lot of heat, too. Bell Laboratories began looking for an alternative to vacuum tubes to stabilize and strengthen the growing national telephone network in the 1930s. In 1945, the lab concentrated on finding a way to take advantage of semiconductors. A semiconductor is a material that can act as both a conductor and an insulator. Conductors are materials that permit the flow of electrons -- they conduct electricity. Insulators have an atomic structure that inhibits electron flow. Semiconductors can do both. The control of the flow of electrons is what makes electronics work. Finding a way to harness the Gordon Moore holds a vacuum tube, the precursor to transistors. unique nature of semiconductors became a high priority for Bell Labs. In 1947, John Bardeen and Walter Brattain built the first working transistor. The transistor is a device designed to control electron flows -- it has a gate that, when closed, prevents electrons from flowing through the transistor. This basic idea is the foundation for the way practically all electronics work. Early transistors were huge compared to the transistors manufacturers produce today. The very first one was half an inch (1.3 centimeters) tall. But once engineers learned how to build a working transistor, the race was on to build them better and smaller. For the first few years, transistors existed only in scientific laboratories as engineers improved the design. In 1958, Jack Kilby made the next huge contribution to the world of electronics: the integr ated circuit. Earlier electric circuits consisted of a series of individual components. Electrical engineers would construct each piece and then attach them to a foundation called a substrate. Kilby experimented with building a circuit out of a single piece of semiconductor material and overlaying the metal parts necessary to connect the different pieces of circuitry on top of it. The result was an integrated circuit. Moore's Observation The physicist Jean Hoerni developed the technique for creating planar transistors in 1959. By 1961, a company called Fairchild Semiconductor produced the first planar integrated circuit. From that moment on, the technology advanced rapidly. Physicists and engineers found new and more efficient ways to create integrated circuits. They refined the processes they used to make components smaller and more compact. This meant they could fit more transistors on a single semiconductor wafer than previous generations of the technology. During this time, the director for research and development at Fairchild was Gordon Moore. Electronics magazine asked Moore to predict what would happen over the next 10 years of development in the field of electronics. Moore wrote an article with the snappy title "Cramming more components onto integrated circuits." The magazine published the article on April 19, 1965. Moore based his prediction upon the rapid development of the industry since the introduction of the integrated circuit. He saw that as techniques improved and components on circuits shrank, the price for producing an individual component dropped. Semiconductor companies had an incentive to refine their production techniques -- not only were the new circuits more powerful, the individual elements were more cost efficient. As long as that relationship held true, Moore said the trend would continue. Moore also pointed out that over a 12-month period, the number of components on a one-inch (2.5-centimeter) diameter semiconductor wafer doubled. Moore attributed this to two major trends: companies were finding ways to make smaller components and they were getting better at arranging the components to conserve space on the wafer. Moore extrapolated his prediction of the next 10 years by looking at this data. At the time of the article, Moore said that the optimal number of components on a circuit was 50. He projected that by 1975, that number would be closer to 65,000. The prediction held -- by 1975, integrated circuits had nearly 65,000 components. THE POWER OF CIRCUITS One loose interpretation of Moore's law disregards the emphasis on how many components go onto a substrate. This version of the law states that processing power for CPU chips doubles every 18 months. That's the neat thing about Moore's Law -- the meaning shifts depending upon whom you ask. Moore said the trend he predicted 10 years earlier would progress at that same rate for at least a few more years. But Moore also said that he believed the semiconductor industry was approaching the limit for some techniques, such as conserving space on a circuit. He called this factor "circuit cleverness." He believed that we'd reach a limit on how clever we could arrange components -- eventually we'd have the optimal use of space. Once that factor is removed from the equation, the rate of advancements must slow down. He said he believed after a few years components would double only every 24 months. While Moore's original observation focused on technological advances and the economics behind producing circuits, many people reduce his observation to the simple statement we call Moore's Law. The most common version of Moore's Law is that the number of transistors on a circuit doubles every 18 (or 24) months. Remarkably, this prediction has held true -- today, Intel's Core i7 microprocessor has 731 million transistors, while its Xeon processor has 1.9 billion transistors [source: Intel]. Cramming more components on an integrated circuit doesn't just mean devices are becoming more powerful -- it also means they're getting smaller. The tiny components on compact integrated circuits power all sorts of portable electronic devices. Even a small microprocessor chip today is as powerful as a full-sized chip was a few years ago. The advances in circuit production make devices like smartphones and netbooks possible. Moore's Law in Action Why have Moore's observations and predictions held true over so many decades? Moore's Law isn't really a law at all -in fact, there's no fundamental law of physics behind it. Moore's Law only holds true because of the actions of human beings. But what keeps the cycle going even as the challenge to make more powerful circuits grows? Much of the reason is psychological and driven by the market. Companies that make integrated circuits are competing against each other and everyone knows about Moore's Law. That means every corporate executive has this in mind: if our company doesn't double the power of our circuits in 18 months, another company will beat us to it. Because companies don't want to give an edge to competitors, they pour a lot of money into research and development (R&D). These R&D divisions work to develop new techniques to create smaller components and arrange them in such a way that maximizes their performance. It costs a lot of money to keep up the cycle of research, but this cost is balanced against the threat of competitors gaining a foothold and dominating the market. Another factor is the simple desire to overcome a challenge. Many people have predicted the end of Moore's Law over the years. Some people thought it would come to an end during the 1980s. Others said the same thing in the mid '90s. It seemed like engineers would eventually bump up against a barrier that would be fundamentally impossible to breach. But engineers somehow manage to find a solution each time, keeping Moore's Law alive. Consumers also drive Moore's Law. The rapid development of electronics has created a sense of expectation among consumers. Every year, faster and more advanced electronics hit the market. From the consumer's point of view, there's no reason not to expect something better next year. Moore's Law and the Nanoscale Today, transistors on integrated circuits have reached a size so small that it would take more than 2,000 of them stacked next to each other to equal the thickness of a human hair. The transistors on Intel's latest chips are only 45 nanometers wide -- the average human hair is about 100,000 nanometers thick [source: National Nanotechnology Initiative]. Creating such narrow transistors is an amazing achievement. Visible light has wavelengths in the range of 400-700 nanometers. Conventional light microscope lenses can only focus on objects half the size of the wavelength of visible light or larger. You have to rely on special equipment like scanning electron microscopes to create an image of something on so small a scale. One thing you must consider when dealing with such small devices is that as you approach the nanoscale, you leave behind the world of classical physics and enter the realm of quantum mechanics. The rules of physics in the quantum world are very different from the way things work on the macro scale. For example, quantum particles like electrons can pass through extremely thin walls even if they don't have the kinetic energy necessary to break through the barriers. Quantum physicists call this phenomenon quantum tunneling. A nanometer is 1 billionth of a meter. The diameter of a helium atom, for example, is about 0.1 nm, and that of a ribosome is about 20 nm. Because electronics depend upon controlling the flow of electrons to work, issues like quantum tunneling create serious problems. These problems force electrical engineers to re-evaluate the way they design circuits. In some cases, shifting to different materials solves the issue. In others, finding a completely new way to build circuits might work. There's even the possibility that someone will come up with a revolutionary idea that makes the transistor and integrated circuit obsolete. While that may sound far-fetched, the fact remains that despite numerous tech pundits and engineers pronouncing the end of Moore's Law, circuit manufacturers are still finding ways to keep it going. As it turns out, the challenges may not be quite as impossible as some believe. The Future of Moore's Law Even Gordon Moore has questioned how long the cycle of innovation and production can keep up the frenzied pace of the last four decades. He has also expressed amazement at the way companies like Intel find new ways to work around what initially seemed like an insurmountable problem. Will there ever be an end to Moore's Law? The answer is yes, but it's difficult to pin down when that might happen. For one, we could hit a technical barrier that prevents engineers from finding a way to make smaller components. But even if we don't encounter a technical barrier, economics could come into the equation. If it's not economically feasible to produce circuits with smaller transistors there may be no reason to pursue further development. Or we could bump up against the fundamental laws of physics -- like the speed of light, for instance. The problem with predicting a specific date when one or more of these barriers will stop progress is that we have to base it on what we know today. But every day engineers are learning new ways to design, build and produce circuits. What we know tomorrow may make the things that seem impossible today completely achievable. Is Moore's Law even relevant today? The era of the personal computer has been dominated by a sense that the consumer needs the latest and greatest machine on the market. But today, some people are questioning that philosophy. Part of that is due to changes in consumer behavior -- many computer owners use their computers for simple tasks like browsing the Web or sending e-mail. These applications don't put a heavy demand on the computer's hardware. Another reason powerful PCs aren't as necessary is the rise in popularity of cloud computing. Cloud computing shifts the burden of processing and storing data to a network of computers. Users can access applications and information using the Internet, so they don't necessarily need a powerful machine of their own to take advantage of cloud computing. As a result, devices like smartphones and netbooks are becoming more popular. These devices don't have the raw processing power of the latest desktop and laptop computers. But they still allow users to access the applications and data they need. Lessons Learned from Moore's Law If consumers continue to purchase devices like smartphones and netbooks, microprocessor manufacturers will have less of an incentive to meet the expectations of Moore's Law. If there's no market for ultra-powerful processors, then we've hit the economic barrier that could bring an end to the cycle. That said, some facilities may still push the limits of integrated circuit production. While the average consumer may not see the value in a powerful PC, research facilities still rely on the fastest processors in production. More powerful microprocessors can aid in everything from weather prediction to cosmological studies. One lesson we can draw from Moore's Law and the semiconductor industry is that pure research can yield beneficial results for society. The engineers at Bell Laboratories had no guarantee that their experimental work with the earliest transistor models would yield positive results. But their research and hard work spawned an industry that changed the way we live. It's an example of how scientific research can have a dramatic impact on our lives even when there's no obvious or immediate benefit. Perhaps the most important lesson we can take from Moore's Law is that we shouldn’t be too quick to say something is impossible. Henry L. Ellsworth, the commissioner of the U.S. Patent Office in 1843, once said "the advancement of the arts, from year to year, taxes our credulity and seems to presage the arrival of that period when human improvement must end" [source: Sass]. Ellsworth was pointing out that the rate of human inventiveness and innovation was so impressive that it was hard to believe. He was not, as some have implied, suggesting that everything that could be invented had been invented already. And as a matter of fact, the rate of innovation has only increased since then. While we know a great deal about building electronics, there may yet be very much we do not know. Moore's Law helps as a motivational device for inventive engineers. They don't want to disappoint Gordon Moore, even if it means they have to find unique solutions to seemingly impossible problems. Henry L. Ellsworth – U.S. Patten Office Commissioner 1843 A Good Paying Job with Exponential Growth Name _____________________________________________ Class Period __________________________ You have been offered two different jobs. Which would you accept? Do the math below and then circle your choice. 1) Your boss agrees to pay you $50 per hour. You can work 8 hours a day for one month. 2) Your boss agrees to pay you 1 cent per day but will double your pay every day that you work. Job 1 : $50 per hour 1. Write a math expression to calculate the number of hours worked in 30 days. __________________________ 2. How much pay would you earn in 30 days at $50 per hour? ________________________________ Job 2 : $0.01 for the first day. Double salary each day. Day 1 Day 2 Day 3 Day 4 Day 5 Day 6 Day 7 Day 8 Day 9 Day 10 0.01 0.02 0.04 0.08 0.16 Day 11 Day 12 Day 13 Day 14 Day 15 Day 16 Day 17 Day 18 Day 19 Day 20 Day 21 Da y22 Day 22 Da y23 Day 23 Da y24 Day 24 Da y25 Day 25 Da y26 1. Since early days of the pay schedule make little difference in the total amount, add only the last 10 days. You may round numbers to make the math easier. What is the estimated total pay? ______________________ 2. Draw a line chart showing how much money you would earn each day. Work backwards from day 30 to day 1. Pay 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 Day Number