logo资料库

Digital Integrated Circuits A Design Perspective.pdf

第1页 / 共513页
第2页 / 共513页
第3页 / 共513页
第4页 / 共513页
第5页 / 共513页
第6页 / 共513页
第7页 / 共513页
第8页 / 共513页
资料共513页,剩余部分请下载后查看
Capitolo1
Capitolo2
Capitolo3
Capitolo4
Capitolo5
Capitolo6
Capitolo7
Capitolo9
Capitolo10
AppendiceA
AppendiceE
EserciziCapitolo3
EserciziCapitolo4
EserciziCapitolo5
EserciziCapitolo6
EserciziCapitolo7
EserciziCapitolo10
EserciziCapitolo11
EserciziCapitolo12
Varie
Indice
chapter1.fm Page 9 Friday, January 18, 2002 8:58 AM C H A P T E R 1 I N T R O D U C T I O N The evolution of digital circuit design n Compelling issues in digital circuit design How to measure the quality of a design n n Valuable references 1.1 A Historical Perspective 1.2 Issues in Digital Integrated Circuit Design 1.3 Quality Metrics of a Digital Design 1.4 Summary 1.5 To Probe Further 9
chapter1.fm Page 10 Friday, January 18, 2002 8:58 AM 10 INTRODUCTION Chapter 1 1.1A Historical Perspective The concept of digital data manipulation has made a dramatic impact on our society. One has long grown accustomed to the idea of digital computers. Evolving steadily from main- frame and minicomputers, personal and laptop computers have proliferated into daily life. More significant, however, is a continuous trend towards digital solutions in all other areas of electronics. Instrumentation was one of the first noncomputing domains where the potential benefits of digital data manipulation over analog processing were recognized. Other areas such as control were soon to follow. Only recently have we witnessed the con- version of telecommunications and consumer electronics towards the digital format. Increasingly, telephone data is transmitted and processed digitally over both wired and wireless networks. The compact disk has revolutionized the audio world, and digital video is following in its footsteps. The idea of implementing computational engines using an encoded data format is by no means an idea of our times. In the early nineteenth century, Babbage envisioned large- scale mechanical computing devices, called Difference Engines [Swade93]. Although these engines use the decimal number system rather than the binary representation now common in modern electronics, the underlying concepts are very similar. The Analytical Engine, developed in 1834, was perceived as a general-purpose computing machine, with features strikingly close to modern computers. Besides executing the basic repertoire of operations (addition, subtraction, multiplication, and division) in arbitrary sequences, the machine operated in a two-cycle sequence, called “store” and “mill” (execute), similar to current computers. It even used pipelining to speed up the execution of the addition opera- tion! Unfortunately, the complexity and the cost of the designs made the concept impracti- cal. For instance, the design of Difference Engine I (part of which is shown in Figure 1.1) required 25,000 mechanical parts at a total cost of £17,470 (in 1834!). Figure 1.1 Working part of Babbage’s Difference Engine I (1832), the first known automatic calculator (from [Swade93], courtesy of the Science Museum of London).
chapter1.fm Page 11 Friday, January 18, 2002 8:58 AM Section 1.1 A Historical Perspective 11 The electrical solution turned out to be more cost effective. Early digital electronics systems were based on magnetically controlled switches (or relays). They were mainly used in the implementation of very simple logic networks. Examples of such are train safety systems, where they are still being used at present. The age of digital electronic computing only started in full with the introduction of the vacuum tube. While originally used almost exclusively for analog processing, it was realized early on that the vacuum tube was useful for digital computations as well. Soon complete computers were realized. The era of the vacuum tube based computer culminated in the design of machines such as the ENIAC (intended for computing artillery firing tables) and the UNIVAC I (the first successful commercial computer). To get an idea about integration density, the ENIAC was 80 feet long, 8.5 feet high and several feet wide and incorporated 18,000 vacuum tubes. It became rapidly clear, however, that this design technology had reached its limits. Reliability problems and excessive power consumption made the implementation of larger engines economically and practically infeasible. All changed with the invention of the transistor at Bell Telephone Laboratories in 1947 [Bardeen48], followed by the introduction of the bipolar transistor by Schockley in 1949 [Schockley49]1. It took till 1956 before this led to the first bipolar digital logic gate, introduced by Harris [Harris56], and even more time before this translated into a set of integrated-circuit commercial logic gates, called the Fairchild Micrologic family [Norman60]. The first truly successful IC logic family, TTL (Transistor-Transistor Logic) was pioneered in 1962 [Beeson62]. Other logic families were devised with higher perfor- mance in mind. Examples of these are the current switching circuits that produced the first subnanosecond digital gates and culminated in the ECL (Emitter-Coupled Logic) family [Masaki74]. TTL had the advantage, however, of offering a higher integration density and was the basis of the first integrated circuit revolution. In fact, the manufacturing of TTL components is what spear-headed the first large semiconductor companies such as Fair- child, National, and Texas Instruments. The family was so successful that it composed the largest fraction of the digital semiconductor market until the 1980s. Ultimately, bipolar digital logic lost the battle for hegemony in the digital design world for exactly the reasons that haunted the vacuum tube approach: the large power con- sumption per gate puts an upper limit on the number of gates that can be reliably integrated on a single die, package, housing, or box. Although attempts were made to develop high integration density, low-power bipolar families (such as I2L—Integrated Injection Logic [Hart72]), the torch was gradually passed to the MOS digital integrated circuit approach. The basic principle behind the MOSFET transistor (originally called IGFET) was proposed in a patent by J. Lilienfeld (Canada) as early as 1925, and, independently, by O. Heil in England in 1935. Insufficient knowledge of the materials and gate stability prob- lems, however, delayed the practical usability of the device for a long time. Once these were solved, MOS digital integrated circuits started to take off in full in the early 1970s. Remarkably, the first MOS logic gates introduced were of the CMOS variety [Wanlass63], and this trend continued till the late 1960s. The complexity of the manufac- turing process delayed the full exploitation of these devices for two more decades. Instead, 1 An intriguing overview of the evolution of digital integrated circuits can be found in [Murphy93]. (Most of the data in this overview has been extracted from this reference). It is accompanied by some of the his- torically ground-breaking publications in the domain of digital IC’s.
chapter1.fm Page 12 Friday, January 18, 2002 8:58 AM 12 INTRODUCTION Chapter 1 the first practical MOS integrated circuits were implemented in PMOS-only logic and were used in applications such as calculators. The second age of the digital integrated cir- cuit revolution was inaugurated with the introduction of the first microprocessors by Intel in 1972 (the 4004) [Faggin72] and 1974 (the 8080) [Shima74]. These processors were implemented in NMOS-only logic, which has the advantage of higher speed over the PMOS logic. Simultaneously, MOS technology enabled the realization of the first high- density semiconductor memories. For instance, the first 4Kbit MOS memory was intro- duced in 1970 [Hoff70]. These events were at the start of a truly astounding evolution towards ever higher integration densities and speed performances, a revolution that is still in full swing right now. The road to the current levels of integration has not been without hindrances, how- ever. In the late 1970s, NMOS-only logic started to suffer from the same plague that made high-density bipolar logic unattractive or infeasible: power consumption. This realization, combined with progress in manufacturing technology, finally tilted the balance towards the CMOS technology, and this is where we still are today. Interestingly enough, power consumption concerns are rapidly becoming dominant in CMOS design as well, and this time there does not seem to be a new technology around the corner to alleviate the problem. Although the large majority of the current integrated circuits are implemented in the MOS technology, other technologies come into play when very high performance is at stake. An example of this is the BiCMOS technology that combines bipolar and MOS devices on the same die. BiCMOS is used in high-speed memories and gate arrays. When even higher performance is necessary, other technologies emerge besides the already men- tioned bipolar silicon ECL family—Gallium-Arsenide, Silicon-Germanium and even superconducting technologies. These technologies only play a very small role in the over- all digital integrated circuit design scene. With the ever increasing performance of CMOS, this role is bound to be further reduced with time. Hence the focus of this textbook on CMOS only. 1.2Issues in Digital Integrated Circuit Design Integration density and performance of integrated circuits have gone through an astound- ing revolution in the last couple of decades. In the 1960s, Gordon Moore, then with Fair- child Corporation and later cofounder of Intel, predicted that the number of transistors that can be integrated on a single die would grow exponentially with time. This prediction, later called Moore’s law, has proven to be amazingly visionary [Moore65]. Its validity is best illustrated with the aid of a set of graphs. Figure 1.2 plots the integration density of both logic IC’s and memory as a function of time. As can be observed, integration com- plexity doubles approximately every 1 to 2 years. As a result, memory density has increased by more than a thousandfold since 1970. An intriguing case study is offered by the microprocessor. From its inception in the early seventies, the microprocessor has grown in performance and complexity at a steady and predictable pace. The transistor counts for a number of landmark designs are collected in Figure 1.3. The million-transistor/chip barrier was crossed in the late eighties. Clock frequencies double every three years and have reached into the GHz range. This is illus-
chapter1.fm Page 13 Friday, January 18, 2002 8:58 AM Section 1.2 Issues in Digital Integrated Circuit Design 1010 1010 109 109 108 108 107 107 106 106 105 105 104 104 i i p p h h c c r r e e p p s s t t i i b b f f o o r r e e b b m m u u N N 13 64 Gbits 64 Gbits 0.08µm* 0.08µm* Human memory Human memory Human memory Human memory Human DNA Human DNA Human DNA Human DNA 4 Gbits 4 Gbits 0.15µm 0.15µm 1 Gbits 1 Gbits 0.15-0.2µm 0.15-0.2µm 256 Mbits 256 Mbits 0.25-0.3µm 0.25-0.3µm 64 Mbits 64 Mbits 0.35-0.4µm 0.35-0.4µm Book Book Book Book 16 Mbits 16 Mbits 0.5-0.6µm 0.5-0.6µm Encyclopedia Encyclopedia Encyclopedia Encyclopedia 2 hrs CD Audio 2 hrs CD Audio 2 hrs CD Audio 2 hrs CD Audio 30 sec HDTV 30 sec HDTV 30 sec HDTV 30 sec HDTV 2000 2000 2010 2010 4 Mbits 4 Mbits 0.7-0.8µm 0.7-0.8µm 1 Mbits 1 Mbits 1.0-1.2µm 1.0-1.2µm 256 Kbits 256 Kbits 1.6-2.4µm 1.6-2.4µm 64 Kbits 64 Kbits 1970 1970 1980 1980 Page Page Page Page 1990 1990 Year Year (a) Trends in logic IC complexity (b) Trends in memory complexity Figure 1.2 Evolution of integration complexity of logic ICs and memories as a function of time. trated in Figure 1.4, which plots the microprocessor trends in terms of performance at the beginning of the 21st century. An important observation is that, as of now, these trends have not shown any signs of a slow-down. It should be no surprise to the reader that this revolution has had a profound impact on how digital circuits are designed. Early designs were truly hand-crafted. Every transis- tor was laid out and optimized individually and carefully fitted into its environment. This is adequately illustrated in Figure 1.5a, which shows the design of the Intel 4004 micro- processor. This approach is, obviously, not appropriate when more than a million devices have to be created and assembled. With the rapid evolution of the design technology, time-to-market is one of the crucial factors in the ultimate success of a component. 100000000 100000000 10000000 10000000 s s r r o o t t s s i i s s n n a a r r T T 1000000 1000000 100000 100000 Pentium 4 Pentium 4 Pentium III Pentium III Pentium II Pentium II Pentium ® Pentium ® 486 486 386 386 286 ™ 286 ™ 8086 8086 10000 10000 4004 4004 1000 1000 8080 8080 8008 8008 1970 1970 1975 1975 1985 1985 1990 1980 1980 1990 Year of Introduction Year of Introduction 1995 1995 2000 2000 Figure 1.3 Historical evolution of microprocessor transistor count (from [Intel01]).
chapter1.fm Page 14 Friday, January 18, 2002 8:58 AM 14 10000 10000 1000 1000 100 100 10 10 1 1 ) ) z z h h M M ( ( y y c c n n e e u u q q e e r r F F Doubles every Doubles every 2 years 2 years P6 P6 Pentium ® proc Pentium ® proc 486 486 386 386 286 286 8086 8086 8085 8085 8080 8080 INTRODUCTION Chapter 1 8008 8008 4004 4004 0.1 0.1 1970 1970 1980 1980 1990 1990 Year Year 2000 2000 2010 2010 Figure 1.4 Microprocessor performance trends at the beginning of the 21st century. Designers have, therefore, increasingly adhered to rigid design methodologies and strate- gies that are more amenable to design automation. The impact of this approach is apparent from the layout of one of the later Intel microprocessors, the Pentium® 4, shown in Figure 1.5b. Instead of the individualized approach of the earlier designs, a circuit is constructed in a hierarchical way: a processor is a collection of modules, each of which consists of a number of cells on its own. Cells are reused as much as possible to reduce the design effort and to enhance the chances for a first-time-right implementation. The fact that this hierar- chical approach is at all possible is the key ingredient for the success of digital circuit design and also explains why, for instance, very large scale analog design has never caught on. The obvious next question is why such an approach is feasible in the digital world and not (or to a lesser degree) in analog designs. The crucial concept here, and the most important one in dealing with the complexity issue, is abstraction. At each design level, the internal details of a complex module can be abstracted away and replaced by a black box view or model. This model contains virtually all the information needed to deal with the block at the next level of hierarchy. For instance, once a designer has implemented a multiplier module, its performance can be defined very accurately and can be captured in a model. The performance of this multiplier is in general only marginally influenced by the way it is utilized in a larger system. For all purposes, it can hence be considered a black box with known characteristics. As there exists no compelling need for the system designer to look inside this box, design complexity is substantially reduced. The impact of this divide and conquer approach is dramatic. Instead of having to deal with a myriad of elements, the designer has to consider only a handful of components, each of which are characterized in performance and cost by a small number of parameters. This is analogous to a software designer using a library of software routines such as input/output drivers. Someone writing a large program does not bother to look inside those library routines. The only thing he cares about is the intended result of calling one of those modules. Imagine what writing software programs would be like if one had to fetch every bit individually from the disk and ensure its correctness instead of relying on handy “file open” and “get string” operators.
chapter1.fm Page 15 Friday, January 18, 2002 8:58 AM Section 1.2 Issues in Digital Integrated Circuit Design 15 (a) The 4004 microprocessor Standard Cell Module Memory Module (b) The Pentium ® 4 microprocessor Figure 1.5 Comparing the design methodologies of the Intel 4004 (1971) and Pentium ® 4 (2000 microprocessors (reprinted with permission from Intel).
chapter1.fm Page 16 Friday, January 18, 2002 8:58 AM 16 INTRODUCTION Chapter 1 Typically used abstraction levels in digital circuit design are, in order of increasing abstraction, the device, circuit, gate, functional module (e.g., adder) and system levels (e.g., processor), as illustrated in Figure 1.6. A semiconductor device is an entity with a + SYSTEM MODULE GATE CIRCUIT G DEVICE D n+ S n+ Figure 1.6 Design abstraction levels in digital circuits. very complex behavior. No circuit designer will ever seriously consider the solid-state physics equations governing the behavior of the device when designing a digital gate. Instead he will use a simplified model that adequately describes the input-output behavior of the transistor. For instance, an AND gate is adequately described by its Boolean expres- sion (Z = A.B), its bounding box, the position of the input and output terminals, and the delay between the inputs and the output. This design philosophy has been the enabler for the emergence of elaborate com- puter-aided design (CAD) frameworks for digital integrated circuits; without it the current design complexity would not have been achievable. Design tools include simulation at the various complexity levels, design verification, layout generation, and design synthesis. An overview of these tools and design methodologies is given in Chapter 8 of this textbook. Furthermore, to avoid the redesign and reverification of frequently used cells such as basic gates and arithmetic and memory modules, designers most often resort to cell libraries. These libraries contain not only the layouts, but also provide complete docu- mentation and characterization of the behavior of the cells. The use of cell libraries is, for
分享到:
收藏