I. Field of the Disclosure
The technology of the disclosure relates generally to computer memories, computer memory design, and related systems and methods for reducing memory power consumption and latency.
II. Background
In a processor-based memory architecture, it is generally desirable to have fast memory access times (i.e. low memory latency). The overall memory latency of a memory may be defined as the worst-case latency to access a memory location in the memory. The resistance of bit and word lines connected between a memory access interface (MAI) and memory cells in memory banks affect memory latency. As length of the bit and word lines increase, so does the resistance, and in turn so does the signal delay on the hit and word lines. Accordingly, memory banks located farther in distance from a memory access interface (MAI) will generally suffer greater resistance delay than memory banks located closer in distance to the memory access interface (MAI). Accordingly, the memory bank located farthest from the memory access interface (MAI) may determine the worst-case latency (i.e. worst case memory access time) of the memory.
In this regard,
Embodiments disclosed in the detailed description include asymmetrically-arranged memories having reduced current leakage and/or latency, and related systems and methods. In this regard in one embodiment, a memory comprises a memory access interface (MAI). The memory further comprises a first memory portion(s) accessible by the MAI. The first memory portion(s) has a first latency and a first current leakage. The memory further comprises a second memory portion(s) accessible by the MAI. The first and second memory portion(s) may be comprised of a memory bank(s) and/or a memory sub-bank(s). To provide an asymmetrical memory arrangement, the first latency of the first memory portion(s) is increased such that the second memory portion(s) has a second latency greater than or equal to the first latency of the first memory portion(s). As a result, the first current leakage of the first memory portion is reduced such that the second memory portion(s) has a second current leakage greater than the first current leakage of the first memory portion(s). In this manner, the overall current leakage of the memory is reduced while not increasing the overall latency of the memory.
As non-limiting examples, the first memory portion(s) may be located a first distance from the MAI, and the second memory portion(s) may be located a second distance greater than the first distance from the MAI. The second latency may be less than the first latency by a first latency differential threshold. The second current leakage may be greater than the first current leakage by a first current leakage differential threshold. The channel length, channel width, and/or threshold voltage (Vt) of memory cell transistors in the first memory portion(s) may be altered to increase latency of the first memory portion(s) and to reduce current leakage in the first memory portion(s) while not increasing the latency of the second memory portion(s) and while also not increasing the overall latency of the memory. In this manner, the overall current leakage of the memory is reduced while the overall latency of the memory is not increased.
In another embodiment, a memory comprises a memory access interface (MAI) means. The memory further comprises a first memory portion(s) means accessible by the MAI means. The first memory portion(s) means has a first latency and a first current leakage. The memory further comprises a second memory portion(s) means accessible by the MAI means. The second memory portion(s) means has a second latency greater than or equal to the first latency and a second current leakage greater than the first current leakage.
In another embodiment, a memory system is provided. The memory system comprises a memory. The memory comprises a MAI. The memory further comprises a first memory portion(s) accessible by the MAI. The first memory portion(s) has a first latency and a first current leakage. The memory further comprises a second memory portion(s) accessible by the MAI. The second memory portion(s) has a second latency greater than or equal to the first latency and a second current leakage greater than the first current leakage. The memory system further comprises a memory controller configured to access the memory through access to the MAI.
In another embodiment, a method of designing a memory is provided. The method comprises providing a memory arrangement. The memory arrangement comprises a MAI. The memory arrangement further comprises symmetric memory banks having symmetric transistor characteristics. The method further comprises measuring latency of a closer memory bank(s) to the MAI. The method further comprises measuring latency of a farther memory bank(s) from the MAI. The method further comprises determining a memory bank latency margin of the closer memory bank(s). The method further comprises, in response to determining that the closer memory bank(s) has a positive memory bank latency margin, modifying transistor characteristics in a memory sub-bank(s) of the closer memory bank(s) to reduce current leakage of the closer memory bank(s).
In another embodiment, a non-transitory computer-readable medium having stored thereon computer-executable instructions is provided. The instructions cause the processor to provide a memory arrangement. The memory arrangement comprises a MAI and symmetric memory portions. The instructions further cause the processor to measure a first latency of a farther memory portion(s) from the MAI. The instructions further cause the processor to measure a second latency of a closer memory portion(s) to the MAI. The instructions further cause the processor to determine latency margin of the closer memory portion(s). The instructions further cause the processor, in response to determining the closer memory portion(s) has positive latency margin, to increase the latency in the closer memory portion(s) to reduce current leakage of the closer memory portion(s).
With reference now to the drawing figures, several exemplary embodiments of the present disclosure are described. The word “exemplary” is used herein to mean “serving as an example, instance, or illustration.” Any embodiment described herein as “exemplary” is not necessarily to be construed as preferred or advantageous over other embodiments.
Embodiments disclosed in the detailed description include asymmetrically-arranged memories having reduced current leakage and/or latency, and related systems and methods. In this regard in one embodiment, a memory comprises a memory access interface (MAI). The memory further comprises a first memory portion(s) accessible by the MAI. The first memory portion(s) has a first latency and a first current leakage. The memory further comprises a second memory portion(s) accessible by the MAI. The first and second memory portion(s) may be comprised of a memory bank(s) or a memory sub-bank(s). To provide an asymmetrical memory arrangement, the first latency of the first memory portion(s) is increased such that the second memory portion(s) has a second latency greater than or equal to the first latency of the first memory portion(s). As a result, the first current leakage of the first memory portion is reduced such that the second memory portion(s) has a second current leakage greater than the first current leakage of the first memory portion(s). In this manner, the overall current leakage of the memory is reduced while not increasing the overall latency of the memory.
In this regard,
With continuing reference to
The memory controller 18 controls the flow of data to and from a memory access interface (MAI) 28(0), 28(X) in the memory chips 20(0)-20(X) via a memory bus 22. In this example, the memory bus 22 includes chip selects (CS(0)-CS(X)) 24(0)-24(X) for each memory chip 20(0)-20(X). The chips selects 24(0)-24(X) are selectively enabled by the memory controller 18 to enable the memory chips 20(0)-20(X) containing the desired memory location to be accessed. The memory bus 22 also includes an address/control bus (ADDR/CTRL) 32 that allows the memory controller 18 to control the memory address accessed through the memory access interfaces (MAIs) 28(0)-28(X) in the memory chips 20(0)-20(X) for either writing or reading data to or from the memory 20. The memory bus 22 also includes a clock signal (CLK) 34 to synchronize timing between the memory controller 18 and the memory chips 20(0)-20(X) for memory accesses.
With continuing reference to
Each memory chip 20(0)-20(X) in this example contains a plurality of memory portions 35. In one embodiment, the memory portions 35 are each memory banks, referred to generally as element 36. A memory bank is a logical unit of memory, in the illustrated example, each memory chip 20(0)-20(X) contains a plurality of memory banks 36(0)-36(Y) (also denoted B0-BY). Each memory bank 36 is organized into a grid-like pattern, with “rows” or memory pages 38 and “columns” 36. The accessed data may be provided by the memory controller 18 over a system bus 46 to another component in a processor-based system. In the illustrated example of
A memory bank 36 may comprise one or more memory “sub-banks” referred to as memory sub-bank(s) 42. A memory sub-bank 42 is comprised of one or more memory pages 38 in a memory bank 36. The memory portions 35 may comprise one or more of the memory sub-bank(s) 36. When a memory bank 36 is comprised of multiple memory sub-banks 42, each memory sub-bank 42 may comprise a same or different number of memory pages 38 than other memory sub-banks 42 of the memory bank 36.
It may be important to conserve power in the memory system 16 in
In this regard,
A “symmetric” or “symmetrically-arranged” memory contains two or more memory portions which have the same or substantially the same internal latency characteristics. The internal latency characteristics of a memory portion are the latency characteristics that are independent of the distance of the memory portion from a MAI. Only by these memory portions being located different distances away from the MAI do memory accesses to these memory portions encounter different memory access latencies. “Internal latency” of a memory portion is the latency caused by the internal latency characteristics of the memory portion.
“Memory access latency” and/or “memory access time” of a memory portion is the latency (i.e. time) for accessing a memory portion though a MAI, which comprises internal latency of the MAI, latency (as a non-limiting example, line delays) due to the distance of the memory portion from the MAI, and internal latency of the memory portion.
With reference back to
Accordingly, in one embodiment memory portion 44(0) has been modified to have an internal latency characteristic greater than the internal latency characteristic of memory portion 44(M) by a first latency differential threshold. In other words, due to the modifications to increase the internal latency of memory portion 44(0), memory portion 44(0) has an increased internal latency compared to memory portion 44(M) by at least the first latency differential threshold. Further due to these modifications, the current leakage of memory portion 44(M) is greater than the current leakage of memory portion 14(0).
There are various methods of increasing internal latency of a memory portion(s) (as a non-limiting example, memory portion 44(0)) to lower current leakage. For example, a transistor characteristic(s) of memory cell transistors of a memory portion(s) may be modified to tradeoff increased internal latency for reduced current leakage. In this regard, TABLE 1 below illustrates various transistor characteristics, which may be modified to affect the current leakage and internal latency of the memory portion(s). TABLE 1 illustrates effects of modifying memory cell transistor channel length (L), memory cell transistor channel width (W), and memory cell transistor threshold voltage (Vt). In addition, TABLE 1 illustrates effects of selecting among HVt, NVt, or LVt memory cell transistors to provide the memory portion(s). TABLE 1 also illustrates the effects of biasing the body (B) terminal of the memory cell transistors. Table 1 illustrates various effects of modifying the above-mentioned characteristics, including: whether the modification increases (+) or decreases (−) drain-source conductance (GDS) of the induced channels of the memory cell transistors of the memory portion(s); whether the modification increases (+) or decreases (−) drain-source resistance (RDS) of the induced channels of the memory cell transistors of the memory portion(s); whether the modification increases (+) or decreases (−) current leakage of the memory portion(s); and whether the modification increases (+) or decreases (−) internal latency of the memory portion(s).
As shown in TABLE 1 above, various transistor characteristics may be modified to provide an increased internal latency and reduced current leakage for first memory portion 44(0). In this regard, memory cell transistors of first memory portion 44(0) may have a greater channel length (L), a reduced channel width (W), and/or a higher threshold voltage (Vt) than memory cell transistors of the second memory portion 44(M). As illustrated by TABLE 1, each of these modifications increases (+) the drain-source resistance (RDS) of the induced channels of the memory cell transistors of the first memory portion 44(0). In this regard, another characterization of memory 51 is that the drain-source resistance (RDS) of the induced channels of the memory cell transistors of the first memory portion 44(0) is greater than the drain-source resistance (RDS) of the induced channels of the memory cell transistors of the second memory portion 44(M). With continuing reference to
Referring now to
In one embodiment, at least one memory portion 58 among memory portions 58(0) through 58(M-1) has been modified to have less current leakage than memory portion 58(M). In this embodiment, the at least one memory portion 58 among memory portions 58(0) through 58(M-1) may also have a greater internal latency than the memory portion 58(M), while not having a memory access latency greater than the worst-case memory access latency for any of the memory portions 58(0-M) for accessing the memory portion 58(M). In this manner, power consumption of the memory 54 is reduced without increasing the memory access latency of memory 54.
With continuing reference to
In this embodiment, the transistors of memory portions 66(0), 66(1), and 66(2) may have a same threshold voltage (Vt). However, in another embodiment, the transistors of memory portions 66(0), 66(1), and 66(2) may also be provided having different threshold voltages (Vt) in accordance with
As illustrated in
Additional components of memory 62 may also be modified to provide a further reduced power consumption memory 62, based on the reduced load and power consumption of the memory banks 68(0) through 68(7) realized in accordance with the herein discussed apparatuses and methods. Because herein discussed apparatuses and methods provide reduced load and reduced power consumption of memory banks 68(0) through 68(7) by using asymmetric memory portions, local input drivers for each of the memory banks 68(0) through 68(7) do not need to be as large as local input drivers provided for symmetric memory banks. Smaller local input drivers consume less power than larger local input drivers. In this regard, local input drivers for each memory bank 68(0), 68(1) may be made smaller than local input drivers for each memory bank 68(2), 68(3), 68(4), 68(5), 68(6), and 68(7). Similarly, local input drivers for each memory bank 68(2), 68(3), and 68(4) may be made smaller than local input drivers for each memory bank 68(5), 68(6), and 68(7). In this manner, the memory 62 may be made to provide further reduced power consumption.
In addition, because herein discussed apparatuses and methods provide reduced load and reduced power consumption of the memory banks 68(0) through 68(7) by using asymmetric memory portions, local memory address decoders for each memory bank 68(0) through 68(7) do not need to be as large as local memory address decoders provided for symmetric memory banks. Smaller memory address decoders consume less power than larger memory address decoders. In this regard, local memory address decoders for each memory bank 68(0), 68(1) may be made smaller than local memory address decoders for each memory bank 68(2), 68(3), 68(4), 68(5), 68(6), and 68(7). Similarly, local memory address decoders for each memory bank 68(2), 68(3), and 68(4) may be made smaller than local memory address decoders for each memory bank 68(5), 68(6), and 68(7). In this manner, the memory 62 may be made to provide even further reduced power consumption. The memory 62 may be used as the memory 20 in the memory system 16 of
If the closer memory portion(s) 44(0) has a negative latency margin which exceeds a negative latency margin threshold (block 90, YES) (which may happen if a latency increase of the closer memory portion(s) 44(0) was overshot (i.e. increased too much) in a previous block), then the closer memory portion(s) 44(0) may be modified to reduce latency and increase current leakage of the closer memory portion(s) 44(0) (block 92). In this regard a transistor characteristic(s) in a memory sub-banks) 42 and/or memory bank(s) 36 of the closer memory portion(s) 44(0) may be modified to reduce latency and increase current leakage of the closer memory portions) 44(0) (block 92).
If the determined latency margin of the closer memory portion(s) 44(0) is less than the positive latency margin threshold and greater than the negative latency margin threshold, then the method continues to block 94.
At this point, an asymmetric memory 20 (as further non-limiting examples 51, 54, 62) has been designed which provides reduced current leakage and reduced latency (i.e. reduced memory latency and increased memory speed) than the symmetric memory (from block 80). This asymmetric memory 20, 51, 54, 62 contains asymmetric memory portions 14, 58, 66 (as non-limiting examples, asymmetric memory sub-banks 42 and/or asymmetric memory banks 36) as herein discussed regarding
In some embodiments, such as
In this regard, referring now to
If the farther memory portion(s) 44(M) has a negative latency margin which exceeds the negative latency margin threshold (block 106, YES) (which may happen if a latency increase of the farther memory portion(s) 44(M) was overshot (i.e. modified too much) in a previous block), then the farther memory portion(s) 44(M) may be modified to reduce latency and increase current leakage of the farther memory portion(s) 44(M) (block 108). In this regard, a transistor characteristic(s) in a memory sub-bank(s) 42 and/or bank(s) 36 of the farther memory portion(s) 44(M) may be modified to reduce latency of the farther memory portion(s) 44(M) (block 108).
If the determined latency margin of the farther memory portion(s) 44(M) is less than the positive latency margin threshold and greater than the negative latency margin threshold, then the method ends (block 110). At this point, an asymmetric memory 20, 51, 54, 62 has been designed which provides reduced current leakage compared to the asymmetric memory of block 94 of
Alternatively, power consumption of memory 20, 51, 54, 62 may be reduced even further by reducing the size of the global bit line driver 72, by reducing the size of local input drivers for each of the memory portions 44, and/or by reducing the size of memory address decoders for each of the memory portions 44, as herein also discussed.
Herein disclosed embodiments discuss reducing current leakage. Reducing current leakage may comprise reducing drain-source current (IDS) of memory cell transistors of the memory while the gate-source voltage (VGS) is lower than the threshold voltage (Vt) of the memory cell transistors of the memory. However, reducing current leakage may also comprise reducing other current leakage of the memory cell transistors of the memory. As a non-limiting example, reducing current leakage may also comprise reducing gate-to-source current leakage through an oxide layer of memory cell transistors of the memory.
The asymmetrically-arranged memories having reduced current leakage and/or latency, and related systems and methods according to embodiments disclosed herein may be provided in or integrated into any processor-based device. Examples, without limitation, include a set top box, an entertainment unit, a navigation device, a communications device, a fixed location data unit, a mobile location data unit, a mobile phone, a cellular phone, a computer, a portable computer, a desktop computer, a personal digital assistant (PDA), a monitor, a computer monitor, a television, a tuner, a radio, a satellite radio, a music player, a digital music player, a portable music player, a digital video player, a video player, a digital video disc (DVD) player, and a portable digital video player.
In this regard,
Other master and slave devices can be connected to the system bus 46. As illustrated in
The CPU(s) 114 may also be configured to access the display controller(s) 132 over the system bus 46 to control information sent to one or more displays 148. The display controller(s) 132 sends information to the display(s) 148 to be displayed via one or more video processors 146, which processes the information to be displayed into a format suitable for the display(s) 148. The display(s) 148 can include any type of display, including but not limited to a cathode ray tube (CRT), a liquid crystal display (LCD), a light emitting diode (LED) display, a plasma display, a two dimensional (2-D) display, a three dimensional (3-D) display, a touch-screen display, etc.
The CPU(s) 114 and the display controller(s) 132 may act as master devices to make memory access requests to one or more memory access interfaces (MAIs) 28 of memory chips 20 of memories 136 over the system bus 46. Different threads within the CPU(s) 114 and the display controller(s) 132 may make requests to access memory to memory controller 18 which in turn accesses memory through the one or more memory access interfaces (MAN) 28 of memory chips 20 of the memory 136. Any memory in the system 112, including memory 136, may be provided as asymmetric memory according to the apparatuses and methods disclosed herein.
Those of skill in the art would further appreciate that the various illustrative logical blocks, modules, circuits, and algorithms described in connection with the embodiments disclosed herein may be implemented as electronic hardware, instructions stored in memory or in another computer-readable medium and executed by a processor or other processing device, or combinations of both. The memories, memory banks, memory sub-banks, memory access interfaces (MAIs), memory controllers, buses, master devices, and slave devices described herein may be employed in any circuit, hardware component, integrated circuit (IC), or IC chip, as examples. Memory disclosed herein may be any type and size of memory and may be configured to store any type of information desired. To clearly illustrate this interchangeability, various illustrative components, blocks, modules, circuits, and steps have been described above generally in terms of their functionality. How such functionality is implemented depends upon the particular application, design choices, and/or design constraints imposed on the overall system. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present invention.
The various illustrative logical blocks, modules, and circuits described in connection with the embodiments disclosed herein may be implemented or performed with a processor, a digital signal processor (DSP), an Application Specific Integrated Circuit (ASIC), an Field Programmable Gate Array (FPGA) or other programmable logic device, discrete gate or transistor logic, discrete hardware components, or any combination thereof designed to perform the functions described herein. A processor may be a microprocessor, but in the alternative, the processor may be any conventional processor, controller, microcontroller, or state machine. A processor may also be implemented as a combination of computing devices, e.g., a combination of a DSP and a microprocessor, a plurality of microprocessors, one or more microprocessors in conjunction with a DSP core, or any other such configuration.
The embodiments disclosed herein may be embodied in hardware and in instructions that are stored in hardware, and may reside, for example, in Random Access Memory (RAM), flash memory, Read Only Memory (ROM), Electrically Programmable ROM (EPROM), Electrically Erasable Programmable ROM (EEPROM), registers, a hard disk, a removable disk, a CD-ROM, or any other form of computer-readable medium known in the art. An exemplary storage medium is coupled to the processor such that the processor can read information from, and write information to, the storage medium. In the alternative, the storage medium may be integral to the processor. The processor and the storage medium may reside in an ASIC. The ASIC may reside in a remote station. In the alternative, the processor and the storage medium may reside as discrete components in a remote station, base station, or server.
It is also noted that the operational steps described in any of the exemplary embodiments herein are described to provide examples and discussion. The operations described may be performed in numerous different sequences other than the illustrated sequences. Furthermore, operations described in a single operational step may actually be performed in a number of different steps. Additionally, one or more operational steps discussed in the exemplary embodiments may be combined. It is to be understood that the operational steps illustrated in the flow chart diagrams may be subject to numerous different modifications as will be readily apparent to one of skill in the art. Those of skill in the art would also understand that information and signals may be represented using any of a variety of different technologies and techniques. For example, data, instructions, commands, information, signals, bits, symbols, and chips that may be referenced throughout the above description may be represented by voltages, currents, electromagnetic waves, magnetic fields or particles, optical fields or particles, or any combination thereof.
The previous description of the disclosure is provided to enable any person skilled in the art to make or use the disclosure. Various modifications to the disclosure will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other variations without departing from the spirit or scope of the disclosure. Thus, the disclosure is not intended to be limited to the examples and designs described herein, but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.
The present application claims priority to U.S. Provisional Patent Application Ser. No. 61/586,867 entitled “ASYMMETRICALLY-ARRANGED MEMORIES HAVING REDUCED CURRENT LEAKAGE AND/OR LATENCY, AND RELATED SYSTEMS AND METHODS” filed on Jan. 16, 2012, which is hereby incorporated herein by reference in its entirety.
Number | Date | Country | |
---|---|---|---|
61586867 | Jan 2012 | US |