This invention relates generally to non-volatile semiconductor memory such as electrically erasable programmable read-only memory (EEPROM) and flash EEPROM and, more specifically, to techniques for handling defects in such memories.
Solid-state memory capable of nonvolatile storage of charge, particularly in the form of EEPROM and flash EEPROM packaged as a small form factor card, has recently become the storage of choice in a variety of mobile and handheld devices, notably information appliances and consumer electronics products. Unlike RAM (random access memory) that is also solid-state memory, flash memory is non-volatile, retaining its stored data even after power is turned off. In spite of the higher cost, flash memory is increasingly being used in mass storage applications. Conventional mass storage, based on rotating magnetic medium such as hard drives and floppy disks, is unsuitable for the mobile and handheld environment. This is because disk drives tend to be bulky, are prone to mechanical failure and have high latency and high power requirements. These undesirable attributes make disk-based storage impractical in most mobile and portable applications. On the other hand, flash memory, both embedded and in the form of a removable card is ideally suited in the mobile and handheld environment because of its small size, low power consumption, high speed and high reliability features.
EEPROM and electrically programmable read-only memory (EPROM) are non-volatile memory that can be erased and have new data written or “programmed” into their memory cells. Both utilize a floating (unconnected) conductive gate, in a field effect transistor structure, positioned over a channel region in a semiconductor substrate, between source and drain regions. A control gate is then provided over the floating gate. The threshold voltage characteristic of the transistor is controlled by the amount of charge that is retained on the floating gate. That is, for a given level of charge on the floating gate, there is a corresponding voltage (threshold) that must be applied to the control gate before the transistor is turned “on” to permit conduction between its source and drain regions.
The floating gate can hold a range of charges and therefore can be programmed to any threshold voltage level within a threshold voltage window. The size of the threshold voltage window is delimited by the minimum and maximum threshold levels of the device, which in turn correspond to the range of the charges that can be programmed onto the floating gate. The threshold window generally depends on the memory device's characteristics, operating conditions and history. Each distinct, resolvable threshold voltage level range within the window may, in principle, be used to designate a definite memory state of the cell.
The transistor serving as a memory cell is typically programmed to a “programmed” state by one of two mechanisms. In “hot electron injection,” a high voltage applied to the drain accelerates electrons across the substrate channel region. At the same time a high voltage applied to the control gate pulls the hot electrons through a thin gate dielectric onto the floating gate. In “tunneling injection,” a high voltage is applied to the control gate relative to the substrate. In this way, electrons are pulled from the substrate to the intervening floating gate.
The memory device may be erased by a number of mechanisms. For EPROM, the memory is bulk erasable by removing the charge from the floating gate by ultraviolet radiation. For EEPROM, a memory cell is electrically erasable, by applying a high voltage to the substrate relative to the control gate so as to induce electrons in the floating gate to tunnel through a thin oxide to the substrate channel region (i.e., Fowler-Nordheim tunneling.) Typically, the EEPROM is erasable byte by byte. For flash EEPROM, the memory is electrically erasable either all at once or one or more blocks at a time, where a block may consist of 512 bytes or more of memory.
The memory devices typically comprise one or more memory chips that may be mounted on a card. Each memory chip comprises an array of memory cells supported by peripheral circuits such as decoders and erase, write and read circuits. The more sophisticated memory devices also come with a controller that performs intelligent and higher level memory operations and interfacing. There are many commercially successful non-volatile solid-state memory devices being used today. These memory devices may employ different types of memory cells, each type having one or more charge storage element.
One simple embodiment of the split-channel memory cell is where the select gate and the control gate are connected to the same word line as indicated schematically by a dotted line shown in
A more refined embodiment of the split-channel cell shown in
When an addressed memory transistor within an NAND cell is read and verified during programming, its control gate is supplied with an appropriate voltage. At the same time, the rest of the non-addressed memory transistors in the NAND cell 50 are fully turned on by application of sufficient voltage on their control gates. In this way, a conductive path is effective created from the source of the individual memory transistor to the source terminal 54 of the NAND cell and likewise for the drain of the individual memory transistor to the drain terminal 56 of the cell. Memory devices with such NAND cell structures are described in U.S. Pat. Nos. 5,570,315, 5,903,495, 6,046,935.
Memory Array
A memory device typically comprises of a two-dimensional array of memory cells arranged in rows and columns and addressable by word lines and bit lines. The array can be formed according to an NOR type or an NAND type architecture.
NOR Array
Many flash EEPROM devices are implemented with memory cells where each is formed with its control gate and select gate connected together. In this case, there is no need for steering lines and a word line simply connects all the control gates and select gates of cells along each row. Examples of these designs are disclosed in U.S. Pat. Nos. 5,172,338 and 5,418,752. In these designs, the word line essentially performed two functions: row selection and supplying control gate voltage to all cells in the row for reading or programming.
NAND Array
Block Erase
Programming of charge storage memory devices can only result in adding more charge to its charge storage elements. Therefore, prior to a program operation, existing charge in a charge storage element must be removed (or erased). Erase circuits (not shown) are provided to erase one or more blocks of memory cells. A non-volatile memory such as EEPROM is referred to as a “Flash” EEPROM when an entire array of cells, or significant groups of cells of the array, is electrically erased together (i.e., in a flash). Once erased, the group of cells can then be reprogrammed. The group of cells erasable together may consist one or more addressable erase unit. The erase unit or block typically stores one or more pages of data, the page being the unit of programming and reading, although more than one page may be programmed or read in a single operation. Each page typically stores one or more sectors of data, the size of the sector being defined by the host system. An example is a sector of 512 bytes of user data, following a standard established with magnetic disk drives, plus some number of bytes of overhead information about the user data and/or the block in with it is stored.
Read/Write Circuits
In the usual two-state EEPROM cell, at least one current breakpoint level is established so as to partition the conduction window into two regions. When a cell is read by applying predetermined, fixed voltages, its source/drain current is resolved into a memory state by comparing with the breakpoint level (or reference current IREF). If the current read is higher than that of the breakpoint level, the cell is determined to be in one logical state (e.g., a “zero” state). On the other hand, if the current is less than that of the breakpoint level, the cell is determined to be in the other logical state (e.g., a “one” state). Thus, such a two-state cell stores one bit of digital information. A reference current source, which may be externally programmable, is often provided as part of a memory system to generate the breakpoint level current.
In order to increase memory capacity, flash EEPROM devices are being fabricated with higher and higher density as the state of the semiconductor technology advances. Another method for increasing storage capacity is to have each memory cell store more than two states.
For a multi-state or multi-level. EEPROM memory cell, the conduction window is partitioned into more than two regions by more than one breakpoint such that each cell is capable of storing more than one bit of data. The information that a given EEPROM array can store is thus increased with the number of states that each cell can store. EEPROM or flash EEPROM with multi-state or multi-level memory cells have been described in U.S. Pat. No. 5,172,338.
In practice, the memory state of a cell is usually read by sensing the conduction current across the source and drain electrodes of the cell when a reference voltage is applied to the control gate. Thus, for each given charge on the floating gate of a cell, a corresponding conduction current with respect to a fixed reference control gate voltage may be detected. Similarly, the range of charge programmable onto the floating gate defines a corresponding threshold voltage window or a corresponding conduction current window.
Alternatively, instead of detecting the conduction current among a partitioned current window, it is possible to set the threshold voltage for a given memory state under test at the control gate and detect if the conduction current is lower or higher than a threshold current. In one implementation the detection of the conduction current relative to a threshold current is accomplished by examining the rate the conduction current is discharging through the capacitance of the bit line.
As can be seen from the description above, the more states a memory cell is made to store, the more finely divided is its threshold window. This will require higher precision in programming and reading operations in order to be able to achieve the required resolution.
U.S. Pat. No. 4,357,685 discloses a method of programming a 2-state EPROM in which when a cell is programmed to a given state, it is subject to successive programming voltage pulses, each time adding incremental charge to the floating gate. In between pulses, the cell is read back or verified to determine its source-drain current relative to the breakpoint level. Programming stops when the current state has been verified to reach the desired state. The programming pulse train used may have increasing period or amplitude.
Prior art programming circuits simply apply programming pulses to step through the threshold window from the erased or ground state until the target state is reached. Practically, to allow for adequate resolution, each partitioned or demarcated region would require at least about five programming steps to transverse. The performance is acceptable for 2-state memory cells. However, for multi-state cells, the number of steps required increases with the number of partitions and therefore, the programming precision or resolution must be increased. For example, a 16-state cell may require on average at least 40 programming pulses to program to a target state.
Factors Affecting Read/Write Performance and Accuracy
In order to improve read and program performance, multiple charge storage elements or memory transistors in an array are read or programmed in parallel. Thus, a logical “page” of memory elements are read or programmed together. In existing memory architectures, a row typically contains several interleaved pages. All memory elements of a page will be read or programmed together. The column decoder will selectively connect each one of the interleaved pages to a corresponding number of read/write modules. For example, in one implementation, the memory array is designed to have a page size of 532 bytes (512 bytes plus 20 bytes of overheads.) If each column contains a drain bit line and there are two interleaved pages per row, this amounts to 8512 columns with each page being associated with 4256 columns. There will be 4256 sense modules connectable to read or write in parallel either all the even bit lines or the odd bit lines. In this way, a page of 4256 bits (i.e., 532 bytes) of data in parallel are read from or programmed into the page of memory elements. The read/write modules forming the read/write circuits 170 can be arranged into various architectures.
Referring to
As mentioned before, conventional memory devices improve read/write operations by operating in a massively parallel manner on all even or all odd bit lines at a time. This architecture of a row consisting of two interleaved pages will help to alleviate the problem of fitting the block of read/write circuits. It is also dictated by consideration of controlling bit-line to bit-line capacitive coupling. A block decoder is used to multiplex the set of read/write modules to either the even page or the odd page. In this way, whenever one set bit lines are being read or programmed, the interleaving set can be grounded to minimize immediate neighbor coupling.
However, the interleaving page architecture is disadvantageous in at least three respects. First, it requires additional multiplexing circuitry. Secondly, it is slow in performance. To finish read or program of memory cells connected by a word line or in a row, two read or two program operations are required. Thirdly, it is also not optimum in addressing other disturb effects such as field coupling between neighboring charge storage elements at the floating gate level when the two neighbors are programmed at different times, such as separately in odd and even pages.
The problem of neighboring field coupling becomes more pronounced with ever closer spacing between memory transistors. In a memory transistor, a charge storage element is sandwiched between a channel region and a control gate. The current that flows in the channel region is a function of the resultant electric field contributed by the field at the control gate and the charge storage element. With ever increasing density, memory transistors are formed closer and closer together. The field from neighboring charge elements then becomes significant contributor to the resultant field of an affected cell. The neighboring field depends on the charge programmed into the charge storage elements of the neighbors. This perturbing field is dynamic in nature as it changes with the programmed states of the neighbors. Thus, an affected cell may read differently at different time depending on the changing states of the neighbors.
The conventional architecture of interleaving page exacerbates the error caused by neighboring floating gate coupling. Since the even page and the odd page are programmed and read independently of each other, a page may be programmed under one set of condition but read back under an entirely different set of condition, depending on what has happened to the intervening page in the meantime. The read errors will become more severe with increasing density, requiring a more accurate read operation and coarser partitioning of the threshold window for multi-state implementation. Performance will suffer and the potential capacity in a multi-state implementation is limited.
United States Patent Publication No. US-2004-0060031-A1 discloses a high performance yet compact non-volatile memory device having a large block of read/write circuits to read and write a corresponding block of memory cells in parallel. In particular, the memory device has an architecture that reduces redundancy in the block of read/write circuits to a minimum. Significant saving in space as well as power is accomplished by redistributing the block of read/write modules into a block read/write module core portions that operate in parallel while interacting with a substantially smaller sets of common portions in a time-multiplexing manner. In particular, data processing among read/write circuits between a plurality of sense amplifiers and data latches is performed by a shared processor.
Therefore there is a general need for high performance and high capacity non-volatile memory. In particular, there is a need for a compact non-volatile memory with enhanced read and program performance having an improved processor that is compact and efficient, yet highly versatile for processing data among the read/writing circuits.
A non-volatile memory circuit including an array of non-volatile memory cells formed along columns of multiple bits, the columns including a plurality of regular columns and one or more redundancy columns, is described. The memory circuit also includes a plurality of latches, each corresponding to one of the regular columns and having a bit whose value indicates if the corresponding column is defective. The memory circuit storing a column redundancy data table whose contents indicate for each redundancy column whether the redundancy column is being used and, for redundancy columns that are being used, a defective regular column to which it corresponds and the bits therein which are defective. The memory circuit stores data corresponding to the defective bits of defective regular columns in the redundancy column portion.
According to an additional set of aspects, a method of operating a non-volatile memory circuit is presented, where the memory circuit includes an array of non-volatile memory cells formed along columns of multiple bits and having a latch associated with each of the columns whose value indicates if the corresponding column has a defect. The method includes: performing a write operation to concurrently program a plurality of memory cells on a corresponding plurality of columns, including one or more columns having an associated latch whose value indicates the corresponding column has a defect; determining the number of the plurality of concurrently programmed memory cells that were not successfully programmed in the write operation, wherein the columns whose latch values indicate the column has a defect are not counted in the determining; and determining whether the number of cells that were not successfully been programmed during the write operation is acceptable.
According to another set of aspects, methods of operating a non-volatile memory circuit having an array of non-volatile memory cells formed along columns of multiple bits, the columns including a plurality of regular columns and one or more redundancy columns are presented. The method includes performing a plurality of column test operations to determine which columns are defective and the individual bits therein which are defective, each of the column tests including: writing and reading back an externally supplied data pattern to the columns; and comparing the externally supplied data pattern as read back with an expected data pattern, wherein said column test operation are performed by circuitry on the memory circuit and each of the column tests uses a different data pattern. The method also includes recording addresses of any of the regular columns determined defective and the individual bits therein which are determined defective in a column redundancy data table stored on the memory circuit; and, for any of the regular columns determined defective, setting a latch associated therewith to a value indicating that the associated column is defective.
In other aspects, a method of operating a non-volatile memory circuit having an array of non-volatile memory cells formed along columns of multiple bits, the columns including a plurality of regular columns and one or more redundancy columns is described. The method includes: storing on the memory circuit a column redundancy data table whose contents indicate for each redundancy column whether the redundancy column is being used and, for redundancy columns that are being used, a defective regular column to which it corresponds and the bits therein which are defective; receiving a set of data to program into the memory array; determining the elements of the set of data assigned to be programmed to defective bits of defective regular columns based upon the column redundancy circuit data table; storing the elements of the set of data determined to be assigned to be programmed to defective bits of defective columns in peripheral latch circuits on the memory circuit; storing the set of data into programming latches for the memory array; performing a programming operation into the regular columns of the memory array from the programming latches; and programming the elements of the data set stored in the peripheral latches into the redundancy columns.
Various aspects, advantages, features and embodiments of the present invention are included in the following description of exemplary examples thereof, which description should be taken in conjunction with the accompanying drawings. All patents, patent applications, articles, other publications, documents and things referenced herein are hereby incorporated herein by this reference in their entirety for all purposes. To the extent of any inconsistency or conflict in the definition or use of terms between any of the incorporated publications, documents or things and the present application, those of the present application shall prevail.
b are examples of data latches that could be used for data compactification.
The control circuitry 310 cooperates with the read/write circuits 370 to perform memory operations on the memory array 300. The control circuitry 310 includes a state machine 312, an on-chip address decoder 314 and a power control module 316. The state machine 312 provides chip level control of memory operations. The on-chip address decoder 314 provides an address interface between that used by the host or a memory controller to the hardware address used by the decoders 330 and 370. The power control module 316 controls the power and voltages supplied to the word lines and bit lines during memory operations.
The entire bank of partitioned read/write stacks 400 operating in parallel allows a block (or page) of p cells along a row to be read or programmed in parallel. Thus, there will be p read/write modules for the entire row of cells. As each stack is serving k memory cells, the total number of read/write stacks in the bank is therefore given by r=p/k. For example, if r is the number of stacks in the bank, then p=r*k. One example memory array may have p=512 bytes (512×8 bits), k=8, and therefore r=512. In the preferred embodiment, the block is a run of the entire row of cells. In another embodiment, the block is a subset of cells in the row. For example, the subset of cells could be one half of the entire row or one quarter of the entire row. The subset of cells could be a run of contiguous cells or one every other cell, or one every predetermined number of cells.
Each read/write stack, such as 400-1, essentially contains a stack of sense amplifiers 212-1 to 212-k servicing a segment of k memory cells in parallel. A preferred sense amplifier is disclosed in United States Patent Publication No. 2004-0109357-A1, the entire disclosure of which is hereby incorporated herein by reference.
The stack bus controller 410 provides control and timing signals to the read/write circuit 370 via lines 411. The stack bus controller is itself dependent on the memory controller 310 via lines 311. Communication among each read/write stack 400 is effected by an interconnecting stack bus 431 and controlled by the stack bus controller 410. Control lines 411 provide control and clock signals from the stack bus controller 410 to the components of the read/write stacks 400-1.
In the preferred arrangement, the stack bus is partitioned into a SABus 422 for communication between the common processor 500 and the stack of sense amplifiers 212, and a DBus 423 for communication between the processor and the stack of data latches 430.
The stack of data latches 430 comprises of data latches 430-1 to 430-k, one for each memory cell associated with the stack The I/O module 440 enables the data latches to exchange data with the external via an I/O bus 231.
The common processor also includes an output 507 for output of a status signal indicating a status of the memory operation, such as an error condition. The status signal is used to drive the gate of an n-transistor 550 that is tied to a FLAG BUS 509 in a Wired-Or configuration. The FLAG BUS is preferably precharged by the controller 310 and will be pulled down when a status signal is asserted by any of the read/write stacks. (The isolation latch IL 529 is discussed in the following section on bad column management.)
The input logic 510 receives data from the PBUS and outputs to a BSI node as a transformed data in one of logical states “1”, “0”, or “Z” (float) depending on the control signals from the stack bus controller 410 via signal lines 411. A Set/Reset latch, PLatch 520 then latches BSI, resulting in a pair of complementary output signals as MTCH and MTCH*.
The output logic 530 receives the MTCH and MTCH* signals and outputs on the PBUS 505 a transformed data in one of logical states “1”, “0”, or “Z” (float) depending on the control signals from the stack bus controller 410 via signal lines 411.
At any one time the common processor 500 processes the data related to a given memory cell. For example,
The PBUS 505 of the common processor 500 has access to the SA latch 214-1 via the SBUS 422 when a transfer gate 501 is enabled by a pair of complementary signals SAP and SAN. Similarly, the PBUS 505 has access to the set of data latches 430-1 via the DBUS 423 when a transfer gate 502 is enabled by a pair of complementary signals DTP and DTN. The signals SAP, SAN, DTP and DTN are illustrated explicitly as part of the control signals from the stack bus controller 410.
In the case of the PASSTHROUGH mode where BSI is the same as the input data, the signals ONE is at a logical “1”, ONEB<0> at “0” and ONEB<1> at “0”. This will disable the pull-up or pull-down but enable the transfer gate 522 to pass the data on the PBUS 505 to the output 523. In the case of the INVERTED mode where BSI is the invert of the input data, the signals ONE is at “0”, ONEB<0> at “1” and ONE<1> at “1”. This will disable the transfer gate 522. Also, when PBUS is at “0”, the pull-down circuit will be disabled while the pull-up circuit is enabled, resulting in BSI being at “1”. Similarly, when PBUS is at “1”, the pull-up circuit is disabled while the pull-down circuit is enabled, resulting in BSI being at “0”. Finally, in the case of the FLOATED mode, the output BSI can be floated by having the signals ONE at “1”, ONEB<0> at “1” and ONEB<1> at “0”. The FLOATED mode is listed for completeness although in practice, it is not used.
One feature of the invention is to constitute the pull-up circuits with PMOS transistors and the pull-down circuits with NMOS transistors. Since the pull by the NMOS is much stronger than that of the PMOS, the pull-down will always overcome the pull-up in any contentions. In other words, the node or bus can always default to a pull-up or “1” state, and if desired, can always be flipped to a “0” state by a pull-down.
In the FLOATED mode, all four branches are disabled. This is accomplished by having the signals PINV=1, NINV=0, PDIR=1, NDIR=0, which are also the default values. In the PASSTHROUGH mode, when MTCH=0, it will require PBUS=0. This is accomplished by only enabling the pull-down branch with n-transistors 535 and 536, with all control signals at their default values except for NDIR=1. When MTCH=1, it will require PBUS=1. This is accomplished by only enabling the pull-up branch with p-transistors 533 and 534, with all control signals at their default values except for PINV=0. In the INVERTED mode, when MTCH=0, it will require PBUS=1. This is accomplished by only enabling the pull-up branch with p-transistors 531 and 532, with all control signals at their default values except for PDIR=0. When MTCH=1, it will require PBUS=0. This is accomplished by only enabling the pull-down branch with n-transistors 537 and 538, with all control signals at their default values except for NINV=1. In the PRECHARGE mode, the control signals settings of PDIR=0 and PINV=0 will either enable the pull-up branch with p-transistors 531 and 531 when MTCH=1 or the pull-up branch with p-transistors 533 and 534 when MTCH=0.
Common processor operations are developed more fully in U.S. patent application Ser. No. 11/026,536, Dec. 29, 2004, which is hereby incorporated in its entirety by this reference.
Bad Column Management with Bit Information
A memory will often have defective portions, either from the manufacturing process or that arise during the operation of the device. A number of techniques exist for managing these defects including error correction coding or remapping portions of the memory, such as described in U.S. Pat. Nos. 7,405,985, 5,602,987, 5,315,541, 5,200,959, and 5,428,621. For instance, a device is generally thoroughly tested before being shipped. The testing may find a defective portion of the memory that needs to be eliminated. Before shipping the device, the information on these defects is stored on the device, for example in a ROM area of the memory array or in a separate ROM, and at power up it is read by a controller and then used so that the controller can substitute a good portion of the memory for the bad. When reading or writing, the controller will then need to refer to a pointer structure in the controller's memory for this remapping.
In previous arrangements for managing bad columns, such as in U.S. Pat. No. 7,405,985, when there is an error in a column, the whole column is typically mapped out, with the corresponding whole byte or word will be marked to be bad. According to the aspects presented in this section, the system can detect when only 1 bit in the byte is bad and bytes with single bit failures can be utilized as long as the single bit is saved elsewhere in the memory. Through the analysis of the any defective columns, it can be determined whether they are in the category where the whole will be treated as bad or whether it only has only single bit failures so that the other bits in the bad columns can be used as good. In an exemplary application, during the die sort, those single bit failures and their column address as well as bit address can be detected and saved in a non-volatile ROM block. When the controller manages these bad columns by this information, the bit information can be used to extract the corresponding bits saved in a column redundancy area. The can consequently enhance the yield so that more defects can be repaired by the column redundancy, since columns with only single bit errors can still be used, rather than mapped out.
More specifically, each column of the memory has an associated isolation latch or register whose value indicates whether the column is defective, but in addition to this information, for columns marked as defective, additional information is used to indicate whether the column as a whole is to be treated as defective, or whether just individual bits of the column are defective. The defective elements can then be re-mapped to a redundant element at either the appropriate bit or column level based on the data. When a column is bad, but only on the bit level, the good bits can still be used for data, although this may be done at a penalty of under programming for some bits, as is described further below. In an exemplary embodiment, the bad column and bad bit information is determined as part of a self contained Built In Self Test (BIST) flow constructed to collect the bit information through a set of column tests. Based on this information, the bad bits can be extracted and re-grouped into bytes by the controller or on the memory, depending on the embodiment, to more efficiently use the column redundancy area. These techniques and structures can be applied to the various memory architectures described above, including NOR architectures, NAND architectures, and even the sort of 3D memory structures described in U.S. patent application Ser. No. 12/414,935. When reference to a specific memory architecture is useful, NAND flash memory will serve as the exemplary embodiment.
Returning briefly to the case of where bad columns are managed without bit information, non-volatile memories usually have redundancy to repair on-chip failures. Column redundancy is used to repair the bad columns, where the repair unit is normally one byte as a unit, or sometimes a word as a unit. Under this arrangement, even for a 1 bit fail in the 1 byte, the whole byte will be marked to be a bad column and the data will be moved to the redundancy area. This is a convenient way to isolate the bad column as a group of bad bitlines, but the penalty is that the redundancy repair unit could be exhausted fairly rapidly. The bad column address is normally saved in the ROM block of the non-volatile memory. In the exemplary embodiments below, there are 13 column addresses, A<13:1>. The format for column redundancy data can then use 2 bytes to remember one column address. There are 2 flag bits to indicate that it is a unused column redundancy, or a used column redundancy, or a Bad column redundancy, as shown in the table of
Bad columns can classified as one of two types: those such as an related to bitline short or open circuit, where there can be multiple bad bit failure, and the whole column is taken as defective; and those such as defects in the data latches or sense amps, which are typically individual bit failures. To keep the physical array structure simple and save on die size, the latch or register that indicates a column is bad (the isolation latch) uses one 1-bit latch per byte. (For architectures that have a top and bottom latch that would be isolated together, then one defect will isolate 2 bytes (1 top, 1 bottom).) If the minimum repair unit is taken as a byte or a word, this could cause inefficiency in the management of bad columns, since, typically, most of the bits in the bad columns are good bits which can be used.
It should be noted that when the isolation latch is set under this arrangement, this does not mean the column is no longer accessible, just that it is marked as “don't care” with respect to program or erase completion. Under this arrangement, columns that are defective on the bit level will have their isolation set and not counted among the good columns; however, even though the bad columns are “isolated”, the cells will get programmed (and erased) and verified. At the end of a program operation, however, at the isolation latch is set, any of their bits that have failed to program (slow bits) will not get counted as part the total failure count. Therefore, these bad columns do not participate in the pseudo-pass criteria for programming (or erase) and there may consequently be some cells that are under-programmed (or under-erased) but un-detected. As these are slow cells in the normal good columns, the number of program (erase) pulses will be applied on the wordline to make sure that the data will be programmed (or erased) successfully. Additionally, as stronger ECC capability is available to the non-volatile memory system, it allows for the system to take care of most of the slow bits.
For example, the system may have an allowance for 40 bits fail during programming. Taking a programming operation as having, say, 9000 bytes, the ratio of failed bits is then 40/(9000*8). If 24 columns have been replaced with redundancy columns, where each byte has 1 bit bad bitline, and with 7 bits per byte programming without detection, then the number of failed programmed bits will be {24*2*7*40/(9000*8)}=4 bits failure. The rest of the bits (24×7), besides the bad bitlines, in the bad column will be programmed correctly and these 4 bits can be managed by the error correction code.
In another embodiment, the mode of failures can be recorded in the bad column information.
According to one aspect presented here, during die sort or the built in self-test (BIST) test flow discussed in the following, the bad columns can be tested bit by bit in multiple column tests and failed bit information will be accumulated into a CRD table such as
Thus, in the arrangement presented here, the number of failed bits can be recorded in the one of these formats, which allows the column redundancy data to record multiple bit failures for a column. The bad column can be managed by the memory circuits as well as controller. For the simplicity of presentation, the description here is mainly given for the case when the controller manages the bad columns. Similar function can also be achieved by the circuits inside the non-volatile memory. During the program process, the controller will load the user program data intro the data latches inside the memory. The location corresponding to the bad bits can be left with user data or filled with “1”, but the copy of the data will also be saved in a good bit location in the redundancy column area. As isolated bad columns with bit errors will have some good data they will going through the program (or erase) process, and so the bad bit can just have their data latched for them as well as in the remapped location. Regardless of the data in the bad bitline, the operations can be done collectively on all cells without increasing the power consumption in NAND flash architecture. In some other architecture, such as, NOR flash or 3D Read/Writable architecture, the bad bitlines are filled with data of non-operation to avoid extra power loss.
The replacement of bad bits with good bits from the redundancy columns can be illustrated schematically using
The Build In Self Test (BIST) mechanism for bad column addresses with bit information referred to above will now be described. This uses an algorithm to determine the bad column with bit information. A state machine on the memory itself (not the controller) can execute the process for externally supplied test sequences and corresponding test data. The flow chart of
To improve robustness, multiple copies of the column redundancy information (
At 707, the next test is begun, with the expected data for this test again compare with the read out data at 709. The stored result from 705 is then fetched at 711 and compared with that from 709 for any address matches between the two. Address match can be done with XOR logic as well, with an exemplary circuit for this is shown in
If there is no match at 713, a new entry is written back at 717. Both 715 and 717 loop back to 709 and the process continues until the current test is done for all columns, after which the flow decides if there are more tests at 719. If so, the flow loops back to 707 and if not, at 721 the stored results from the series of test are fetched and the isolated latches set for the columns found defective. The bad column information will also be written into the designated ROM block in the non-volatile memory. In some cases, the test flow could be broken into tests done at different times. The test result can be stored in the ROM block for first few tests, and then the data will be read back from the ROM block and continue with the subsequent tests following same test algorithm as described above. Although the embodiment presented above is for an initial sort based upon externally provided tests, alternate embodiments could be performed to dynamically update the defect information, based on tests executed, for example, by the controller or sophisticated tester.
Considering the data in process further, this can be taken as the steps of:
The data out process will need undo the data in process and can be taken as the steps of:
The on-chip implementation of the bad bit packing and un-packing may use a large number of registers, possibly increasing die size. One to implement the process using a relatively small die area and a limited number of registers is to divide the bad bytes into several groups. Each time, a group of bad columns will be packed or unpacked with fixed number registers to handle address and data information. The algorithm for packing or un-packing can still be the same as described above. For example, if the memory have 40 bad bytes, it can process 10 bytes at a time and finish the bad byte processing in 4 groups. After instance of packing, the packed bytes can be put into the extended column area data latches. After each instance of un-packing, the un-packed bits (or bytes) can be sorted back to their original data place. More details of such an implementation, in a slightly different context, are presented in U.S. patent application Ser. No. 12/414,935.
The techniques described above for the applications of bad column with bad bit information. The bit information will enhanced device yield since more bad columns with bad bits can be repaired with the fixed number column redundancies typically available on a device. Besides the normal operations, it also benefits the bad column management in the devices incorporating an internal folding algorithm, such as that described in U.S. patent application Ser. No. 12/478,997.
The bad bits can be arranged in the column redundancy area as shown in the example of
The reason to set the bad column isolation latch is that some failures could cause detection fail if the detection is done collectively and simultaneously, but these failure bits should not be counted as they are already repaired by the redundancy. This could lead to overly strict criteria to pass program (or erase) and make the operations return with failed status. For example, if there are 20 bad column repaired by the redundancy columns, these 20 bad columns will cause 40 bits failures. If the program pseudo-pass criteria is set to be 40, then there will be 0 failures allowed for the whole page program. If the program pseudo-pass criteria is set to be less than 40, the page program will always fail. When such situations occur, the status will not reflect the real situation as to whether the write operation has succeeded or not. In order to make sure that the program status reflect the real program situation, the bad columns should be masked out or isolated. If the had bits are counted serially by toggling the data out one byte (or a word) at a time, then the isolation latch is not necessary.
This sort of bit level management can be particularly advantageous for incorporating an internal folding, as that described in U.S. patent application Ser. No. 12/478,997. Briefly, data is initially written to a memory in binary form, folded into a multi-state format in the memory latches, and then rewritten back into the non-volatile memory. To take a 3-bit per cell example, three pages would initially be written onto three physical pages in binary form and then rewritten in 3-bit per cell format onto a single physical wordline. In the case of a bad column, this defect will need to be reflected in the columns with which it is folded, leading to a corresponding increase in number of redundant columns used.
This process can be illustrated with
Because of this, a bad column will need to be reflected in the other columns with which it is folded. Consequently, in an N-bit per cell folding process, each bad column may be magnified by a factor of N, which could quickly exhaust the available number of redundant columns. Because of this, the use of bit information for bad column can be particularly advantageous in system that use such folding. Even though the folding process will create more failed bits during the process of folding, the bad bits management will reduce the impact of wasting too many redundancy columns because of folding.
Although the various aspects of the present invention have been described with respect to certain embodiments, it is understood that the invention is entitled to protection within the full scope of the appended claims.
This application is a divisional of U.S. patent application Ser. No. 12/498,220 filed Jul. 6, 2009, which is incorporated in its entirety herein by this reference.
Number | Name | Date | Kind |
---|---|---|---|
3710348 | Craft | Jan 1973 | A |
3895360 | Cricchi et al. | Jul 1975 | A |
4357685 | Daniele et al. | Nov 1982 | A |
4426688 | Moxley | Jan 1984 | A |
4720815 | Ogawa | Jan 1988 | A |
4757477 | Nagayama et al. | Jul 1988 | A |
5070032 | Yuan et al. | Dec 1991 | A |
5095344 | Harari | Mar 1992 | A |
5172338 | Mehrotra et al. | Dec 1992 | A |
5200959 | Gross et al. | Apr 1993 | A |
5270979 | Harari et al. | Dec 1993 | A |
5313421 | Guterman et al. | May 1994 | A |
5315541 | Harari et al. | May 1994 | A |
5343063 | Yuan et al. | Aug 1994 | A |
5380672 | Yuan et al. | Jan 1995 | A |
5386390 | Okitaka | Jan 1995 | A |
5418752 | Harari et al. | May 1995 | A |
5428621 | Mehrotra et al. | Jun 1995 | A |
5430679 | Hiltebeitel et al. | Jul 1995 | A |
5430859 | Norman et al. | Jul 1995 | A |
5442748 | Chang et al. | Aug 1995 | A |
5479370 | Furuyama et al. | Dec 1995 | A |
5485425 | Iwai et al. | Jan 1996 | A |
5570315 | Tanaka et al. | Oct 1996 | A |
5595924 | Yuan et al. | Jan 1997 | A |
5602987 | Harari et al. | Feb 1997 | A |
5642312 | Harari | Jun 1997 | A |
5657332 | Auclair et al. | Aug 1997 | A |
5661053 | Yuan | Aug 1997 | A |
5663901 | Wallace et al. | Sep 1997 | A |
5699300 | Akamatsu et al. | Dec 1997 | A |
5712180 | Guterman et al. | Jan 1998 | A |
5768192 | Eitan | Jun 1998 | A |
5774397 | Endoh et al. | Jun 1998 | A |
5783958 | Lysinger | Jul 1998 | A |
5822245 | Gupta et al. | Oct 1998 | A |
5848009 | Lee et al. | Dec 1998 | A |
5862080 | Harari et al. | Jan 1999 | A |
5890192 | Lee et al. | Mar 1999 | A |
5903495 | Takeuchi et al. | May 1999 | A |
5930167 | Lee | Jul 1999 | A |
5936971 | Harari et al. | Aug 1999 | A |
6011725 | Eitan | Jan 2000 | A |
6021463 | Belser | Feb 2000 | A |
6038167 | Miwa et al. | Mar 2000 | A |
6038184 | Naritake | Mar 2000 | A |
6046935 | Takeuchi et al. | Apr 2000 | A |
6091666 | Arase et al. | Jul 2000 | A |
6134148 | Kawahara et al. | Oct 2000 | A |
6151248 | Harari et al. | Nov 2000 | A |
6222762 | Guterman et al. | Apr 2001 | B1 |
6230233 | Lofgren et al. | May 2001 | B1 |
6252800 | Chida | Jun 2001 | B1 |
6266273 | Conley et al. | Jul 2001 | B1 |
6282624 | Kimura et al. | Aug 2001 | B1 |
6353553 | Tamada et al. | Mar 2002 | B1 |
6426893 | Conley et al. | Jul 2002 | B1 |
6456528 | Chen | Sep 2002 | B1 |
6480423 | Toda et al. | Nov 2002 | B2 |
6510488 | Lasser | Jan 2003 | B2 |
6512263 | Yuan et al. | Jan 2003 | B1 |
6522580 | Chen et al. | Feb 2003 | B2 |
6523132 | Harari et al. | Feb 2003 | B1 |
6560146 | Cernea | May 2003 | B2 |
6581142 | Jacobs | Jun 2003 | B1 |
6594177 | Matarrese et al. | Jul 2003 | B2 |
6657891 | Shibata et al. | Dec 2003 | B1 |
6771536 | Li et al. | Aug 2004 | B2 |
6813184 | Lee | Nov 2004 | B2 |
6853596 | Cheung | Feb 2005 | B2 |
6870768 | Cernea et al. | Mar 2005 | B2 |
6967873 | Hamilton et al. | Nov 2005 | B2 |
6990018 | Tanaka et al. | Jan 2006 | B2 |
6996017 | Scheuerlein et al. | Feb 2006 | B2 |
7027330 | Park | Apr 2006 | B2 |
7039781 | Iwata et al. | May 2006 | B2 |
7057939 | Li et al. | Jun 2006 | B2 |
7058818 | Dariel | Jun 2006 | B2 |
7076611 | Steere et al. | Jul 2006 | B2 |
7110294 | Kawai | Sep 2006 | B2 |
7158421 | Li et al. | Jan 2007 | B2 |
7170802 | Cernea et al. | Jan 2007 | B2 |
7206230 | Li et al. | Apr 2007 | B2 |
7224605 | Moogat | May 2007 | B1 |
7257689 | Baird | Aug 2007 | B1 |
7299314 | Lin et al. | Nov 2007 | B2 |
7310347 | Lasser | Dec 2007 | B2 |
7345928 | Li | Mar 2008 | B2 |
7405985 | Cernea et al. | Jul 2008 | B2 |
7411846 | Terzioglu | Aug 2008 | B2 |
7420847 | Li | Sep 2008 | B2 |
7426623 | Lasser | Sep 2008 | B2 |
7447070 | Cernea | Nov 2008 | B2 |
7490283 | Gorobets et al. | Feb 2009 | B2 |
7493457 | Murin | Feb 2009 | B2 |
7502254 | Murin et al. | Mar 2009 | B2 |
7502259 | Gorobets | Mar 2009 | B2 |
8144512 | Huang et al. | Mar 2012 | B2 |
20010000023 | Kawahara et al. | Mar 2001 | A1 |
20020118574 | Gongwer et al. | Aug 2002 | A1 |
20030021169 | Beer et al. | Jan 2003 | A1 |
20030182317 | Kahn et al. | Sep 2003 | A1 |
20030223274 | Cernea | Dec 2003 | A1 |
20040060031 | Cernea | Mar 2004 | A1 |
20040109357 | Cernea et al. | Jun 2004 | A1 |
20050073884 | Gonzalez et al. | Apr 2005 | A1 |
20050141387 | Cernea et al. | Jun 2005 | A1 |
20050144360 | Bennett et al. | Jun 2005 | A1 |
20050144365 | Gorobets et al. | Jun 2005 | A1 |
20050213393 | Lasser | Sep 2005 | A1 |
20060126390 | Gorobets | Jun 2006 | A1 |
20060136656 | Conley et al. | Jun 2006 | A1 |
20060161728 | Bennett et al. | Jul 2006 | A1 |
20060256631 | Nobunaga et al. | Nov 2006 | A1 |
20070061502 | Lasser | Mar 2007 | A1 |
20070065119 | Pomerantz | Mar 2007 | A1 |
20070091677 | Lasser et al. | Apr 2007 | A1 |
20070103977 | Conley et al. | May 2007 | A1 |
20070159652 | Sato | Jul 2007 | A1 |
20070180346 | Murin | Aug 2007 | A1 |
20070186032 | Sinclair et al. | Aug 2007 | A1 |
20070211530 | Nakano | Sep 2007 | A1 |
20070220197 | Lasser | Sep 2007 | A1 |
20070237006 | Murin et al. | Oct 2007 | A1 |
20070260808 | Raines et al. | Nov 2007 | A1 |
20070268745 | Lasser | Nov 2007 | A1 |
20070283081 | Lasser | Dec 2007 | A1 |
20070285980 | Shimizu et al. | Dec 2007 | A1 |
20080104309 | Cheon et al. | May 2008 | A1 |
20080104312 | Lasser | May 2008 | A1 |
20080147996 | Jenkins et al. | Jun 2008 | A1 |
20080159012 | Kim | Jul 2008 | A1 |
20080181000 | Lasser | Jul 2008 | A1 |
20080209112 | Yu et al. | Aug 2008 | A1 |
20080244338 | Mokhlesi et al. | Oct 2008 | A1 |
20080244367 | Chin et al. | Oct 2008 | A1 |
20080250220 | Ito | Oct 2008 | A1 |
20080250300 | Mokhlesi et al. | Oct 2008 | A1 |
20080279005 | France | Nov 2008 | A1 |
20080294814 | Gorobets | Nov 2008 | A1 |
20080301532 | Uchikawa et al. | Dec 2008 | A1 |
20080310224 | Roohparvar et al. | Dec 2008 | A1 |
20090067244 | Li et al. | Mar 2009 | A1 |
20090089481 | Kapoor et al. | Apr 2009 | A1 |
20090089520 | Saha et al. | Apr 2009 | A1 |
20090094482 | Zilberman | Apr 2009 | A1 |
20090172247 | Bar-Or et al. | Jul 2009 | A1 |
20090172498 | Schlick et al. | Jul 2009 | A1 |
20090303814 | Fekih-Romdhane | Dec 2009 | A1 |
20090310408 | Lee et al. | Dec 2009 | A1 |
20100107004 | Bottelli et al. | Apr 2010 | A1 |
20100157641 | Shalvi et al. | Jun 2010 | A1 |
20100172179 | Gorobets et al. | Jul 2010 | A1 |
20100172180 | Paley et al. | Jul 2010 | A1 |
20100174845 | Gorobets et al. | Jul 2010 | A1 |
20100174846 | Paley et al. | Jul 2010 | A1 |
20100174847 | Paley et al. | Jul 2010 | A1 |
20100287217 | Borchers et al. | Nov 2010 | A1 |
20100325351 | Bennett | Dec 2010 | A1 |
20100329007 | Chibvongodze et al. | Dec 2010 | A1 |
20110063909 | Komatsu | Mar 2011 | A1 |
20110096601 | Gavens et al. | Apr 2011 | A1 |
20110099460 | Dusija et al. | Apr 2011 | A1 |
20110153913 | Huang et al. | Jun 2011 | A1 |
Number | Date | Country |
---|---|---|
1549133 | Nov 2004 | CN |
61292747 | Dec 1986 | JP |
1128297 | May 1989 | JP |
01128297 | May 1989 | JP |
06150666 | May 1994 | JP |
WO 9844420 | Oct 1998 | WO |
WO 0049488 | Aug 2000 | WO |
WO 03025939 | Mar 2003 | WO |
WO 2006064318 | Jun 2006 | WO |
WO 2007141783 | Dec 2007 | WO |
Entry |
---|
“Numonyx Sector-Based Compact File System (SCFS) Software is a Feature-Rich Flash Solution,” Numonyx, Nov. 3, 2009, 2 pages. |
Choudhuri et al., “Performance Improvement of Block Based NAND Flash. Translation Layer,” CODES + ISSS '07, Salzburg, Austria, Sep. 2007, pp. 257-262. |
Chang et al., “Real-Time Garbage Collection for Flash-Memory Storage Systems of Real-Time Embedded Systems,” Nov. 2004, ACM, ACM Transactions on Embedded Computing Systems, vol. 3, pp. 837-863. |
Lee et al., “Block Recycling Schemes and Their Cost-Based Optimization in NAND Flash Memory Based Storage System,” Oct. 2007, ACM., EMSOFT '07, pp. 174-182. |
Kang et al., “A Superblock-Based Flash Translation Layer for NAND Flash Memory,” EMSOFT'06, Oct. 2006, pp. 161-170. |
Im et al., “Storage Architecture and Software Support for SLC/MLC Combined Flash Memory,” Mar. 2009, ACM, SAC'09, pp. 1664-1669. |
Notification Concerning Transmittal of International Preliminary Report for Int'l Application No. PCT/US2010/040828 dated Jan. 19, 2012, 16 pages. |
“Numonyx Sector-Based Compact File System (SCFS) Software is a Feature-Rich Flash Solution,” Numonyx, 2009, 2 pages. |
“SanDisk, Toshiba Develop 32-Nanometer NAND Flash Technology,” SanDisk Corporation and Toshiba Corporation, Feb. 11, 2009, www.physorg.com/news153597019.html, 9 pages. |
Choudhuri et al., “Performance Improvement of Block Based NAND Flash Translation Layer,” CODES + ISSS '07, Salzburg, Austria, pp. 257-262. |
Communication Relating to the Results of the Partial International Search, International Appl. No. PCT/US2010/040828 mailed Oct. 20, 2010, 1 page. |
Eitan et al., “NROM: A Novel Localized Trapping, 2-Bit Nonvolatile Memory Cell,” IEEE Electron Device Letters, vol. 21, No. 11, Nov. 2000, pp. 543-545. |
Notification of Transmittal of the International Search Report and then Written Opinion of the International Searching Authority, or the Declaration for International Application No. PCT/US2010/040828 dated Jan. 10, 2011, 18 pages. |
U.S. Office Action for U.S. Appl. No. 12/498,220 mailed Jun. 10, 2011, 17 pages. |
U.S. Appl. No. 11/026,536 entitled “Non-Volatile Memory and Method with Shared Processing for an Aggregate of Read/Write Circuits,” filed Dec. 29, 2004, 50 pages. |
U.S. Appl. No. 12/051,462 entitled “Adaptive Algorithm in Cache Operation with Dynamic Data Latch Requirements,” filed Mar. 19, 2008, 20 pages. |
U.S. Appl. No. 12/051,492 entitled “Different Combinations of Wordline Order and Look-Ahead Read to Improve Flash Memory Performance,” filed Mar. 19, 2008, 20 pages. |
U.S. Appl. No. 12/348,819 entitled “Wear Leveling for Non-Volatile Memories: Maintenance of Experience Count and Passive Techniques,” filed Jan. 5, 2009, 73 pages. |
U.S. Appl. No. 12/348,825 entitled “Spare Block Management in Non-Volatile Memories,” filed Jan. 5, 2009, 76 pages. |
U.S. Appl. No. 12/348,891 entitled “Non-Volatile Memory and Method With Write Cache Partitioning,” filed Jan. 5, 2009, 151 pages. |
U.S. Appl. No. 12/348,895 entitled “Nonvolatile Memory With Write Cache Having Flush/Eviction Methods,” filed Jan. 5, 2009, 151 pages. |
U.S. Appl. No. 12/348,899 entitled “Non-Volatile Memory and Method With Write Cache Partition Management Methods,” filed Jan. 5, 2009, 149 pages. |
U.S. Appl. No. 12/478,997 entitled “Folding Data Stored in Binary Format into Multi-State Format Within Non-Volatile Devices,” filed Jun. 5, 2009, 52 pages. |
U.S. Appl. No. 12/642,740 entitled “Atomic Program Sequence and Write Abort Detection,” filed Dec. 18, 2009, 60 pages. |
U.S. Appl. No. 61/142,620 entitled “Non-Volatile Memory and Method With Improved Block Management System,” filed Jan. 5, 2009, 144 pages. |
Raz et al., “Implementing MLC NAND Flash for Cost-Effective, High-Capacity Memory,” Sep. 2003., M-Systems, 13 pages. |
U.S. Office Action for U.S. Appl. No. 12/498,220 mailed Oct. 29, 2010, 25 pages. |
Number | Date | Country | |
---|---|---|---|
20120297245 A1 | Nov 2012 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 12498220 | Jul 2009 | US |
Child | 13293494 | US |