This subject matter is related generally to access and management of managed non-volatile memory (NVM).
Flash memory is a type of electrically erasable programmable read-only memory (EEPROM). Because flash memories are non-volatile and relatively dense, they are used to store files and other persistent objects in handheld computers, mobile phones, digital cameras, portable music players, and many other devices in which other storage solutions (e.g., magnetic disks) are inappropriate.
NAND is a type of flash memory that can be accessed like a block device, such as a hard disk or memory card. Each block consists of a number of pages (e.g., 64-128 pages). A typical page size is 4 KB-8 KB bytes. A NAND device can have multiple dies each having 4096-8192 blocks. Associated with each page are a number of bytes that are used for storage of error detection and correction checksums. Reading and programming is performed on a page basis, erasure is performed on a block basis, and data in a block can only be written sequentially. NAND relies on Error Correction Code (ECC) to compensate for bits that may flip during normal device operation. When performing erase or program operations, the NAND device can detect blocks that fail to program or erase and mark the blocks as bad in a bad block map. The data can be written to a different, good block, and the bad block map updated.
Managed NAND devices combine raw NAND with a memory controller to handle error correction and detection, as well as memory management functions of NAND memory. Managed NAND is commercially available in Ball Grid Array (BGA) packages, or other Integrated Circuit (IC) package which supports standardized processor interfaces, such as Multimedia Memory Card (MMC) and Secure Digital (SD) card. A managed NAND device can include a number of NAND devices or dies, which can be accessed using one or more chip select signals. A chip select is a control line used in digital electronics to select one chip out of several chips connected to the same bus. The chip select is typically a command pin on most IC packages, which connects the input pins on the device to the internal circuitry of that device. When the chip select pin is held in the inactive state, the chip or device ignores changes in the state of its input pins. When the chip select pin is held in the active state, the chip or device responds as if it is the only chip on the bus.
The Open NAND Flash Interface Working Group (ONFI) has developed a standardized low-level interface for NAND flash chips to allow interoperability between conforming NAND devices from different vendors. ONFI specification version 1.0 specifies: a standard physical interface (pin-out) for NAND flash in TSOP-48, WSOP-48, LGA-52, and BGA-63 packages; a standard command set for reading, writing, and erasing NAND flash chips; and a mechanism for self-identification. ONFI specification version 2.0 supports dual channel interfaces, with odd chip selects (also referred to as chip enable or “CE”) connected to channel 1 and even CEs connected to channel 2. The physical interface shall have no more than 8 CEs for the entire package.
While the ONFI specifications allow interoperability, the current ONFI specifications do not take full advantage of managed NAND solutions.
A read command initiates reads of pages or portions of pages of non-volatile memory using a memory address that specifies a row, column and length. A host controller can use the read command with a read operation or status request. In some implementations, the memory address further specifies a die or plane and a block.
In some implementations, a method comprises: transmitting a partial page read command to a non-volatile memory device, the partial page read command specifying a subset of a page of non-volatile memory; and receiving data retrieved by the non-volatile memory device from the specified subset of the page in response to the partial page read command.
In some implementations, a method comprises: transmitting a status request command to a non-volatile memory device, the status request command specifying a subset of a page of non-volatile memory; and responsive to the status request command, receiving status data from the non-volatile memory device, the status data associated with the specified subset of the page.
Other implementations are disclosed related to apparatuses, systems, methods and computer-readable mediums.
In some implementations, the NVM package 104 can include a controller 106 for accessing and managing the NVM devices 108 over internal channels using internal chip select signals. An internal channel is a data path between the controller 106 and a NVM device 108. The controller 106 can perform memory management functions (e.g., wear leveling, bad block management) and can include an error correction (ECC) engine 110 for detecting and correcting data errors (e.g., flipped bits). In some implementations, the ECC engine 110 can be implemented as a hardware component in the controller 106 or as a software component executed by the controller 106. In some implementations, the ECC engine 110 can be located in the NVM devices 108.
In some implementations, the host controller 102 and NVM package 104 can communicate information (e.g., control commands, addresses, data) over a communication channel visible to the host (“host channel”). The host channel can support standard interfaces, such as raw NAND interfaces or dual channel interfaces, such as is described in ONFI specification version 2.0. The host controller 102 can also provide a host chip enable (CE) signal. The host CE is visible to the host controller 102 to select the host channel.
In the exemplary memory system 100, the NVM package 104 supports CE hiding. CE hiding allows the single host CE to be used for each internal channel in the NVM package 104, thus reducing the number of signals required to support the interface of the NVM package 104. Memory accesses can be mapped to internal channels and the NVM devices 108 using an address space and mapping scheme, as described in reference to
The address mapping will now be described with respect to the exemplary memory architecture of
The address mapping shown in
The NVM package 104 defines a CAU as an area that can be accessed (e.g., moving data from the NAND memory cells to an internal register) simultaneous to, or in parallel with other CAUs. In this exemplary architecture, it is assumed that all CAUs include the same number of blocks. In other implementations, CAUs can have a different numbers of blocks. Table I below describes an exemplary row address format for accessing a page in a CAU.
Referring to Table I, an exemplary n-bit (e.g., 24 bits) row address can be presented to a controller in the NAND device in the following format: [CAU: Block: Page]. CAU is a number (e.g., an integer) that represents a die or plane. Block is a block offset in the CAU identified by the CAU number, and Page is a page offset in the block identified by Block. For example, in a device with 128 pages per block, 8192 blocks per CAU and 6 CAUs: X will be 7 (27=128), Y will be 13 (213=8192) and Z will be 3 (22<6<23).
The exemplary NVM package 104 shown in
The NVM package includes an NVM controller 202, which communicates with the CAUs through control bus 208 and address/data bus 210. During operation, the NVM controller 202 receives commands from the host controller (not shown) and in response to the command asserts control signals on the control bus 208 and addresses or data on the address/data bus 210 to perform an operation (e.g., read, program, or erase operation) on one or more CAUs. In some implementations, the command includes a row address having the form [CAU: Block: Page], as described in reference to
In the example shown in
In an implementation that uses CAUs, the memory address can be [CAU: Block: Page: Col: Length], where CAU determines a die or plane containing the data to be read, Block identifies a block in the CAU containing the data to be read, Page specifies a page in the block containing the data to be read, Col specifies an offset into the Page where the data to be read is stored and Length specifies a number of bytes of the data to be read.
Following the row address is the code 37h, which is a command that commits the read operation by the device. The 37H code is followed by the code 77h, which indicates the beginning of a read operation status command. Following the code 77h is the row address for which the status is to be performed on, represented by bytes R1, R2, R3. The row address is followed by the code 40h, which is a mask value that is being waited for on the bus by the host controller to indicate that the remaining bits of the status byte being read are valid and that the operation is complete, allowing progress to continue and the final portion of the operation, the actual data transfer, to commence. Following the code 40h is the actual reading of data from memory which is represented by R-d0 . . . R-dN, where N is the number of pages to be read starting from the page address specified by R1, R2, R3 in the multipage read operation.
While this specification contains many specifics, these should not be construed as limitations on the scope of what being claims or of what may be claimed, but rather as descriptions of features specific to particular embodiments. Certain features that are described in this specification in the context of separate embodiments can also be implemented in combination in a single embodiment. Conversely, various features that are described in the context of a single embodiment can also be implemented in multiple embodiments separately or in any suitable subcombination. Moreover, although features may be described above as acting in certain combinations and even initially claimed as such, one or more features from a claimed combination can in some cases be excised from the combination, and the claimed combination may be directed to a subcombination or variation of a subcombination.
Similarly, while operations are depicted in the drawings in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order, or that all illustrated operations be performed, to achieve desirable results. In certain circumstances, multitasking and parallel processing may be advantageous. Moreover, the separation of various system components in the embodiments described above should not be understood as requiring such separation in all embodiments, and it should be understood that the described program components and systems can generally be integrated together in a single software product or packaged into multiple software products.
Thus, particular embodiments have been described. Other embodiments are within the scope of the following claims.
This application claims the benefit of priority from U.S. Provisional Patent Application No. 61/176,087 filed May 6, 2009, which is incorporated by reference herein in its entirety.
Number | Date | Country | |
---|---|---|---|
61176087 | May 2009 | US |