The present invention relates generally to a storage control system, and more particularly to a system for data management and control.
Recently, forms of long term storage other than electromechanical hard disks have become feasible for use in computers. One of these is flash electrically erasable programmable read only memory (EEPROM). Flash EEPROM memory is comprised of a large plurality of floating-gate field effect transistors arranged as memory cells in typical row and column fashion with circuitry for accessing individual cells and placing the memory transistors of those cells in one of two memory conditions.
A flash memory cell, like a typical EEPROM cell but in contrast to dynamic random access memory (DRAM), retains information when power is removed. Flash EEPROM memory has a number of characteristics, which adapt it to use as long term memory. It is light in weight, occupies very little space, and consumes less power than electromechanical disk drives. More importantly, it is especially rugged. It will withstand without adverse effects repeated drops each of which would destroy a typical electromechanical hard disk drive. A problem exhibited by flash memory is that it tends to have a limited life in use.
Thus, a need still remains for better data management devices. In view of the increasing demand for data management devices, it is increasingly critical that answers be found to these problems. In view of the ever-increasing commercial competitive pressures, along with growing consumer expectations and the diminishing opportunities for meaningful product differentiation in the marketplace, it is critical that answers be found for these problems. Additionally, the need to reduce costs, improve efficiencies and performance, and meet competitive pressures adds an even greater urgency to the critical necessity for finding answers to these problems.
Solutions to these problems have been long sought but prior developments have not taught or suggested any solutions and, thus, solutions to these problems have long eluded those skilled in the art.
The present invention provides a method of operation of a storage control system including: providing a memory controller; accessing a volatile memory table by the memory controller; writing a non-volatile semiconductor memory for persisting changes in the volatile memory table; and restoring a logical-to-physical table in the volatile memory table, after a power cycle, by restoring a random access memory with a logical-to-physical partition from a most recently used list.
The present invention provides a storage control system including: a memory controller; a volatile memory table accessed by the memory controller; a non-volatile semiconductor memory written for persisting changes in the volatile memory table; and a logical-to-physical table in the volatile memory table, restored after a power cycle, by a random access memory restored with a logical-to-physical partition from a most recently used list.
Certain embodiments of the invention have other steps or elements in addition to or in place of those mentioned above. The steps or element will become apparent to those skilled in the art from a reading of the following detailed description when taken with reference to the accompanying drawings.
The following embodiments are described in sufficient detail to enable those skilled in the art to make and use the invention. It is to be understood that other embodiments would be evident based on the present disclosure, and that system, process, or mechanical changes may be made without departing from the scope of the present invention.
In the following description, numerous specific details are given to provide a thorough understanding of the invention. However, it will be apparent that the invention may be practiced without these specific details. In order to avoid obscuring the present invention, some well-known circuits, system configurations, and process steps are not disclosed in detail.
The drawings showing embodiments of the system are semi-diagrammatic and not to scale and, particularly, some of the dimensions are for the clarity of presentation and are shown exaggerated in the drawing FIGs. Similarly, although the views in the drawings for ease of description generally show similar orientations, this depiction in the FIGs. is arbitrary for the most part. Generally, the invention can be operated in any orientation.
The term “K” refers to decimal numerical value of one thousand and twenty four. The term “bit” refers to a decimal numeric value of zero or one, represented, and expressed in a binary numeral system or base-2 number system. The term “B”, also referred to as a byte, refers to an eight bit representation having a decimal numeric value between zero and two hundred and fifty five, inclusive. The term “KB” refers to one thousand and twenty four bytes.
Where multiple embodiments are disclosed and described having some features in common, for clarity and ease of illustration, description, and comprehension thereof, similar and like features one to another will ordinarily be described with similar reference numerals. The embodiments have been numbered first embodiment, second embodiment, etc. as a matter of descriptive convenience and are not intended to have any other significance or provide limitations for the present invention.
Referring now to
For illustrative purposes, the storage control system 100 is shown having the storage controller 102, the memory controller 104, the volatile memory table 108, at least the one non-volatile semiconductor memory 112, and the non-volatile memory space 114. It is understood that the storage control system 100 can have any number of similar or different semiconductor devices. For example, the memory controller 104 could be physically integrated in the storage controller 102 and the non-volatile memory space 114 could physically be in a portion of one or more of the non-volatile semiconductor memory 112.
The non-volatile semiconductor memory 112 can also be referred to as a NVS memory 112. The non-volatile memory space 114 can also be referred to as a NVM space 114. Hardware, software, or a combination thereof can be used to handle, control, and refresh the semiconductor devices or the storage control system 100.
Any software used to control or monitor the semiconductor devices can also be referred to as a flash layer. Portions of the flash layer can be implemented in hardware to optimize product cost or performance.
For illustrative purposes, the NVM space 114 is shown as a separate module. The NVM space 114 can be physically integrated in one of the semiconductor devices or outside the storage control system 100. For example, the NVM space 114 can physically be a reserved portion of the NVS memory 112. In another example, the NVM space 114 can be a semiconductor chip physically located outside the storage control system 100 and electrically connected to the storage control system 100 using a separate interface.
Contents of the NVM space 114 is used to reconstruct or rebuild all or portions of contents of the volatile memory table 108. The volatile memory table 108 can also be referred to as a VM table 108 and can include ram, DRAM, DDRAM, SRAM, SDRAM, or any volatile semiconductor memory device. The contents of the VM table 108 must have current information to ensure proper access to data in the NVS memory 112. The NVM space 114 is updated any time a change or update is made to the VM table 108 to enable the NVM space 114 to have the capability to restore or initialize the VM table 108.
In addition to the semiconductor devices, the storage control system 100 includes at least one system connector 118. The system connector 118 can provide electrical connectivity between the storage controller 102 and a host (not shown), such as a computer, test machinery, a cell phone, a video device, or any component having electronics requiring storage or retrieval of customer data using a customer data bus 122.
The storage controller 102 forwards a customer data address on a data address bus 116 received from the host to the memory controller 104 to electrically connect to or access physical locations within the NVS memory 112. The storage controller 102 can receive a read command and respond by sending customer data on the customer data bus 122, previously stored in the NVS memory 112, such as such as flash, NAND flash, NOR flash, NVRAM, or any non-volatile semiconductor memory device, to the host through the system connector 118.
The storage controller 102 can receive a write command and the customer data from the host to be written or stored in to the NVS memory 112, such as such as flash, NAND flash, NOR flash, or any non-volatile semiconductor memory device. The storage controller 102 can receive other commands that result on the reading, writing, or a combination of read and write operations to the NVS memory 112.
The storage controller 102 can also include an allocation unit, referred to as an AU, containing data, error correction code (ECC), and metadata and always processed as one unit. The AU cannot be subdivided in such a way to access components within it as the ECC applies to all of the components in the AU as a group. There may be one or more AUs stored in a physical page but the controller processes elements on AU boundaries.
The data address bus 116 can be referred to as a DA bus 116. The memory controller 104 receives control signals and the customer data address from the storage controller 102. The memory controller 104 forwards the customer data address on the DA bus 116 from the storage controller 102 to the VM table 108 using a logical address bus 124, also referred to as a LA bus 124.
The VM table 108 contains a table that effectively maps or translates the customer data address from the LA bus 124 to electrical signals 128 sent from the VM table 108 to the NVS memory 112. The electrical signals 128 are used to connect, control, or access one or more physical memory locations of the NVS memory 112 for reading or writing of data, including the customer data.
The electrical signals 128 can include signals representing flash addresses, physical memory chip address, semiconductor chip selects, page addresses, subpage addresses, erase block controls, channel selects, multiplexor circuit controls, write inhibits, user defined address banks, user defined selects, or any combination thereof.
The control signals from the storage controller 102 to the memory controller 104 can be used to indicate whether a pending memory operation will access a physical memory location without modifying any physical memory location for the customer data in the NVS memory 112 or modify the physical memory location designated for the customer data in the NVS memory 112.
An example of a pending memory operation that would not require any modifications to any physical memory location for the customer data is a normal read operation, free of any disturb related action, from a memory location the NVS memory 112. Examples of a pending memory operation that would require memory modifying operations are a write, an erasure, a re-write, or a reprogram operation, to contents of a memory location in the NVS memory 112.
The VM table 108 is not updated when the pending memory operation is not modifying any physical memory location for the customer data. Also, if the VM table 108 is not to be updated, the memory controller 104 does not temporarily save the customer data address on the logical address bus 124 or the DA bus 116 that is associated with the pending memory operation.
The VM table 108 is updated when the pending memory operation is to modify a physical memory location for the customer data. Also, the memory controller 104 can temporarily save the customer data address from the logical address bus 124 or the DA bus 116 that is associated with the pending memory operation into a local buffer 132.
The customer data address saved in the local buffer 132 until change events have been detected by the memory controller 104. Each of the change events are the result of the pending memory operation completion and includes an update to the VM table 108 that includes the VM table 108 sending a different or displaced physical memory location on the electrical signals 128 to the NVS memory 112.
The different or displaced physical memory location stored in the VM table 108 and sent to the NVS memory 112 is captured and stored in the local buffer 132 with the customer data address previously saved. The memory controller can either immediately write the contents of the local buffer 132 or provide several pairs of displaced physical memory locations and corresponding customer data address before writing into the NVM space 114 using a non-volatile memory interface or NVM interface 136.
The NVM interface 136 can be a serial or parallel interface and include a serial peripheral interface (SPI), a firewire interface (P1394), or a user defined interface capable of accessing the non-volatile semiconductor memory of the storage control system 100. The memory controller 104 is persistent in continued updates to the NVM space 114 in response to any of the change events ensuring that the contents of the NVM space 114 represents the current contents of the VM table 108. The contents of the NVM space 114, having current information representing the current contents of the VM table 108, is always available to be used to restore or initialize the VM table 108.
It has been discovered that the NVM space 114 substantially reduces the latency time to initialize the VM table 108 of semiconductor storage devices.
It has been discovered that the NVM space 114 substantially reduces the latency from a power-up or a system recovery until start of customer applications for semiconductor storage devices.
It has been discovered that the NVM space 114 can be updated in parallel with the updating of the VM table 108 to provide zero latency for the NVM space 114 backups and no degradation to the semiconductor storage device performance.
It has been discovered that persistent updates to the NVM space 114 based on the change events substantially reduces the amount of data, the power-down latency, the power-up latency, or the need for additional operational resources from hardware or software to provide maximum semiconductor storage device performance.
It has been discovered that the updates to the NVM space 114 due to the change events can continue indefinitely and does not require any snapshots after the NVM space 114 has been initialized resulting in reliability improvements of the NVM space 114.
Referring now to
For illustrative purposes, the flash page 202 is shown having two logical flash pages 204. It is understood that the flash page 202 can have any number of logical pages. For example, the flash page 202 could have been defined have four logical pages.
Also for illustrative purposes, the flash page 202 is shown having a flash page size 206 of 8K bytes, formed from logical pages with a size of 4K bytes. It is understood that the logical pages and the flash page can be any size. For example, the logical pages could have a size of 6K bytes resulting in the flash page having a size of 12K bytes. Flash pages can be of fixed size set forth by the manufacturer as part to the architecture and design of the chip. The variable size of the host pages can make it advantageous to put more or less logical pages in a physical page. NAND flash page sizes have historically been a power of 2. For example, the flash pages can be 1K, 2K, 4K, 8K, or 16K in size. There are also mechanisms to access multiple physical flash pages as one unit. For example, one or more flash pages can be written a time providing an increased data bandwidth for the device.
Referring now to
An example, of a logical page contains 4K bytes of 4K-aligned and contiguous user data for a 512 Byte sector sized semiconductor storage system is discussed. It is understood that the logical page can have other sizes. For example, the logical page could have a size of 8K bytes.
The logical page would be larger than 4K byte if the host is using a 520 or 528 byte sector size. For example, with a 4 KB logical page, the L2P table 302 can be 200 MB to exemplify a logical-to-physical mapping table organized for a 200 GB semiconductor storage system.
The L2P table 302 is organized having a one-directional logical-to-physical mapping to access the NVS memory 112. The L2P table 302 can have a structure that includes logical page numbers 304, logical page entries 306, logical-to-physical identifiers 308, and a physical identifier length 312.
The logical page numbers 304, also referred to as LP offsets 304, identify physical locations of the logical page entries 306 in the L2P table 302 and are not stored in the L2P table 302. The logical page entries 306 can also be referred to as LP entries 306 and contain the logical-to-physical identifiers 308 that can include the physical memory location indirectly used for the generation of the electrical signals 128 of
Each of the L2P IDs 308 in the LP entries 306 have the physical identifier length 312, predetermined and based on a range or size of a specific flash memory system (not shown). The physical identifier length 312 can be referred to as a PIL 312.
For illustrative purposes, the PIL 312 can have a length of four bytes. It is understood that the PIL 312 can be of any length. For example, the PIL 312 can have a length of twenty four bits. In another example, the PIL 312 can have a length of twenty eight bits for a different flash memory system (not shown).
It has been discovered that the NVM space 114 of
It has been discovered that the L2P table 302 can be structured to access any size flash page configuration in the NVS memory 112 to provide a compact sized backup data structure for the NVM space 114.
It has been discovered that the entire L2P table 302 can be rebuilt from the most recent version of the NVM space 114 to provide the storage control system 100 recover from any product power cycle event, in which power is removed and then restored at a later time.
It has been discovered that the most recent version of the L2P table 302 can be determined by either traversing through the NVM space 114 or saving any changes to the structure of the L2P table 302 during a system power down.
It has been discovered that the saving of any changes to the structure of the L2P table 302 during power down can be eliminated by using random-access non-volatile memory for the NVM space 114 resulting in a minimized power down latency.
It has been discovered that storing the entire L2P table 302 having a total storage capacity equal to one tenth of a percent, or less for different logical page size organizations, of the logical drive capacity of a memory system in DDRAM provides optimal random-read performance over memory systems without the L2P table 302 of the present invention.
It has been discovered that contents of the NVM space 114 can be used to identify regions of the data structure that are seldom used and can be made available for providing additional redundant methods or structures used to rebuild the L2P table 302 resulting in a high reliability and availability semiconductor storage system.
It has been discovered that the combination of the NVM space 114 and the VM table 108 provide the storage control system 100 with the capability to rebuild the data structure of the VM table 108 only as need, minimizing flash wear and read disturb errors resulting in a extremely reliable semiconductor storage system.
Referring now to
For illustrative purposes, two of the logical flash pages 204 are labeled and identified as logical page 399 and logical page 9012, respectively. It is understood that the labels and identifiers are arbitrary and the customer data in the NVS memory 112 is often non-sequential and often physically distributed and relocated randomly in the NVS memory 112. For example, the logical page 9012 could be physically located at the first location of the first flash page 202 of the NVS memory 112 in a subsequent operation from the host. In another example, the logical page 9012 could be further physically relocated to the last flash page 202 of the NVS memory 112.
An application of the host can request data from a customer data address that corresponds to one of the LP offsets 304 that identifies one of the LP entries 306 physically in the L2P table 302. The L2P IDs 308 stored in the one of the LP entries 306 is translated and interpreted to generate the electrical signals 128 of
In this example, the application of the host requests to read from a customer data address, known to the application only as having an address of nine thousand and twelve. The L2P table 302 in the VM table 108 can be indirectly used for purposes of generating the electrical signals 128 to access the physical location of the logical flash page 202 in the NVS memory 112 that corresponds to the data, labeled as the logical page 9012 in this example.
The storage control system 100 of
Referring now to
In this example, the application of the host is performing a write operation to update the customer data address known as nine thousand and twelve by the application. The storage control system 100 of
The storage control system 100 records and flags the L2P IDs 308 in the LP entries 306 corresponding to the customer data address known as nine thousand and twelve by the application as invalid. The invalidation may occur before, during, or after the write operation. The storage control system 100 updates the L2P IDs 308 at the LP entries 306 corresponding to the physical location of nine thousand and twelve of the VM table 108 with a different physical location previously identified, flagged, and chosen by the storage control system 100 as available for immediate use.
The write operation is executed and the newly updated L2P IDs 308 in the VM table 108 is translated and interpreted to generate the electrical signals 128 of
It has been discovered that maintaining the NVM space 114 of
Referring now to
Each of the physical location of entries in the L2P table 302 contains multiple logical-to-physical-page IDs referred to as L2P page IDs 608. Each of the L2P page IDs 608 have a physical identifier length, referred to as a PIL 612 in a manner similar to the PIL 312 of
Each of the L2P page ids 608 can be used to generate the electrical signals 120 of
Each individual L2P page IDs 608 within one of the L2P partition offsets 604 can be physically located by an offset from a starting location at the beginning of each of the L2P partition offsets 604 in a manner similar to the LP offsets 304 of
For purposes of discussion, each of the L2P partition offsets 604 can be formed having the PIL 612 equal to 512-bytes and each of the L2P partition offsets 604 can have one hundred twenty eight L2P page IDs 608 on a 200 GB drive. It is understood that the L2P partition offsets 604 or the quantity of the L2P page IDs 608 can be changed, resulting a larger or smaller quantity of storage space needed in the volatile memory of the VM table 108 of
Referring now to
This section describes how changes to the VM table 108 can be captured using the L2P update log 702 in the NVM space 114. Super log blocks 704 can be referenced to determine how the VM table 108 has been updated. The L2P update log 702 can span multiple of the super log blocks 704. The super log blocks 704 of the L2P update log 702 can be scattered throughout the drive and there is no restriction on the use of the super blocks. The super log blocks 704, when in chronological order, form a sequential log. This sequential log operates in read only mode during drive initializations.
Referring now to
The log super block page 802, can be referred to as a L_SB page 802. The L_SP page 802 can contain up to sixteen 512-byte partitions. Each of the byte partitions can also be referred to as L2P partitions 806. The 512-byte partitions contain the actual data from the VM table 108 of
Metadata pages 812 contain an index of all the metadata 808 in each of the 8 KB_LPs 804 within the super log blocks 704. The index includes all of the partition numbers of all partitions that are stored within the 8 KB_LPs 804 of each of the super log blocks 704. The metadata pages 812 provide a function similar to the logical page identified as a Page 63 used to manage, recycle, or group the metadata 808 for rebuilding, building, and tracking across tabling information during power cycles and used to recover from memory corruptions during normal operations. The metadata pages 812 enables efficient and quick access to the metadata 808 of the 8 KB_LPs 804 without having to individually read each and every one of the 8 KB_LPs 804.
It has been discovered that the metadata 808 provides a redundancy in the event of loss or corruption of the metadata pages 812 and can be used to rebuild the metadata pages 812, provide superior fault tolerance performance, or improve reliability of solid state storage devices.
It has been discovered that the metadata pages 812 provides a redundancy in the event of loss or corruption of the metadata 808 and can be used to rebuild any of the metadata 808 of the 8 KB_LPs 804, provide superior fault tolerance performance, or improve reliability to solid state storage devices.
It has been discovered that the metadata pages 812 provides the flash layer or hardware of solid state storage devices the ability to quickly access the metadata 808 for rapid analysis, initialization, or rebuild of the VM table 108 of
It has been discovered that the metadata pages 812 provides the flash layer or hardware of solid state storage devices the ability to quickly retrieve the metadata 808 and provides the flash layer or the hardware with a redundant or backup copy of the metadata 808.
Referring now to
For illustrative purposes, a logical structural organization 900 of the L_SB page 802 is shown. Note that the exact layout or structure of the L_SB page 802 of the 8 KB log pages 800 is optimized for the best overlay of the layout or the structure with a host data page. It is understood that the logical structural organization 900 could be different. For example, the logical structural organization 900 could have been represented having consecutive pairs of partitions separated by consecutive pairs of metadata, such as partition 0, partition 1, metadata 0, metadata 1, partition 2, partition 3, and etc.
Also shown is an example of first four elements 902 of one of the metadata pages 812 of
It has been discovered that the metadata pages 812 and the metadata 808 with robust redundancy protection can be used to provide additional user defined information for the initialization of the VM table 108, power-up, or power-down of the semiconductor storage device.
Referring now to
The L2P update log 702 can require more writes for random small-block sized host writes than for large and/or sequential host writes. The number of 8 KB_LPs 804 or the super log blocks 704 written depends on how the drive is being used and the number of L2P page IDs 608 of
Whenever the flash page 202 of
The L2P update log 702 can be used to rebuild the L2P table 302 during initialization. The initialization process occurs in two steps: first step is to rebuild the L2P table 302 and second step is to rebuild the L2P page IDs 608.
The main focus during initialization is to rebuild the L2P table 302. The L2P table 302 identifies the location of each partition in the flash, such as the L2P page IDs 608. The L2P update log 702 can be used to load the portion of the L2P table 302 first accessed before the entire L2P table 302 has been rebuilt to service host commands before the L2P table 302 has been fully loaded into the volatile memory table 108 or non-volatile memory.
In order to initialize the L2P table 302, the memory controller 104 determines the most recent version of each of the L2P partition 806 of
L2P update log 702 is traversed to determine the most recent version of each of the L2P partition 806 of
An example of an initialization of a data structure includes identifying the super log blocks 704 that comprise the L2P update log 702, sorting the super log blocks 704 in reverse chronological order using a sequence number of super log blocks 704, and evaluating the metadata pages 812 of
For example, a 420 GB drive with 1 KB partitions using the L2P update log 702 with a size equal to, five times greater than, or ten times greater than a size of the L2P table 302 can have total initialization times of four hundred and thirty, six hundred and two, and eight hundred and seventeen milliseconds, respectively. The total initialization times assume reads of the metadata pages 812 can be overlapped with writes to different flash devices, such as the NVS memory 112 of
It is assumed that the metadata pages 812 of the super log blocks 704 can be read and partition entries of the super log blocks 704, such as partition entries, can be processed in parallel. Also, the total initialization times assume that writing the L2P table 302 will take no more than one microsecond and skipping an invalid entry will take no more than one hundred nanoseconds and can change based on the read latency of the NAND flash. For example, the total initialization times can improve as the read latency of NAND flash improves
After initializing the L2P table 302, none, some, or all of the L2P table 302 of
If the entries of the L2P page IDs 608 match do not match entries of the location in the super block 702, the entries of the L2P page IDs 608 are patch up with corrected entries and a need to schedule the 8 KB_LPs 804 to be written to the super log blocks 704.
Any remaining initialization time can be used to start building the rest of the L2P table 302 or for other initialization tasks such as initializing lists of the super log blocks 704. As an example, the most-recent version of the logical pages has been determined in the super log blocks 704 that are open without any need to flush out any writes to the L2P update log 702.
The L2P update log 702 can represent the state from the last time a Page 63 has been written to one of the super log blocks 704. Changes in the super log blocks 704 that are open do need to be captured because the L2P partitions 806 can be loaded and patched up as described earlier and can eliminate the need to capture all updates, such as changes in the super log blocks 704 that are not open, in the L2P update log 702.
Furthermore, as long as the super log blocks 704 contain the metadata/partition information, there is no requirement that all logging super blocks must have the metadata pages 812. For example, for the L2P update log 702 having a size that is ten times greater than a size of the L2P table 302, the ordered list is 2120 bytes or 530 super blocks times 4 bytes long.
The latency or efficiency from using a SPI or serial peripheral interface to write the ordered list could be improved by writing the ordered list to non-volatile memory and sending the location of the ordered list in the non-volatile memory, not the ordered list itself, using the SPI. If the L2P update log 702 is used in a circular manner, the super log blocks 704 can be continually logged with the SPI and the L2P tabling operations can be handled without any special power-down sequence requirements. It is understood that the non-volatile memory can include NOR flash, NAND flash, or any available non-volatile memory device capable of communicating with the SPI.
If the location of the ordered list in NAND is lost during a power cycle, the location can be recovered by searching through the Page 63s combined metadata pages held within a super block, locate the super log blocks 704, and identify the super log blocks 704 that contain the active section of the open super log blocks 704 that are open. After locating the active section of the super log blocks 704, sequence numbers of the super log blocks 704 can be sorted to determine the ordered list in NAND.
Referring now to
Referring now to
Referring now to
The flash translation layer (FTL) can delay writing to the L2P update log 702 of
If a multiple pages within the L2P update log 702 are written, multiple changes can be captured for one write to one of the super log blocks 704 of
The cache 1302 consists of all modified L2P partitions 806 that have not yet been flushed out to the L2P update log 702. The cache 1302 is rebuilt during initialization. The size and complexity of the cache 1302 may depend on available memory and initialization time. The cache 1302 can be used for random writes.
Referring now to
The super log blocks 704 come from a resource pool 1402 of available super blocks. Once designated as an erased super block, the erased super block will be allocated as one of the super log blocks 704 to store the 8 KB_LPs 804 of
The super log blocks 704 can be written to and allowed to grow before any of the super log blocks 704 need to be recycled. The oldest of the super log blocks 704 can be recycled first. The L2P update log 702 is like a journal and can include multiple obsolete versions of the super log blocks 704.
The L2P update log 702 can be a fully compacted log containing no obsolete entries and having a size similar to a size of the L2P table 302 of
The amount of growth the L2P update log 702 can expand is limited by the amount of memory allocated for the L2P update log 702. The size of the L2P table 302 and the compacted size of the L2P update log 702 can each be approximately one-tenth of a percent of the drive capacity.
If the L2P update log 702 is allowed to grow ten times in size, it would consume approximately one percent or less of the total capacity of the semiconductor storage device. Increasing the size of the L2P update log 702 will increase the initialization time for the L2P table 302 while provide the capability to save more old versions of the super log blocks 704 and reduce the erase and recycle activities.
It has been discovered that the L2P update log 702 and the size of memory allocated for the L2P update log 702 can be used to optimize the design of any semiconductor storage device for a cost, performance, size, or a combination thereof.
Referring now to
The L2P update log 702 having multiples of the super log blocks 704 require that eventually one or more of the super log blocks 704 be recycled as they become obsolete. Also, the oldest of the super log blocks 704 in the L2P update log 702 will need to be recycled as contents of the L2P update log 702 increase in size.
The 8 KB_LPs 804 of
The partitions do not need to be read out of the 8 KB_LPs 804 that are recycled because the L2P update log 702 contains the most recent version, assuming that the partition has been previously loaded. The valid partitions can be located by reading the metadata pages 812 of
The worst-case write amplification caused by the writes to the L2P update log 702 are a function of the size of the L2P update log 702. For example, a the L2P update log 702 having a size ten times larger, would result in a worst-case write amplification of 1/10.
The circular wear leveling approach is forced to move log entries having lots of static read only partitions during recycling. The recycling process of the super log blocks 704 can naturally identify potential regions of static data. Separating static data from dynamic data reduces write amplification. The worst write amplification occurs when many of the 8 KB_LPs 804 are obsolete, such as dynamic pages, are evenly distributed throughout the super log blocks 704.
Recycling of the super log blocks 704 could be implemented using a static log for static partitions and a dynamic log for dynamic partitions. The oldest of the super log blocks 704 in the L2P update log 702 is recycled in to the static log. The static log will grow and require recycling, albeit at a much slow pace than the dynamic log. Since the static log does not require any read scrubbing since the static log is read only during initialization.
The L2P partitions 806 entries that are static and their associated logical pages can be grouped together in the static log to prevent interspersion with the dynamic data and increasing the write amplification. The static data can be marked as static and aggregated when moving a valid page during recycling.
For example, a partition marked as static can be moved to a super block having or collecting static data on a recycling of the partition. The L2P partitions 806 of the 8 KB_LPs 804 or the metadata pages 812 that are rarely modified compared to other of the other L2P partitions need to be moved to clean up the oldest of the super log blocks 704 in the L2P update log 702.
These L2P partitions may contain static data and used to identify static host data that can be grouped and separated from dynamic host data to further improve performance and reduce write amplification. Write amplification is greatest with obsolete dynamic pages are evenly distributed throughout super blocks.
It has been discovered that the L2P partitions of the 8 KB_LPs 804 or the metadata pages 812 can be used to identify and isolate static host or static customer data from dynamic flash pages to provide improved wear leveling characteristics that result in superior reliability and product life.
It has been discovered that moving static data for wear-leveling and read-scrubbing purposes to a higher than average program/erase count location results in significant wear leveling characteristic that result in superior reliability and product life.
Referring now to
Writes to a flash page log having large sequential block operations provides significantly smaller write amplifications. For example, writing 128 KB blocks randomly to the 8 KB physical flash page with the 1 KB partition size results in a write amplification of one-two hundred and fifty sixth or one-one hundred and twenty eighth when writes are within single partitions and writes are across two partitions, respectively.
A recycle log, similar to the L2P update log 702, can influence write amplification when recycled. For example, movement related to updates of the L2P partitions 806 can result in page movement within the recycle log.
Assuming eight partitions can be written into one log write operation, the write amplification is at most ⅛th. As described above, the write amplification is significantly less for large block or sequential operations. The write amplification associated with the recycle operations for a log ten times greater is at most one-eightieth when compared to the original host write of one-eighth.
The recycle writes themselves can trigger their own L2P log write and L2P log recycle writes, which can increase the write amplification for the writes by a maximum of one-eighth.
It has been discovered that the L2P update log 702 with the super log blocks 704 for hosts writes enables large blocks of data to be kept together, aligned, or sequential and recycled to provide minimized write amplification by enabling writes to one or two partitions of NAND flash at a time, resulting in superior drive performance and reliability.
Referring now to
The logical-to-physical log pages, also referred to as L2P log pages 1704, are used to rebuild a logical-to-physical table 1708, similar to the L2P table 302 of
The L2P partition table 1702 is smaller than the logical-to-physical table 1708 and can be physically located in a portion of the VM table 108 of
The L2P partition table 1702 identifies the most-recent version of each partition in the L2P log pages 1704. The L2P partition table 1702 enables the option of deferring or staggering the build of the logical-to-physical table 1708. The L2P Partition table 1702 is used to identify valid partitions when recycling super log blocks, such as any of the super log blocks 704 of
Since the L2P update log 702 of
For purposes of illustration, the L2P partition table 1702 for a 300 GB drive would be four hundred and nine thousand and six hundred partitions. Each of the L2P log pages 1704 could store 16 partitions. Selecting a size for the L2P partition table 1702 largely depends on the resources available for initializing the L2P partition table 1702. In another example, a 200 GB drive with four hundred nine thousand five hundred and ninety nine partitions and with each of the partitions having one hundred and twenty eight pages would require 1.6 MB of volatile memory.
For example, a 420 GB drive with a partition size of 2K-bytes and two hundred fifteen thousand and forty partitions, would have an initialization time of five tenths of a second. The 420 GB drive with a partition size of 1K-bytes and four hundred thirty thousand and eighty partitions, would have an initialization time of one second, for example.
The 420 GB drive with a partition size of 512-bytes and eight hundred sixty thousand, one hundred and sixty partitions, would have an initialization time of two seconds, for example. Note the initialization times are inversely proportional to the partition size and are estimations based on the L2P update log 702 having a size that is ten times the size of the logical-to-physical table 1708.
Referring now to
An increase in size of a solid state drive (SSD) results in an increase in size of the logical-to-physical array, such as the logical-to-physical table 302 of
It has been discovered that the logical-to-physical table 302 pointing to an address of flash page 1802 containing the logical-to-physical pieces 1804 can be used to minimize the required size of the RAM, simplify architecture of a SSD, or reduce interface traffic between the flash page 1802 and the RAM to improve performance of the SSD.
Referring now to
Referring now to
Size of the L2P pool can range from about one quarter of the size of the L2P table. For example, the size of a L2P table for a 480 GB drive would be over 480 MB. Memory size can be limited with drives having a capacity over 480 GB. Also, the larger the L2P pool size, the higher probability of identifying data in RAM as the drive remains powered on. Data structures can be manipulated in the volatile memory table 108 faster than manipulating the data structures in NAND. This also effective reduces the number of NAND operations.
Referring now to
The oldest entries can be removed when a pool is empty and a new entry is allocated. The MRU list can be doubly-linked to facilitate removal from any point in the list. Removal of a partition from the MRU list can result in the corresponding partition pointer in the partition table set to NULL to indicate that the L2P partition is no longer in RAM.
Referring now to
The metadata 2204 can be traversed so that each of the L2P log pages 1704 can be used to determine if a Partition Table Entry 2206 points to the flash page just read. If the flash page just read is in the log page, then it is still the most up-to-date partition, so it can be linked to the most recently used (MRU) list.
It has been discovered that the traversal of the metadata 2204 to the partition table entry 2206 will prevent additional reads of the flash page to access another instance of the Partition 2202 in the same page.
It has been further discovered that the traversal of the metadata 2204 can be efficiently implemented using the linked-list to read each 1K data into the partition table entry 2206 from a pool of updated or unused instances of the partition table entry 2206.
It has been discovered that the traversal of the metadata 2204 removes the need to back-fill the full L2P table. An estimated time for filling 480 MB of DDRAM using a DMA would take at least 150 milliseconds. The traversal of the metadata 2204 also removes need to read the entire L2P table, which is would require reading 480 MB from NAND flash and is estimated to take at least 600 milliseconds.
It has been discovered that the initialization time can be reduce by the storage control system 100 of
Referring now to
The components of
Referring now to
The changes in the 100 GB L2P table 2402 can be persistently captured and stored in non-volatile memory using logs in a manner similar to the L2P update log 702 of
Referring now to
L2P update log 2504, is an example of a data structure in the NVM space 114. The L2P update log 2504 must be updated in a persistent manner to reflect any updates to the VM table 108 of
Referring now to
Referring now to
The resulting method, process, apparatus, device, product, and/or system is straightforward, cost-effective, uncomplicated, highly versatile, accurate, sensitive, and effective, and can be implemented by adapting known components for ready, efficient, and economical manufacturing, application, and utilization.
Another important aspect of the present invention is that it valuably supports and services the historical trend of reducing costs, simplifying systems, and increasing performance.
These and other valuable aspects of the present invention consequently further the state of the technology to at least the next level.
While the invention has been described in conjunction with a specific best mode, it is to be understood that many alternatives, modifications, and variations will be apparent to those skilled in the art in light of the aforegoing description. Accordingly, it is intended to embrace all such alternatives, modifications, and variations that fall within the scope of the included claims. All matters hithertofore set forth herein or shown in the accompanying drawings are to be interpreted in an illustrative and non-limiting sense.
This application claims the benefit of U.S. Provisional Patent Application Ser. No. 61/440,400 filed Feb. 8, 2011, and the subject matter thereof is incorporated herein by reference thereto.
Number | Name | Date | Kind |
---|---|---|---|
4048481 | Bailey, Jr. et al. | Sep 1977 | A |
5311395 | McGaha et al. | May 1994 | A |
5479638 | Assar et al. | Dec 1995 | A |
5930504 | Gabel | Jul 1999 | A |
5949785 | Beasley | Sep 1999 | A |
5963983 | Sakakura et al. | Oct 1999 | A |
6091652 | Haehn et al. | Jul 2000 | A |
6275436 | Tobita et al. | Aug 2001 | B1 |
6345367 | Sinclair | Feb 2002 | B1 |
6356447 | Scafidi | Mar 2002 | B2 |
6381670 | Lee et al. | Apr 2002 | B1 |
6412080 | Fleming et al. | Jun 2002 | B1 |
6552581 | Gabara | Apr 2003 | B1 |
6587915 | Kim | Jul 2003 | B1 |
6618249 | Fairchild | Sep 2003 | B2 |
6728913 | Parker | Apr 2004 | B1 |
6763424 | Conley | Jul 2004 | B2 |
6775792 | Ulrich et al. | Aug 2004 | B2 |
6778387 | Fairchild | Aug 2004 | B2 |
6850443 | Lofgren et al. | Feb 2005 | B2 |
6854070 | Johnson et al. | Feb 2005 | B2 |
6903972 | Lasser et al. | Jun 2005 | B2 |
6906961 | Eggleston et al. | Jun 2005 | B2 |
6975028 | Wayburn et al. | Dec 2005 | B1 |
7082495 | DeWhitt et al. | Jul 2006 | B2 |
7107389 | Inagaki et al. | Sep 2006 | B2 |
7139864 | Bennett et al. | Nov 2006 | B2 |
7233497 | Simon et al. | Jun 2007 | B2 |
7243186 | Liang et al. | Jul 2007 | B2 |
7330927 | Reeve et al. | Feb 2008 | B1 |
7333364 | Yu et al. | Feb 2008 | B2 |
7355896 | Li et al. | Apr 2008 | B2 |
7434122 | Jo | Oct 2008 | B2 |
7441067 | Gorobets et al. | Oct 2008 | B2 |
7516267 | Coulson et al. | Apr 2009 | B2 |
7613871 | Tanaka et al. | Nov 2009 | B2 |
7620769 | Lee et al. | Nov 2009 | B2 |
7639532 | Roohparvar et al. | Dec 2009 | B2 |
7661054 | Huffman et al. | Feb 2010 | B2 |
7679948 | Park et al. | Mar 2010 | B2 |
7738502 | Chang et al. | Jun 2010 | B2 |
7743216 | Lubbers et al. | Jun 2010 | B2 |
7818525 | Frost et al. | Oct 2010 | B1 |
7827348 | Lee et al. | Nov 2010 | B2 |
7830164 | Earle et al. | Nov 2010 | B2 |
7979614 | Yang | Jul 2011 | B1 |
8001135 | Perlmutter et al. | Aug 2011 | B2 |
8010738 | Chilton et al. | Aug 2011 | B1 |
8028123 | Kilzer et al. | Sep 2011 | B2 |
8046645 | Hsu et al. | Oct 2011 | B2 |
8051241 | Feldman et al. | Nov 2011 | B2 |
8072805 | Chou et al. | Dec 2011 | B2 |
8095724 | Ji et al. | Jan 2012 | B2 |
8095765 | Asnaashari et al. | Jan 2012 | B2 |
8117396 | Fair et al. | Feb 2012 | B1 |
8127202 | Cornwell et al. | Feb 2012 | B2 |
8145984 | Sommer et al. | Mar 2012 | B2 |
8154921 | Mokhlesi et al. | Apr 2012 | B2 |
8169825 | Shalvi et al. | May 2012 | B1 |
8219724 | Caruso et al. | Jul 2012 | B1 |
8219776 | Forhan et al. | Jul 2012 | B2 |
8228701 | Sokolov et al. | Jul 2012 | B2 |
8245101 | Olbrich et al. | Aug 2012 | B2 |
8254172 | Kan | Aug 2012 | B1 |
8259506 | Sommer et al. | Sep 2012 | B1 |
8289801 | Smith et al. | Oct 2012 | B2 |
8332578 | Frickey, III et al. | Dec 2012 | B2 |
8363413 | Paquette et al. | Jan 2013 | B2 |
8369141 | Sommer et al. | Feb 2013 | B2 |
8386700 | Olbrich et al. | Feb 2013 | B2 |
8407409 | Kawaguchi | Mar 2013 | B2 |
8464106 | Filor et al. | Jun 2013 | B2 |
8612804 | Kang et al. | Dec 2013 | B1 |
20020159285 | Morley et al. | Oct 2002 | A1 |
20030046603 | Harari et al. | Mar 2003 | A1 |
20030074592 | Hasegawa | Apr 2003 | A1 |
20030163633 | Aasheim et al. | Aug 2003 | A1 |
20040080985 | Chang et al. | Apr 2004 | A1 |
20040252670 | Rong et al. | Dec 2004 | A1 |
20050021904 | Iaculo et al. | Jan 2005 | A1 |
20050038792 | Johnson | Feb 2005 | A1 |
20050073884 | Gonzalez et al. | Apr 2005 | A1 |
20060020745 | Conley et al. | Jan 2006 | A1 |
20060136682 | Haridas et al. | Jun 2006 | A1 |
20060143365 | Kikuchi | Jun 2006 | A1 |
20060253641 | Gatzemeier et al. | Nov 2006 | A1 |
20060256624 | Eggleston et al. | Nov 2006 | A1 |
20060282644 | Wong | Dec 2006 | A1 |
20060294574 | Cha | Dec 2006 | A1 |
20070061511 | Faber | Mar 2007 | A1 |
20070083779 | Misaka et al. | Apr 2007 | A1 |
20070234004 | Oshima et al. | Oct 2007 | A1 |
20070260811 | Merry, Jr. et al. | Nov 2007 | A1 |
20070263444 | Gorobets et al. | Nov 2007 | A1 |
20070276973 | Tan et al. | Nov 2007 | A1 |
20080046630 | Lasser | Feb 2008 | A1 |
20080052446 | Lasser et al. | Feb 2008 | A1 |
20080082736 | Chow et al. | Apr 2008 | A1 |
20080183918 | Dhokia et al. | Jul 2008 | A1 |
20080313505 | Lee et al. | Dec 2008 | A1 |
20090019321 | Radke | Jan 2009 | A1 |
20090089485 | Yeh | Apr 2009 | A1 |
20090138654 | Sutardja | May 2009 | A1 |
20090157948 | Trichina et al. | Jun 2009 | A1 |
20090164702 | Kern | Jun 2009 | A1 |
20090172262 | Olbrich et al. | Jul 2009 | A1 |
20090228761 | Perlmutter et al. | Sep 2009 | A1 |
20090259819 | Chen et al. | Oct 2009 | A1 |
20090259896 | Hsu et al. | Oct 2009 | A1 |
20090323419 | Lee et al. | Dec 2009 | A1 |
20090327581 | Coulson | Dec 2009 | A1 |
20090327591 | Moshayedi | Dec 2009 | A1 |
20100017650 | Chin et al. | Jan 2010 | A1 |
20100023674 | Aviles | Jan 2010 | A1 |
20100050053 | Wilson et al. | Feb 2010 | A1 |
20100138592 | Cheon | Jun 2010 | A1 |
20100169541 | Freikorn | Jul 2010 | A1 |
20100174845 | Gorobets et al. | Jul 2010 | A1 |
20100217915 | O'Connor et al. | Aug 2010 | A1 |
20100262792 | Hetzler et al. | Oct 2010 | A1 |
20100262795 | Hetzler et al. | Oct 2010 | A1 |
20100262875 | Hetzler et al. | Oct 2010 | A1 |
20100287328 | Feldman et al. | Nov 2010 | A1 |
20100293367 | Berke et al. | Nov 2010 | A1 |
20100312954 | Jeon et al. | Dec 2010 | A1 |
20100318719 | Keays et al. | Dec 2010 | A1 |
20100332726 | Wang | Dec 2010 | A1 |
20110055468 | Gonzalez et al. | Mar 2011 | A1 |
20110066788 | Eleftheriou et al. | Mar 2011 | A1 |
20110131365 | Zhang et al. | Jun 2011 | A1 |
20110131447 | Prakash et al. | Jun 2011 | A1 |
20110145473 | Maheshwari | Jun 2011 | A1 |
20110190963 | Glassl et al. | Aug 2011 | A1 |
20110191522 | Condict et al. | Aug 2011 | A1 |
20110191649 | Lim et al. | Aug 2011 | A1 |
20110238892 | Tsai et al. | Sep 2011 | A1 |
20110320687 | Belluomini et al. | Dec 2011 | A1 |
20120047320 | Yoo et al. | Feb 2012 | A1 |
20120047409 | Post et al. | Feb 2012 | A1 |
20120124046 | Provenzano | May 2012 | A1 |
20120124273 | Goss et al. | May 2012 | A1 |
20120151260 | Zimmermann et al. | Jun 2012 | A1 |
20120216085 | Weingarten et al. | Aug 2012 | A1 |
20120239858 | Melik-Martirosian | Sep 2012 | A1 |
20120266048 | Chung et al. | Oct 2012 | A1 |
20120331207 | Lassa et al. | Dec 2012 | A1 |
20130007380 | Seekins et al. | Jan 2013 | A1 |
20130073788 | Post et al. | Mar 2013 | A1 |
20130080691 | Weingarten et al. | Mar 2013 | A1 |
20130100600 | Yin et al. | Apr 2013 | A1 |
20130124792 | Melik-Martirosian et al. | May 2013 | A1 |
20140108891 | Strasser et al. | Apr 2014 | A1 |
Number | Date | Country |
---|---|---|
1 990 921 | Nov 2008 | EP |
Entry |
---|
IBM Corporation, Systems Management Contrilling System Shutdown Using a Power-Handling Program, Version 5, Release 4, 9th Edition, pp. 1-21, Feb. 2006. |
Texas Instruments, Power Management IC for Digital Set Top Boxes, SLVSA 10A, pp. 1-22, Sep. 2009. |
International Search Report and Written Opinion dated Dec. 20, 2013, received in PCT/US2013/045282, which corresponds to U.S. Appl. No. 13/493,949, 7 pages (Ellis). |
Cooke, “Introduction to Flash Memory (T1A),” Flash Memory Summit, Aug. 22, 2008, Micron Technology, Inc., 102 pages. |
Gal et al., “Algorithms and Data Structures for Flash Memories.” ACM Computing Surveys, Jun. 2005, Vol, 37, No. 2, 30 pages. |
O'Brien, “SMART Storage Systems Optimus SAS Enterprise SSD Review,” SMART Storage Systems. Oct. 9, 2012, 44 pages. |
Spanjer, “Flash Management—Why and How?” Smart Modular Technologies. Nov. 2009, http://www.scantec.de/fileadmin/pcf/Smart—Modular/Flash-Management.pdf, 14 pages. |
International Search Report and Written Opinion dated Jun. 12, 2014, received in PCT/US2014/018972, which corresponds to U.S. Appl. No. 13/779,352, 12 pages (Schmier). |
International Search Report and Written Opinion dated May 14, 2014, received in International Patent Application no. PCT/US2014/017168, which corresponds to U.S. Appl. No. 14/076,115, 6 pages (Fitzpatrick). |
International Search Report and Written Opinion dated May 14, 2014, received in International Patent Application No. PCT/US2014/017169, which corresponds to U.S. Appl. No. 14/076,148, 6 pages (Fitzpatrick). |
Number | Date | Country | |
---|---|---|---|
20120203958 A1 | Aug 2012 | US |
Number | Date | Country | |
---|---|---|---|
61440400 | Feb 2011 | US |