System and method for improving a data redundancy scheme in a solid state subsystem with additional metadata

Information

  • Patent Grant
  • 9110835
  • Patent Number
    9,110,835
  • Date Filed
    Monday, March 17, 2014
    10 years ago
  • Date Issued
    Tuesday, August 18, 2015
    9 years ago
Abstract
In one embodiment of the invention, a flash-based/solid-state storage system with an implemented data redundancy scheme such as RAID is configured to hold parity data in a volatile memory such as RAM and write such parity data to the non-volatile flash media when a full stripe of data has been written to the media. Other embodiments in certain situations force an early write of the parity for a partial stripe that has not been fully written to the non-volatile media. Those situations may include a data access error on data in a partial stripe and a detected power loss event with a partial stripe present. Embodiments are directed to writing additional data with the parity data for the partial stripe and then later using the additional data in data recovery. This approach allows the controller to easily detect the presence of a partial stripe and handle such a stripe accordingly.
Description
BACKGROUND

Many data storage components such as hard disks and solid state drives have certain advertised reliability guarantees that the manufacturers provide to customers. For example, certain solid state drive manufacturers guarantee a drive failure rate of 10−16 or 10−17. To increase data reliability, a data redundancy scheme such as RAID (Redundant Arrays of Independent Disks) is used to increase storage reliability. The redundancy may be provided by combining multiple storage elements within the storage device into groups providing mirroring and/or error checking mechanisms. For example, various memory pages/blocks of a solid state storage device may be combined into data stripes in which user data is stored.





BRIEF DESCRIPTION OF THE DRAWINGS

Systems and methods which embody the various features of the invention will now be described with reference to the following drawings, in which:



FIG. 1 is a block diagram illustrating a storage subsystem implementing a data redundancy scheme with a mechanism for storing metadata with parity data to enhance data recovery according to one embodiment.



FIG. 2 is a flow diagram showing a process of handling parity data according to one embodiment.



FIGS. 3A-3B illustrate an example of a partial data stripe according to one embodiment.



FIG. 4A is a flow diagram showing a process of parity marking according to one embodiment



FIG. 4B is a flow diagram showing a process of using marked parity according to one embodiment.



FIG. 5 is block diagram showing an example of writing additional metadata with parity data according to one embodiment.



FIG. 6 is block diagram showing an example of using metadata for data recovery according to one embodiment.





DETAILED DESCRIPTION OF SPECIFIC EMBODIMENTS

Overview


A common approach to overcome storage element failure is to use redundant RAID (mirroring, parity, etc.) to allow data recovery should one or more failures (e.g., a read failure) occur. Typically, a target number of storage elements (e.g., pages, blocks, etc.) per stripe is chosen to achieve a desired reliability at a given cost in storage overhead. In one embodiment of the present invention, a flash-based/solid-state storage system is configured to hold parity data in a temporary volatile memory such as RAM (Random Access Memory) and write such parity data to the non-volatile flash media when the full stripe's worth of data has been written to the non-volatile flash media.


However, there are situations when the parity for an open stripe is written to the flash media before the stripe is fully written. (Since the writing of the parity data “closes” a stripe, a stripe prior to the writing of its parity data is said to be an “open” stripe.) For example, in various embodiments of the invention, it may be appropriate to force a write of the parity data associated with an open stripe upon the detection of an uncorrectable data access error (e.g., a read or program error) on data in the stripe. In another example, the parity data may be written to the flash media when there is a detected power loss event and an open stripe present. Whatever the original cause that may have triggered the early write of parity data, embodiments of the invention are directed to writing additional data with the parity data to the flash media and then later using the additional data in data recovery operations. This approach allows the storage subsystem to easily detect the presence of a partial stripe and handle such a stripe accordingly. In one embodiment, the additional data is written in a spare area that is typically reserved for various system metadata and is used to indicate validity of pages in the partial stripe.


System Overview



FIG. 1 is a block diagram illustrating a storage subsystem with a mechanism for storing metadata with parity data to enhance data recovery. As shown, a storage subsystem 140 includes a controller 150 and one or more non-volatile solid-state memory arrays 160. The arrays 160 may comprise a plurality of solid-state storage devices coupled to the controller 150. The arrays may comprise, for example, flash integrated circuits, Chalcogenide RAM (C-RAM), Phase Change Memory (PC-RAM or PRAM), Programmable Metallization Cell RAM (PMC-RAM or PMCm), Ovonic Unified Memory (OUM), Resistance RAM (RRAM), NAND memory, NOR memory, EEPROM, Ferroelectric Memory (FeRAM), or other discrete NVM (non-volatile memory) chips. The solid-state storage devices may be physically divided into planes, blocks, pages, and sectors, as is known in the art. Other forms of storage (e.g., battery backed-up volatile DRAM or SRAM devices, magnetic disk drives, etc.) may additionally or alternatively be used.


The controller 150 in one embodiment in turn includes a RAID module 158 and a volatile memory 164, which may be implemented in, for example, RAM such as a DRAM or SRAM. The controller may alternatively be implemented in-whole or in-part as an ASIC, FPGA, or other device, which may but need not execute firmware. In another embodiment, the volatile memory 164 is outside of the controller 150 in the storage subsystem 140. In one embodiment, the RAID module 158 is configured to execute data access commands to maintain a data redundancy scheme in the storage subsystem. For example, the RAID module 158 may maintain data on which storage elements are assigned to which RAID stripes and determine how data are arranged in the data redundancy scheme (e.g., grouped into stripes with parity). In another embodiment, the various processes described herein may be executed by the RAID module 158 within the controller 150, by one or more components within the controller 150, or by a combination of both.


In one embodiment, the controller 150 of the storage subsystem 140 is configured to receive and execute commands from a storage interface 132 in a host system 130. The memory commands from the storage interface 132 may include write and read commands issued by the host system 130. As further shown in FIG. 1, in one embodiment, the controller 150 executes the received commands in one or more non-volatile solid-state memory arrays 160. As will be explained below, in one embodiment, the volatile memory 164 is configured to temporarily hold parity data for data stripes (e.g., RAID stripes) that are yet to be completed/closed. In one embodiment, when a stripe is complete (e.g., all pages of a stripe are programmed in the non-volatile memory arrays 160), the parity data is then transferred to the non-volatile memory arrays 160. The volatile memory 164 may also be configured to temporarily hold “user” data to be written to the memory arrays 160, and such user data may include data received from the host system 130 and/or other data related to internal storage subsystem operations such as garbage collection and wear leveling. In one embodiment, the user data may be held in a separate volatile memory than that which is used to hold the parity data, or in a different portion of the volatile memory 164 than that which is used for the parity data. The controller 150 may also execute internal memory commands such as those needed for garbage collection and wear leveling, etc. and write data related to those internal memory commands to the non-volatile solid-state memory arrays 160 (via the volatile memory or directly to the memory arrays).


Although this disclosure uses RAID as an example, the systems and methods described herein are not limited to the RAID redundancy schemes and can be used in any data redundancy configuration that utilizes striping and/or grouping of storage elements for mirroring or error checking purposes. In addition, although RAID is an acronym for Redundant Array of Independent Disks, those skilled in art will appreciate that RAID is not limited to storage devices with physical disks and is applicable to a wide variety of storage devices including the non-volatile solid state devices described herein.


Parity Handling Process



FIG. 2 is a flow diagram showing a process of handling parity data according to one embodiment. In block 202, the controller 150 handles the creation of data stripes as data is received from the host (or via internal commands such as garbage collection). At the same time, the controller may be processing read commands on data stripes that have already been written to the non-volatile memory array. Blocks 204 and 206 illustrate two example conditions under which parity for a partial stripe may be written to the non-volatile memory. In block 204, it is determined whether an uncorrectable error (e.g., uncorrectable by ECC) has occurred as a result of a read operation on an open stripe (i.e., a data stripe that has not been completely written to the non-volatile memory array and for which parity data has not been written to the non-volatile memory array). In one embodiment, the controller employs an ECC (error-correcting code) mechanism to attempt to correct any reading error, but the error is deemed uncorrectable if the ECC mechanism cannot correct it. Note that if the error is correctable, the condition is not met in block 204 and there is no need to force an early write of the parity data. Although an ECC read error example is described here, the same process is applicable to a program error as well. If an uncorrectable data access error has occurred, the controller would force a write of parity data for that open stripe (block 208). In one embodiment, this means that the parity data (for the open stripe) that is currently held in the volatile memory is written to the non-volatile memory array.


In block 206, the controller may also determine whether there is a power failure event occurring with an open stripe present. If so, the parity data for the open stripe is written to the non-volatile memory as part of the power failure handling process (block 208). In either case, the controller in block 210 may execute a recovery process using the newly written parity data. In one embodiment where RAID is implemented, the process is a standard RAID data recovery procedure. Although not shown, there are other situations besides those illustrated in blocks 204 and 206 where the controller may write the parity for an open stripe. For example, certain data transfer interface standards support a Force Unit Access (F.U.A.) command, which may require an immediate write to the non-volatile memory and trigger a corresponding writing of parity data.


Partial Stripe with Forced Parity Write Example



FIGS. 3A-3B illustrate an example of a partial data stripe according to one embodiment. FIG. 3A shows four example stripes (Stripes 1-4). Stripes 1-3 are all full stripes in the sense that each has 15 data pages covered by one parity page. Stripes 1-3 have all been written to the non-volatile memory array. These three stripes represent stripes created under normal operation in which no uncorrectable (e.g., uncorrectable by ECC) data errors or a power failure were detected in the process of stripe creation. In one embodiment, for these completed stripes with parity already written, any error not corrected by ECC can be corrected by using the normal RAID recovery procedure using the parity data. The fourth stripe, Stripe 4, is an open stripe, since it has only 6 pages written to the non-volatile memory array, and the parity data held in the volatile memory is not yet written to the non-volatile memory since parity in this example is typically written after the 15th page is written to the non-volatile memory.



FIG. 3B shows the state of the non-volatile memory after the forced write of the parity data. The process shown in FIG. 3B may be performed in whole or in part by the controller shown in FIG. 1, or by a component of the controller such as the RAID module. As previously shown in FIG. 2, the forced write of parity data occurs: (1) when an uncorrectable data access error is encountered in one of the pages in an open stripe for which parity data has not been written, (2) when a power failure event is detected while there is an open stripe, or (3) when some other condition occurs that necessitates an early write of the parity data.


In this example, an uncorrectable ECC error occurred while a read was being performed on one of the six pages in Stripe 4, prompting the writing of the parity data to the non-volatile memory ahead of its scheduled write time. The result, as shown in FIG. 3B, is that Stripe 4 is now a partial stripe within the non-volatile memory, and recovery from that uncorrectable data error proceeds with the partial parity data that was written. In one embodiment, after the storage subsystem recovers from that data access error through the use of parity data, the subsystem may move on to create other data stripes, leaving Stripe 4 as a partial stripe. In one embodiment, the data recovery may additional use the meta data written with the parity data, as described below.


Parity Marking


While FIG. 3B shows an example of writing parity data for a partial stripe, in various embodiments additional data is written along with the parity data for a partial stripe to “mark” the parity data. FIG. 4A is a flow diagram showing a process of parity marking according to one embodiment. In block 402, it is determined whether the current write operation is writing a parity page of a truncated stripe (partial stripe). If so, in block 404, the spare/metadata area of the parity page is marked with information concerning which of the pages in the associated stripe are valid. As shown in the figure, a parity page for a regular (non-truncated) stripe would not have metadata on valid pages information (since presumably all pages are valid).



FIG. 4B is a flow diagram showing a process of using marked parity according to one embodiment. In block 412, the controller determines whether it is reading a page with metadata mismatch. If not, this indicates that the controller is reading a user data page, and the controller would continue with regular operation (block 416). If it is determined that the there is a metadata mismatch, the expected result is that the controller is reading a parity page for a partial stripe. However, if it is determined in block 414 that it is not reading a parity page, the controller would initiate special error handling in block 418. Otherwise, in block 420, the controller reads the metadata to find out which pages in the associated stripe are valid. In a RAID based embodiment, for example, the valid pages and parity are XOR'ed to generate data in a recovery process.



FIG. 5 is block diagram showing an example of writing additional metadata with parity data according to one embodiment. In one embodiment, each page, whether it is used for storing user data or parity data, includes a spare area for metadata. The spare area may include metadata such as that used for ECC purposes or another system data. In addition, in one embodiment, each page includes within its spare area an indication that it is either a user data page or a parity page (indicators 508 and 518). The indicator is used, for example, by the process depicted in FIG. 4B to indicate to a controller reading the page whether the page contains user data or parity data. As shown, a user data page 502 includes both a user data area 504 and a spare area 506. Similarly, a regular parity data page 512 includes both a regular parity data area 514 and a spare area 516.



FIG. 5 also shows a truncated parity page 522 including truncated parity data 524 and a spare area for metadata 526. The metadata includes an indicator 528 indicating that this is a parity page. In addition, the example truncated parity page 522 includes validity bits 530 within the spare area 526. In this example, the validity bits show that six of the 15 pages of the associated stripe contain valid data. Thus, if a recovery process needs to be executed to recover the data on that partial stripe, the validity bits can indicate which of pages in the stripe are valid, and the controller can thus handle this partial stripe accordingly.


Use of Metadata


In one embodiment, with the metadata in place, once a partial stripe is written with parity it does not need to be moved to a new location until the controller needs to execute a data relocation operation such as garbage collection. This is because the metadata allows the controller to detect the presence of a partial stripe and provides information to the controller to help it decipher which of the pages within the partial stripe are valid. The elimination of the need to further process the partial stripe once written (e.g., by re-writing the partial strip into a full stripe) saves time and thus does not introduce additional write amplification (additional write operations to accommodate write commands issued by a host system). By storing the metadata in the spare data area of a page, most of the time the controller can execute read/write operations on regular stripes normally without having to incur the additional penalty of looking up and check to see if the stripe is truncated. In one embodiment, as shown in FIG. 4B, the additional work is incurred only when a metadata mismatch is encountered in the course of normal reading operations.


In one embodiment, with the afore-mentioned truncated stripe marking, the hardware can detect the marking and report back to firmware for special handling during a garbage collection read operation. Therefore, the normal read operation performance is not affected. The marking will indicate to the garbage collection process which pages are valid in a truncated stripe should an error arises. In one embodiment, the parity page is the last page in a stripe by default, which simplifies both normal read and write operations as well as defect management. The truncated page will be detected and handled accordingly.



FIG. 6 illustrates an example data recovery process that uses metadata. The figure illustrates a series of pairings of metadata (610, 612, and 614) with the corresponding data stripe in the non-volatile memory (650, 652, and 654). Stripe 650 shows a normal stripe with seven user data pages (D1-D7) associated with one parity page (P). Metadata 610 for stripe 650 shows a “1” bit for each of the valid pages D1-D7 and an indicator “P” for the parity page. Stripe 652, on the other hand, depicts a partial stripe created by an early write of parity data as described above. Stripe 652 includes three valid data pages (D1-D3) followed by a parity page (P). The rest of the pages in stripe 652 are not valid as indicated by the markings “X.” As such, metadata 612 for stripe 652 indicates the validity of the pages D1-D3 (“1”), the parity page location (“P”), and those pages that do not have valid data (“0”).


Stripe 654 illustrates the result of a data recovery caused by a data access error on page D2 of stripe 652. The controller, upon encountering such an error, used the parity and data from pages D1 and D3 to recover D2. As shown, the recovered D2 is written after the parity page, as indicated by the label “D2 REC.” Accordingly, metadata 612 is updated (shown by metadata 614) to reflect the fact that the old D2 page is no longer valid (shown by a “0” bit in the second position) and that the valid D2 page is now after the old parity page (shown by a “1” bit in the fifth position). Note that metadata 614 also reflects the location of the new parity page for the stripe.


The updated metadata enables the controller to track the locations of the valid pages. Thus, for example, if a data access error occurs on D3, the controller will need to read the parity page and pages D1 and D2 to recover D3. The updated metadata 614 may serve as a bit mask to point the controller to the right location of the valid D2 page (which is in the fifth position rather than in the second position as originally written) as well as the new parity page. This enables the controller to quickly locate the needed data to recover D3.


CONCLUSION

While certain embodiments of the inventions have been described, these embodiments have been presented by way of example only, and are not intended to limit the scope of the inventions. Indeed, the novel methods and systems described herein may be embodied in a variety of other forms. Furthermore, various omissions, substitutions and changes in the form of the methods and systems described herein may be made without departing from the spirit of the inventions. The accompanying claims and their equivalents are intended to cover such forms or modifications as would fall within the scope and spirit of the inventions. For example, those skilled in the art will appreciate that in various embodiments, the actual steps taken in the processes shown in FIGS. 2, 4A, and 4B may differ from those shown in the figures. Depending on the embodiment, certain of the steps described in the example above may be removed, others may be added, and the sequence of steps may be altered and/or performed in parallel. Although this disclosure describes examples in which data are stored in units of pages, those skilled in the art will recognize that the various embodiments disclosed are applicable to systems with other memory measurement units such as blocks, sectors, etc. Also, the features and attributes of the specific embodiments disclosed above may be combined in different ways to form additional embodiments, all of which fall within the scope of the present disclosure. Although the present disclosure provides certain preferred embodiments and applications, other embodiments that are apparent to those of ordinary skill in the art, including embodiments which do not provide all of the features and advantages set forth herein, are also within the scope of this disclosure. Accordingly, the scope of the present disclosure is intended to be defined only by reference to the appended claims.

Claims
  • 1. A data storage system, comprising: a non-volatile memory array configured to store a plurality of data stripes; anda controller configured to: in response to detecting a failure event, create a partial data stripe that comprises (1) metadata that specifies which of data elements of the partial data stripe are valid, (2) parity data, and (3) both valid and invalid data elements; anduse the metadata to determine which data elements of the partial data stripe are valid.
  • 2. The data storage system of claim 1, wherein the controller is further configured to write the partial stripe to the non-volatile memory array.
  • 3. The data storage system of claim 1, wherein the controller is further configured to: recover at least one invalid data element using a valid data element and parity data; andupdate the metadata to indicate that the at least one invalid data element has been recovered.
  • 4. The data storage system of claim 3, wherein the controller is further configured to recover the at least one invalid data element using a plurality of valid data elements and parity data.
  • 5. The data storage system of claim 3, wherein the controller is further configured to: generate updated parity data based at least in part on the recovered data element; andupdate the metadata to indicate that updated parity data has been generated.
  • 6. The data storage system of claim 5, wherein the metadata comprises location information of data in the non-volatile memory array and the controller is configured to update the metadata to indicate that the parity data has been updated by indicating in the metadata location information of where the updated parity data is stored.
  • 7. The data storage system of claim 1, wherein the metadata comprises a bit mask.
  • 8. The data storage system of claim 1, wherein the failure event comprises at least one of a power failure event and a data access error from a data operation on the non-volatile memory array.
  • 9. The data storage system of claim 8, wherein the controller is further configured to use the parity data and metadata to recover from the data access error.
  • 10. The data storage system of claim 1, wherein the controller is further configured to use the metadata in a data relocation operation to determine which of the data elements in the partial data stripe are valid.
  • 11. A method of recovering data stored in a data storage system, the method comprising: in response to detecting a failure event, creating a partial data stripe that comprises (1) metadata that specifies which of data elements of the partial data stripe are valid, (2) parity data, and (3) both valid and invalid data elements;writing the partial data stripe to a non-volatile memory array; andusing the metadata to determine which data elements of the partial data stripe are valid,wherein the method is performed by a controller of the data storage system.
  • 12. The method of claim 11, further comprising: recovering at least one invalid data element using a valid data element and parity data; andupdating the metadata to indicate that the at least one invalid data element has been recovered.
  • 13. The method of claim 12, wherein the metadata comprises location information of data in the non-volatile memory array and the method further comprises updating the metadata to indicate that the at least one invalid data element has been recovered by indicating in the metadata location information of where the recovered data element is stored.
  • 14. The method of claim 12, wherein recovering the at least one invalid data element comprises using a plurality of valid data elements and parity data.
  • 15. The method of claim 12, further comprising: generating updated parity data based at least in part on the recovered data element; andupdating the metadata to indicate that updated parity data has been generated.
  • 16. The method of claim 15, wherein the metadata comprises location information of data in the non-volatile memory array and the method further comprises updating the metadata to indicate that the parity data has been updated by indicating in the metadata location information of where the updated parity data is stored.
  • 17. The method of claim 11, wherein the metadata comprises a bit mask.
  • 18. The method of claim 11, wherein the failure event comprises at least one of a power failure event and a data access error from a data operation on the non-volatile memory array.
  • 19. The method of claim 18, further comprising using the parity data and metadata to recover from the data access error.
  • 20. The method of claim 11, further comprising using the metadata in a data relocation operation to determine which of the data elements in the partial data stripe are valid.
CROSS REFERENCE TO RELATED APPLICATIONS

This application is a continuation of U.S. patent application Ser. No. 13/044,400, filed on Mar. 9, 2011, which is hereby incorporated by reference in its entirety.

US Referenced Citations (157)
Number Name Date Kind
5408644 Schneider et al. Apr 1995 A
5499337 Gordon Mar 1996 A
6467022 Buckland et al. Oct 2002 B1
6523087 Busser Feb 2003 B2
6856556 Hajeck Feb 2005 B1
6959413 Humlicek et al. Oct 2005 B2
7069382 Horn et al. Jun 2006 B2
7126857 Hajeck Oct 2006 B2
7200715 Kleiman et al. Apr 2007 B2
7409492 Tanaka et al. Aug 2008 B2
7430136 Merry, Jr. et al. Sep 2008 B2
7447807 Merry et al. Nov 2008 B1
7502256 Merry, Jr. et al. Mar 2009 B2
7509441 Merry et al. Mar 2009 B1
7596643 Merry, Jr. et al. Sep 2009 B2
7640390 Iwamura et al. Dec 2009 B2
7653778 Merry, Jr. et al. Jan 2010 B2
7685337 Merry, Jr. et al. Mar 2010 B2
7685338 Merry, Jr. et al. Mar 2010 B2
7685374 Diggs et al. Mar 2010 B2
7733712 Walston et al. Jun 2010 B1
7765373 Merry et al. Jul 2010 B1
7779294 Corrado et al. Aug 2010 B2
7856528 Frost et al. Dec 2010 B1
7898855 Merry, Jr. et al. Mar 2011 B2
7912991 Merry et al. Mar 2011 B1
7936603 Merry, Jr. et al. May 2011 B2
7962792 Diggs et al. Jun 2011 B2
8078918 Diggs et al. Dec 2011 B2
8090899 Syu Jan 2012 B1
8095851 Diggs et al. Jan 2012 B2
8108692 Merry et al. Jan 2012 B1
8122185 Merry, Jr. et al. Feb 2012 B2
8127048 Merry et al. Feb 2012 B1
8135903 Kan Mar 2012 B1
8151020 Merry, Jr. et al. Apr 2012 B2
8161227 Diggs et al. Apr 2012 B1
8166245 Diggs et al. Apr 2012 B2
8243525 Kan Aug 2012 B1
8254172 Kan Aug 2012 B1
8261012 Kan Sep 2012 B2
8296625 Diggs et al. Oct 2012 B2
8312207 Merry, Jr. et al. Nov 2012 B2
8316176 Phan et al. Nov 2012 B1
8321597 Yu et al. Nov 2012 B2
8341339 Boyle et al. Dec 2012 B1
8375151 Kan Feb 2013 B1
8392635 Booth et al. Mar 2013 B2
8397107 Syu et al. Mar 2013 B1
8402217 Burd Mar 2013 B2
8407449 Colon et al. Mar 2013 B1
8423722 Deforest et al. Apr 2013 B1
8433858 Diggs et al. Apr 2013 B1
8443167 Fallone et al. May 2013 B1
8447920 Syu May 2013 B1
8458435 Rainey, III et al. Jun 2013 B1
8478930 Syu Jul 2013 B1
8489854 Colon et al. Jul 2013 B1
8503237 Horn Aug 2013 B1
8521972 Boyle et al. Aug 2013 B1
8549236 Diggs et al. Oct 2013 B2
8583835 Kan Nov 2013 B1
8601311 Horn Dec 2013 B2
8601313 Horn Dec 2013 B1
8612669 Syu et al. Dec 2013 B1
8612804 Kang et al. Dec 2013 B1
8615681 Horn Dec 2013 B2
8638602 Horn Jan 2014 B1
8639872 Boyle et al. Jan 2014 B1
8683113 Abasto et al. Mar 2014 B2
8700834 Horn et al. Apr 2014 B2
8700950 Syu Apr 2014 B1
8700951 Call et al. Apr 2014 B1
8706985 Boyle et al. Apr 2014 B1
8707104 Jean Apr 2014 B1
8713066 Lo et al. Apr 2014 B1
8713357 Jean et al. Apr 2014 B1
8719531 Strange et al. May 2014 B2
8724422 Agness et al. May 2014 B1
8725931 Kang May 2014 B1
8745277 Kan Jun 2014 B2
8751728 Syu et al. Jun 2014 B1
8769190 Syu et al. Jul 2014 B1
8769232 Suryabudi et al. Jul 2014 B2
8775720 Meyer et al. Jul 2014 B1
8782327 Kang et al. Jul 2014 B1
8788778 Boyle Jul 2014 B1
8788779 Horn Jul 2014 B1
8788880 Gosla et al. Jul 2014 B1
8793429 Call et al. Jul 2014 B1
20040015657 Humlicek et al. Jan 2004 A1
20040123032 Talagala et al. Jun 2004 A1
20050066124 Horn et al. Mar 2005 A1
20050086429 Chatterjee et al. Apr 2005 A1
20050177672 Rao Aug 2005 A1
20060004957 Hand et al. Jan 2006 A1
20060236029 Corrado et al. Oct 2006 A1
20070268905 Baker et al. Nov 2007 A1
20070283079 Iwamura et al. Dec 2007 A1
20070294565 Johnston et al. Dec 2007 A1
20070297265 Kim et al. Dec 2007 A1
20080133969 Manoj Jun 2008 A1
20080141054 Danilak Jun 2008 A1
20080155160 McDaniel Jun 2008 A1
20080229148 Forhan et al. Sep 2008 A1
20080276124 Hetzler et al. Nov 2008 A1
20090073762 Lee et al. Mar 2009 A1
20090083504 Belluomini et al. Mar 2009 A1
20090150599 Bennett Jun 2009 A1
20090172335 Kulkarni et al. Jul 2009 A1
20090204852 Diggs et al. Aug 2009 A1
20090210744 Kamalavannan Aug 2009 A1
20090248998 Sato et al. Oct 2009 A1
20090327604 Sato et al. Dec 2009 A1
20090327803 Fukutomi et al. Dec 2009 A1
20100049914 Goodwin Feb 2010 A1
20100064111 Kunimatsu et al. Mar 2010 A1
20100088557 Weingarten et al. Apr 2010 A1
20100088579 Hafner et al. Apr 2010 A1
20100115175 Zhuang et al. May 2010 A9
20100122115 Olster May 2010 A1
20100169543 Edgington et al. Jul 2010 A1
20100174849 Walston et al. Jul 2010 A1
20100250793 Syu Sep 2010 A1
20100262773 Borchers et al. Oct 2010 A1
20100281202 Abali et al. Nov 2010 A1
20110035548 Kimmel et al. Feb 2011 A1
20110099323 Syu Apr 2011 A1
20110126045 Bennett May 2011 A1
20110173484 Schuette et al. Jul 2011 A1
20110191649 Lim et al. Aug 2011 A1
20110219259 Frost et al. Sep 2011 A1
20110264843 Haines et al. Oct 2011 A1
20110283049 Kang et al. Nov 2011 A1
20110314218 Bert Dec 2011 A1
20120079318 Colgrove et al. Mar 2012 A1
20120110376 Dreifus et al. May 2012 A1
20120151253 Horn Jun 2012 A1
20120173790 Hetzler et al. Jul 2012 A1
20120233406 Igashira et al. Sep 2012 A1
20120246403 McHale et al. Sep 2012 A1
20120260020 Suryabudi et al. Oct 2012 A1
20120278531 Horn Nov 2012 A1
20120284460 Guda Nov 2012 A1
20120324191 Strange et al. Dec 2012 A1
20130132638 Horn et al. May 2013 A1
20130145106 Kan Jun 2013 A1
20130290793 Booth et al. Oct 2013 A1
20140059405 Syu et al. Feb 2014 A1
20140101369 Tomlin et al. Apr 2014 A1
20140115427 Lu Apr 2014 A1
20140133220 Danilak et al. May 2014 A1
20140136753 Tomlin et al. May 2014 A1
20140149826 Lu et al. May 2014 A1
20140157078 Danilak et al. Jun 2014 A1
20140181432 Horn Jun 2014 A1
20140223255 Lu et al. Aug 2014 A1
Foreign Referenced Citations (1)
Number Date Country
2010049928 May 2010 WO
Non-Patent Literature Citations (3)
Entry
Notice of Allowance dated Dec. 6, 2013 from U.S. Appl. No. 13/044,400, 10 pages.
Office Action dated Jul. 2, 2013 from U.S. Appl. No. 13/044,400, 19 pages.
Interview Summary dated Oct. 29, 2013 from U.S. Appl. No. 13/044,400, 3 pages.
Continuations (1)
Number Date Country
Parent 13044400 Mar 2011 US
Child 14216919 US