The present invention relates to the use of a flash memory-based cache for data stored in a hard disk-based storage device and, in particular, to techniques for integrating cache management operations with write coalescing and providing other efficiencies in such a caching device.
As most computer users know, data storage is of paramount importance. Different forms of data storage devices have been developed to address different needs. For example, some data storage devices are optimized to allow very rapid read and write accesses, so as not to present a bottleneck to other processing operations involving the data being read from or written to the storage device. Usually, these high speed read/write storage devices can only accommodate a limited amount of data and/or are expensive. Other storage devices are designed to accommodate large volumes of data (e.g., terabytes of data), but operate at much slower speeds. Such devices are usually intended for applications where the cost of high speed storage devices is not justified.
A popular form of storage system is one that uses rapidly rotating disks coated with a magnetic material to store data in the form of magnetically encoded information elements. These so-called hard disk drives (HDD), or simply hard disks, are found in many personal computers and dedicated storage appliances. Hard disks can offer significant available storage space (e.g., on the order of terabytes), but the speed at which data can be read from such devices is limited by physical properties such as the size of the disk(s) on which the data is stored, the speed at which the disk(s) rotate, and the time required for the read head to be maneuvered into the correct position to read the requested data information elements (the so-called seek time).
So-called solid state storage devices, typically those that employ flash memory arrays as the storage medium, offer improved read times, in part because there are no moving parts associated with such a device. Write times, however, are often worse than those associated with hard disks because flash arrays can only be written in relatively large “erase block” sizes (e.g., typically 128 KB-512 KB), which must be erased and rewritten in their entirety even if only a small amount of data within the block needs to be updated.
To address the inefficiencies inherent with writes to a flash array, flash memory controllers typically employ a process known as write coalescing. This allows the flash controllers to deliver acceptable performance for random writes (i.e., writes to random, non-sequential addresses within the flash array). Write coalescing uses principles that were first developed for log structured file systems. In essence, this technique bundles together or coalesces a group of random writes so that the data associated with those applications is written to a physically contiguous region of flash memory, called a “segment” (in flash, a segment should be an integral multiple of the erase block size).
An associated process performed by the flash controller, known as “garbage collection”, ensures that large segments of the flash array are kept available for the contiguous writes required for proper write coalescing. As an application updates data at arbitrary logical addresses, and those data blocks are written to new physical locations in the flash array, any preexisting versions of the data in previously written portions of the array are marked as “obsolete”, meaning that these versions are no longer needed. Note, the data blocks referred to immediately above are best understood as units for writing to the flash and are different than the erase blocks referred to previously. These data blocks are typically much smaller than the erase blocks, e.g., on the order of 4 KB-8 KB, depending on the flash controller. Herein, the term block, when used by itself, should be understood as referring to these data blocks. The term erase block will be used when referring specifically to erase blocks.
The obsolete blocks tend to be scattered about the flash array, due to the nature of the random updates performed by the application making use of the data, but a garbage collection routine running on the flash controller periodically regenerates entire segments by copying non-obsolete blocks of data in previously written segments of the array into a smaller number of new segments and then erasing the old segments.
Today, new forms of storage devices that employ both flash memory and hard disks are being marketed. In some instances, the flash memory portion of these devices is being used as a cache for data stored on the hard disk. A cache is generally regarded to be a storage area that holds a subset of the data stored on a larger, generally slower, device. Here, the flash memory cache provides faster read access than the hard disk and so data stored in the cache portion of the device can be delivered more rapidly than if the data had to be accessed from the hard disk. Of course, while a flash memory-based cache offers advantages for reads, the problems inherent with random writes must still be addressed.
Embodiments of the present invention provide systems and methods in which a flash memory controller of a flash array configured as a cache for a disk-based storage device, consolidates information concerning blocks of obsolete data stored in the flash array with information concerning blocks of data stored in the flash array that a cache management routine running on the controller has determined may be removed from the cache; and erases (or instructs erasure of) segments containing one or more of the blocks of obsolete data and the blocks of data that the cache management routine has determined may be removed from the cache to produce reclaimed segments of the flash array. Each segment may include multiple blocks of the flash array. In various embodiments, the blocks of obsolete data may include data that has been made obsolete as a result of more recent versions of said data being written to the flash array. Further, the blocks of data that the cache management routine has determined may be removed from the cache may include data that is less frequently and/or recently used by an application accessing the cache than other data stored in the cache.
Prior to erasing segments containing one or more of the blocks of obsolete data and the blocks of data that the cache management routine has determined may be removed from the cache, the controller may write (or instruct writing) valid data from the segments to other segments in the flash array. In some cases, fewer than all of the segments containing one or more of the blocks of obsolete data and the blocks of data that the cache management routine has determined may be removed from the cache are erased. Further, the segments containing one or more of the blocks of obsolete data and the blocks of data that the cache management routine has determined may be removed from the cache may be ranked, and the segments erased according to their respective rankings. For example, the segments may be ranked according to a total number of blocks of obsolete data and the blocks of data that the cache management routine has determined may be removed from the cache each contains. Alternatively, the segments may be ranked according to a percentage of a total number of blocks of obsolete data and the blocks of data that the cache management routine has determined may be removed from the cache each contains to a total number of blocks in each segment.
The controller may further maintain a direct mapping of block addresses used by an application accessing data stored in the storage device with physical locations of said data within the flash memory.
These and other embodiments of the invention are discussed further below.
The present invention is illustrated by way of example, and not limitation, in the figures of the accompanying drawings in which:
Described herein are techniques for improving the efficiency of flash memory storage devices when used as a data cache for hard disks.
As mentioned above, when compared to hard disks, flash memory provides relatively fast reads, but is slow when it comes to small writes (i.e., those which do not involve large amounts of data) at random addresses. Therefore, flash memory-based storage devices typically coalesce multiple writes destined for random addresses into contiguous regions in the flash array. However, such write coalescing requires concomitant garbage collection activities, which impose a significant processing load and impact application performance.
While one could, in general, implement a flash memory-based cache for a hard disk-based storage system simply by layering some cache management functions (e.g., functions for maintaining data likely to be used by an application in the cache while moving other data out of the cache) over a flash device that incorporates write coalescing, several inefficiencies would likely result. Accordingly, the present invention provides techniques that integrate cache management and write coalescing to yield efficiencies for such a device. More specifically, the present invention provides methods and systems which offer improved reductions in time and space overhead when looking up data in the cache, and which reduce the processing load imposed by garbage collection routines, thereby reducing the impact on application performance.
Controller 106A is, in some embodiments of the invention, configured such that cache management operations are integrated with write coalescing and garbage collection operations. This integration is achieved by providing a single mapping that maps backend block addresses (BBA) used by applications seeking to access data stored on storage device 100 to physical locations (PL) in the flash array. This single level of mapping avoids the need to first translate BBAs to cache block addresses (CBAs) and then translate the CBAs to PLs as would be required if cache management operations remained distinct from write coalescing operations. The map may be stored in a dynamic random access memory (DRAM) 110 accessible to controller 106A.
To better understand the above, consider first how a conventional cache operates. As noted above, a cache generally stores a subset of the data stored on a larger (in terms of storage capacity), usually slower (in terms of read time), device. Thus, using a cache inherently requires a mapping from the large address space of the slower, backend device to the smaller address space of the cache. Such cache maps typically map a block address of the backend storage device (the BBA) to a block address in the cache (the CBA). If a block is not in the cache, the cache map does not have an entry for its BBA.
The CBA is not, however, sufficient in terms of identifying the actual storage location of data within the flash array. Recall that because a flash array cannot be written at random, write coalescing is used to provide certain efficiencies. Referring to
As noted, the present invention eliminates the need for separate BBA-to-CBA and CBA-to-PL maps and instead provides a single BBA-to-PL mapping for a flash memory-based cache. This is illustrated in
But this efficient mapping process is not the only advantage offered by the present invention. A second efficiency concerns the ability of storage devices configured in accordance with the invention to reduce the load presented by garbage collection processes by leveraging the “temperature” of blocks of data. To better understand this notion, consider the management operations performed in a conventional cache.
In order to take better advantage of the read speed afforded by a cache, many devices employ adaptive routines that seek to ensure data that is used frequently by the host (i.e., applications running on the host) is always available in the cache. Typically, this involves making predictions about which data is likely to be needed (so-called “hot” data) and moving that data into the cache while also moving data less likely to be needed (so-called “cold” data) out of the cache to make room for the hot data. This movement of hot and cold data is referred to as cache management and while various techniques can be employed to make decisions about which data should be classified as hot or cold, such details are not critical to the present invention. That is, the present invention can be employed with any cache management routine. For example, cache management routines may determine the likelihood of access (or “temperature”) of a particular block of data using heuristics based on how recently and/or frequently the block was accessed over a period of time, and such routines are well suited for use in connection with the present invention.
Returning to
As should be apparent from this example, if the flash device were to continue writing new segments as the data is modified, the entire addressable storage space of the flash array would quickly be filled. To prevent this occurrence, the flash device reclaims storage space by identifying segments that have obsolete versions of data blocks and erasing them. Before the erase can be made, however, the non-obsolete blocks in the subject segments must be identified and “moved” (i.e., copied) to new segments. This is shown in illustration 2(e).
In this example, segments 2021, 2022 and 2023 were identified as having obsolete blocks (i.e., blocks 8, 4 and 6). The remaining blocks in these segments (i.e., blocks 5, 2, 1, 3, 7, 9, 4 and 10) were collected and copied into as-yet unused segments 2025 and 2026. As a result, segments 2024, 2025 and 2026 now contain valid data, while segments 2021, 2022 and 2023 all contain only obsolete data. As illustration 2(f) indicates, the segments containing the obsolete data are then erased, making them available for reuse.
This process of cleaning up segments that store obsolete data is known as garbage collection and, as should be apparent from the foregoing discussion, it involves a significant internal read and write load on the controller of the flash device. This reduces the available time for processing reads and writes requested by applications running on a host, thereby reducing the overall performance of the device. Further, the extra writes required for garbage collection reduce the effective write endurance of the flash.
The present invention improves the efficiency of the garbage collection operations by leveraging block temperature. Instead of dealing solely with data blocks that are garbage due to the existence of more recently written versions thereof, the garbage collection operations of flash devices configured in accordance with the present invention also treat “cold” data blocks as if they were garbage and make them eligible for reclamation as part of the garbage collection operation. As indicated above, this technique is independent of the actual criteria used to determine which blocks are sufficiently “cold” to be treated as obsolete, making it applicable to any cache management process.
This time, however, the garbage collection routine is provided with cache management information concerning cold block 2 in segment 4021, cold block 3 in segment 4022, cold blocks 9 and 10 in segment 4023, and cold blocks 12 and 11 in segment 4024. The cold nature of these blocks is indicated by cross hatching in the opposite direction to that used to highlight obsolete blocks. The information regarding the temperature of these blocks is provided to the garbage collection routine in accordance with the present invention and is obtained from the cache management process that oversees the caching of data.
When the garbage collection routine now searches for segments to free up by rewriting non-obsolete blocks to new segments and erasing the previously used segments, the routine seeks out those segments that have the greatest number or highest percentage of garbage blocks, considering cold blocks as garbage blocks, and treats them as candidates for reclamation. In this example, as shown in illustration 4(b), segments 4021, 4022 and 4023 were selected. Blocks 5 and 1 from segment 4021, block 7 from segment 4022 and block 4 from segment 4023 were rewritten to segment 4025 and the original copies of these blocks then flagged as obsolete. As shown in illustration 4(c), this allows three segments, 4021, 4022 and 4023, to be erased, freeing up new storage space for the application to write to.
Notice that by integrating the cache management and garbage collection operations a significant improvement in “write amplification” has been achieved. Write amplification is a measure of the processing load imposed by the garbage collection operations and can be expressed as a ratio of the total number of blocks written within the flash array (because of either application updates or garbage collection) to the total number of blocks updated by the application. In the stable state, the number of blocks written by the application is equal to the number of blocks freed by the garbage collection (GC) operation; therefore, the write amplification may be expressed as:
Write Amplification=(# of blocks copied by GC+# of blocks freed by GC)/# of blocks freed by GC
In the example shown in
Write Amplification=(8+4)/4=3
In general write amplification is the inverse of the average fraction of obsolete or garbage blocks in segments collected by the garbage collection routine.
In comparison, the process illustrated in
Write Amplification=(4+8)/8=1.5
This represents a move of data from three segments to one segment to free up two segments and provides a write amplification one-half that of the scenario depicted in
This reduction in garbage collection load is not possible where cache management is simply layered over a device that performs write coalescing and garbage collection. In such a scheme, the garbage collection routine would not have information concerning cold blocks and so no efficiencies would be obtained. Instead, as provided by the present invention, this significant improvement requires that the cache management operations be integrated with the garbage collection and write coalescing operations of the flash device.
Further expediency simplifications could be achieved if a flash memory device can be permitted to actually drop some non-garbage blocks of data instead of having to rewrite same to other segments during the garbage collection operations. That is, in situations where it is not critical that data be retained in the flash memory, for example where the flash memory serves as a cache for another persistent storage device or medium, the garbage collection routine could be configured to simply treat as garbage all of the blocks of segments that are comprised of mostly obsolete or cold blocks, even where no other copies of some of the blocks in that segment exist in the cache. This may be useful, for example, where space must be freed on an urgent basis or in other circumstances where it is more important to ensure space is available for new data than to ensure that old data is available in the cache. Segment 4022 in illustration 4(a) illustrates an example of a segment that could be treated in this fashion should the need arise.
It is also worth noting that the garbage collection routine need not address all possible segments having obsolete or cold blocks each time segments are to be reclaimed (i.e., for each garbage collection operation, only a portion of the total number of available segments are reclaimed, generally those having the highest number or percentage of obsolete/cold blocks). Notice in
In a related aspect of the present invention, extensions can be made to an ATA or SCSI “trim” command. The trim command has been proposed as a means of allowing applications to indicate that a filesystem can “trim” certain blocks of storage space which are not currently allocated within the filesystem. In accordance with the present invention, this technique can be extended to provide a command that notifies a storage device of changes in a block's temperature (e.g., from hot to cold or vice versa). This would allow the storage device to take actions similar to those discussed above with reference to
The present invention thus provides efficiencies by consolidating the mapping of block addresses to physical addresses into a single map, without the need for an intermediate cache block map, and further by integrating cache management with write coalescing and garbage collection.
At 502, a garbage collection routine in accordance with the invention is initiated. As part of this routine, at 504, obsolete blocks are identified. The identification may be done as in conventional garbage collection routines.
At 506, information from a cache management routine (which may also be executing on the controller) regarding cold blocks is obtained. The information may, for example, be read from a table maintained by the cache management routine and indicate which blocks are considered to be cold. Note that the determination as to whether a block is hot or cold may be made in any conventional fashion, for example on the basis of which blocks are least recently used by an application, or other criteria.
At 508, the process identifies those segments of the flash array having obsolete and/or cold blocks and, optionally at 510, may select a subset of those segments to be reclaimed. This is an optional step because in some implementations a garbage collection routine may simply reclaim a pre-established number of segments each time it executes, for example any number of segments from one to all eligible segments may be reclaimed (although in most cases it will be best to limit reclamation to some number of segments less than all eligible segments in order to reduce the processing burden imposed by the garbage collection routine).
In determining which segments to reclaim, the garbage collection routine may rank the eligible segments (a segment is eligible if it has at least one obsolete or cold block) in terms of the number of obsolete/cold blocks they contain (or, conversely, the number of valid blocks they contain). For example, segments having more (e.g., either literally or on a percentage of blocks-to-segment basis) obsolete/cold blocks may be ranked higher than those having fewer obsolete/cold blocks. The garbage collection routine may operate to reclaim segments in order of rank from such a list, for example, each time reclaiming some pre-determined number of the ranked segments or some pre-determined fraction thereof. Alternatively the number of segments to be reclaimed may be determined dynamically based on a present processing load on the controller.
Regardless of how the segments identified for reclamation are determined, at 512 the valid data, if any, from those segments are rewritten to fresh segments as discussed above. This allows the copies of those previously valid blocks to now be identified as obsolete in the segments from which they were copied and, at 514, those segments are then erased. That is, the segments are reclaimed for use by an application running on a host that is accessing the storage device in which the flash device is operating, for example as a cache.
At 516, the process checks to see if it is done with the reclamation process and, if so, quits at 518. Otherwise, the process repeats from 504 by identifying new segments having obsolete blocks. In many instances, process 500 will execute continually in the background on the controller.
Thus, techniques for improving the efficiency of flash memory, especially when used as a data cache for hard disks, have been described. In the foregoing description, several examples were set forth and discussed in detail, but these examples should not be read as limiting the present invention, which is best measured in terms of the claims that follow.
The present application is a Continuation of U.S. patent application Ser. No. 15/218,299 filed Jul. 25, 2016, which is a Continuation of U.S. patent application Ser. No. 15/052,531 filed Feb. 24, 2016 (now issued as U.S. Pat. No. 9,430,377), which is a Continuation of U.S. patent application Ser. No. 14/873,520 filed Oct. 27, 2015 (now issued as U.S. Pat. No. 9,304,909), which is a Continuation of U.S. patent application Ser. No. 14/750,734 filed Jun. 25, 2015 (now issued as U.S. Pat. No. 9,176,874), which is a Continuation of U.S. patent application Ser. No. 14/656,568 filed Mar. 12, 2015 (now issued as U.S. Pat. No. 9,098,405), which is a Continuation of U.S. patent application Ser. No. 14/221,894 filed Mar. 21, 2014 (now issued as U.S. Pat. No. 9,015,406), which is a Continuation of U.S. patent application Ser. No. 13/595,211 filed Aug. 27, 2012 (now issued as U.S. Pat. No. 8,719,488), which is a Continuation of U.S. patent application Ser. No. 12/636,693 filed Dec. 11, 2009 (now issued as U.S. Pat. No. 8,285,918), all of which are incorporated herein by reference.
Number | Name | Date | Kind |
---|---|---|---|
5519853 | Moran et al. | May 1996 | A |
5530850 | Ford et al. | Jun 1996 | A |
8285918 | Maheshwari | Oct 2012 | B2 |
8527467 | Shmueli | Sep 2013 | B2 |
8555018 | Rohr et al. | Oct 2013 | B1 |
8719488 | Maheshwari | May 2014 | B2 |
8856442 | Kim et al. | Oct 2014 | B2 |
9015406 | Maheshwari | Apr 2015 | B2 |
9043530 | Sundaram et al. | May 2015 | B1 |
9098405 | Maheshwari | Aug 2015 | B2 |
9176874 | Maheshwari | Nov 2015 | B1 |
9223843 | Madhavarapu et al. | Dec 2015 | B1 |
9304909 | Maheshwari | Apr 2016 | B2 |
9430377 | Maheshwari | Aug 2016 | B2 |
9665497 | Maheshwari | May 2017 | B2 |
9697133 | Maheshwari | Jul 2017 | B2 |
9727481 | Shetty et al. | Aug 2017 | B2 |
20050144396 | Eschmann et al. | Jun 2005 | A1 |
20070005928 | Trika et al. | Jan 2007 | A1 |
20070186065 | Lee et al. | Aug 2007 | A1 |
20080034175 | Traister | Feb 2008 | A1 |
20080307192 | Sinclair | Dec 2008 | A1 |
20090204765 | Gupta et al. | Aug 2009 | A1 |
20090327804 | Moshayedi | Dec 2009 | A1 |
20100174845 | Gorobets | Jul 2010 | A1 |
20100274982 | Mehr et al. | Oct 2010 | A1 |
20100333116 | Prahlad et al. | Dec 2010 | A1 |
20110071989 | Wilson et al. | Mar 2011 | A1 |
20130246711 | Testardi et al. | Sep 2013 | A1 |
20130290598 | Fiske et al. | Oct 2013 | A1 |
20150006596 | Fukui et al. | Jan 2015 | A1 |
20150199138 | Ramachandran et al. | Jul 2015 | A1 |
Number | Date | Country |
---|---|---|
104573119 | Apr 2015 | CN |
Entry |
---|
Amendment and Response filed Mar. 20, 2017, from U.S. Appl. No. 15/218,272 (filed Jul. 25, 2016), 8 Pages. |
Amendment and Response to Ex Parte Quayle Action filed Jun. 1, 2016, from U.S. Appl. No. 15/052,531 (filed Feb. 24, 2016), 7 pages. |
Amendment filed Mar. 14, 2017, for U.S. Appl. No. 15/218,299 (filed Jul. 25, 2016), 7 pages. |
Andrew Case, “New Paper: In Lieu of Swap: Analyzing Compressed RAM in Mac OS X and Linux;” Aug. 14, 2014, pp. 1-3 [online], Retrieved from the Internet on Nov. 16, 2015 at URL: <votatility-labs.blogspot.com/2014/08/new-paper-in-lieu-of-swap-analyzing.html>. |
Ex Parte Quayle Action dated Apr. 27, 2016, from U.S. Appl. No. 15/052,531 (filed Feb. 24, 2016), 8 pages. |
Final Office Action dated Jun. 15, 2012, from U.S. Appl. No. 12/636,693 (filed Dec. 11, 2009), 11 pages. |
Golden G. Richard III and Andrew Case, “In lieu of swap: Analyzing compressed RAM in Mac OS X and Linux,” Digital Investigation 11, 2014, pp. S4-S12, Elsevier. |
Information Week, “EMC Documentum Reveals ILM Offering,” Jul. 23, 2004, pp. 1-4 [online], Retrieved from the Internet on Mar. 6, 2018 at URL: <informationweek.com/software/information-management/emc-documentum-reveals-ilm-offering/d/d-id/1026362>. |
John Rydning et al., “Technology Assessment: Cold Storage Is Hot Again Finding the Frost Point,” May 2013, pp. 1-15, vol. 1, IDC. |
KIPO, International Search Report and Written Opinion, International Application No. PCT/US2015/058461, dated Jul. 25, 2016, pp. 1-8. |
Non-Final Office Action dated Dec. 1, 2016, for U.S. Appl. No. 15/218,299 (filed Jul. 25, 2016), 9 pages. |
Notice of Allowance dated Aug. 28, 2012, from U.S. Appl. No. 12/636,693 (filed Dec. 11, 2009), 8 pages. |
Notice of Allowance dated Aug. 7, 2015, from U.S. Appl. No. 14/750,734 (filed Jun. 25, 2015), 9 pages. |
Notice of Allowance dated Dec. 23, 2013, from U.S. Appl. No. 13/595,211 (filed Aug. 27, 2012), 8 pages. |
Notice of Allowance dated Jan. 20, 2016, from U.S. Appl. No. 14/873,520 (filed Oct. 27, 2015), 9 pages. |
Notice of Allowance dated Jun. 16, 2016, from U.S. Appl. No. 15/052,531 (filed Feb. 24, 2016), 5 pages. |
Notice of Allowance dated Jun. 24, 2015, from U.S. Appl. No. 14/656,568 (filed Mar. 12, 2015), 10 pages. |
Notice of Allowance dated Mar. 2, 2015, from U.S. Appl. No. 14/221,894 (filed Mar. 21, 2014), 12 pages. |
Notice of Allowance dated Apr. 4, 2017, from U.S. Appl. No. 15/218,272 (filed Jul. 25, 2016), 5 pages. |
Notice of Allowance dated Mar. 30, 2017, for U.S. Appl. No. 15/218,299 (filed Jul. 25, 2016), 5 pages. |
Office Action dated Apr. 20, 2015, from U.S. Appl. No. 14/656,568 (filed Mar. 12, 2015), 10 pages. |
Office Action dated Aug. 25, 2014, from U.S. Appl. No. 14/221,894 (filed Mar. 21, 2014), 12 pages. |
Office Action dated Dec. 21, 2011, from U.S. Appl. No. 12/636,693 (filed Dec. 11, 2009), 8 pages. |
Office Action dated Mar. 20, 2013, from U.S. Appl. No. 13/595,211 (filed Aug. 27, 2012), 8 pages. |
Office Action dated Nov. 30, 2016, from U.S. Appl. No. 15/218,272 (filed Jul. 25, 2016), 10 pages. |
Response to Final Office Action filed Aug. 15, 2012, from U.S. Appl. No. 12/636,693 (filed Dec. 11, 2009), 10 pages. |
Response to Non-Final Office Action of Jun. 30, 2017, filed Sep. 28, 2017, from U.S. Appl. No. 15/612,803 (filed Jun. 2, 2017), 3 pages. |
Response to Office Action dated Jun. 2, 2015, from U.S. Appl. No. 14/656,568 (filed Mar. 12115), 8 pages. |
Response to Office Action filed Apr. 13, 2012, from U.S. Appl. No. 12/636,693 (filed Dec. 11, 2009), 13 pages. |
Response to Office Action filed Nov. 25, 2014, from U.S. Appl. No. 14/221,894 (filed Mar. 21, 2014), 9 Pages. |
Response to Office Action filed Oct. 10, 2013, from U.S. Appl. No. 13/595,211 (filed Aug. 27, 2012),5 pages. |
Seagate, “Gassing Up Your SSD: Top Off the Tank for Higher Performance,” Apr. 4, 2013, <https://web.archive.org/web/20150111042909/https://blog.seagate.com/intelligent/gassing-up-your-ssd/>(3 pages). |
Supplemental Response to Office Action filed Apr. 16, 2012, from U.S. Appl. No. 12/636,693 (filed Dec. 11, 2009), 9 pages. |
USPTO, Non-Final Office Action dated Jun. 30, 2017, from U.S. Appl. No. 15/612,803 (filed Jun. 2, 2017), 7 pages. |
USPTO, Notice of Allowance and List of References cited (892), dated Mar. 24, 2017, in U.S. Appl. No. 15/019,909 (filed Feb. 9, 2016), 8 pages. |
USPTO, Notice of Allowance dated Oct. 23, 2017, from U.S. Appl. No. 15/612,803 (filed Jun. 2, 2017), 5 pages. |
Number | Date | Country | |
---|---|---|---|
20180165216 A1 | Jun 2018 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15612803 | Jun 2017 | US |
Child | 15893150 | US | |
Parent | 15218299 | Jul 2016 | US |
Child | 15612803 | US | |
Parent | 15052531 | Feb 2016 | US |
Child | 15218299 | US | |
Parent | 14873520 | Oct 2015 | US |
Child | 15052531 | US | |
Parent | 14750734 | Jun 2015 | US |
Child | 14873520 | US | |
Parent | 14656568 | Mar 2015 | US |
Child | 14750734 | US | |
Parent | 14221894 | Mar 2014 | US |
Child | 14656568 | US | |
Parent | 13595211 | Aug 2012 | US |
Child | 14221894 | US | |
Parent | 12636693 | Dec 2009 | US |
Child | 13595211 | US |