This invention relates to caching data on a solid-state storage device and, in particular, to maintaining metadata pertaining to the cache and the solid-state storage device for use in managing the cache and performing grooming operations.
Cache devices are storage devices that are used to allow data to be accessed more quickly in a system. Caches can significantly improve performance in systems by reducing the input/output (I/O) time for operations that use the data in the cache. Generally, the cache is implemented in front of another storage device which has greater storage capacity, but slower I/O times, than the cache device. The benefits of caches are well understood and caches have been implemented advantageously in a variety of contexts and scales ranging from the caches in CPUs to caches in storage area networks (SANs).
The nature of the cache may also vary based on the context; for example, in a single computer (such as a desktop or a laptop) the cache may be implemented in volatile memory, such as static random access memory (SRAM), dynamic random access memory (DRAM), or others. Currently, cache grooming is inefficient because it is separate from other memory management operations.
A solid-state storage device may operate as a nonvolatile cache device for a nonvolatile backing store. The solid-state storage device may be transparent to the clients using the logical space presented by the solid-state storage device. In some embodiments, the solid-state storage device may be configured to present a logical size to the client that is the same as the physical size of the backing store, such that the solid-state storage device presents to clients a logical storage space that is the same size as the backing store, regardless of the physical storage size of the solid-state storage device. Alternatively, or in addition, the solid-state storage device may present a logical address space to the client that exceeds a physical storage capacity of both the solid-state storage device and the backing store.
The solid-state storage device may maintain metadata pertaining to cache operations and storage operations performed on the solid-state storage device. The storage metadata may include, but is not limited to metadata pertaining to sequential storage operations performed on the solid-state storage device (e.g., associations between client logical addresses and storage units on the device, storage log metadata, etc.), maintenance operations pertaining to the sequential storage operations (e.g., storage recovery, garbage collection, error correction, etc.), the status of the storage media (e.g., wear levels, reliability, integrity, etc.), and so on. The storage metadata may comprise validity indicators to distinguish between storage units comprising valid data from those comprising invalid data. A storage unit may refer to any addressable storage location on the solid-state storage device.
The solid-state storage device may cache data in storage units of the solid-state storage device. The cache metadata may comprise metadata pertaining to the cache stored on the solid-state storage device, cache operations performed on the solid-state storage device (e.g., access patterns, etc.), cache management (e.g., cache eviction and/or retention policy, cache attributes, cache directives, etc.), and so on. The cache metadata may comprise discardability indicators to identify data that can be discarded. Discardable data may be cache data that has been backed up (e.g., stored) on a backing store.
The metadata maintained by the solid-state storage device may comprise cache entries, which may represent data cached on the device (e.g., data cached in the storage units of the solid-state storage device). A cache entry may associate a logical address (of a storage client) with one or more storage locations (e.g., storage units) on identifying where the data is stored on the solid-state storage device; a cache entry may provide an index or mapping between client logical addresses, and solid-state storage device storage locations (e.g., storage units). In some embodiments, a cache entry may have a variable size, and as such, may reference a portion of a storage unit, a single storage unit, and/or multiple storage units on the solid-state storage device. For instance, a first cache entry may comprise a first amount of data on a single storage unit (e.g., 256 bytes on a portion of a single storage unit), whereas a second cache entry may comprise a second, different amount of data on multiple storage units (e.g., 4 kilobytes on a plurality of storage units, or virtual storage unit).
The cache entries may be indexed by logical address, which may enable fast and efficient access to the cache entries responsive to client storage requests. The cache entries may be used to store (and/or reference) shared metadata, including metadata related to sequential storage operations of the solid-state storage device, as well as the cache. For example, a groomer (described below) of the solid-state storage device may access the metadata pertaining to the cache to efficiently perform storage recovery and/or maintenance operations. Similarly, a cache manager (described below) operating on and/or in conjunction with the solid-state storage device may access the metadata to efficiently manage the cache (e.g., make cache policy decisions pertaining to cache eviction and/or retention, and so on). In some embodiments, the solid-state storage device may provide a storage interface (or “interface” generally), which may expose portions of the metadata to a storage client. Using the interface, a storage client may configure the solid-state storage device (e.g., set cache attributes, cache directives, issue TRIM commands, and so on). In addition, the interface may expose a “sparse” or “thinly provisioned” logical address space through the interface, which may allow the storage client to interact with the solid-state storage device without a redundant address translation layer (e.g., within a full 64 bit address space).
In some embodiments, a backup engine may be configured to backup data cached on the solid-state storage device to a backing store. When stored on the backing store, metadata associated with the data (e.g., associated with the cache entry and/or corresponding storage units) may be updated to indicate that the data is discardable. The backup engine may iterate through the storage units of the solid-state storage device chronologically (from oldest to newest). The traversal may be enabled by the shared metadata maintained by the solid-state storage device (e.g., the reverse index and/or sequence indicators of the storage units). The operation of the backup engine may be informed by the storage metadata; for example, the backup engine may ignore data associated with cache entries (and/or in associated storage units) that the metadata indicates are invalid (e.g., data has been moved to a different storage location due to a recovery operation, a sequential write, modification, or the like).
In some embodiments, the backup engine operates at a relatively low priority (e.g., as a background process or operation). The backup engine may be configured to operate at a higher priority when needed. For example, the metadata may indicate a physical storage capacity, a target occupancy, a maximum occupancy, and/or a current occupancy of the solid-state storage device. When the current occupancy exceeds a threshold (e.g., exceeds the target occupancy, the maximum occupancy, or some other threshold), the background engine may be configured to operate at a higher priority.
Storage recovery and maintenance operations may be similarly informed by the metadata. For example, a groomer process configured to recover storage divisions of the solid-state storage device may be prioritized using the metadata (e.g., priority increased when the current occupancy exceeds a threshold). The groomer may access metadata related to the cache in performing its data recovery operations. For example, the groomer may select storage divisions for recovery based on the metadata, may select storage units in a storage division to be retained (e.g., moved to a different storage division) using the metadata, and so on.
Accordingly, in some embodiments, a method for caching data on a solid-state storage device may comprise: maintaining metadata pertaining to cache operations performed on a solid-state storage device and to storage operations of the solid-state storage device; accessing the metadata to manage storage operations of the solid-state storage device; and accessing the metadata to manage a cache stored on the solid-state storage device. The metadata pertaining to cache operations and the metadata pertaining to storage operations may be stored in a shared datastructure.
The metadata may comprise discardability indicators identifying storage units that have been stored on a backing store, and the indicators may be used to select storage divisions for recovery. Cache eviction metadata (e.g., cold, hot, warm, etc.) may be used to select storage divisions for recovery and/or to determine whether storage units in a storage division selected for recovery should be retained. The metadata may be associated with a logical address of the cache entry and, as such, may remain associated with the cache entry when the storage location of the corresponding data changes on the solid-state storage device.
The metadata pertaining to the cache (e.g., shared cache metadata) may be used to manage sequential storage operations of the solid-state storage device, such as storage division recovery, garbage collection, error management, sequential storage (e.g., log-based storage operations), and so on. Metadata pertaining to sequential storage operations (e.g., shared storage metadata) may be used to manage the cache. For example, cache operations such as backup, cache eviction, occupancy, cache data movement, and the like may be informed by the shared, storage metadata.
The metadata may comprise a plurality of cache entries, each associating a logical address with storage units on the solid-state storage device comprising the data of the cache entry. Client storage requests may be serviced using the cache entries. If a cache entry for a particular logical address exists in the metadata, the corresponding data may be available on the solid-state storage device; otherwise, the data may not be available (and a cache miss may occur). Client storage requests may be serviced quickly and efficiently, since the cache entries are indexed by logical address (e.g., in a B-Tree datastructure). The metadata may be used to determine a current occupancy of the solid-state storage device. The client entries may be modified responsive to cache and/or storage operations. When proximate logical addresses are stored, the corresponding cache entries may be merged, which may reduce the number of entries that must be searched to find a particular logical address. Similarly, when logical addresses are removed, the corresponding cache entries may (or may not) be split. The cache entries may be of variable size and, as such, may comprise a single logical address, a range of logical addresses, a set of logical addresses, or the like.
Reference throughout this specification to features, advantages, or similar language does not imply that all of the features and advantages that may be realized with the present invention should be or are in any single embodiment of the invention. Rather, language referring to the features and advantages is understood to mean that a specific feature, advantage, or characteristic described in connection with an embodiment is included in at least some embodiments of the present invention. Thus, discussion of the features and advantages, and similar language, throughout this specification may, but do not necessarily, refer to the same embodiment.
Furthermore, the described features, advantages, and characteristics of the invention may be combined in any suitable manner in one or more embodiments. One skilled in the relevant art will recognize that the invention may be practiced without one or more of the specific features or advantages of a particular embodiment. In other instances, additional features and advantages may be recognized in some embodiments that may not be present in all embodiments of the invention.
These features and advantages of the present invention will become more fully apparent from the following description and appended claims, or may be learned by the practice of the invention as set forth hereinafter.
In order that the advantages of the invention will be readily understood, a more particular description of the invention briefly described above will be rendered by reference to specific embodiments that are illustrated in the appended drawings. Understanding that these drawings depict only typical embodiments of the invention and are not therefore to be considered to be limiting of its scope, the invention will be described and explained with additional specificity and detail through the use of the accompanying drawings, in which:
Many of the functional units described in this specification have been labeled as modules, in order to more particularly emphasize their implementation independence. For example, a module may be implemented as a hardware circuit comprising custom VLSI circuits or gate arrays, off-the-shelf semiconductors such as logic chips, transistors, or other discrete components. A module may also be implemented in programmable hardware devices such as field programmable gate arrays, programmable array logic, programmable logic devices or the like.
Modules may also be implemented in software for execution by various types of processors. An identified module of executable code may, for instance, comprise one or more physical or logical blocks of computer instructions which may, for instance, be organized as an object, procedure, or function. Nevertheless, the executables of an identified module need not be physically located together, but may comprise disparate instructions stored in different locations which, when joined logically together, comprise the module and achieve the stated purpose for the module.
Indeed, a module of executable code may be a single instruction, or many instructions, and may even be distributed over several different code segments, among different programs, and across several memory devices. Similarly, operational data may be identified and illustrated herein within modules, and may be embodied in any suitable form and organized within any suitable type of data structure. The operational data may be collected as a single data set, or may be distributed over different locations including over different storage devices, and may exist, at least partially, merely as electronic signals on a system or network. Where a module or portions of a module are implemented in software, the software portions are stored on one or more computer readable media.
Reference throughout this specification to “one embodiment,” “an embodiment,” or similar language means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment of the present invention. Thus, appearances of the phrases “in one embodiment,” “in an embodiment,” and similar language throughout this specification may, but do not necessarily, all refer to the same embodiment.
Reference to a computer readable medium may take any form capable of storing machine-readable instructions on a digital processing apparatus. A computer readable medium may be embodied by a transmission line, a compact disk, digital-video disk, a magnetic tape, a Bernoulli drive, a magnetic disk, a punch card, flash memory, integrated circuits, or other digital processing apparatus memory device.
Furthermore, the described features, structures, or characteristics of the invention may be combined in any suitable manner in one or more embodiments. In the following description, numerous specific details are provided, such as examples of programming, software modules, user selections, network transactions, database queries, database structures, hardware modules, hardware circuits, hardware chips, etc., to provide a thorough understanding of embodiments of the invention. One skilled in the relevant art will recognize, however, that the invention may be practiced without one or more of the specific details, or with other methods, components, materials, and so forth. In other instances, well-known structures, materials, or operations are not shown or described in detail to avoid obscuring aspects of the invention.
The schematic flow chart diagrams included herein are generally set forth as logical flow chart diagrams. As such, the depicted order and labeled steps are indicative of one embodiment of the presented method. Other steps and methods may be conceived that are equivalent in function, logic, or effect to one or more steps, or portions thereof, of the illustrated method. Additionally, the format and symbols employed are provided to explain the logical steps of the method and are understood not to limit the scope of the method. Although various arrow types and line types may be employed in the flow chart diagrams, they are understood not to limit the scope of the corresponding method. Indeed, some arrows or other connectors may be used to indicate only the logical flow of the method. For instance, an arrow may indicate a waiting or monitoring period of unspecified duration between enumerated steps of the depicted method. Additionally, the order in which a particular method occurs may or may not strictly adhere to the order of the corresponding steps shown.
As used herein, metadata may refer to any data pertaining to the operation of the solid-state storage device. Accordingly, metadata may refer to cache metadata and/or storage metadata. In some embodiments, the solid-state storage device may be configured to maintain a shared datastructure comprising cache and/or storage metadata. The datastructure may be accessible by a cache manager, a groomer, and/or other management modules, which may use the metadata to efficiently manage caching, storage, and/or maintenance operations (e.g., grooming) of the solid-state storage device.
As used herein, cache metadata may refer to any data pertaining to a cache stored on the solid-state storage device and/or to cache operations performed on the solid-state storage device. Cache metadata may include, but is not limited to: discardability indicators to identify cache entry storage units that have been backed up to a backing store, cache eviction metadata, cache attributes (e.g., PIN, BLACKLIST), cache directives (e.g., Flush), and so on. For example, cache eviction metadata may comprise data access patterns, such as least recently used patterns, access frequency, and so on. The cache eviction metadata may be used to make cache eviction decisions regarding particular cache entries. For example, the cache eviction metadata may be used to identify a cache entry that is “hot,” “warm,” and/or cold, or the like. Cold cache entries may be candidates for eviction, whereas “hot” entries may be retained. The cache eviction metadata may also include cache attributes, such as PIN, BLACKLIST, or the like, which may be used to manage cache operations (e.g., a PIN directive may prevent data from being evicted from the solid-state storage device). Similarly, the cache eviction metadata may respond to cache directives, such as a “FLUSH” directive to flush or evict a cache entry from the cache.
As used herein, storage metadata may refer to any data pertaining to solid-state storage media of the solid-state storage device and/or storage operations performed on the solid-state storage device. Storage metadata may include, but is not limited to: validity indicators to identify valid and/or invalid data on the solid-state storage device, forward indexing metadata to provide efficient logical address to storage unit lookups, reverse indexing metadata to provide efficient storage unit to logical address lookups, occupancy metadata (e.g., data stored on the solid-state storage device, occupancy threshold, etc.), storage media status (e.g., wear level, reliability, error rate, etc.), sequential storage metadata (e.g., logging, etc.), and so on.
As used herein, a storage unit may refer to any location for storing data on the solid-state storage device. Therefore, a storage unit may comprise a block of a solid-state storage element, a virtual block comprising blocks, or portions thereof, on a plurality of solid-state storage elements, a page, a virtual page comprising pages, or portions thereof of a plurality of solid-state storage elements, an storage division, a virtual storage division, an erase block, a virtual erase block, or the like.
A storage unit may be part of a storage division. As used herein, a storage division may refer to an erasable division of the solid-state storage device. Accordingly, a storage division may refer to an erase block of a solid-state storage element, a virtual erase block comprising storage divisions, or portions thereof, on a plurality of solid-state storage elements, or the like.
In some embodiments, a cache may comprise a plurality of “cache entries.” A cache entry may comprise a logical address (or range of logical addresses, the range of logical addresses may include ordered addresses, unordered addresses, contiguous addresses, and noncontiguous addresses) and one or more storage units on which data corresponding to the logical addresses is stored on the solid-state storage device. The data of a cache entry may be stored on a portion of a storage unit, a single storage unit, and/or multiple storage units. A cache entry may have a variable size. As discussed above, a cache entry may be associated with a logical address of a client. The associations between logical addresses and respective cache entries may be maintained by the solid-state storage device 114 (e.g., in metadata 117). The data of a cache entry may be stored within a storage unit of the solid-state storage device (e.g., a page, a virtual page, an erase block, and/or erase block). Alternatively, the data of a cache entry may cross storage boundaries (e.g., be stored on a plurality of pages, virtual pages, erase blocks, and/or virtual erase blocks).
The metadata maintained by solid-state storage device may comprise a plurality of cache entries, each associating a logical address with corresponding storage units on the solid-state storage device. These associations may be maintained in a map or index (referred to as a “forward map” or “forward index”). As used herein, a forward map or forward index (or index generally) may refer to any datastructure capable of associating cache entry logical addresses (e.g., logical block addresses from a client) with storage units of the solid-state storage device. In some embodiments, the index may comprise a tree datastructure structure, such as a B-tree or the like. The tree datastructure may comprise cache entries indexed by logical address, or set of logical addresses (e.g., as a range and/or length). Accordingly, the index may be referred to as a Range B-Tree or (B-Tree generally). Examples of datastructures comprising a plurality of cache entries are described below in conjunction with
The forward index may be used to determine whether data corresponding to a logical address is stored on the solid-state storage device. For example, if a node associated with the address is in the index, then valid data for the address is available on the solid-state storage device; otherwise, if no node exists in the index, data corresponding to the logical address is not stored on the solid-state storage device).
The metadata may further comprise a reverse index, which may index storage units by their physical address (as opposed to indexing by logical address as in the forward map). Accordingly, the reverse index may provide fast and efficient lookups of particular storage units and/or provide for a sequential traversal through a physical address space of the solid-state storage device. In some embodiments, a reverse index may link to the same nodes (leaf nodes) as the forward index (e.g., the reverse index may mirror the forward index, differing in that leaf nodes are indexed by physical as opposed to logical addresses). Alternatively, the reverse index may comprise a separate datastructure, such as a tree, map, array, table, bitmap, or the like. In some embodiments, the reverse index may provide associations between storage units and storage unit metadata, such as a validity indicator, discardability indicator (e.g., whether the data of the storage unit has been stored on a backing store), storage unit wear level, error rate, access patterns, and the like.
As used herein, cache eviction metadata, may refer to metadata used to make cache eviction (or retention) decisions. Cache eviction metadata may be associated with a cache entry, and as such, may be indexed by logical address, as opposed to a particular storage location on the solid-state storage device. This may allow the cache eviction metadata to be maintained despite changes to the physical storage location of the data (e.g., due to sequential storage operations). Cache eviction metadata may identify cache entries that are “hot,” “warm” and/or “cold.” As used herein, “hot,” “warm,” and/or “cold” data may refer to data having a particular access and/or use pattern, physical spatial relationship to other data (in other words the “temperature” of a set of data may be determined or influenced by closely the data is physically on the media or logically in the logical address space to other data used in the cache). For instance, “hot” or “warm” data may refer to data that has been recently accessed and/or has been accessed at a particular frequency (as indicated by the cache eviction metadata). On the other hand, “cold” data may refer to data that has not been accessed within a particular time threshold, at a particular frequency, or the like (also as indicated by the cache eviction metadata). Various different techniques and/or data access patterns known in the art may be used to identify “hot,” “warm,” and/or “cold” data, any of which may be implemented under the teachings of this disclosure. In addition, cache eviction metadata may comprise cache attributes and/or directives. The cache attributes may inform cache management; for example, a “PIN” cache directive may prevent a particular cache entry from being evicted from the solid-state storage device, and a “BLACKLIST” attribute and/or “FLUSH” directive may prevent data from being retained in the cache.
In some embodiments, the solid-state storage device may store data sequentially. As used herein, sequential storage may refer to an out-of-place storage scheme in which data is stored at one or more append points. Accordingly, when data stored on the solid-state storage device is modified (or re-stored for some other reason), the data may be stored at different storage units of the solid-state storage device (e.g., as opposed to be re-written to the same storage unit where the data was originally stored).
Sequential storage may also refer to a mechanism for maintaining storage metadata (e.g., log-based storage) by the solid-state storage device. The storage metadata may allow the sequence of storage operations performed on the solid-state storage device to be reconstructed or replayed. The storage metadata may include storage division timestamps, sequence indicators, packets headers, forward and/or reverse indexes and so on. For example, when data corresponding to a first logical address is stored on the solid-state storage device, logging data may be maintained. The metadata may comprise indexing the data in the forward index (e.g., B-Tree) disclosed above. For example, a cache entry corresponding to the logical address may be added to the index, which may be configured to reference storage units on the solid-state storage device comprising data corresponding to the logical address. As discussed above, the logical address may comprise an address and length or range, and/or may comprise a set of logical address. When the data is modified by a client (or other entity), the modified data may be stored sequentially (e.g., “out-of-place”) on the solid-state storage device. Accordingly, the storage units comprising the valid data associated with the logical addresses may change. The cache entry may be updated to reflect the new storage unit(s), and the old storage units may be invalidated (e.g., in a reverse index, or other data structure). This sequence of storage operations may be replayed using the forward map, sequence indicators stored on the solid-state storage device, and/or other storage metadata (e.g., a reverse index, or the like). Other sequential storage operations may include, but are not limited to: storage division recovery, garbage collection, error management, virtual storage management, sequential storage (e.g., log-based storage operations), and so on. As used herein, a sequential storage operation may refer to any storage operation described above.
The solid-state storage device 114 may be configured to provide caching services to the clients 110a-c and/or the backing store 118. Using the solid-state storage device 114 as a cache may provide a number of different benefits. For example, the access time for clients 110a-c reading from the solid-state storage device 114 and writing to the solid-state storage device 114 may be less than the access times for clients 110a-c reading and writing data directly from/to the backing store 118, the solid-state storage device 114 allows the clients 110a-c to access data more quickly. In some embodiments, the solid-state storage device 114 may present a logical interface to the clients 110a-c, which may comprise a logical address space that exceeds a physical storage capacity of the solid-state storage device 114. The logical address space may correspond to the physical storage capacity of the backing store 118, which may allow the clients to interact with the solid-state storage device 114 transparently (e.g., as if they were accessing the backing store 118 directly). Alternatively, or in addition, the solid-state storage device 114 may present a logical address space that exceeds the physical storage capacity of both the solid-state storage device 114, and the backing store 118 (e.g., a “sparse” or “thin” logical address space). The logical address space may comprise a 64 bit address space, a 32 bit address space, or another addressing scheme used by the clients 110a-c.
The solid-state storage device 114 may implement a logical to physical index to map logical addresses of the clients 110a-c to storage units on the solid-state storage device 114, which may remove a redundant translation layer implemented by the clients 110a-c, simplify client 110a-c access to the storage devices 114 and 118, and/or provide a flexible storage infrastructure for the clients 110a-c (e.g., a “sparse” or “thinly provisioned” storage services).
The solid-state storage device 114 may comprise a solid-state storage controller 115, which may be configured to maintain metadata 117 pertaining to a cache stored on the solid-state storage device 114 (and/or pertaining to a cache operations performed on the device 114), as well as storage metadata pertaining to storage operations performed on the solid-state storage device 114, the status of storage units on the solid-state storage device 114, and so on. The metadata may comprise a validity map and a discardability map. The validity map may identify storage units of the solid-state storage device 114 that comprise valid data (data that is up-to-date and is associated with a logical address of a client 110a-c). Accordingly, the validity map may comprise the forward index discussed above. An invalidity map may be derived from the validity map (e.g., all storage units not in the validity map may be identified as comprising invalid data) and/or may be maintained in a separate datastructure, such as a reverse index, or the like. The discardability map may identify cache entry storage units that have been stored on the backing store 118.
The solid-state storage device 114 and/or controller 115 may comprise a cache management apparatus 130, which may use the metadata to identify cache entries in response to clients storage requests. In some embodiments, the cache management apparatus 130 may be implemented using a device driver (e.g., a Linux device driver).
While
The solid-state storage device 114 functions as a cache for the backing store 118. In some embodiments, the solid-state storage device 114 and the backing store 118 are connected by a communication bus 116. The communication bus 116 may be a local communication bus (e.g., a communication bus of a computer system), a distributed bus, a, local area network (“LAN”), a wide area network (“WAN”), or other type of communication network, and may be implemented using various protocols. In some embodiments, the solid-state storage device 114 is directly attached to the backing store 118. The solid-state storage device 114 and the backing store 118 may also be connected by a bus.
The solid-state storage device 114 may be configured to act as a look aside cache that sees data access requests substantially simultaneously with the backing store 118. In other embodiments, the solid-state storage device 114 is configured to act as a look through cache that sees the data access requests before the backing store 118. In some embodiments, where the solid-state storage device 114 is a look through cache, the solid-state storage device 114 passes the data access request to the backing store 118 if the attempted access on the solid-state storage device 114 is a miss. The solid-state storage device 114 may also be implemented as a write-back cache. In some embodiments, the current caching mode may be available in the metadata maintained by the solid-state storage device 114. The caching mode may be modifiable by the clients 110a-c via a storage interface, an enhanced storage interface, or the like. The metadata may further indicate cache attributes such as PIN, BLACKLIST, or the like and/or cache directives (e.g., FLUSH). The cache attributes and/or directives may be stored in the metadata maintained by the solid-state storage device 114 for access by the cache management apparatus 130 and/or solid-state storage controller 115.
The backing store 118 may provide storage for the clients 110a-c, which read and write data from/to the backing store 118. The backing store 118 may include one or more nonvolatile storage devices such as solid state storage, hard disk drives, tape, some other nonvolatile data storage medium, or a combination of the preceding examples. In some embodiments, the backing store 118, or at least certain storage mediums in the backing store 118, may have access and/or throughput performance that is lower than that of the solid-state storage device 114.
In some embodiments, the system 100 is implemented within a computing device. For example, the solid-state storage device 114 may be attached to a motherboard of a computing device, such as a desktop and act as a cache for backing store 118 that is a hard disk drive. The CPU of the computing device may use the solid-state storage device 114 to read and write data and take advantage of the faster read/write times of the solid-state storage device 114 in comparison to the backing store 118. The system 100, in some embodiments, also includes additional solid-state storage devices 114. For example, the system 100 may implement a multi-tiered cache with multiple solid-state storage devices 114 organized in a cache hierarchy.
In some embodiments, the backing store 118 provides physical storage space on which the data is actually stored by the file system 210. The amount of space made available by the backing store 118 can vary widely depending on the system 200; 500 GB may be typical for a desktop computer, while Terabytes (“TB”) or Petabytes (“PB”) of backing store 118 may be provided in a distributed storage system, such as a Storage Area Network (SAN). Typically, the backing store 118 is made up of storage devices that are block storage devices; however, other types of backing storage devices may be used in the backing store 118.
The solid-state storage device 114 may have a pre-determined physical storage capacity. In some embodiments, the solid-state storage device 114 may be logically and/or physically partitioned, such that a portion of the physical space of the solid-state storage device 114 is dedicated to act as a cache for the backing store 118, and a part of the solid-state storage device 114 is configured to operate as storage (e.g., part of the backing store 118, or other storage service). For example, a solid-state storage device 114 may be formatted such that 200 GB of physical space is dedicated to act as a cache, while the remaining 120 GB is treated as part of the backing store 118.
In some embodiments, the solid-state storage device 114 presents a cache logical space 220 to the file system 210 that is the same size as the physical space of the backing store 118 regardless of the physical size of the solid-state storage device 114. For example, the solid-state storage device 114 may be formatted to be the same size as the backing store 118. In typical embodiments, the physical storage capacity of the backing store 118 (e.g., 1 PB) is larger than the physical storage capacity of the solid-state storage device 114 (e.g., 80 GB).
In some embodiments, the solid-state storage device 114 may present a logical address space 220 to the file system 210 that is larger than the physical space of the backing store 118 (as well as the solid-state storage device 114 itself). A user may be able to add physical space to the backing store 118 without changing the size of the cache logical space 220. The solid-state storage device 114 may maintain associations between logical addresses and the “sparse” or “thinly provisioned” storage units in the device 114 and/or backing store 118 using a forward map as described above. In some embodiments, the solid-state storage device 114 may be configured to request additional storage space (e.g., addition backing store 118) when an occupancy of the solid-state storage device 114 and/or backing store 118 exceeds a threshold. The solid-state storage device 114 may present an interface by which a client (e.g., file system 210) may issue queries to determine the actual physical capacity of the solid-state storage device 114, the backing store 118, the remaining storage capacity, current occupancy levels (e.g., the amount of valid data currently stored on the solid-state storage device), and so on. The interface may further provide calls for the client (e.g., file system 210) to request storage capacity, reserve storage capacity, and so on.
In some embodiments, the solid-state storage device 114 may dynamically adjust the size of the cache logical space 220 based on the changing physical storage of the backing store 118. For example, in the event that a new storage device is added to the backing store 118, the solid-state storage device 114 may dynamically increase the size of the cache logical space 220 presented to the file system 210.
In some embodiments, the solid-state storage device 114 may be transparent to the file system 210. The file system 210 may only see the cache logical space 220 presented to it by the solid-state storage device 114. In some embodiments, the solid-state storage device 114 acts as a controller for the backing store 118. Thus, the file system 210 directs data access events (such as read and write requests) to the cache logical space 220 that it is presented with by the solid-state storage device 114. The file system 210 may be unaware that the solid-state storage device 114, which it is dealing with does not have a physical storage space that is as large as the cache logical space 220. In some embodiments, the solid-state storage device 114 stripes data to the backing store 118 in a RAID configuration. In some embodiments, the solid-state storage device 114 may be configured with other solid-state storage devices in a RAID configuration.
Thus, the solid-state storage device 114 may implement a sparsely-mapped 64-bit address space. In some embodiments, when the file system 210 requests data (for example, a read request) and provides an address for the requested data (such as a logical block address, or LBA), the solid-state storage device 114 (and/or controller 115) may look up the address in the metadata 117 (an index, such as the forward index described above) maintained by the solid-state storage device 114 and, if data associated with the logical address is available on the solid-state storage device 114, initiate a read at the corresponding storage units. The solid-state storage device 114 may generate an error if the data is not at the specified address and/or the logical address is not in the index. The error may be interpreted as a cache miss, which may be handled by the solid-state storage device 114 reading requested data from the backing store 118 and caching the data in one or more storage units. The cache miss may further comprise updating the metadata to reference the logical address (e.g., add a leaf node to index), and so on.
In embodiments where the solid-state storage device 114 presents a cache logical space 220 that is larger than the physical space of the solid-state storage device 114, the cache management apparatus 130 manages the solid-state storage device 114 to ensure that the solid-state storage device 114 does not exceed its physical storage capacity. Exemplary approaches for managing the physical storage capacity of the solid-state storage device 114 while it presents a larger logical space 220 are described below.
In such embodiments, the solid-state storage device 114 may be transparent to the file system 210 and other entities seeking access to the backing store 118. The complexities of managing the solid-state storage device 114 and looking for data may be hidden from the file system 210. Thus, the file system 210 may not need to check in the solid-state storage device 114 for data to read, or write to the solid-state storage device 114 instead of writing to the backing store 118. Instead, the file system 210 simply accesses data in the cache logical space 220 that is presented by the solid-state storage device 114 and allows the solid-state storage device 114 to transparently determine the most efficient way to handle the data. The physical storage arrangement underlying the cache logical space 220 (i.e., the solid-state storage device 114 and its relationship with the backing store 118) is hidden from the file system 210. In addition, by allowing the solid-state storage device 114 to manage the caching of data, the cache management apparatus 130 may make intelligent choices about cache maintenance using the metadata 117 maintained by the solid-state storage device 114 that is normally not available to the file system 210.
In the depicted embodiment, the cache management apparatus 130 comprises a cache controller 312, and a backup engine 314. The groomer 310 and cache management apparatus 130 may have access to metadata 117, which may comprise, inter alia, storage metadata 316, which may comprise a validity map, and storage metadata 318, which may comprise a discardability map. The groomer 310 and/or cache management apparatus 130 may be implemented as firmware, software, hardware, or some combination thereof. As illustrated in
The groomer 310 moves data stored on the solid-state storage media 320 as part of maintenance operations, such as storage division recovery (e.g., erase block recovery), data refresh operations, data error correction, error handling, and so on. The groomer 310 may move data between storage units in the solid-state storage media 320 in a manner that is transparent to the file system 210 and the backing store 118. For example, the groomer 310 may perform garbage collection operations on the data in a solid state storage device that is operating as a solid-state storage device 114 in order to reclaim storage space on the solid-state storage device 114. The groomer 310 may also defragment files that are heavily fragmented in the solid-state storage device 114. Exemplary grooming operations are described in U.S. patent application Ser. No. 11/952,101 for David Flynn, et al., filed Dec. 6, 2007, entitled “Apparatus, System, and Method for Storage Space Recovery in Solid-state Storage,” which is hereby incorporated by reference in its entirety. Similarly, the groomer 310 may move data in order to protect the data from read disturbs and/or as part of wear leveling operations. Those skilled in the art will appreciate other grooming operations which may require moving data within a storage device for data maintenance and space maintenance purposes. Metadata pertaining to the groomer's operation, such as the storage divisions selected for recovery, a validity map, storage division error rates, storage division wear levels, access patterns, and so on, may be maintained in the metadata 117 (e.g., in the storage metadata 316). The metadata 117 (e.g., storage metadata 316) may further comprise an index comprising associations between logical addresses provided by storage clients to storage units on the solid-state storage media 320 where valid data pertaining to the logical addresses is stored. As described above, the index may comprise a B-Tree comprising length and/or range encoded nodes indexed by logical address.
The cache controller 312 coordinates the exchange of data between storage clients (such as the clients 110a-c, the file system 210, or others) and the backing store 118 in relation to the solid-state storage device 114. In some embodiments, the cache controller 312 is responsible for maintaining an eviction policy that specifies when and how data is evicted from the solid-state storage device 114. In some embodiments, the eviction policy may be based, inter alia, upon cache eviction metadata, which may comprise data access patterns (e.g., whether the data is “hot,” “warm,” “cold” or the like). The eviction metadata may correspond to a last access time (least recently used), access frequency, a ratio of cache entry size to access time and/or access frequency, and so on.
In some embodiments, cache eviction metadata may comprise cache attributes and/or directives, which may be used to make eviction decisions. For example, a PIN cache attribute may cause the associated data to be retained in the cache, even if the data is “cold.” Cache directives (e.g., a FLUSH directive) may cause specified cache entries to be flushed from the solid-state storage device 114. Cache attributes and/or directives may be provided by a storage client (e.g., client 110a-c) via an interface. For example, the solid-state storage device 114 may provide an interface by which a client may specify data to retain in a cache (e.g., PIN attribute), to remove from the cache (e.g., BLACKLIST or FLUSH), or the like. In some embodiments, cache eviction decisions may be based on the storage metadata 316. As discussed above, the storage metadata 316 may comprise a forward map associating cache entries, each comprising a respective logical addresses (or logical address set or range) with respective storage units on the solid-state storage device 114. The index may be used maintain associations between access pattern metadata and a corresponding logical addresses (e.g. counts in a tree index may track how often certain nodes, edges, or leaves are visited and for which types of operations (read, write, trim)). These associations may be retained even when the underlying data is moved on the solid-state storage device (e.g., moved from a first storage unit to a second storage unit in response to a modification, during a storage division recovery operation, or the like). Accordingly, when making an eviction decisions, the cache controller 312 (or other entity) may access the storage metadata 316 to obtain relevant access pattern metadata.
In some embodiments, cache eviction policy may be based upon spatial and/or temporal proximity. Spatial proximity assumes that if a logical address is “hot,” spatially proximate logical addresses are likely to be “hot” as well (even if they haven't been accessed yet). The “spatial locality” between cache entries may be maintained in the storage metadata 316 (e.g., using the index or forward map). As discussed above, a forward index of the storage metadata 316 may associate logical addresses to physical storage units of the solid-state storage device 114. The logical addresses may be arbitrarily mapped to the storage units and, due to out-of-place writing and other sequential storage operations, the physical location of data may change over time. Therefore, it may be impractical to determine logical proximity using storage unit addressing along. The cache controller 312 (or other entity) may access the storage metadata 316 to identify spatial proximity for use in making eviction decisions (e.g., data that is spatially proximate to “hot” data in the logical address space may be less preferred for eviction, while data that is not spatially proximate to the “hot” data may be more preferred for eviction).
In some embodiments, the solid-state storage controller 115 is configured with a maximum amount of physical storage space available on the solid-state storage device 114 for use in support of caching. An administrator (or other entity) may provide this information when configuring the solid-state storage device 114. In other embodiments, the cache controller 312 discovers the maximum amount of storage space for itself (e.g., by identifying a physical storage capacity of the solid-state storage device 114). The maximum occupancy (and the thresholds derived therefore), may change over time. For example, certain storage locations of the solid-state storage device 114 may become unreliable and/or fail, which may reduce the physical storage capacity of the device 114. The maximum occupancy may, therefore, be reduced. In some embodiments, the reductions may occur transparently to the clients (e.g., the solid-state storage device 114 may continue operation and/or present the same logical address space to the clients). Alternatively, issue a notification informing the clients (e.g., clients 110a-c) of the reduction in storage capacity. Similarly, the physical storage capacity of the solid-state storage device 114 may increase (e.g., when additional solid-state storage media is added). The occupancy thresholds may be adjusted in accordance with the additional storage capacity.
Using the maximum occupancy information, the cache controller 312 may derive a maximum occupancy threshold that specifies the maximum amount of data that the cache management apparatus 130 may allow to be cached in the solid-state storage device 114. In some embodiments, the maximum occupancy threshold does not use all the useable physical capacity and instead may reserve a portion of the physical capacity for the solid-state storage device 114 to perform maintenance operations such as storage space recovery and/or replacement of worn out or unreliable storage units. In some embodiments, the cache controller 312 (or other entity) may dynamically determine a maximum occupancy threshold. The maximum occupancy threshold may be based upon the metadata 117, such as data access patterns, and the like. For example, if the solid-state storage device 114 comprises a large amount of “hot” data, which is likely to cause a subsequent cache miss if evicted, the occupancy threshold may be dynamically reduced. Conversely, if the solid-state storage device 114 comprises a large amount “cold” data that could be evicted with a low chance of a subsequent cache miss, the maximum occupancy threshold may be increased. Alternatively, or in addition, the solid-state storage device 114 may be configured to track cache performance in the metadata 117 to heuristically determine ideal thresholds. For example, a high cache miss rate for recently evicted cache entries may indicate that additional cache storage space is required (or that the cache eviction policy should be adjusted), and so on. Alternatively, if the cache miss rate is low, the thresholds may be reduced, and excess storage space of the solid-state storage device may be used for other purposes (e.g., as a cache for another backing store, as storage space, or the like).
The maximum occupancy threshold may be, for example, the amount of storage provided by the solid-state storage device 114 minus the amount of storage occupied by software or other data used by the solid-state storage device 114 and stored thereon. In other embodiments, the maximum occupancy threshold may be some percentage of the amount of storage available in the solid-state storage device 114. In some embodiments, the maximum occupancy threshold is determined by the solid-state storage device 114 based on performance criteria, such as a quality of service (QOS) attribute, or the like. The maximum occupancy threshold may be changed dynamically based on demand, job priorities, or administrative control. Other criteria may also be used to adjust, tune, or otherwise change the maximum occupancy threshold. In some embodiments, the solid-state storage device 114 may provide an interface whereby an occupancy threshold may be set. Alternatively, or in addition, the interface may allow a client to set one or more cache attributes and/or directives (e.g., a QOS, PIN, BLACKLIST, or the like), from which a maximum occupancy threshold may be determined.
In some embodiments, the cache controller 312 queries the metadata 117 to determine the amount of remaining physical storage capacity available on the solid-state storage device 114 to ensure that the solid-state storage device 114 does not exceed the maximum occupancy threshold. The cache controller 312 (or other module) may determine the remaining physical capacity using the metadata 117 prior to each insertion of data into the solid-state storage device 114. In other embodiments, the cache controller 312 may make this determination at varying frequencies depending on the amount of physical storage capacity that remains in the solid-state storage device 114. For example, if the solid-state storage device 114 has over half of its physical storage capacity available, the cache controller 312 may not query remaining physical storage capacity prior to each insertion; the cache controller 312 may instead wait until a certain number of cache insertions have occurred before making the determination. As the solid-state storage device 114 becomes increasingly full, the frequency of the queries to determine the remaining physical storage capacity may increase to the point where the cache controller 312 queries prior to each insertion into the solid-state storage device 114.
In some embodiments, once the solid-state storage device 114 reaches maximum occupancy as defined by the maximum occupancy threshold the cache controller 312 prevents any cache insertions into the solid-state storage device 114 until additional storage capacity becomes available. The cache controller 312 may allow reads to occur on data in the solid-state storage device 114 even when the solid-state storage device 114 reaches maximum occupancy. The solid-state storage device 114 may thus ensure that the solid-state storage device 114 does not overfill and provide protection for data stored within the solid-state storage device 114.
In some embodiments, the cache controller 312 also defines a target occupancy threshold that defines an amount of physical storage capacity used on the solid-state storage device 114. In some embodiments, the cache controller 312 may take actions to increase the rate at which data is evicted from the solid-state storage device 114 once the solid-state storage device 114 has reached the target occupancy threshold. In some embodiments, the cache management apparatus 130 only begins evicting data from the solid-state storage device 114 once the target occupancy threshold is reached. In other embodiments, the cache controller 312 may increase a priority of the backup engine 314 (e.g., configure the backup engine 314 to operate as a foreground process or otherwise increase the priority and/or the relative bandwidth allocated to the backup engine 314) when the target occupancy threshold is reached. In such embodiments, the backup engine 314 operates as a background process (for example, by giving the backup engine 314 a lower priority or a lower bandwidth allocation) until the target occupancy threshold is reached. The current physical storage capacity, thresholds, or the like may be maintained in the metadata 117 for access by the cache controller 312 and/or groomer 310.
As discussed above, the metadata 117 may comprise storage metadata 316, which may include a validity map. As used herein, a map may refer to any associative datastructure associating of a collection of unique keys with respective values. Looking up a unique key in the map returns the associated value. The validity map may associate storage units of the solid-state storage device 114 with a validity indicator that specifies that the data associated with the unit is either valid or invalid. The validity indicator may be a bit in the validity map, with one state representing valid data and the other state representing invalid data. The validity map 316 may be a bit map, a table, a list, or other data structure known in the art. For example, the validity map may be a data structure suited for managing a very large and potentially sparsely populated domain such as an address space, comprising representations of only valid or invalid storage units (and/or storage unit sets or ranges). For example, a validity map may comprise a sparse array of entries of each storage unit comprising valid data. An invalidity may be derived from the validity map (if a location is not in the validity map, then the location is invalid) or vice versa.
The unit or storage unit of the solid-state storage device 114 may refer to an addressable, physical storage unit on the solid-state storage media 320. In some embodiments, the metadata 117 may maintain a one to one ratio of logical block addresses (LBAs) to storage units; that is, one LBA may be associated with one storage unit or unit. Alternatively, or in addition, an LBA (or index node) may be associated with a length or range of storage units. In some embodiments, the storage units are physical blocks addressed by a physical block address (“PBA”) in the solid state storage media 320, which may comprise Flash memory or other non-volatile storage media. In another embodiment, the storage units may refer to cylinder head sector (“CHS”) locations on a hard disk drive. The nature of the storage unit may vary depending on the particular storage medium used by the solid-state storage device 114. The solid-state storage device 114 may simultaneously support multiple types of storage units.
Data associated with the storage units of the solid-state storage device 114 that are marked valid in the validity map 316 may be retained on the solid-state storage device 114, while data associated with invalid units (either marked and/or inferred) may be removed from the solid-state storage device 114. Invalid storage units may be candidates for removal in a storage division reclamation process (e.g., garbage collection). Valid data in a storage division to be recovered may be retained on the solid-state storage device 114 by writing the valid data to other storage divisions of the solid-state storage device 114.
The cache metadata 316 may comprise a discardability map, which may comprise a bit map, a table, a list, or other data structure known in the art. The discardability map may associate a storage unit of the solid-state storage device 114 with a discardability indicator that specifies whether the data associated with the unit is discardable or nondiscardable. The data associated with the unit is discardable if it has been stored in the backing store 118. If the data associated with the unit has not been stored in the backing store 118, it is typically nondiscardable.
In the
The cache management apparatus 130 may comprise a backup engine 314, which may be configured to identify cache entries to backup on the backing store 118 using the metadata 117. The backup engine 314 may cause storage units that are marked as valid in the storage metadata 316 and nondiscardable in the cache metadata 318 to be stored on the backing store 118. The backing store 118 stores the data such that it can be properly retrieved out of backing store 118. For example, the backup engine 314 may identify, using the metadata 117, a storage unit that is marked as valid and nondiscardable. The data associated with the unit may comprise the LBA and the data that corresponds to that LBA (as stored on the solid-state storage media 320). The backup engine 314 may read the data and cause the data to be stored at the backing store 118. As a result, future read requests for the LBA that are directed to the backing store 118 properly receive the data associated with the LBA. After determining that the data has been successfully stored on the backing store 118, the backup engine 314 may mark the storage unit as discardable.
In some embodiments, the backup engine 314 operates as a background process on the solid-state storage device 114. As such, the backup engine 314 may run with a low priority and only consume spare cycles and bandwidth on the solid-state storage device 114. In some embodiments, the cache controller 312 may increase the priority of the backup engine 314 or otherwise move the backup engine 314 into the foreground. In some embodiments, the cache controller 312 increases the priority of the backup engine 314 when the solid-state storage device 114 reaches the target occupancy threshold. The cache controller 312 may further increase the priority of the cache controller 312 as the solid-state storage device 114 approaches the maximum occupation threshold. In some embodiments, the cache controller 312 gives the backup engine 314 the highest priority once the solid-state storage device 114 reaches the maximum occupation threshold if there are no units marked as discardable in the solid-state storage device 114, or where there are insufficient units marked discardable in the solid-state storage device 114.
In some embodiments, the backup engine 314 begins with the units in the solid-state storage device 114 associated with the oldest data and proceeds chronologically through the units in the solid-state storage device 114 performing backup operations, as discussed in greater detail in connection with
The metadata 117 may comprise indicators (e.g., a log) to track the age of the data in the storage units of the solid-state storage device 114. The metadata may comprise indicators of when data was written to a particular storage unit and/or when data was read from a particular storage unit. U.S. patent application Ser. No. 11/952,091 to David Flynn, et al., filed Dec. 6, 2007, entitled “Apparatus, System, and Method for Managing Data Using a Data Pipeline,” which is hereby incorporated by reference, describes an approach to storing data serially within a solid state storage device. In such a device, the order in which data is stored in the storage units, and thus, the age of the data associated with those storage units, may in most cases, be derived from the sequence of the data in the storage device. In embodiments where the solid-state storage device 114 serially stores data as described in the above application, the cache controller 312 may derive information about the age of the data from the order in which the units are stored in the solid-state storage device 114 and/or sequence indicators associated a storage division of the storage unit. In such an embodiment, the cache management apparatus 130 may derive or estimate the age of data without the associated cost in memory and computation of maintaining a separate log to track when writes occurred.
In some embodiments, the metadata 117 may be used to track the number of reads on the data in a particular unit of the solid-state storage device 114. The metadata 117 may also comprise indicators of when those reads were issued. This metadata 117 may be maintained by the solid-state storage controller 115, the groomer 310, or other processes of the solid-state storage device 114. The cache controller 312 may access the metadata 117 to track when the last read for a particular storage unit was issued (determine the least recently used storage unit). Alternatively, or in addition, the cache controller 312 may use the metadata 117 to track storage unit access patterns and take the frequency or timing of accesses into consideration in determining whether a particular unit contains hot or cold data.
When the file system 210 has data to be stored in persistent storage (e.g., as a file), it typically determines one or more appropriate LBAs to use. The file system 210 may choose an LBA from a pool of available LBAs that have no data associated with them. For example, when the file system 210 writes data D, the file system selects an LBA, such as LBA 0 for the data D. The file system writes the data D to the solid state storage device acting as a solid-state storage device 114. In some embodiments, as discussed above, the file system 210 is interacting with the logical space presented by the solid-state storage device 114 and is unaware of the existence of a solid-state storage device 114.
The solid state storage device has a storage unit 402 that is initialized and ready for use. The solid-state storage device 114 may store metadata that associates the LBA with a storage unit on the solid-state storage device 114. The storage unit 402 may be addressable on the solid-state storage device 114; in some embodiments, the address of the storage unit 402 is a physical block address (“PBA”). In particular embodiments, there is a one-to-one mapping of LBAs to PBAs in the solid-state storage device 114.
In the metadata 117, the storage unit 402 may be associated with a validity indicator (represented by the “valid” column) and a discardability indicator (represented by the “discard” column). In some embodiments, the metadata may be “sparse,” comprising only entries for which valid data exists. Accordingly, the validity indicator may be inferred from inclusion in the metadata 117 (e.g., if an LBA and/or storage unit is in the metadata, the storage unit comprises valid data; otherwise, the storage unit comprises invalid data). In some embodiments, the storage unit 402 may be mapped to the validity indicator and the discardability indicator using a physical address (e.g., a PBA) for the storage unit 402 as the key (e.g., in a reverse index as described above).
When the solid state solid-state storage device 114 receives the write request from the file system 210, the solid state solid-state storage device 114 selects an initialized unit (such as storage unit 402) where data referenced in the request will be stored. The request may comprise the data, may comprise a portion of the data, and/or may be substantially free of the data. As described above, selection of a storage unit 402 may be made according to a sequential writing mechanism. For example, the metadata 117 may maintain one or more append points (not shown) referencing a “next” available storage unit. As data is written to the solid-state storage device 114, the append point may be sequentially incremented to a next storage unit of the solid-state storage device 114. This sequential writing may provide, inter alia, wear leveling benefits and may reduce write amplification.
The solid-state storage device 114 stores the data of the storage request (for example, logical address LBA 0 and the data D) in the selected storage unit 402 and updates the metadata 117 accordingly. Updating the metadata 117 may comprise updating an index 410 to associate the LBA 0 with a reference (e.g., physical address) of the storage unit 402. The updating may comprise adding a leaf node to an B-Tree, or other forward map datastructure. The entry may be indexed by the LBA (LBA 0) to enable fast and efficient LBA to storage unit associations.
Updating the metadata 117 may further comprise setting a validity and discardability indicators for the storage unit 402. These operations are represented in
As shown in
In response to the storage request, the solid-state storage device 114 may write the modified data D′ sequentially to the solid-state storage media. As discussed above, sequential storage may comprise storing the modified data D′ “out-of-place” at a different storage unit of the solid-state storage device. Sequential writing may provide, inter alia, wear leveling benefits, and may reduce write amplification (e.g., storing the modified data D′ on the same storage unit 402 may require erasing the storage division of the storage unit 402 and writing D′ along with other valid data on the storage division).
Storing the modified data D′ may comprise the solid-state storage device 114 selecting a “next” storage unit (storage unit 404) to store the modified data D′. The selection may be made according to the sequential storage mechanisms described above (e.g., may be a “next” storage unit at one or more append points). The metadata 117 pertaining to the storage division 402 may be updated to indicate that the data is no longer valid, and is discardable (or is in a do not care state). Metadata 117 associating logical addresses to storage units may be updated to associate the LBA 0 with the selected storage division 404. In addition, metadata pertaining to the selected storage unit 404 may be updated to indicate that the storage unit 404 comprises valid, nondiscardable data.
A groomer 310 may, at some point, look at the validity indicator for the unit 402, see that it is set to invalid, and perform a garbage collection operation (GC 1) on unit 402. The garbage collection operation may put the storage unit 402 back into an initialized state, at which point the unit 402 may again be used to store data.
In some embodiments, the backup operation occurs in a manner that is transparent to the file system; that is, the file system is unaware that the data D′ is now in both the solid-state storage device 114 and the hard disk drive. Similarly, the grooming operations, such as GC 1, may be performed transparently to the file system.
The update 2 also causes LBA 0 and the data D″ to be sequentially written to a new unit 406, which, as above, has the validity indicator set to valid. Because the data D″ has not been moved to the hard disk drive and is not otherwise in the hard disk drive, the discardability indicator is set to nondiscardable. The update 2 may further comprising modifying the index 410 of the metadata 117 to associate the LBA 0 with the storage unit 406. Once the backup engine 314 flushes the data D″ to the hard disk drive as shown (BU2), the discardability indicator is changed to discardable.
The file system may also issue a TRIM command to the solid state storage device acting as a solid-state storage device 114. A TRIM command allows a file system to tell a solid state storage device (such as a solid-state storage device 114) which data storage units (logical block addresses) are no longer in use. This ensures that the solid-state storage device 114 knows which storage units are in use and which are free and ensures that the solid state storage device does not retain data that the file system knows no longer needs and/or is invalid. The TRIM command specification is currently being standardized as part of the AT Attachment (“ATA”) interface standard led by Technical Committee T13.
As shown in
In some embodiments, the solid state storage device acting as a solid-state storage device 114 may not support TRIM. In some embodiments, both the solid-state storage device 114 and the backing store 118 may support TRIM. In such embodiments, the solid-state storage device 114 may execute the TRIM command and then pass the TRIM command to the backing store 118 as shown in
In a related embodiment, the solid-state storage device 114 may maintain an indicator that an LBA in the backing store 118 has been trimmed and utilize this information to correctly respond to a client file system even though the HDD does not support TRIM. Thus, for example, an operation that attempts to access data in the trimmed LBA on the hard disk drive is blocked by the solid-state storage device 114 even though the data for the trimmed LBA still exists on the hard disk drive. The solid-state storage device 114 may also overwrite or otherwise sanitize the data related to the LBA on the HDD to ensure that the data is not recoverable. The solid-state storage device 114 may perform these operations transparently from the perspective of the file system and the HDD.
In response to the cache miss, the solid-state storage device 114 redirects the read request to the hard disk drive, which accesses the data A at LBA 1. The data A is retrieved from the hard disk drive and returned to the file system 210. The cache controller 312 of the solid-state storage device 114 also executes an inject SSD event (inject 1) to store the data A in a selected storage unit 502 along with the logical address of LBA 1. The solid-state storage device 114 may update the metadata 117 to associate the LBA 1 with the storage unit 502 in the index 410.
As shown in
When the data A is stored in unit 502 as part of the inject 1 operation, the solid-state storage device 114 sets the validity indicator to valid and also sets the discardability indicator to discardable to reflect that the data A is properly stored in and associated with LBA 1 in the hard disk drive. Thus, the data A may be removed from the solid-state storage device 114 without data corruption. For example, a cache management apparatus operating on the solid-state storage device 114 may apply an eviction policy and determine that the data A should be evicted (e.g., based on access patterns, grooming considerations, or the like). Since the discardability indicator is set to “1” in the metadata 117, the eviction may occur without a backup on unit 502 prior to the eviction.
The solid-state storage device 114 may then garbage collect the unit 504 to set the unit 504 back to an initialized state. In other embodiments, the solid-state storage device 114 garbage collects the unit 504 directly as part of the eviction process instead of putting the unit 504 into the intermediate state shown at evict 1. In other embodiments, the groomer 310 evicts the data A′ as part of a grooming operation, even though the data A′ is marked as valid, as discussed in greater detail below.
In some embodiments, the clients 110a-c communicate with the solid-state storage device 114 through a file server 210. However, the disclosure is not limited in this regard, and the clients 110a-c could communicate with the solid-state storage device 114 directly and/or through other processes or entities, such as a database server, a web server, a storage protocol, or the like. The method 600 begins with a user making 602 the solid-state storage device 114 available in the system 100 and making 604 the backing store 118 available in the system 100.
In some embodiments, at step 606, the solid-state storage device may be configured to provide a logical address space having the same physical storage capacity as the backing store 118. If the backing store 118 has a larger physical storage capacity than the physical storage capacity of the solid-state storage device 114, the solid-state storage device 114 may be configured to support sparse addressing, such that the solid-state storage device 114 may present a larger logical address space than is physically available on the solid-state storage device 114. For example, the solid-state storage device 114 may be built with the a FUSION_SPARSEFS_SUPPORT flag set. In some embodiments, the configuration of step 606 may comprise the solid-state storage device presenting a logical address space that exceeds a physical storage capacity of both the solid-state storage device 114 and the backing store 118.
At step 608, with both the solid-state storage device 114 and the backing store 118 present in the system, the solid-state storage device 114 may be configured act as a cache for the backing store 118. The method 600 may also include setting a maximum occupancy threshold 610 that defines the maximum amount of physical space on the solid-state storage device 114 available for use as a cache. The maximum occupancy threshold may be provided to the solid-state storage device 114 by an external entity (e.g., an administrator), or the cache management apparatus 130 may determine an appropriate threshold itself.
At step 612 a target occupancy threshold for the solid-state storage device 114 may be set as described above (e.g., by an entity and/or by the solid-state storage device 114), and the method 600 ends. In some embodiments, the cache management apparatus 130 attempts to keep occupancy in the solid-state storage device 114 below the target occupancy threshold. The cache management apparatus 130 may do so by starting cache evictions when the solid-state storage device 114 exceeds the target occupancy threshold.
The method 700 begins with the solid-state storage device 114 receiving 702 a read request that includes an address for the data such as an LBA. At step 704, the method 700 determines whether data associated with the address is available on the solid-state storage device 114. The determination may be made using metadata 117 maintained by the solid-state storage device; in particular, an index comprising associations between logical addresses and associated storage units. If the address is present in the index, the method 700 may determine that the data is available, and the flow may continue to step 706; otherwise, the flow may continue to step 710.
The requested data may be accessed on one or more storage units at step 706. The storage units may be identified in the index as described above. At step 708, metadata pertaining to the storage operation may be updated to reflect the data access. The metadata may be used to identify access patters, such as “hot” and/or “cold” logical addresses, determine access times, and so on. In some embodiments, the storage metadata may be accessible to a groomer 310 to perform storage division recovery operations and/or to a cache management apparatus 130 to manage caching operations (e.g., determine eviction policy). The flow may then end.
At step 710, the method 700 redirects the read request to the address in the backing store 118. The cache controller 312 reads 712 the data and returns the data to the entity that generated the read request.
The cache controller 312 also inserts 714 the data into the solid-state storage device 114 and associates the data with the specified address (in an index or other metadata). In some embodiments, the data is written sequentially and/or in accordance with the write method described in connection with
The cache controller 312 may also update the metadata (indexes) such that future read requests retrieve the data from the solid-state storage device 114 until the data is removed from the solid-state storage device 114. In addition, the cache controller 312 updates the metadata 117 to indicate that the storage unit comprising the data is valid and discardable. Step 716 may further comprise the cache controller 312 updating logging metadata to indicate when the data was inserted into the solid-state storage device 114, and the method 700 ends. In other embodiments, the solid-state storage device 114 may store data sequentially and the age of the data is derived from its position within the solid-state storage device 114.
The cache management apparatus 130 may configure the solid-state storage device 114 as a write back cache where data is first written to the solid-state storage device 114 by the clients and later moved to the backing store 118. The method 800 begins with the cache management apparatus 130 receiving 802 a write request. The write request may reference data to be stored in persistent storage. The request may include the data and/or may be substantially free of the data. The write request may be associated with a logical address (e.g., an LBA).
The cache controller 312 determines 804 whether there is space available on the solid-state storage device 114. The cache controller 312 performs this check to ensure that the solid-state storage device 114 does not exceed its maximum capacity threshold. In some embodiments, the cache controller 312 performs the determination 804 prior to writing data into the solid-state storage device 114. In other embodiments, as described above, the frequency of the determination 804 varies.
The determination may comprise accessing metadata maintained by the solid-state storage device 114. The metadata may comprise an index associating logical block addresses to storage units. Storage units that are occupied by valid data may not be available for storage, while storage units that are not may be available for use. The metadata may further comprise a reverse index, which may associate storage units of the solid-state storage device 114 with respective validity/discardability indicators. Storage units that are invalid and/or discardable may be considered to be available for storage.
If there is space available in the solid-state storage device 114, the cache controller 312 writes 806 the data into a unit of the solid-state storage device 114 and marks the unit as nondiscardable. The writing of step 806 may comprise a sequential storage operation as described above. Sequential writing may comprise selecting storage units to store the data (e.g., using an append point or other sequential indicator), writing the data on the selected storage units, updating metadata of the solid-state storage device 114 to associate the logical address(es) of the write request with the selected storage units, and so on.
The backup engine 314 may then backup 808 the data to the backing store 118 such that the data is stored in the backing store 118. Backing up the storage units may comprise updating metadata of the solid-state storage device to indicate that the storage units are discardable. The backup engine 314 may operate at a selected priority level and, as such, may wait until an opportune time to back 808 the data to the backing store 118. Following step 808, the method 800 may end.
If there is no space for the data on the solid-state storage device 114, the cache management apparatus 130 pauses 810 the write request until the cache management apparatus 130 can make room for the data in the solid-state storage device 114. In other embodiments, the cache management apparatus 130 may cause the data to be written to the backing store 118, and perform a read operation that results in the data being inserted into the solid-state storage device 114 once sufficient space is generated in the solid-state storage device 114.
After pausing 810 the write request, the cache management apparatus 130 may determine 812 if there are storage units in the solid-state storage device 114 that are marked discardable and/or invalid (using the metadata 117). The check may further comprise evaluating a “do not care” state for the valid/discardable metadata. In some embodiments, nondiscardable storage units may be retained, even if marked as invalid (e.g., may be retained for a snap shot, backup, or the like). If there are discardable storage units, in some embodiments, the cache management device 130 finds 820 the units that are discardable and evicts them from the solid-state storage device 114 and returns to determine 804 if space is available on the solid-state storage device 114. The cache controller 312 may, for example, mark all units that are discardable as invalid. In another embodiment, the groomer 310 may groom the solid-state storage device 114 without preserving data in units that are marked discardable. If, after the evictions, there is room on the solid-state storage device 114, the data is stored in the solid-state storage device 114.
If there are no discardable storage units in the solid-state storage device 114, the cache management apparatus 130 may find 814 storage units in the cache that are marked nondiscardable in the discardability map and back up 816 the units to the backing store 118. In some embodiments, the cache management apparatus 130 does this by increasing a priority of the backup engine 314. Once the data is moved into the backing store 118, the cache management apparatus 130 updates the metadata 117 to mark 818 the units as discardable and finds 820 units that are discardable and evicts them from the solid-state storage device 114. The cache management apparatus 130 returns and again determines 804 if space is available on the solid-state storage device 114 to write the data.
a shows an exemplary embodiment of a backup engine 314 performing backup operations on storage units 914 through 926 of a solid-state storage device 114.
In some embodiments, the data in the units 914 through 926 is written sequentially; in such an embodiment, the units 914 through 926 have physical addresses on the solid-state storage device 114, which increase incrementally. For example, the unit 914 may have a physical address “0”, the unit 916 may have a physical address “1”, and so on. Where the data is written sequentially, the data is written at the lowest available physical address. When the file system begins writing the data “A” to the solid-state storage device 114, it writes the data “A” to unit 914 with physical address 0, and then writes the next data “B” to unit 916 with physical address 1, and so on.
In such embodiments, the physical organization of the data on the solid-state storage device 114 can be used as a log indicating which data is oldest and which is newest. For example, one might assume that the data written in unit 914 is older than the data written in unit 920. In other embodiments, the log is a data structure that specifies the age of the data in the solid-state storage device 114 absolutely (i.e., through the use of timestamps), relative to other data in the solid-state storage device 114, or both. In some embodiments, the physical organization of the data on the solid-state storage device 114 is used in conjunction with a log to determine the age of the units in the solid-state storage device 114.
As shown in
In some embodiments, the backup engine 314 first determines whether the data for the particular unit 914-926 is valid or invalid using the metadata (e.g., the reverse index or log), which may comprise a validity indicator 930a-g as described above. If the validity indicator 930a-g is set to invalid, the backup engine 314 may choose not to backup the data in the unit 914-926. For example, as shown in
In
Where the discardability indicator 940b is set to nondiscardable, the backup engine 314 may store the data B into the backing store 118. In some embodiments, the backup engine 314 initiates a write request directed to the backing store 118 for the data B. In some embodiments, as discussed previously, the data includes both the data and the LBA the data is associated with. In such embodiments, the backup engine 314 may extract the LBA from the data in the unit 916 and request that the data be written to that LBA in the backing store 118.
a shows that, initially, the data B in unit 916 is nondiscardable. When the data B is written to the backing store 118, the backup engine 314 updates the metadata to change the discardability indicator 940b to discardable (a “1” in
As discussed above, the backup engine 314 may determine that the data C in unit 918 is invalid and thus move on without backing up the data C. The backup engine 314 may then backup the data D in unit 920 and change the discardability indicator 940d to discardable.
In certain instances, the backup engine 314 may encounter a unit, such as unit 924, with data C that is marked as both valid and discardable. If the discardability indicator 940f is set to discardable (indicating that the data C has already been backed up) the backup engine 314 may move to the next unit without backing up the data C.
In some embodiments, the backup engine 314 continuously scans over the units 914-926 in the solid-state storage device 114. In other embodiments, the backup engine 314 completes an iteration through the units 914-926 in the solid-state storage device 114 and waits a period of time before beginning a second iteration. In some embodiments, the backup engine 314 acts as a background process on the solid-state storage device 114 and operates when spare cycles (whether CPU cycles, bus cycles, or others) are available. The backup engine 314 may also be given a set schedule; for example, an administrator may configure a solid-state storage device 114, such that the backup engine 314 performs backing operations at a certain time, such as during the night when traffic is low.
In some embodiments, the backup engine 314 will search through the validity and discardability indicators to determine where action is required and queue backup activities and commands to be executed when resources are available. For example, the backup engine 314 may go through the logic of
The backup engine 314 may prioritize backup operations based on metadata 117 associated with the storage units. For example, the metadata 117 may indicate that a storage unit comprising nondiscardable data has a high level of wear and/or a high error rate. Based upon the metadata 117, the backup engine 314 may increase the priority of the backup of the identified storage unit. Increasing the priority may comprise inserting the backup operation at the head of a queue, skipping other nondiscardable storage unit backups, or the like.
In some embodiments, as mentioned above, a cache controller 312 may increase the priority of the operations of the backup engine 314. For example, the cache controller 312 may give the backup engine 314 the highest priority if the solid-state storage device 114 is full and there are no units 914-926 (or not enough units 914-926) in the solid-state storage device 114 that are marked as discardable. In some embodiments, the cache controller 312 may incrementally increase the priority of the backup engine 314 as the solid-state storage device 114 begins to fill and the available storage space on the solid-state storage device 114 decreases.
In some embodiments, the backup engine 314 may be invoked at opportune times to supplement the chronological backup process. The backup engine 314 may be configured to proceed chronologically through units in the solid-state storage device 114 yet be available to be invoked out of sequence at opportune times. For example, the solid-state storage device 114 may invoke the backup engine 314 to back up data in one or more units when that data is being read and when the backup resources are available. The read may be initiated by a client accessing the data. The read may be initiated by the groomer 310 while performing garbage collection, scrubbing the data, recovering the data from a worn or failing block, or other grooming operation.
The read may be initiated by a reduplication process. The read might also be initiated by a RAID rebuild or a progressive RAID operation as described in U.S. patent application Ser. No. 11/952,115 to David Flynn, et. al., filed Dec. 6, 2007, entitled “Apparatus, System, and Method for Reliable, High Performance Storage of Data with Progressive RAID,” which application is hereby incorporated by reference in its entirety. The backup engine 314 may work in conjunction with other processes to reduce system overhead. One skilled in the art will recognize other processes that might be opportunistically used to initiate backup.
The backup engine 314 may also be invoked by the groomer 310. In some embodiments, the groomer 310 invokes the backup engine 314 to free storage divisions (e.g. erase blocks of the solid-state storage device 114). For example, the groomer 310 may access the metadata 117 to determine that a particular storage division contains a significant amount of nondiscardable data and that the data is cold. The groomer 310 may invoke the backup engine 314 to back up the units of the particular erase block and then groom the erase block, treating the newly discardable units as if they were invalid, and thus freeing space on the solid-state storage device 114.
The groomer 310 may perform the operations to perform the decisions listed in
The groomer 310 then accesses metadata 117 to identify 1006 storage units that contain valid data contain data that are discardable (e.g., using a discardability map). In some embodiments, those units that the groomer 310 determines are nondiscardable are moved 1010 to the new storage division as part of the grooming operation, and the method 1000 ends. The groomer 310 may initially move the valid and nondiscardable data into a buffer in preparation for moving the data to a new storage division. In some embodiments, the groomer 310 may also invoke the backup engine 314 to back up the data in these units, such that the units become discardable. In some embodiments, the groomer 310 may invoke the backup engine 314 and pause the grooming operation until the backup engine 314 moves the data into the backing store 118. The groomer 310 may then resume after the backup operation is complete and treat the newly discardable units as discardable data.
If the groomer 310 determines 1006 that the particular units are discardable, the groomer 310 may make a determination 1008 as to whether or not the data associated with the units that are discardable can be evicted (e.g., are cold). The determination may be made by a cache management apparatus 130 using, inter alia, metadata pertaining to storage operations performed on the solid-state storage device 114, such as access patterns, frequency, and so on. If the groomer 310 determines 1008 that the data is evictable, the groomer 310 does not move 1012 the data to a new erase block and the method 1000 ends. If the groomer 310 determines 1008 that the data is not cold, the groomer 310 moves 1010 the data to a new erase block and the method 1010 ends.
In some embodiments, the groomer 310 may assume that all units that have been backed up, and thus have a discardability indicator set to discardable, are cold. In such embodiments, all of the units that have a discardability indicator set to discardable are not moved during the grooming operation. Such an embodiment may result in the groomer 310 aggressively reclaiming space in the solid-state storage device 114 and may be particularly useful when the solid-state storage device 114 is at or is approaching maximum capacity. In addition, in embodiments where the backup engine 314 performs backup operations chronologically, it may be a reasonable assumption that units that are marked discardable contain cold data, at least from a write perspective, as the older data is more likely than newer data to be marked discardable.
In other embodiments, the groomer 310 may make a determination, using the metadata 117, as to whether the individual storage units are evictable (per a caching policy of the cache management apparatus 130, or the like). In some embodiments, the groomer 310 makes the evictability determination as to an individual storage unit and treats a set of contiguous units similarly. For example, a set of storage units associated with a logical address range (e.g., a particular node in index or B-Tree), may be evicted (or retained) as a group. The groomer 310 may implement the necessary instructions to determine evictability may consult another entity, such as the cache controller 312, as to which units contain cold data.
In some embodiments, the cache controller 312 may access metadata pertaining to storage operations to determine which storage units contain cold data and which contain warm data. In some embodiments, the metadata 117 comprises a read log (or timestamps) to tracks reads on storage units (or logical addresses), as well as writes to storage units, and the reads are also considered in determining whether or not the data in the unit is cold. The cache controller 312 may deem any data that is at the start of the log to be cold and, once the data is backed to the backing store 118, evictable.
In some embodiments, the cache controller 312 does not deem data to be cold until the target occupancy threshold is reached. Thus, while the used storage of the solid-state storage device 114 is below the target occupancy threshold, the cache controller 312 may choose not to evict any data from the solid-state storage device 114. In such embodiments, the backup engine 314 may still perform backup operations on data; however, the cache controller 312 may instruct the groomer 310 to treat data having the discardability indicator set to discardable as not cold until the target occupancy threshold is reached. The cache controller 312 may instruct the groomer 310 to treat the data that has the discardability indicator set to discardable once the target occupancy threshold is reached.
In some embodiments, the cache controller 312 may have multiple criteria for determining whether data is cold. For example, the cache controller 312 may deem any data that is older than a certain time period (for example, 4 hours) to be cold and a candidate for eviction regardless of the occupancy of the solid-state storage device 114. The cache controller 312 may also have new rules for designating data to be cold once the target occupancy threshold is reached, as described above.
In some embodiments, the cache controller 312 dynamically modifies the rules for what data is considered cold based on parameters. For example, the cache controller 312 may learn when there is a high write workload and prepare the solid-state storage device 114 by increasing the amount of data that is deemed cold (and thus the evictions), thereby freeing space on the solid-state storage device 114. The cache controller 312 may empty the solid-state storage device 114 at the end of every day such that each day begins with an empty solid-state storage device 114.
In some embodiments, the cache controller 312 may direct the backup engine 314 to back data of a storage unit to the backing store 118 only if the data of the unit is cold. The cache controller 312 may determine what data is cold using approaches similar to those described above. In such an embodiment, the groomer 310 may assume that all units having a discardability indicator set to discardable are cold.
In some embodiments, the data from more than one storage unit will be part of a logical data unit. These logical-to-storage unit associations may be maintained in the metadata 117 (e.g., in the forward and/or reverse indexes). For example, a particular file may be too large to store in a single storage unit; thus, the file may be stored in multiple storage units, with each storage unit in the solid-state storage device 114 corresponding to a unique storage unit LBA. Where the cache controller 312 has access to metadata describing this context information (such as a map) that specifies how the data in LBAs (and thus how the data in various units) are related, the backup engine 314 and the cache controller 312 may operate at a logical level to enable more efficient operation.
For example, the backup engine 314 may back up the data for a particular unit, determine whether there are additional storage units in the solid-state storage device 114 that hold that is part of the same logical data unit, and back up those storage units as well. Similarly, if the groomer 310 determines that a particular storage unit contains data that is valid, discardable, and cold, the groomer 310 may use the context information to determine which other units contain associated data. In some embodiments, the groomer 310 ensures that the storage units containing associated data are also backed up and sets the invalidity indicator to invalid. If the storage unit is not backed up, the groomer 310 may call the backup engine 314 to backup the data in the associated units.
In some embodiments, the solid-state storage device 114 may not have access to context information that allows the cache management apparatus 130 to definitively determine the associations between the data in various storage units of the solid-state storage device 114. In some embodiments, the cache controller 312 makes intelligent guesses as to which units belong together based on the metadata 117 (e.g., storage metadata 316, such as a sequential storage log). For example, the cache controller 312 may assume a set of storage units that were written contiguously constitute a logical data unit such as a file. Similarly, the cache controller 312 may assume that a set of units that are consistently read together constitute a logical data unit. The cache controller 312 may then direct the backup engine 314 and the groomer 310 to treat the set of units as a logical data unit as described above. In related embodiments, multiple files and datasets that are typically accessed together may logically be associated and operated upon by the solid-state storage device 114 as a group.
In some embodiments, once the cache controller 312 determines that a set of units hold data forming a logical data unit, the cache controller manages the backup engine 314 to ensure that the units are backed up together. These units need not be stored contiguously in the solid-state storage device 114. Many hard disk drives and other media in the backing store 118 read data more efficiently when the data is presented as a stream; as such, the cache controller 312 may cause the solid-state storage device 114 to stream the data from units forming a logical data unit to the hard disk drive. Similarly, the cache controller 312, in response to a cache miss, may receive data as a stream from the backing store 118.
As shown in
Although in the
Each cache entry in the index 1104 may associate one or more logical addresses with respective storage unit(s) of the solid-state storage device 114. For example, the entry 1114 may associate the logical address range 072-083 with storage units 95-106. Those of skill in the art will recognize that the physical addresses of the storage units may be stored in the index 1104 or an offset into one or more storage structures of the solid-state storage media. As discussed above, the storage units of the entry 1114 may change due to modifications to the underlying data (e.g., due to modify, recovery operations, or the like). The cache entries may further comprise and/or reference metadata 1119, which may comprise cache eviction metadata (e.g., cache entry access patterns), cache entry age, size, cache attributes, cache directive, storage metadata (e.g., address of the storage units 1117), and so on. Since the metadata 1119 is associated with the cache entries, which are indexed by logical address (e.g., address 1115), the metadata 1119 may remain associated with the cache entry 1114 regardless of changes to the location where the underlying data is stored (e.g., changes to the storage unit addresses 1117).
The index 1104 may be used to efficiently determine whether the solid-state storage device 114 comprises data referenced in a client storage request and/or to identify a location of the data on the device 114. For example, the solid-state storage device 114 may receive a storage request 1102 for a particular logical address. The request may be directed to a backing store, or may be directed to the solid-state storage device 114. The request 1102 may specify a logical address and a length (e.g., request 3 units of data starting from logical address 074). Alternatively, the request 1102 may comprise a set of logical addresses, logical address ranges (continuous or discontinuous), or the like.
The solid-state storage device 114 may determine whether a cache entry corresponding to the requested logical addresses is available in the index 1104 using a search operation as described above. If a cache entry comprising the requested logical addresses is found in the index 1104, the solid-state storage device 114 comprises the requested data (a cache hit); otherwise, a cache miss occurs.
In the example depicted in
If the cache miss references logical addresses proximate to an cache entry already in the index 1104 (e.g., references addresses 084-088), a merge operation may occur. In a merge operation, an existing cache entry may be “merged” with one or more other cache entries. For instance, a new cache entry for logical addresses 084-088 may be merged with entry 1114. The merge may comprise modifying the logical address 1115 of the cache entry to include the new addresses (e.g., 072-088) and/or may the storage units 1117 to include the storage units on which the data wrote stored.
A similar operation may occur when a partial cache miss occurs. A partial cache miss may occur when the index 1104 includes a portion, but not all, of the requested data. For example, the storage request 1102 may request addresses 078-088. In this case, the entry 1114 may reference a portion of the requested data, but not all of the data. To service the request, the solid-state storage unit may read the data available on the solid-state storage device (e.g., up to logical address 083), and access the remaining (e.g., missing) data from the backing store 118. The data read from the backing store may be stored on the solid-state storage device 114 and included in the index 1104 as a separate cache entry and/or within the cache entry 1114.
Although the cache entries in the index 1104 are shown as comprising references to storage unit addresses (e.g., addresses 1117), the disclosure is not limited in this regard. In other embodiments, the cache entries comprise reference or indirect links to the storage units. For example, the cache entries may include a storage unit identifier (or reference to the reverse index 1122).
The metadata of
The metadata 1136 may include, but is not limited to, metadata pertaining to sequential storage operations performed on the storage units. For example, the metadata 1136 may include a sequence indicator (e.g., timestamp), which may indicate a sequence in which the storage units were stored (e.g., as well as an “age” of the storage units and so on). The metadata may further include metadata pertaining to the storage media, such as wear level, reliability, error rate, disturb status, and so on. As described above, the storage metadata may be used to efficiently implement cache management operations, such as cache eviction decisions, cache data movement (e.g., if data of a “hot” cache entry is stored in a less reliable storage unit), and so on.
The reverse map 1122 may be organized according to storage divisions (e.g., erase blocks) of the solid-state storage device 114. In this example, the entry 1120 that corresponds to cache entry 1118 is located in erase block n 1138. Erase block n 1138 is preceded by erase block n−1 1140 and followed by erase block n+1 1142 (the contents of erase blocks n−1 and n+1 are not shown). An erase block may comprise a predetermined number of storage units. An erase block may refer to an area in the solid-state storage device 114 that is erased together in a storage recovery operation.
The validity and/or discardability metadata 1130 and 1132 may be shared by a cache management backup process and groomer (e.g., storage recovery and maintenance). A backup process may use the reverse map 1122 to iterate over the storage units in chronological order (from oldest to youngest). The backup process may backup storage units to the backing store that are valid and nondiscardable (as indicated by metadata 1130 and 1132). The backup process may prioritize backup using the metadata 1136; for example, the backup engine may increase the backup priority of data stored on storage units that are unreliable, have a high amount of wear, or the like as indicated by the metadata 1136.
Similarly, a groomer may use the metadata 1130, 1132, and/or 1136 to perform sequential storage operations. For example, when recovering a storage division, the groomer may be required to retain (e.g., rewrite) valid data to a next storage division. However, if the discardability indicator 1132 indicates that the data is discardable, the groomer may elect to note restore the data, which may allow the groomer to operate more efficiently (e.g., reduce write amplification). Alternatively, or in addition, the groomer may use the logical address 1134 associations to access the cache metadata (e.g., metadata 1119) to identify storage units that should be retained on the solid-state storage device during storage division recovery operations (retain discardable data if it is “hot” per the metadata 1119). The groomer may use the cache metadata (e.g., metadata 1119) for other purposes including, but not limited to: performing data refresh operations, performing data moves (e.g., move “hot” data from an unreliable storage unit to another storage unit), and so on.
Typically, the reverse map 1122 is arranged by storage division (e.g. erase blocks) or erase region so that traversing a section of the map associated with an erase block (e.g. erase block n 1138) to efficiently make storage recovery and/or refresh decisions. Arranging an index into a forward map 1104 that can be quickly searched to identify a physical address 1126 from a virtual address 1106 and a reverse map 1122 that can be quickly searched to identify valid data and quantity of valid data in an erase block 1138 is beneficial because the reverse map 1122 may be optimized for searches and storage recovery operations. One of skill in the art will recognize other benefits of an index with an index 1104 and a reverse map 1122.
At step 1210, a solid-state storage device may be initialized for use. The initialization may comprise allocating resources for the solid-state storage device (e.g., solid-state storage device 114), such as communications interfaces (e.g., bus, network, and so on), allocating volatile memory, accessing solid-state storage media, and so on. The initialization may further comprise initializing a shared metadata datastructure, such as the data structures described above in conjunction with
At step 1220, the solid-state storage device may service storage requests from one or more clients. The storage requests may be directed to the backing store (when the solid-state storage device is configured to cache data transparently) and/or may be directed to the solid-state storage device itself.
At step 1230, the solid-state storage device may maintain shared metadata (e.g., metadata 117) pertaining to the storage requests. The shared metadata may comprise storage metadata pertaining to storage operations performed on the solid-state storage device and/or to the solid-state storage media (e.g., metadata 316). Accordingly, the storage metadata may comprise associations between logical addresses and solid-state storage units, a validity map, storage reclamation metadata (e.g., wear level, reliability, append point, and so on), and the like.
The shared metadata maintained at step 1230 may further comprise cache metadata pertaining to a cache stored on the solid-state storage device (e.g., pertaining to cache operations performed on the solid-state storage device). The cache metadata may share the logical address to storage unit index of the storage metadata. The shared metadata may, therefore, comprise an index (e.g., index 1104) comprising a plurality of cache entries, each associating a logical address to one or more storage units of the solid-state storage device. The cache metadata may further comprise cache eviction metadata, such as data access patterns, cache attributes, cache directives, and so on (e.g., stored in metadata 1119 of
At step 1240, the method 1200 may access to the shared metadata to perform storage operations on the solid-state storage device. The sequential storage operations may include, but are not limited to, performing sequential (e.g., out-of-place) storage operations, maintaining logical-to-storage unit associations, maintaining storage unit validity indicators, selecting storage divisions for recovery (e.g., garbage collection), determining a current occupancy level of the solid-state storage device, selecting storage units to retain during a storage division recovery operations, and so on.
At step 1260, the method 1200 may access to the shared metadata to manage the cache stored on the solid-state storage device. Managing the cache may comprise, but is not limited to: backing up storage divisions to the backing store, maintaining cache eviction metadata, selecting cache entries for eviction, and so on.
At steps 1310, 1320, and 1330, the method 1300 may be initialized, service storage requests, and maintain shared metadata, as described above.
At step 1340, the method 1300 may access the shared metadata to manage sequential storage operations of the solid-state storage device.
At step 1342, the method 1300 may use the cache metadata to select a storage division for recovery. Step 1342 may comprise accessing cache metadata identifying storage units that have been stored on the backing store (e.g., discardable storage units). The selection of a storage division may be based on a threshold of valid data in the storage division that is to be retained on the storage device (e.g., rewritten to another storage location). The amount of valid data may be determined using the discardability indicators; if a storage unit is valid and discardable, the storage unit may be erased during the recovery operation without losing and/or corrupting data.
Step 1342 may further comprise accessing cache metadata to determine whether a storage unit is associated with a cache entry that should be retained on the solid-state storage device (e.g., according to an eviction policy or other metadata). For example, if a cache entry is “hot” (likely to result in a cache miss if removed), the storage unit may be retained during the recovery operation even though it is identified as discardable.
At step 1344, the method 1300 may implement a recovery operation on the selected storage division. The recovery operation may access the cache metadata (in the shared metadata) to perform the recovery operation. The recovery operation may comprise identifying nondiscardable storage units on the storage division (using the cache metadata) and causing the identified storage units to be stored on another storage division and/or causing the nondiscardable storage units to be stored on the backing store. The operation may further comprise accessing the metadata to identify storage units associated with cache entries that should be retained on the solid-state storage device (e.g., according to cache eviction metadata associated with the cache entries). The identified storage units may be stored on another storage division of the solid-state storage device.
At step 1346, the method 1300 may implement a data grooming operation to refresh and/or manage the storage units (the data grooming operation may also include storage recovery operations, as described above in conjunction with steps 1342 and 1344). The grooming operation may access the shared metadata pertaining to the cache operations. For example, the grooming operation may identify storage units associated with “hot” cache entries or cache entries having a particular cache attribute, such as a “PIN” attribute (e.g., using cache eviction metadata), storage units that are nondiscardable, or the like. The grooming operation may also consult storage metadata to determine whether the storage unit should be moved to another storage division based on the storage division wear level, reliability (e.g., error rate), read disturbs, age, and so on.
At step 1360, the method 1300 may access the shared metadata to manage the cache stored on the solid-state storage device.
At step 1362, the method 1300 may select storage units to store on the backing store. The selection may be based on a chronology of when the storage units were written (from oldest to newest), which may be determined using storage metadata pertaining to the sequence of storage operations performed on the solid-state storage device. In some embodiments, the selection may be based on the cache eviction metadata (e.g., access patterns, such as access frequency, and so on). Depending on the use case, cache that have a high rate of modification may not be selected for back up in order to optimize use of bandwidth and the latency costs of performing the backup. The selection of step 1362 may comprise accessing a validity indicator; data marked as invalid may not be backed up regardless of its age and/or discardability indicator.
In some embodiments, step 1362 may comprise prioritizing backup operations. The prioritization may be based on storage metadata, such as storage division wear level, reliability, occupancy, and so on. For example, a storage unit stored on an unreliable (or high wear) storage division (as indicated by the storage metadata) may be backed up before a storage unit stored on a more reliable (or less wear) storage division.
At step 1364, the method 1300 may prioritize storage division backup operations based upon inter alia, storage metadata. The storage metadata may be used to determine a current occupancy of the solid-state storage device. The priority of the backup operations may be determined based on the current occupancy. When the occupancy is low (as compared to an occupancy threshold), backup may be configured to operation in the background. When occupancy is high, the priority of the backup operations may be increased (e.g., moved into the foreground).
At step 1366, the method 1300 may identify cache entries for eviction using the shared metadata (e.g., storage metadata). As described above, cache entries may be associated with cache eviction metadata, which may comprise access patterns, cache attributes, cache directives, and so on. This metadata may pertain to the cache entry data. However, due to the nature of sequential storage operations (e.g., write out of place, etc.), associations between storage units and cache metadata may be lost (e.g., when the physical storage location changes). The storage metadata, however, may maintain logical-to-storage unit associations (cache entries of
The identification of step 1366 may access other types of storage metadata. For example, cache entries that are stored on high-wear and/or unreliable storage divisions may be preferred for eviction over similar entries stored on more reliable storage divisions, and so on.
At step 1368, the method 1300 may perform cache maintenance operations using the shared metadata (e.g., storage metadata). A cache maintenance operation may comprise identifying a “hot” cache entry and, using the storage metadata, causing the “hot” cache entry to be moved to another storage division of the solid-state storage device. The move may be performed for wear leveling purposes, to prevent excessive read disturbs, and the like. The selection may be made according to storage metadata that maintains relevant storage statistics (e.g., read count, age, and so on), maintained at the storage unit, storage division, virtual storage division, or other granularity level.
A cache maintenance operation of step 1368 may further comprise moving storage units that are of a certain age and/or are not backed up (e.g., are nondiscardable). Storage units that are nondiscardable are not stored on a backing store and, as such, if lost may cause data loss and/or corruption. These storage units may be moved from unreliable and/or high-wear storage divisions to prevent such loss. Similarly, nondiscardable storage units may be moved from proximity of “hot” storage entries to prevent errors caused by read disturbs, or other conditions.
At steps 1410, 1420, and 1430, the method 1400 may be initialized, service storage requests, and maintain shared metadata, as described above. The shared metadata may comprise a plurality of cache entries, each associating a logical address with one or more storage units comprising data of the cache entry. In some embodiments, the metadata may comprise a forward index, such as the B-Tree datastructure depicted in
At step 1440, the method 1400 may receive a request for data at a particular logical address.
At step 1450, the method 1400 may determine whether data corresponding to the particular logical address is stored on the solid-state storage device. In some embodiments, step 1450 may comprise searching the forward index of the shared metadata. If a cache entry comprising the particular logical address exists in the forward index, the method 1400 may determine that the data is available, and the flow may continue at step 1460; otherwise, the flow may continue at step 1470.
At step 1460, the storage request may be serviced. Servicing the storage request may comprise accessing the cache entry identified at step 1450 to determine the one or more storage units comprising the requested data. The storage units may be read, and the resulting data may be returned to the requester.
At step 1470, a cache miss may occur. A cache miss may be serviced by accessing data corresponding to the particular logical address in another data store, such as the backing store 118.
At step 1472, the data may be stored sequentially on the solid-state storage device. At step 1474, the metadata may be updated to indicate that the data is available on the solid-state storage device, and at step 1476, the data may be provided to the requester.
Step 1474 may comprise adding a new cache entry to the metadata. The new cache entry may associate the particular logical address with storage units comprising the data read from the backing store.
In some embodiments, step 1474 may comprise updating an existing cache entry. The particular logical address may be proximate to a logical address in the metadata. Accordingly, rather than adding a new, separate cache entry, the existing cache entry may be updated to include the particular logical address and to reference the storage units comprising the data read from the backing store. If the particular logical address falls between two or more existing cache entries (e.g., fills a gap between the cache entries), the updating of step 1474 may comprise merging the cache entries to include the new range (e.g., logical addresses of the original cache entries and the particular logical address), and to reference the original storage units and the storage units comprising the data read from the backing store.
In some embodiments, a cache entry may comprise a portion of the requested data, but not all of the data. In this case, at steps 1470 and 1472 the method 1400 may read and store only the missing portion of data. At step 1474, a new cache entry may be added to associate the particular logical address with the storage units comprising the data read from the backing store. Alternatively, step 1474 may comprise updating an existing cache entry as described above.
At steps 1510, 1520, and 1530, the method 1500 may be initialized, service storage requests, and maintain shared metadata as described above. At step 1540, the method 1500 may receive a storage request to remove data stored on the solid-state storage device. The request may reference the data by logical address.
At step 1550, the method 1500 may determine if data corresponding to the logical address is stored on the solid-state storage device as described above in conjunction with
At step 1560, the method may remove the data from the solid-state storage device. Removing the data may comprise removing a cache entry associated with the logical address from the metadata and/or invalidating the storage units of the cache entry. If the cache entry includes data that is to be retained on the solid-state storage device, the cache entry may be updated to de-associate the cache entry with the removed logical addresses and/or the invalidated storage units. The de-association may comprise splitting a cache entry as described above. Step 1560 may be performed in response to a TRIM command (e.g., issued through a storage interface provided by the method 1500) or the like.
At step 1570, the storage request may be passed through to a backing store, to allow the backing store to remove the data.
At steps 1610, 1620, and 1630, the method 1600 may be initialized, service storage requests, and maintain shared metadata as described above. The metadata maintained at step 1630 may comprise a forward index associating logical addresses with storage units of the solid-state storage device. The metadata may further comprise validity and/or discardability metadata.
At step 1640, the maximum occupancy of the solid-state storage device may be calculated using the shared metadata 117 (e.g., the storage metadata 316). The storage metadata 316 may include metadata pertaining to the solid-state storage media, such as error rates, reliability, wear levels, and so on. The storage metadata 316 may include indicators of storage divisions or storage units that are no longer usable (due to unacceptable error rates, or other conditions). The maximum occupancy of the solid-state storage device may be calculated by subtracting the unusable storage divisions from an original maximum occupancy (e.g., physical storage capacity, or a factor thereof). The physical storage space available to cache data may similarly change over time as storage units become unavailable due to wear or other error conditions. The solid-state storage device may continue to operate as a cache (presenting the same size or similar sized logical address space) despite the reduction of available physical storage capacity, due to the unavailable storage divisions and/or storage units. This reduction in available physical storage capacity may continue until the solid-state storage device no longer has available physical capacity to cache data and the input/output rates resemble those for accessing the backing store directly.
At step 1650, the current occupancy of the solid-state storage device may be calculated using the shared metadata 117. In some embodiments, the current occupancy may be calculated as comprising all valid data on the solid-state storage device. The valid data may be identified by traversing the forward index (e.g., index 1104 of
At step 1660, the method 1600 may provide the maximum occupancy and/or current occupancy to a requester, such as a backup engine, groomer, cache manager, or the like.
In some embodiments, the method 1600 may present a logical address space that exceeds the address space of the backing store. At step 1670, the method 1600 may service a storage request associated with a logical address that is outside of a logical address range of the backing store. The solid-state storage device may service the request may storing data on the solid-state storage device. However, the data may not be backed up to the backing store (since its address falls outside of the backing store's logical address). Accordingly, the storage units of the cache entry associated with the logical address may be marked as nondiscardable and may remain nondiscardable until additional logical or physical backing storage capacity is available. In some embodiments, step 1670 may comprise requesting additional backing storage and/or issuing a notification that the logical storage space has been exceeded.
Reference throughout this specification to “some embodiments” or an “embodiment” means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment. Thus, the appearance of the phrases “in one embodiment” or “in some embodiments” in various places throughout this specification are not necessarily all referring to the same embodiment. Furthermore, an “embodiment” may be a system, a device, an article of manufacture, a method, and/or a process.
The phrases “connected to,” “networked,” and “in communication with” “communicatively coupled to” refer to any form of interaction between two or more entities, including mechanical, electrical, magnetic, and electromagnetic interactions. Two components may be connected to each other even though they are not in direct physical contact with each other and even though there may be intermediary devices between the two components.
Some of the infrastructure that can be used with embodiments disclosed herein is already available, such as: general-purpose processors, programming tools and techniques, non-transitory, computer- and/or machine-readable storage media, and communications networks and interfaces. A computing device, such as an IED and/or monitoring device, may include a processor, such as a microprocessor, microcontroller, logic circuitry, or the like. The processor may include a special purpose processing device such as an ASIC, PAL, PLA, PLD, FPGA, or other processing means. The computing device may also include a computer-readable storage medium such as non-volatile memory, static RAM, dynamic RAM, ROM, CD-ROM, disk, tape, magnetic, optical, flash memory, or other non-transitory machine-readable storage medium.
The above description provides numerous specific details for a thorough understanding of the embodiments described herein. However, those of skill in the art will recognize that one or more of the specific details may be omitted, or other methods, components, or materials may be used. In some cases, operations are not shown or described in detail.
While specific embodiments and applications of the disclosure have been illustrated and described, it is to be understood that the disclosure is not limited to the precise configuration and components disclosed herein. Various modifications, changes, and variations apparent to those of skill in the art may be made in the arrangement, operation, and details of the methods and systems of the disclosure without departing from the spirit and scope of the disclosure.
This application claims priority to U.S. Provisional Application No. 61/240,573, entitled “Apparatus, System, and Method for an Improved Nonvolatile Caching Device,” filed on Sep. 8, 2009, to U.S. Provisional Application No. 61/240,966, entitled, “Apparatus, System, and Method for Allocating Solid-State Storage,” filed on Sep. 9, 2009, and to U.S. Provisional Application No. 61/373,271, entitled “Apparatus, System, and Method for Caching Data,” filed on Aug. 12, 2010, each of which is incorporated by reference in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
4571674 | Hartung | Feb 1986 | A |
5193184 | Belsan et al. | Mar 1993 | A |
5261068 | Gaskins et al. | Nov 1993 | A |
5291496 | Andaleon et al. | Mar 1994 | A |
5307497 | Feigenbaum et al. | Apr 1994 | A |
5313475 | Cromer et al. | May 1994 | A |
5325509 | Lautzenheiser | Jun 1994 | A |
5392427 | Barrett et al. | Feb 1995 | A |
5404485 | Ban | Apr 1995 | A |
5438671 | Miles | Aug 1995 | A |
5469555 | Ghosh et al. | Nov 1995 | A |
5499354 | Aschoff et al. | Mar 1996 | A |
5504882 | Chai et al. | Apr 1996 | A |
5535399 | Blitz et al. | Jul 1996 | A |
5551003 | Mattson et al. | Aug 1996 | A |
5553261 | Hasbun et al. | Sep 1996 | A |
5559988 | Durante et al. | Sep 1996 | A |
5586291 | Lasker et al. | Dec 1996 | A |
5594883 | Pricer | Jan 1997 | A |
5596736 | Kerns | Jan 1997 | A |
5598370 | Nijima et al. | Jan 1997 | A |
5603001 | Sukegawa et al. | Feb 1997 | A |
5651133 | Burkes | Jul 1997 | A |
5680579 | Young et al. | Oct 1997 | A |
5682497 | Robinson | Oct 1997 | A |
5682499 | Bakke et al. | Oct 1997 | A |
5701434 | Nakagawa | Dec 1997 | A |
5734861 | Cohn et al. | Mar 1998 | A |
5745792 | Jost | Apr 1998 | A |
5754563 | White | May 1998 | A |
5757567 | Hetzler et al. | May 1998 | A |
5787486 | Chin et al. | Jul 1998 | A |
5802602 | Rahman et al. | Sep 1998 | A |
5809527 | Cooper et al. | Sep 1998 | A |
5809543 | Byers et al. | Sep 1998 | A |
5845313 | Estakhri et al. | Dec 1998 | A |
5845329 | Onishi et al. | Dec 1998 | A |
5860083 | Sukegawa | Jan 1999 | A |
5907856 | Estakhri et al. | May 1999 | A |
5924113 | Estakhri et al. | Jul 1999 | A |
5930815 | Estakhri et al. | Jul 1999 | A |
5957158 | Volz et al. | Sep 1999 | A |
5960462 | Solomon et al. | Sep 1999 | A |
5961660 | Capps, Jr. et al. | Oct 1999 | A |
6000019 | Dykstal et al. | Dec 1999 | A |
6014724 | Jenett | Jan 2000 | A |
6073232 | Kroeker et al. | Jun 2000 | A |
6101601 | Matthews et al. | Aug 2000 | A |
6128695 | Estakhri et al. | Oct 2000 | A |
6141249 | Estakhri et al. | Oct 2000 | A |
6145051 | Estakhri et al. | Nov 2000 | A |
6170039 | Kishida | Jan 2001 | B1 |
6170047 | Dye | Jan 2001 | B1 |
6172906 | Estakhri et al. | Jan 2001 | B1 |
6173381 | Dye | Jan 2001 | B1 |
6185654 | Van Doren | Feb 2001 | B1 |
6209088 | Reneris | Mar 2001 | B1 |
6223308 | Estakhri et al. | Apr 2001 | B1 |
6230234 | Estakhri et al. | May 2001 | B1 |
6236593 | Hong et al. | May 2001 | B1 |
6240040 | Akaogi et al. | May 2001 | B1 |
6256642 | Krueger et al. | Jul 2001 | B1 |
6279069 | Robinson et al. | Aug 2001 | B1 |
6289413 | Rogers et al. | Sep 2001 | B1 |
6330688 | Brown | Dec 2001 | B1 |
6336174 | Li et al. | Jan 2002 | B1 |
6356986 | Solomon et al. | Mar 2002 | B1 |
6370631 | Dye | Apr 2002 | B1 |
6385710 | Goldman et al. | May 2002 | B1 |
6393513 | Estakhri et al. | May 2002 | B2 |
6404647 | Minne′ | Jun 2002 | B1 |
6412080 | Fleming et al. | Jun 2002 | B1 |
6418478 | Ignatius et al. | Jul 2002 | B1 |
6507911 | Langford | Jan 2003 | B1 |
6516380 | Kenchammana-Hoskote et al. | Feb 2003 | B2 |
6523102 | Dye et al. | Feb 2003 | B1 |
6564285 | Mills et al. | May 2003 | B1 |
6567889 | DeKoning et al. | May 2003 | B1 |
6587915 | Kim | Jul 2003 | B1 |
6601211 | Norman | Jul 2003 | B1 |
6625685 | Cho et al. | Sep 2003 | B1 |
6629112 | Shank | Sep 2003 | B1 |
6658438 | Moore et al. | Dec 2003 | B1 |
6671757 | Multer et al. | Dec 2003 | B1 |
6675349 | Chen | Jan 2004 | B1 |
6715027 | Kim et al. | Mar 2004 | B2 |
6715046 | Shoham et al. | Mar 2004 | B1 |
6728851 | Estakhri et al. | Apr 2004 | B1 |
6751155 | Gorobets | Jun 2004 | B2 |
6754774 | Gruner et al. | Jun 2004 | B2 |
6757800 | Estakhri et al. | Jun 2004 | B1 |
6775185 | Fujisawa et al. | Aug 2004 | B2 |
6779088 | Benveniste et al. | Aug 2004 | B1 |
6779094 | Selkirk et al. | Aug 2004 | B2 |
6785776 | Arimilli et al. | Aug 2004 | B2 |
6785785 | Piccirillo et al. | Aug 2004 | B2 |
6801979 | Estakhri et al. | Oct 2004 | B1 |
6804755 | Selkirk et al. | Oct 2004 | B2 |
6871257 | Conley | Mar 2005 | B2 |
6877076 | Cho et al. | Apr 2005 | B1 |
6880049 | Gruner et al. | Apr 2005 | B2 |
6883069 | Yoshida | Apr 2005 | B2 |
6883079 | Priborsky | Apr 2005 | B1 |
6910170 | Choi et al. | Jun 2005 | B2 |
6912537 | Selkirk et al. | Jun 2005 | B2 |
6912618 | Estakhri et al. | Jun 2005 | B2 |
6938133 | Johnson et al. | Aug 2005 | B2 |
6957158 | Hancock et al. | Oct 2005 | B1 |
6959369 | Ashton et al. | Oct 2005 | B1 |
6977599 | Widmer | Dec 2005 | B2 |
6978342 | Estakhri et al. | Dec 2005 | B1 |
6981070 | Luk et al. | Dec 2005 | B1 |
6996676 | Megiddo et al. | Feb 2006 | B2 |
7010652 | Piccirillo et al. | Mar 2006 | B2 |
7010662 | Aasheim et al. | Mar 2006 | B2 |
7013376 | Hooper, III | Mar 2006 | B2 |
7013379 | Testardi | Mar 2006 | B1 |
7035974 | Shang | Apr 2006 | B2 |
7043599 | Ware et al. | May 2006 | B1 |
7047366 | Ezra | May 2006 | B1 |
7050337 | Iwase et al. | May 2006 | B2 |
7058769 | Danilak | Jun 2006 | B1 |
7076599 | Aasheim et al. | Jul 2006 | B2 |
7076723 | Saliba | Jul 2006 | B2 |
7082495 | DeWhitt et al. | Jul 2006 | B2 |
7082512 | Aasheim et al. | Jul 2006 | B2 |
7085879 | Aasheim et al. | Aug 2006 | B2 |
7089391 | Geiger et al. | Aug 2006 | B2 |
7093101 | Aasheim et al. | Aug 2006 | B2 |
7096321 | Modha | Aug 2006 | B2 |
7111140 | Estakhri et al. | Sep 2006 | B2 |
7130956 | Rao | Oct 2006 | B2 |
7130957 | Rao | Oct 2006 | B2 |
7143228 | Iida et al. | Nov 2006 | B2 |
7149947 | MacLellan et al. | Dec 2006 | B1 |
7167953 | Megiddo | Jan 2007 | B2 |
7171536 | Chang et al. | Jan 2007 | B2 |
7173852 | Gorobets et al. | Feb 2007 | B2 |
7178081 | Lee et al. | Feb 2007 | B2 |
7181572 | Walmsley | Feb 2007 | B2 |
7194577 | Johnson et al. | Mar 2007 | B2 |
7194740 | Frank et al. | Mar 2007 | B1 |
7197657 | Tobias | Mar 2007 | B1 |
7203815 | Haswell | Apr 2007 | B2 |
7215580 | Gorobets | May 2007 | B2 |
7219238 | Saito et al. | May 2007 | B2 |
7234082 | Lai et al. | Jun 2007 | B2 |
7243203 | Scheuerlein | Jul 2007 | B2 |
7246179 | Camara et al. | Jul 2007 | B2 |
7254686 | Islam | Aug 2007 | B2 |
7275135 | Coulson | Sep 2007 | B2 |
7280536 | Testardi | Oct 2007 | B2 |
7293183 | Lee et al. | Nov 2007 | B2 |
7305520 | Voigt et al. | Dec 2007 | B2 |
7310711 | New et al. | Dec 2007 | B2 |
7328307 | Hoogterp | Feb 2008 | B2 |
7337201 | Yellin et al. | Feb 2008 | B1 |
7340558 | Lee et al. | Mar 2008 | B2 |
7340566 | Voth et al. | Mar 2008 | B2 |
7356651 | Liu et al. | Apr 2008 | B2 |
7360015 | Matthews et al. | Apr 2008 | B2 |
7360037 | Higaki et al. | Apr 2008 | B2 |
7366808 | Kano et al. | Apr 2008 | B2 |
7392365 | Selkirk et al. | Jun 2008 | B2 |
7395384 | Sinclair et al. | Jul 2008 | B2 |
7398348 | Moore et al. | Jul 2008 | B2 |
7424593 | Estakhri et al. | Sep 2008 | B2 |
7437510 | Rosenbluth et al. | Oct 2008 | B2 |
7441090 | Estakhri et al. | Oct 2008 | B2 |
7447847 | Louie et al. | Nov 2008 | B2 |
7450420 | Sinclair et al. | Nov 2008 | B2 |
7464221 | Nakamura et al. | Dec 2008 | B2 |
7480766 | Gorobets | Jan 2009 | B2 |
7487235 | Andrews et al. | Feb 2009 | B2 |
7487320 | Bansal et al. | Feb 2009 | B2 |
7500000 | Groves et al. | Mar 2009 | B2 |
7516267 | Coulson et al. | Apr 2009 | B2 |
7523249 | Estakhri et al. | Apr 2009 | B1 |
7526614 | van Riel | Apr 2009 | B2 |
7529905 | Sinclair | May 2009 | B2 |
7536491 | Kano et al. | May 2009 | B2 |
7549013 | Estakhri et al. | Jun 2009 | B2 |
7552271 | Sinclair et al. | Jun 2009 | B2 |
7580287 | Aritome | Aug 2009 | B2 |
7603532 | Rajan et al. | Oct 2009 | B2 |
7610348 | Kisley et al. | Oct 2009 | B2 |
7620773 | Nicholson et al. | Nov 2009 | B2 |
7631138 | Gonzales et al. | Dec 2009 | B2 |
7640390 | Iwamura et al. | Dec 2009 | B2 |
7644239 | Ergan et al. | Jan 2010 | B2 |
7660911 | McDaniel | Feb 2010 | B2 |
7660941 | Lee et al. | Feb 2010 | B2 |
7664239 | Groff et al. | Feb 2010 | B2 |
7669019 | Fujibayashi et al. | Feb 2010 | B2 |
7676625 | Cassiday et al. | Mar 2010 | B2 |
7721059 | Milly et al. | May 2010 | B2 |
7725628 | Phan et al. | May 2010 | B1 |
7831783 | Pandit et al. | Nov 2010 | B2 |
7853772 | Chang et al. | Dec 2010 | B2 |
7856528 | Frost et al. | Dec 2010 | B1 |
7873782 | Terry | Jan 2011 | B2 |
7873803 | Cheng | Jan 2011 | B2 |
7882305 | Moritoki | Feb 2011 | B2 |
7904647 | El-Batal et al. | Mar 2011 | B2 |
7913051 | Todd et al. | Mar 2011 | B1 |
7917803 | Stefanus et al. | Mar 2011 | B2 |
7941591 | Aviles | May 2011 | B2 |
7984230 | Nasu et al. | Jul 2011 | B2 |
8019938 | Flynn et al. | Sep 2011 | B2 |
8046526 | Yeh | Oct 2011 | B2 |
8055820 | Sebire | Nov 2011 | B2 |
8127103 | Kano et al. | Feb 2012 | B2 |
8135900 | Kunimatsu et al. | Mar 2012 | B2 |
8151082 | Flynn et al. | Apr 2012 | B2 |
8171204 | Chow et al. | May 2012 | B2 |
20020053009 | Selkirk et al. | May 2002 | A1 |
20020069317 | Chow | Jun 2002 | A1 |
20020069318 | Chow | Jun 2002 | A1 |
20020103819 | Duvillier | Aug 2002 | A1 |
20020161855 | Manczak et al. | Oct 2002 | A1 |
20020181134 | Bunker et al. | Dec 2002 | A1 |
20020194451 | Mukaida et al. | Dec 2002 | A1 |
20030061296 | Craddock et al. | Mar 2003 | A1 |
20030093741 | Argon et al. | May 2003 | A1 |
20030140051 | Fujiwara et al. | Jul 2003 | A1 |
20030145230 | Chiu et al. | Jul 2003 | A1 |
20030149753 | Lamb | Aug 2003 | A1 |
20030163630 | Aasheim et al. | Aug 2003 | A1 |
20030198084 | Fujisawa | Oct 2003 | A1 |
20040003002 | Adelmann | Jan 2004 | A1 |
20040093463 | Shang | May 2004 | A1 |
20040117586 | Estakhri et al. | Jun 2004 | A1 |
20040148360 | Mehra et al. | Jul 2004 | A1 |
20040186946 | Lee | Sep 2004 | A1 |
20040268359 | Hanes | Dec 2004 | A1 |
20050002263 | Iwase | Jan 2005 | A1 |
20050015539 | Horii | Jan 2005 | A1 |
20050027951 | Piccirillo et al. | Feb 2005 | A1 |
20050055497 | Estakhri et al. | Mar 2005 | A1 |
20050076107 | Goud et al. | Apr 2005 | A1 |
20050120177 | Black | Jun 2005 | A1 |
20050132259 | Emmot et al. | Jun 2005 | A1 |
20050141313 | Gorobets et al. | Jun 2005 | A1 |
20050144361 | Gonzales et al. | Jun 2005 | A1 |
20050149618 | Cheng | Jul 2005 | A1 |
20050149819 | Hwang | Jul 2005 | A1 |
20050177672 | Rao | Aug 2005 | A1 |
20050177687 | Rao | Aug 2005 | A1 |
20050193166 | Johnson et al. | Sep 2005 | A1 |
20050216653 | Aasheim et al. | Sep 2005 | A1 |
20050229090 | Shen et al. | Oct 2005 | A1 |
20050240713 | Wu et al. | Oct 2005 | A1 |
20050246510 | Retnamma et al. | Nov 2005 | A1 |
20050257017 | Yagi | Nov 2005 | A1 |
20050257213 | Chu et al. | Nov 2005 | A1 |
20050273476 | Wertheimer et al. | Dec 2005 | A1 |
20050276092 | Hansen et al. | Dec 2005 | A1 |
20060004955 | Ware et al. | Jan 2006 | A1 |
20060020744 | Sinclair et al. | Jan 2006 | A1 |
20060026339 | Rostampour | Feb 2006 | A1 |
20060059326 | Aasheim et al. | Mar 2006 | A1 |
20060075057 | Gildea et al. | Apr 2006 | A1 |
20060085626 | Roberson et al. | Apr 2006 | A1 |
20060090048 | Okumoto et al. | Apr 2006 | A1 |
20060106968 | Wooi Teoh | May 2006 | A1 |
20060129778 | Clark et al. | Jun 2006 | A1 |
20060136657 | Rudelic et al. | Jun 2006 | A1 |
20060143396 | Cabot | Jun 2006 | A1 |
20060149893 | Barfuss et al. | Jul 2006 | A1 |
20060152981 | Ryu | Jul 2006 | A1 |
20060179263 | Song et al. | Aug 2006 | A1 |
20060184722 | Sinclair | Aug 2006 | A1 |
20060190552 | Henze et al. | Aug 2006 | A1 |
20060224849 | Rezaul Islam et al. | Oct 2006 | A1 |
20060236061 | Koclaines | Oct 2006 | A1 |
20060248387 | Nicholson | Nov 2006 | A1 |
20060265636 | Hummler | Nov 2006 | A1 |
20070016699 | Minami | Jan 2007 | A1 |
20070033325 | Sinclair | Feb 2007 | A1 |
20070033326 | Sinclair | Feb 2007 | A1 |
20070033327 | Sinclair | Feb 2007 | A1 |
20070033362 | Sinclair | Feb 2007 | A1 |
20070043900 | Yun | Feb 2007 | A1 |
20070050571 | Nakamura et al. | Mar 2007 | A1 |
20070061508 | Zweighaft | Mar 2007 | A1 |
20070086260 | Sinclair | Apr 2007 | A1 |
20070088666 | Saito | Apr 2007 | A1 |
20070118676 | Kano et al. | May 2007 | A1 |
20070118713 | Guterman et al. | May 2007 | A1 |
20070124474 | Margulis | May 2007 | A1 |
20070124540 | van Riel | May 2007 | A1 |
20070136555 | Sinclair | Jun 2007 | A1 |
20070143532 | Gorobets et al. | Jun 2007 | A1 |
20070143560 | Gorobets | Jun 2007 | A1 |
20070143566 | Gorobets | Jun 2007 | A1 |
20070143567 | Gorobets et al. | Jun 2007 | A1 |
20070150689 | Pandit et al. | Jun 2007 | A1 |
20070156998 | Gorobets | Jul 2007 | A1 |
20070162830 | Stek et al. | Jul 2007 | A1 |
20070168698 | Coulson et al. | Jul 2007 | A1 |
20070198770 | Horii et al. | Aug 2007 | A1 |
20070204128 | Lee et al. | Aug 2007 | A1 |
20070204197 | Yokokawa | Aug 2007 | A1 |
20070208790 | Reuter et al. | Sep 2007 | A1 |
20070233455 | Zimmer et al. | Oct 2007 | A1 |
20070233937 | Coulson et al. | Oct 2007 | A1 |
20070245217 | Valle | Oct 2007 | A1 |
20070250660 | Gill et al. | Oct 2007 | A1 |
20070260608 | Hertzberg et al. | Nov 2007 | A1 |
20070261030 | Wadhwa | Nov 2007 | A1 |
20070263514 | Iwata et al. | Nov 2007 | A1 |
20070266037 | Terry et al. | Nov 2007 | A1 |
20070271468 | McKenney et al. | Nov 2007 | A1 |
20070271572 | Gupta et al. | Nov 2007 | A1 |
20070274150 | Gorobets | Nov 2007 | A1 |
20070276897 | Tameshige et al. | Nov 2007 | A1 |
20070300008 | Rogers et al. | Dec 2007 | A1 |
20080005465 | Matthews | Jan 2008 | A1 |
20080005748 | Mathew et al. | Jan 2008 | A1 |
20080010395 | Mylly et al. | Jan 2008 | A1 |
20080034153 | Lee et al. | Feb 2008 | A1 |
20080043769 | Hirai | Feb 2008 | A1 |
20080052377 | Light | Feb 2008 | A1 |
20080059752 | Serizawa | Mar 2008 | A1 |
20080091876 | Fujibayashi et al. | Apr 2008 | A1 |
20080098159 | Song et al. | Apr 2008 | A1 |
20080120469 | Kornegay et al. | May 2008 | A1 |
20080126507 | Wilkinson | May 2008 | A1 |
20080126700 | El-Batal et al. | May 2008 | A1 |
20080126852 | Brandyberry et al. | May 2008 | A1 |
20080133963 | Katano et al. | Jun 2008 | A1 |
20080137658 | Wang | Jun 2008 | A1 |
20080140737 | Garst et al. | Jun 2008 | A1 |
20080140819 | Bailey et al. | Jun 2008 | A1 |
20080183965 | Shiga et al. | Jul 2008 | A1 |
20080201535 | Hara | Aug 2008 | A1 |
20080205286 | Li et al. | Aug 2008 | A1 |
20080209090 | Kano et al. | Aug 2008 | A1 |
20080229045 | Qi | Sep 2008 | A1 |
20080235443 | Chow et al. | Sep 2008 | A1 |
20080243966 | Croisettier et al. | Oct 2008 | A1 |
20080263259 | Sadovsky et al. | Oct 2008 | A1 |
20080263305 | Shu et al. | Oct 2008 | A1 |
20080263569 | Shu et al. | Oct 2008 | A1 |
20080276040 | Moritoki | Nov 2008 | A1 |
20080294847 | Maruyama et al. | Nov 2008 | A1 |
20090043952 | Estakhri et al. | Feb 2009 | A1 |
20090070526 | Tetrick | Mar 2009 | A1 |
20090070541 | Yochai | Mar 2009 | A1 |
20090083478 | Kunimatsu et al. | Mar 2009 | A1 |
20090083485 | Cheng | Mar 2009 | A1 |
20090089485 | Yeh | Apr 2009 | A1 |
20090125650 | Sebire | May 2009 | A1 |
20090144496 | Kawaguchi | Jun 2009 | A1 |
20090150605 | Flynn et al. | Jun 2009 | A1 |
20090150641 | Flynn et al. | Jun 2009 | A1 |
20090157956 | Kano | Jun 2009 | A1 |
20090172250 | Allen et al. | Jul 2009 | A1 |
20090204750 | Estakhri et al. | Aug 2009 | A1 |
20090228637 | Moon et al. | Sep 2009 | A1 |
20090235017 | Estakhri et al. | Sep 2009 | A1 |
20090248763 | Rajan et al. | Oct 2009 | A1 |
20090276588 | Murase | Nov 2009 | A1 |
20090276654 | Butterworth et al. | Nov 2009 | A1 |
20090287887 | Matsuki et al. | Nov 2009 | A1 |
20090292861 | Kanevsky et al. | Nov 2009 | A1 |
20090300277 | Jeddeloh | Dec 2009 | A1 |
20090307424 | Galloway et al. | Dec 2009 | A1 |
20090313453 | Stefanus et al. | Dec 2009 | A1 |
20090327602 | Moore et al. | Dec 2009 | A1 |
20090327804 | Moshayedi | Dec 2009 | A1 |
20100005228 | Fukutomi et al. | Jan 2010 | A1 |
20100017556 | Chin et al. | Jan 2010 | A1 |
20100023674 | Aviles | Jan 2010 | A1 |
20100023676 | Moon et al. | Jan 2010 | A1 |
20100023682 | Lee et al. | Jan 2010 | A1 |
20100030946 | Kano et al. | Feb 2010 | A1 |
20100076936 | Rajan | Mar 2010 | A1 |
20100077194 | Zhao et al. | Mar 2010 | A1 |
20100095059 | Kisley et al. | Apr 2010 | A1 |
20100169542 | Sinclair | Jul 2010 | A1 |
20100174870 | Banerjee | Jul 2010 | A1 |
20100205231 | Cousins | Aug 2010 | A1 |
20100205335 | Phan et al. | Aug 2010 | A1 |
20100211737 | Flynn et al. | Aug 2010 | A1 |
20100235597 | Arakawa | Sep 2010 | A1 |
20100262738 | Swing et al. | Oct 2010 | A1 |
20100262740 | Borchers et al. | Oct 2010 | A1 |
20100262757 | Sprinkle et al. | Oct 2010 | A1 |
20100262758 | Swing et al. | Oct 2010 | A1 |
20100262759 | Borchers et al. | Oct 2010 | A1 |
20100262760 | Swing et al. | Oct 2010 | A1 |
20100262761 | Borchers et al. | Oct 2010 | A1 |
20100262762 | Borchers et al. | Oct 2010 | A1 |
20100262766 | Sprinkle et al. | Oct 2010 | A1 |
20100262767 | Borchers et al. | Oct 2010 | A1 |
20100262773 | Borchers et al. | Oct 2010 | A1 |
20100262894 | Swing et al. | Oct 2010 | A1 |
20100262979 | Borchers et al. | Oct 2010 | A1 |
20110016260 | Lomelino et al. | Jan 2011 | A1 |
20110022819 | Post et al. | Jan 2011 | A1 |
20110055458 | Kuehne | Mar 2011 | A1 |
20110258391 | Atkisson et al. | Oct 2011 | A1 |
20110258512 | Flynn et al. | Oct 2011 | A1 |
Number | Date | Country |
---|---|---|
1771495 | May 2006 | CN |
1100001 | May 2001 | EP |
1418502 | May 2004 | EP |
1418502 | May 2004 | EP |
1814039 | Aug 2007 | EP |
1814039 | Mar 2009 | EP |
0123416 | Sep 2001 | GB |
8153014 | Jun 1996 | JP |
2000259525 | Sep 2000 | JP |
4242848 | Jan 2009 | JP |
2009122850 | Jun 2009 | JP |
WO9419746 | Sep 1994 | WO |
WO9419746 | Sep 1994 | WO |
WO9518407 | Jul 1995 | WO |
WO9518407 | Jul 1995 | WO |
WO9612225 | Apr 1996 | WO |
WO9612225 | Apr 1996 | WO |
WO0131512 | May 2001 | WO |
WO0201365 | Jan 2002 | WO |
WO0201365 | Jan 2002 | WO |
WO2004099989 | Nov 2004 | WO |
WO2005103878 | Nov 2005 | WO |
WO2006062511 | Jun 2006 | WO |
WO2006065626 | Jun 2006 | WO |
WO2008130799 | Mar 2008 | WO |
WO2008070173 | Jun 2008 | WO |
WO2008073421 | Jun 2008 | WO |
WO2011106394 | Sep 2011 | WO |
Entry |
---|
EEEL 6892, “Virtual Computers, Lecture 18”, http://www.acis.ufl.edu/˜ming/lectures/eel6892lec18.pdf, visited Mar. 1, 2010. |
Chen, “Hystor: Making SSDs the Survival of the Fittest in High Performance Storage Systems”, Proceedings of 25th ACM International Conference on Supercomputing (ICS 2011), Tucson, Arizona, May 31-Jun. 4, 2011. |
Cambridge Computer Services, Inc., 2008 ©, http://www.clusteredstorage.com/clustered—storage—solutions.html, visited Feb. 16, 2010. |
Annonymous, “Method to Improve Reliability to SSD Arrays”, http://ip.com, IP.com No. IPCOM000189338D, Nov. 5, 2009. |
Stow: Spatially and Temporally Optimized Write Caching Algorithm, Gill, Ko, Debnath, Belluomini, http://www.usenix.org/event/usenix09/tech/slides/gill.pdf. Presented Usenix Nov. 2009. Accessed Apr. 2010. |
Leventhal, “Flash Storage Memory, Communications of the ACM”, http://arstechnica.com/hardware/news/2009/09/intels-new-flash-tech-to-bring-back-turbo-memory-for-real, ACM, vol. 51, No. 7, p. 47-51, Jul. 2008. |
Stokes, Intel's New Flash Tech to Bring Back Turbo Memory, for Real, Sep. 2009, http://arstechnica.com/hardware/news/2009/09/intels-new-flash-tech-to-bring-back-turbo-memory-for-real.ars. |
Wikipedia, “Adaptive Replacement Cache”, last modified Jun. 25, 2012, http://en.wikipedia.org/wiki/Adaptive—l replacement—cache. |
Coburn, “NV-Heaps: Making Persistent Objects Fast and Safe with Next-Generation, Non-Volatile Memories”, ACM 978-1-4503-0266-1/11/0, published Mar. 5, 2011. |
Volos, “Mnemosyne: Lightweight Persistent Memory”, ACM 978-1-4503-0266-1/11/03, published Mar. 5, 2011. |
CIPO, “Office Action for CN Application No. 200780050973.4”, issued Jan. 26, 2011. |
EPO, “Office Action for EP Application No. 07867661.6”, issued Oct. 6, 2011. |
Kawaguchi, “A Flash-Memory Based File System”, Advanced Research Laboratory, Hitachi, Ltd. (1995). |
Plank, A Tutorial on Reed-Solomon Coding for Fault Tolerance in RAID-like System, Department of Computer Science, University of Tennessee, 1999. |
Johnson, “An Introduction to Block Device Drivers”, 1995. |
ASINE®, “ASPMC-660 Rugged IDE Flash Drive PMC Module”, http://www.asinegroup.com/products/aspmc660.html, copyright 2002. |
AGIGATECH, “Bulletproof Memory for RAID Servers, Parts 1-3”, http://agigatech.com/blog/bulletproof-memory-for-raid-servers-part-1/, 2009. |
Casey, “Disk I/O Performance Scaling: the File Caching Solution”, Solid Data Systems, Inc., Paper #528, Mar. 2001, pp. 1-8. |
Casey, Solid State File-Caching for Performance and Scalability, SolidData, Quarter 1 2000, http://www.storagesearch.com/3dram.html, accessed May 15, 2012. |
Bandulet, “Object-Based Storage Devices”, Oracle®, http://developers.sun.com/solaris/articles/osd.html, Jul. 2007. |
Van Hensbergen, “Dynamic Policy Disk Caching for Storage Networking”, IBM Research Division, RC24123 (W0611-189), Nov. 2006. |
Wu, “eNVy: A Non-Volatile, Main Memory Storage System”, ACM 0-89791-660-3/94/0010, 1994. |
Novell, “File System Primer”, http://wiki.novell.com/index.php/File—System—Primer, 2006, visited Oct. 18, 2006. |
Gill, “WOW: Wise Ordering for Writes—Combining Spatial and Temporal Locality in Non-Volatile Caches”, IBM, Fast '05: 4th USENIX Conference on File and Storage Technologies, 2005. |
Microsoft, “How NTFS Works”, http://technet.microsoft.com/en-us/library/cc781134(WS.10).aspx , 2003. |
Slashdot, “Slashdot Hynix 48-GB Flash MCP”, “News for Nerds, stuff that matters”, http://hardware.slashdot.org/article.pl?sid=07/09/06/146218&from=rss, Jul. 9, 2006. |
Dan, “Implementing MLC NAND Flash for Cost-Effective, High-Capacity Memory”, M-Systems, White Paper, 91-SR-014-02-8L, Rev 1.1, Sep. 2003. |
INCITS, Information Technology—SCSI Object-Based Storage Device Commands (OSD), Reference No. ISO/IEC 14776-391:200x, Revision 10, 2004. |
Samsung Electronics, “Introduction to Samsungs Linux Flash File System—RFS Application Note”, Nov. 2006, Version 1.0. |
Morgenstern, David, “Is There a Flash Memory RAID in your Future?”, http://www.eweek.com—eWeek, Ziff Davis enterprise Holdings Inc., Nov. 8, 2006. |
Woodhouse, “JFFS: The Journaling Flash File System”, Ottawa Linux Symposium, http://sources.redhat.com/jffs2/jffs2.pdf, Jul. 2001, visited May 15, 2012. |
SOLIDATA, Best Practices Guide, Sybase: Maximizing Performance through Solid State File-Caching, May 2000. |
Anonymous, “Method for Fault Tolerance in Nonvolatile Storage”, ip.com, IP.com No. IPCOM000042269D, 2005. |
MICRON Technology, Inc., “NAND Flash 101: An Introduction to NAND Flash and How to Design It In to Your Next Product”, TN-29-19, 2006. |
Mesnier, “Object-Based Storage”, IEEE Communications Magazine, Aug. 2003. |
PCT, “International Preliminary Report on Patentability for PCT/US2007/025049”, mailed Mar. 11, 2009. |
PCT, “International Search Report and Written Opinion for PCT/US2007/025049”, mailed May 14, 2008. |
ARI, Performance Boosting and Workload Isolation in Storage Area Networks with SanCache, Hewlett Packard Laboratories, Proceedings of the 23rd IEEE / 14th NASA Goddard Conference on Mass Storage Systems and Technologies (MSST 2006), May 2006, pp. 263-273. |
Pivot 3, “Pivot3 RAIGE Storage Cluster”, Technology Overview, White Paper, Jun. 2007. |
Casey, “San Cache: SSD In the San”, Storage Inc., 2000. |
MICRON Technology, Inc., “Technical Note: Design and Use Considerations for NAND Flash Memory” (TN-29-17), 2006. |
MICRON Technology, Inc., “Technical Note: Hamming Codes for NAND Flash Memory Devices” (TN-29-08), 2005. |
Seagate, “The Advantages of Object-Based Storage—Secure, Scalable, Dynamic Storage Devices”, Seagate Research Technology Paper TP-536, Apr. 2005. |
Rosenblum, “The Design and Implementation of a Log-Structured File System,” 1992. |
WACHA, “Improving RAID-Based Storage Systems with Flash Memory”, First Annual ISSDM/SRL Research Symposium, Oct. 2009. |
Wang, “OGFS: A File System for Object-based Storage Devices”, 21st IEEE/12th NASA Goddard Conference on Mass Storage Systems and Technologies, Apr. 2004. |
DataDirect Networks, White Paper: S2A9550 Overview, © 2007. |
Microsoft Corporation, Windows PC Accelerators, XP-002476842, Nov. 2006. |
Yerrick, “Block Device”, http://www.pineight.com/ds/block/, 2005-2006, visited Mar. 1, 2010. |
Actel, “Actel Fusion FPGAs Supporting Intelligent Peripheral Management Interface (IPMI) Applications,” http://www.actel.com/documents/Fusion—IPMI—AN.pdf, Oct. 1, 2006, visited Mar. 11, 2010. |
Barrall et al., U.S. Appl. No. 60/625,495, “Dynamically EXpandable and Contractible Fault-Tolerant Storage System Permitting Variously Sized Storage Devices and Method,” filed Nov. 5, 2004. |
Barrall et al., U.S. Appl. No. 60/718,768, “Dynamically Adaptable Fault-Tolerant Storage System,” filed Sep. 20, 2005. |
Bitmicro, “BiTMICRO Introduces E-Disk PMC Flash Disk Module at Military & Aerospace Electronics East 2004,” http://www. bitmicro.com/press.sub, published May 18, 2004, visited Mar. 8, 2011. |
Brandon, Jr., “Sparse Matrices in CS Education,” Journal of Computing Sciences in Colleges, vol. 24 Issue 5, May 2009, pp. 93-98. |
ELNEC, “NAND Flash Memories and Programming NAND Flash Memories Using ELNEC Device Programmers, Application Note,” published Mar. 1, 2007. |
European Patent Office, Office Action, EP Application No. 07865345.8, issued Nov. 17, 2010. |
Gal, “A Transactional Flash File System for Microcontrollers,” 2005 USENIX Annual Technical Conference, published Apr. 10, 2009. |
Garfinkel, “One Big File Is Not Enough: A Critical Evaluation of the Dominant Free-Space Sanitization Technique,” 6th Workshop on Privacy Enhancing Technologies. Cambridge, United Kingdom, published Jun. 1, 2006. |
Gutmann, “Secure Deletion of Data from Magnetic and Solid-State Memory”, UseniX, 14 pages, San Jose, CA, published Jul. 1, 1996. |
Huffman, “Non-Volatile Memory Host Controller Interface,” Apr. 14, 2008, 65 pgs. |
Hynix Semiconductor, Intel Corporation, Micron Technology, Inc. Phison Electronics Corp., Sony Corporation, Spansion, STMicroelectronics, “Open NAND Flash Interface Specification,” Revision 2.0, Feb. 27, 2008. |
Information Technology, “SCSI Object-Based Storage Device Commands,” 2 (OSD-2), Project T10/1729-D, Revision 4, published Jul. 30, 2004, printed Jul. 24, 2008. |
Intel, “Non-Volatile Memory Host Controller Interface (NVMHCI) 1.0,” Apr. 14, 2008. |
Johnson, “An Introduction to Block Device Drivers,” Jan. 1, 1995. |
Porter, “Operating System Transactions,” ACM 978-1-60558-752-3/09/10, published Oct. 1, 2009. |
Probert, “Windows Kernel Internals Cache Manager,” Microsoft Corporation, http://www.i.u.-tokyo.co.jp/edu/training/ss/lecture/new-documents/Lectures/15- CacheManager/CacheManager.pdf, printed May 15, 2010. |
Ranaweera, 05-270RO, SAT: Write Same (10) command (41h), T10/05, Jul. 7, 2005, www.t10.org/ftp/t10/document.05/05-270r0.pdf, last visited Apr. 11, 2013. |
Sears, “Stasis: FleXible Transactional Storage,” OSDI '06: 7th USENIX Symposium on Operating Systems Design and Implementation, published Nov. 6, 2006. |
Seltzer, “File System Performance and Transaction Support”, University of California at Berkeley, published Jan. 1, 1992. |
Seltzer, “Transaction Support in a Log-Structured File System”, Harvard University Division of Applied Sciences, published Jan. 1, 1993 (Chapter 5, pp. 52-69). |
Seltzer, “Transaction Support in Read Optimized and Write Optimized File Systems,” Proceedings of the 16th VLDB Conference, Brisbane, Australia, published Jan. 1, 1990. |
Shimpi, Anand, The SSD Anthology: Understanding SSDs and New Drives from OCZ, Mar. 18, 2009, 69 pgs. |
Shu, “Data Set Management Commands Proposals for ATA8-ACS2,” Dec. 12, 2007, http://www.t13.org.Documents/UploadedDocuments/docs2008/e07154r6-Data—Set—Management—Proposal—for—ATA-ACS2.pdf, printed Apr. 5, 2010. |
Spansion, “Data Management Software (DMS) for AMD Simultaneous Read/Write Flash Memory Devices”, published Jul. 7, 2003. |
Spillane “Enabling Transactional File Access via Lightweight Kernel EXtensions”, Stony Brook University, IBM T. J. Watson Research Center, published Feb. 25, 2009. |
State Intellectual Property Office, Office Action, CN Application No. 200780050970.0, issued Oct. 28, 2010. |
State Intellectual Property Office, Office Action, CN Application No. 200780051020.X, issued Nov. 11, 2010. |
State Intellectual Property Office, Office Action, CN Application No. 200780050983.8, issued May 18, 2011. |
State Intellectual Property Office, Office Action, CN Application No. 200780051020.X, issued Jul. 6, 2011. |
State Intellectual Property Office, Office Action, CN Application No. 200780051020.X, issued Nov. 7, 2011. |
State Intellectual Property Office, Office Action, CN Application No. 200780050970.0, issued Jan. 5, 2012. |
Terry et al., U.S. Appl. No. 60/797,127, “Filesystem-aware Block Storage System, Apparatus, and Method,” filed May 3, 2006. |
USPTO, Notice of Allowance for U.S. Appl. No. 12/986,117, Mailed Apr. 4, 2013. |
USPTO, Notice of Allowance for U.S. Appl. No. 12/986,117 Mailed Jun. 5, 2013. |
USPTO, Office Action for U.S. Appl. No. 12/879,004 mailed Feb. 25, 2013. |
USPTO, Office Action for U.S. Appl. No. 13/607,486 mailed Jan. 10, 2013. |
USPTO, Office Action for U.S. Appl. No. 11/952,113, mailed Dec. 15, 2010. |
USPTO, Office Action for U.S. Appl. No. 12/711,113, mailed Jun. 6, 2012. |
USPTO, Office Action for U.S. Appl. No. 12/711,113, mailed Nov. 23, 2012. |
USPTO, Office Action for U.S. Appl. No. 13,607,486 mailed May 2, 2013. |
USPTO, Office Action for U.S. Appl. No. 13/118,237 mailed Apr. 22, 2013. |
USPTO, Notice of Allowance, U.S. Appl. No. 11/952,109, issued May 1, 2013. |
USPTO, Office Action, U.S. Appl. No. 11/952,109, mailed Nov. 29, 2011. |
WIPO, International Preliminary Report of Patentability for PCT/US2007/086691, mailed Feb. 16, 2009. |
WIPO, International Preliminary Report on Patentability for PCT/US2007/086688, mailed Mar. 16, 2009. |
WIPO, International Preliminary Report on Patentability for PCT/US2007/086701, mailed Mar. 16, 2009. |
WIPO, International Preliminary Report on Patentability for PCT/US2007/086687, mailed Mar. 18, 2009. |
WIPO, International Preliminary Report on Patentability for PCT/US2007/025048, mailed Jun. 10, 2009. |
WIPO, International Preliminary Report on Patentability for PCT/US2010/048325, mailed Mar. 13, 2012. |
WIPO, International Search Report and Written Opinion for PCT/US2007/025048, mailed May 27, 2008. |
WIPO, International Search Report and Written Opinion for PCT/US2007/086701, mailed Jun. 5, 2008. |
WIPO, International Search Report and Written Opinion for PCT/US2007/086687, mailed Sep. 5, 2008. |
WIPO, International Search Report and Written Opinion for PCT/US2010/048325, mailed Jun. 1, 2011. |
WIPO, International Search Report and Written Opinion for PCT/US2011/025885, mailed Sep. 28, 2011. |
Wright, “EXtending ACID Semantics to the File System”, ACM Transactions on Storage, vol. 3, No. 2, published May 1, 2011, pp. 1-40. |
Zhang et al., “De-indirection for Flash-Based SSDs with Nameless Writes, ” UseniX InFast 2012 (Feb. 14, 2012). |
Number | Date | Country | |
---|---|---|---|
20110066808 A1 | Mar 2011 | US |
Number | Date | Country | |
---|---|---|---|
61240573 | Sep 2009 | US | |
61240966 | Sep 2009 | US | |
61373271 | Aug 2010 | US |