Managing read and write requests targeting a failed storage region in a storage system

Information

  • Patent Grant
  • 10838834
  • Patent Number
    10,838,834
  • Date Filed
    Tuesday, February 19, 2019
    6 years ago
  • Date Issued
    Tuesday, November 17, 2020
    4 years ago
Abstract
A system and method for efficiently distributing data among multiple storage devices. A data storage array receives read and write requests from multiple client computers. The data storage array includes multiple storage devices, each with multiple allocation units (AUs). A storage controller within the data storage array determines a RAID layout for use in storing data. In response to determining a failure of a first AU, the storage controller begins reconstructing in a second AU the data stored in the first AU. For read and write requests targeting data in the first AU, the request is serviced by the first AU responsive to determining no error occurs when accessing the first AU.
Description
TECHNICAL FIELD

This invention relates to computer networks and, more particularly, to efficiently distributing data among multiple storage devices.


BACKGROUND ART

As computer memory storage and data bandwidth increase, so does the amount and complexity of data that businesses daily manage. Large-scale distributed storage systems, such as data centers, typically run many business operations. A distributed storage system may be coupled to client computers interconnected by one or more networks. If any portion of the distributed storage system has poor performance or becomes unavailable, company operations may be impaired or stopped completely. A distributed storage system therefore is expected to maintain high standards for data availability and high-performance functionality. As used herein, storage disks may be referred to as storage devices as some types of storage technologies do not include disks.


To protect against data loss, storage devices often include error detection and correction mechanisms. Often these mechanisms take the form of error correcting codes, which are generated by the devices and stored within the devices themselves. In addition, distributed storage systems may also utilize decentralized algorithms to distribute data among a collection of storage devices. These algorithms may also include error detection and correction algorithms such as RAID type algorithms (e.g., RAID5 and RAID6). Although data protection is offered, latencies are added for servicing received input/output (I/O) requests from multiple clients. For example, the RAID type algorithms include data reconstruction steps prior to performing steps for the received request. Therefore, performance is reduced until failures are fully recovered.


In view of the above, systems and methods for efficiently distributing data among multiple storage devices are desired.


SUMMARY OF THE INVENTION

Various embodiments of a computer system and methods for efficiently distributing data among multiple storage devices are disclosed.


In various embodiments, a computer system includes multiple client computers that send read and write requests over a network to one or more data storage arrays. The data storage array includes multiple storage devices, each with multiple allocation units (AUs). The AU is a basic unit of storage in the storage device. A size of an AU is chosen to provide both sufficiently large sized units and a relatively low number of units to reduce overhead tracking of the AUs. Each AU includes multiple data storage locations. In various embodiments, the data storage locations are referred to as pages.


In various embodiments, the storage devices are configured in a redundant array of independent drives (RAID) arrangement for data storage and protection. The data storage subsystem further comprises a data storage controller that determines the RAID layout for use in storing data. In response to determining a failure of a first AU, the storage controller begins reconstructing or rebuilding the data stored in the first AU in a second AU. During the reconstruction, read and write requests targeting a given portion (e.g., a page) in the first AU, an attempt is made to service the request using the first AU. In some embodiments, the second AU is simultaneously accessed with the first AU. In such cases, the request may be serviced by the AU that responds first. In various embodiments, the second AU is accessed after an error or a miss occurs in the first AU for the targeted page.


These and other embodiments will become apparent upon consideration of the following description and accompanying drawings.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a generalized block diagram of one embodiment of handling requests in a storage subsystem with a failed region of storage.



FIG. 2 is a generalized block diagram of another embodiment of handling requests in a storage subsystem with a failed region of storage.



FIG. 3 is a generalized block diagram illustrating one embodiment of a network architecture.



FIG. 4 is a generalized block diagram of one embodiment of a storage subsystem.



FIG. 5 is a generalized block diagram of one embodiment of a device unit.



FIG. 6 is a generalized flow diagram illustrating one embodiment of a method for handling requests in a storage subsystem with a failed region of storage.





While the invention is susceptible to various modifications and alternative forms, specific embodiments are shown by way of example in the drawings and are herein described in detail. It should be understood, however, that drawings and detailed description thereto are not intended to limit the invention to the particular form disclosed, but on the contrary, the invention is to cover all modifications, equivalents and alternatives falling within the spirit and scope of the present invention as defined by the appended claims.


DETAILED DESCRIPTION

In the following description, numerous specific details are set forth to provide a thorough understanding of the present invention. However, one having ordinary skill in the art should recognize that the invention might be practiced without these specific details. In some instances, well-known circuits, structures, signals, computer program instruction, and techniques have not been shown in detail to avoid obscuring the present invention.


Referring to FIG. 1, a generalized block diagram of one embodiment of handling requests in a storage subsystem with a failed region of storage is shown. The data storage 20 receives read and write requests 10 from multiple client computers (not shown). The data storage 20 includes multiple storage devices, which are not shown for ease of illustration. Each of the storage devices includes multiple allocation units (AUs) such as AU 12. Some of the AUs among the storage devices are shown in the illustrated example as Data 1, Data 2, Bad Data 3, Repairing Data 3, and Parity 1.


The AU is a basic unit of storage in a storage device. A size of an AU is chosen to provide both sufficiently large sized units and a relatively low number of units to reduce overhead tracking of the AUs. Each of the Data 1, Data 2, Bad Data 3, Repairing Data 3, and Parity 1 represents an AU. A redundant array of independent drives (RAID) arrangement is used for data storage and protection on the shown AUs. Although three data allocation units with a single parity is shown for data storage and protection, other numbers of allocation units and other parity schemes are possible and contemplated.


The data layout being shown is a logical representation. In some examples, the physical arrangement of the stored data may be a random access arrangement, such as used for solid-state devices. In some examples, each of the AUs shown is located on a separate storage device within the data storage 20. In other examples, two or more of the AUs are located on a same storage device.


Each AU includes multiple data storage locations. As shown, each of the AUs includes N storage locations, wherein N is an integer. The storage locations are logically arranged as rows. In some examples, each storage location is a page, such as a 4 kilo-byte (KB) storage space. The data stored in each of the storage locations in a given AU may be associated with data 22 in a free or unallocated location, data 24 in an allocated and valid location, or with data 26 in an allocated and invalid location. The data 26 may be associated with stored data where accesses result in a miss or an error. The particular storage location may be located in a problematic region on a storage device. For example, the region may be overused, may be relatively old, may include fabrication defects and so forth. In various embodiments, in response to detection of errors to one or more portions of an AU, the storage system may be configured to relocate all of the data within the AU to a new AU. In various embodiments, the AU detected to have errors may or may not be removed from further use. In some embodiments, a decision may be made to move data from one AU to another that is not directly based on detection of an error. The methods and mechanisms described herein are applicable regardless of the reasons data is being moved to a new AU.


Control logic within a data storage controller (not shown) may have determined the allocation unit Bad Data 3 satisfies one or more conditions for having an AU state of bad, invalid, failed, or a similar state. The conditions may include at least one miss or one error occurred for an access for a storage location in the allocation unit, a number of misses or errors occurred for accesses to a same storage location in the allocation unit exceeding a threshold, a number of misses or errors occurred for accesses for any storage location in the allocation unit exceeding a threshold, and a miss rate or an error rate for accesses for any storage location in the allocation unit exceeding a threshold. The storage controller selected an unallocated or free AU to replace the allocation unit Bad Data 3. The replacement allocation unit Repairing Data 3 may be located on a different storage device. Alternatively, the replacement allocation unit Repairing Data 3 may be located on a same storage device but in a different region of memory on the same storage device.


Using the RAID layout, the storage controller initiates reconstruction of data stored in the allocation unit Bad Data 3 to the replacement allocation unit Repairing Data 3. For example, data stored in row 2 and in the allocation units Data 1 and Data 2 in addition to data stored in row 2 and in the allocation unit Parity 1 are used to reconstruct the data stored in row 2 and in allocation unit Bad Data 3. The reconstructed data is stored in row 2 in the allocation unit Repairing Data 3. Again, the data storage is shown as a logical representation of the data stored among the allocation units on the multiple storage devices.


As the rebuilding or reconstructing occurs for the data stored on the allocation unit Bad Data 3, requests are concurrently serviced. As shown, each of the allocation unit Bad Data 3 and Repairing Data 3 is not used for servicing requests. Rather, the RAID layout is used to reconstruct data used for the requests. In the example shown, the control logic 30 performs the reconstruction of data using the data read from a requested row in each of the allocation units Data 1, Data 2 and Parity 1. Therefore, each request includes an added latency for the reconstruction of data.


Turning now to FIG. 2, a generalized block diagram of another embodiment of handling requests in a storage subsystem with a failed region of storage is shown. Similar to the data storage 20, the data storage 40 receives read and write requests 10. The data storage 40 includes the allocation units and logic 30 shown earlier for data storage 20. However, the data storage 40 also includes a multiplexer 34. Additionally, control logic 30 is depicted that is configured to reconstruct data as will be described below. It is to be understood that the multiplexor 34 and control logic 30 are depicted as such for purposes of discussion. Those skilled in the art will appreciate that the functions performed by the multiplexor and control logic 30 may be achieved in any of a variety of ways using hardware and/or software. In the example shown, the allocation unit Repairing Data 3 is used to store reconstructed data in order to restore the data originally stored in the allocation unit Bad Data 3. Therefore, detection of failure for the allocation unit Bad Data 3 has already occurred. As the rebuilding occurs for the data stored on the allocation unit Bad Data 3, received requests are concurrently serviced. However, in some embodiments, an attempt to service a read directed to Bad Data 3 is first made using the allocation units Bad Data 3 and Repairing Data 3, rather than simply waiting for rebuilt data using the allocations units Data 1, Data 2 and Parity 1.


The multiplexer 34 receives data stored in locations of a requested row from each of the allocation units Bad Data 3, Repairing Data 3 and Parity 1. In addition, logic 30 receives data from Data 1, Data 2, and Parity 1 that can be used to reconstruct data of Bad Data 3. In some embodiments, the allocation unit Repairing Data 3 is only accessed if an access of the allocation unit Bad Data 3 is not successful. For example, the data in row 2 may be targeted by a read request. The storage location at row 2 in the allocation unit Bad Data 3 may be accessed and result in an error. Although it was already determined that the allocation unit Bad Data 3 fails and needs reconstruction of data stored at another location, the allocation unit Bad Data 3 is still accessed for servicing requests. While an access to a particular row may result in an error, there may be rows within the allocation unit Bad Data 3 that do not fail for accesses. Therefore, the allocation unit Bad Data 3 is accessed. If this access fails, the storage location at row 2 in the allocation unit Repairing Data 3 may be accessed. If the data has been reconstructed and stored in the allocation unit Repairing Data 3, then the data may be selected by the multiplexer 32.


Similarly, the data in row 4 may be targeted by a read request. The storage location at row 4 in the allocation unit Bad Data 3 may be accessed and result in an error-free hit. Again, although it was already determined that the allocation unit Bad Data 3 fails and needs reconstruction of data performed and stored at another location, the allocation unit Bad Data 3 is still accessed for servicing requests. The data in the storage location in row 4 in the allocation unit Bad Data 3 may be selected by the multiplexer 34. In various embodiments, the data stored in the allocation units Bad Data 3 and Repairing Data 3 are simultaneously accessed. In some embodiments, the data in the storage location in row 4 in the allocation unit Bad Data 3 may be copied to row 4 of the allocation unit Repairing Data 3, rather than have the data reconstructed and stored in row 4 of the allocation unit Repairing Data 3. In various embodiments, an attempt may be made to copy as many portions of a failed AU as possible to a new location before reconstruction is performed.


In various embodiments, the control logic 30 performs reconstruction of data according to the RAID layout. The multiplexer 34 selects between the output of the logic 30 and the output of the multiplexer 32. For a read request resulting in an access hit in the allocation unit Bad Data 3, the resulting data in the storage location of the allocation unit Bad Data 3 is sent through the multiplexer 34. For a read request resulting in an access miss or error in the allocation unit Bad Data 3, but an access hit in the allocation unit Repairing Data 3, the resulting data in the storage location of the allocation unit Repairing Data 3 is sent through the multiplexer 34. In various embodiments, logic 30 may begin reconstruction of data simultaneously with accesses to Bad Data 3 and Repairing Data 3. Whichever successfully completes first is then selected for servicing the read (e.g., using a multiplexor 34 type mechanism). Any suitable combination of the ordering of accesses to each of Bad Data 3, Repairing Data 3, and reconstructed data using logic 30, are possible and are contemplated. Some may precede others, while others are concurrent, etc.


As seen in the above, valid data stored in either of the allocation units Bad Data 3 and Repairing Data 3 may be read out to service the read request with no added latency normally associated with reconstructing data. Similar steps may be used and simply modified for write requests for particular technologies used within the storage devices. For example, storage devices using solid-state technology, such as Flash memory, may only use the above steps for read requests as write requests are always sent to different physical storage locations.


Referring to FIG. 3, a generalized block diagram of one embodiment of a network architecture 100 is shown. As described further below, one embodiment of network architecture 100 includes client computer systems 110a-110b interconnected to one another through a network 180 and to data storage arrays 120a-120b. Network 180 may be coupled to a second network 190 through a switch 140. Client computer system 110c is coupled to client computer systems 110a-110b and data storage arrays 120a-120b via network 190. In addition, network 190 may be coupled to the Internet 160 or otherwise outside network through switch 150.


It is noted that in alternative embodiments, the number and type of client computers and servers, switches, networks, data storage arrays, and data storage devices is not limited to those shown in FIG. 3. At various times one or more clients may operate offline. In addition, during operation, individual client computer connection types may change as users connect, disconnect, and reconnect to network architecture 100. Further, while the present description generally discusses network attached storage, the systems and methods described herein may also be applied to directly attached storage systems and may include a host operating system configured to perform one or more aspects of the described methods. Numerous such alternatives are possible and are contemplated. A further description of each of the components shown in FIG. 3 is provided shortly. First, an overview of some of the features provided by the data storage arrays 120a-120b is described.


In the network architecture 100, each of the data storage arrays 120a-120b may be used for the sharing of data among different servers and computers, such as client computer systems 110a-110c. In addition, the data storage arrays 120a-120b may be used for disk mirroring, backup and restore, archival and retrieval of archived data, and data migration from one storage device to another. In an alternate embodiment, one or more client computer systems 110a-110c may be linked to one another through fast local area networks (LANs) in order to form a cluster. Such clients may share a storage resource, such as a cluster shared volume residing within one of data storage arrays 120a-120b.


Each of the data storage arrays 120a-120b includes a storage subsystem 170 for data storage. Storage subsystem 170 may comprise one or more storage devices 176a-176m. These storage devices 176a-176m may provide data storage services to client computer systems 110a-110c. Each of the storage devices 176a-176m uses a particular technology and mechanism for performing data storage. The type of technology and mechanism used within each of the storage devices 176a-176m may at least in part be used to determine the algorithms used for controlling and scheduling read and write operations to and from each of the storage devices 176a-176m. For example, the algorithms may locate particular physical locations corresponding to the operations. In addition, the algorithms may perform input/output (I/O) redirection for the operations, removal of duplicate data in the storage subsystem 170, and support one or more mapping tables used for address redirection and deduplication.


The logic used in the above algorithms may be included in one or more of a base operating system (OS) 132, a volume manager 134, within the storage subsystem controller 174, control logic within each of the storage devices 176a-176m, or otherwise. Additionally, the logic, algorithms, and control mechanisms described herein may comprise hardware and/or software.


Each of the storage devices 176a-176m may be configured to receive read and write requests and comprise a plurality of data storage locations, each data storage location being addressable as rows and columns in an array. In one embodiment, the data storage locations within the storage devices 176a-176m may be arranged into logical, redundant storage containers or RAID arrays (redundant arrays of inexpensive/independent disks).


In some embodiments, each of the storage devices 176a-176m may include or be further coupled to storage consisting of solid-state memory to store persistent data. In one embodiment, the included solid-state memory comprises solid-state drive (SSD) technology. A Solid-State Disk (SSD) may also be referred to as a Solid-State Drive.


Storage array efficiency may be improved by creating a storage virtualization layer between user storage and physical locations within storage devices 176a-176m. In one embodiment, a virtual layer of a volume manager is placed in a device-driver stack of an operating system (OS), rather than within storage devices or in a network. A volume manager or a disk array manager is used to support device groups 173a-173m.


In one embodiment, one or more mapping tables may be stored in the storage devices 176a-176m, rather than memory, such as RAM 172, memory medium 130 or a cache within processor 122. The storage devices 176a-176 may be SSDs utilizing Flash memory. The low read access and latency times for SSDs may allow a small number of dependent read operations to occur while servicing a storage access request from a client computer. The dependent read operations may be used to access one or more indexes, one or more mapping tables, and user data during the servicing of the storage access request.


Network architecture 100 includes client computer systems 110a-110c interconnected through networks 180 and 190 to one another and to data storage arrays 120a-120b. Networks 180 and 190 may include a variety of techniques including wireless connection, direct local area network (LAN) connections, wide area network (WAN) connections such as the Internet, a router, storage area network, Ethernet, and others. Networks 180 and 190 may comprise one or more LANs that may also be wireless. Switch 140 may utilize a protocol associated with both networks 180 and 190. The network 190 may interface with a set of communications protocols used for the Internet 160 such as the Transmission Control Protocol (TCP) and the Internet Protocol (IP), or TCP/IP. Switch 150 may be a TCP/IP switch.


Client computer systems 110a-110c are representative of any number of stationary or mobile computers such as desktop personal computers (PCs), servers, server farms, workstations, laptops, handheld computers, servers, personal digital assistants (PDAs), smart phones, and so forth. Each of the client computer systems 110a-110c may include a hypervisor used to support virtual machines (VMs).


Each of the data storage arrays 120a-120b may be used for the sharing of data among different servers, such as the client computer systems 110a-110c. Each of the data storage arrays 120a-120b includes a storage subsystem 170 for data storage. Storage subsystem 170 may comprise one or more storage devices 176a-176m. Each of these storage devices 176a-176m may be an SSD. A random-access memory (RAM) 172 may be used to batch operations, such as received write requests. In various embodiments, when batching write operations (or other operations) non-volatile storage (e.g., NVRAM) may be used. The controller 174 may comprise logic for handling received read/write requests. The controller 174 may also include the RAID engine 178 to determine data layout among the devices 176a-176m and to add protection for the stored data.


The base OS 132, the volume manager 134 (or disk array manager 134), any OS drivers (not shown) and other software stored in memory medium 130 may provide functionality providing access to files and the management of these functionalities. The base OS 132 and the OS drivers may comprise program instructions stored on the memory medium 130 and executable by processor 122 to perform one or more memory access operations in storage subsystem 170 that correspond to received requests. Each of the data storage arrays 120a-120b may use a network interface 124 to connect to network 180. Similar to client computer systems 110a-110c, in one embodiment, the functionality of network interface 124 may be included on a network adapter card.


In addition to the above, the storage controllers 174 within the data storage arrays 120a-120b may support storage array functions such as snapshots, replication and high availability. In addition, the storage controller 174 may support a virtual machine environment that comprises a plurality of volumes with each volume including a plurality of snapshots. In one example, the storage controller 174 may support hundreds of thousands of volumes, wherein each volume includes thousands of snapshots. In one embodiment, a volume may be mapped in fixed-size sectors, such as a 4-kilobyte (KB) page within storage devices 176a-176m. In another embodiment, a volume may be mapped in variable-size sectors such as for write requests. A volume ID, a snapshot ID, and a sector number may be used to identify a given volume.


Turning now to FIG. 4, a generalized block diagram of one embodiment of a storage subsystem is shown. Each of the one or more storage devices 176a-176m may be partitioned in one of one or more device groups 173a-173m. Other device groups with other devices may be present as well. One or more corresponding operation queues and status tables for each storage device may be included in one of the device units 400a-400w. These device units may be stored in RAM 172. A corresponding RAID engine 178a-178m may be included for each one of the device groups 173a-173m. Each one of the RAID engines 178a-178m may include a monitor 410 that tracks statistics for each of the storage devices included within a corresponding device group. Data layout logic 420 may determine an amount of space to allocate within a corresponding storage device for user data, inter-device redundancy data and intra-device redundancy data. The storage controller 174 may comprise other control logic 430 to perform at least one of the following tasks: wear leveling, garbage collection, I/O scheduling, deduplication and protocol conversion for incoming and outgoing packets.


Turning now to FIG. 5, a generalized block diagram of one embodiment of a device unit is shown. A device unit may comprise a device queue 510 and tables 520. Device queue 510 may include a read queue 512, a write queue 514 and one or more other queues such as other operation queue 516. Each queue may comprise a plurality of entries for storing one or more corresponding requests 530a-530d. For example, a device unit for a corresponding SSD may include queues to store at least read requests, write requests, trim requests, erase requests and so forth.


Tables 520 may comprise one or more state tables 522a-522b, each comprising a plurality of entries for storing state data, or statistics, 530. It is also noted that while the queues and tables are shown to include a particular number of entries in this and other figures, the entries themselves do not necessarily correspond to one another. Additionally, the number of queues, tables, and entries may vary from that shown in the figure and may differ from one another.


Turning now to FIG. 6, a generalized flow diagram of one embodiment of a method 600 for handling requests in a storage subsystem with a failed region of storage is shown. For purposes of discussion, the steps in this embodiment are shown in sequential order. However, some steps may occur in a different order than shown, some steps may be performed concurrently, some steps may be combined with other steps, and some steps may be absent in another embodiment. For example, as discussed above, simultaneous accesses may be made to a failed AU and an AU selected to replace the failed AU. Additionally, reconstruction could begin at the same time as such accesses. All such possible combinations discussed above are contemplated.


Responsive to receiving input/output (I/O) requests, such as read and write requests, to multiple allocation units (AUs) within one or more storage devices, the I/O requests are processed (block 602). Each of the AUs includes multiple pages. In some embodiments, each page is an allocation space that is the smallest allocation size to access for an I/O request. In some examples, the page may be a 4 KB storage space in an AU. The read requests have requested data returned to a requesting client computer, such as one or more targeted pages within a specified AU. Additionally, the write requests have request data written to a targeted storage location, such as one or more pages within an AU.


A given AU may be determined to fail based on multiple conditions. As described earlier, the conditions may include at least one miss or one error occurred for an access for a storage location in the given AU, a number of misses or errors occurred for accesses to a same storage location in the given AU exceeding a threshold, a number of misses or errors occurred for accesses for any storage location in the given AU exceeding a threshold, and a miss rate or an error rate for accesses for any storage location in the given AU exceeding a threshold. Alternatively, an AU may simply be selected to have its data relocated to another AU for any another reason.


If it is determined a first AU fails (conditional block 604), then in block 606, an unallocated second AU is selected to replace the failed first AU. To reconstruct the data originally stored in the first AU, the data may be reconstructed using a supported RAID layout and stored in the second AU. The data originally stored in the pages of the first AU are reconstructed in the second AU (block 608).


For an I/O request targeting the first AU while the AU is being replaced, the targeted page is accessed in the first AU (block 610). If the access in the first AU fails (conditional block 612), and an access for the same page in the second AU fails (conditional block 614), then in block 616, the requested page is reconstructed or rebuilt using AUs other than the first AU or the second AU. For example, a RAID layout with one or more parity pages may be used to reconstruct the requested page. In block 618, the I/O request is serviced. For example, the reconstructed data is sent to a requesting client computer. As already discussed, multiple approaches to obtaining data to service a read request may be attempted simultaneously. In such embodiments, the first approach to successfully obtain the data or otherwise have it ready is used to service the read.


It is noted that the above-described embodiments may comprise software. In such an embodiment, the program instructions that implement the methods and/or mechanisms may be conveyed or stored on a computer readable medium. Numerous types of media which are configured to store program instructions are available and include hard disks, floppy disks, CD-ROM, DVD, flash memory, Programmable ROMs (PROM), random access memory (RAM), and various other forms of volatile or non-volatile storage.


In various embodiments, one or more portions of the methods and mechanisms described herein may form part of a cloud-computing environment. In such embodiments, resources may be provided over the Internet as services according to one or more various models. Such models may include Infrastructure as a Service (IaaS), Platform as a Service (PaaS), and Software as a Service (SaaS). In IaaS, computer infrastructure is delivered as a service. In such a case, the computing equipment is generally owned and operated by the service provider. In the PaaS model, software tools and underlying equipment used by developers to develop software solutions may be provided as a service and hosted by the service provider. SaaS typically includes a service provider licensing software as a service on demand. The service provider may host the software, or may deploy the software to a customer for a given period of time. Numerous combinations of the above models are possible and are contemplated.


Although the embodiments above have been described in considerable detail, numerous variations and modifications will become apparent to those skilled in the art once the above disclosure is fully appreciated. It is intended that the following claims be interpreted to embrace all such variations and modifications.

Claims
  • 1. A method comprising: detecting a failure associated with a first storage region;in response to detecting the failure, reconstructing, at a second storage region, data stored at the first storage region;receiving an I/O request targeting the first storage region during reconstruction of the data stored at the first storage region;in response to receiving the I/O request targeting the first storage region during reconstruction of the data, accessing both the first storage region and the second storage region in order to service the I/O request; andservicing the received I/O request with the requested data from the storage region that is identified as responding first with the requested data.
  • 2. The method as recited in claim 1, further comprising: in response to determining an error occurs when accessing the given data storage location, using reconstructed data to both service the received I/O request and copy data in the given data storage location in the second storage region.
  • 3. The method as recited in claim 1, further comprising: in response to determining an error occurs when accessing the given data storage location, servicing the received I/O request using the second storage region.
  • 4. The method as recited in claim 1, wherein detecting the failure associated with a first storage region comprises determining a number of errors when accessing data in the first storage region exceeds a threshold.
  • 5. The method as recited in claim 1, wherein detecting the failure associated with a first storage region comprises determining an error rate when accessing data in the first storage region exceeds a threshold.
  • 6. An apparatus including a computer processor coupled to computer memory, the computer memory including computer program instructions that when executed by the computer processor cause the apparatus to: detect a failure associated with a first storage region;in response to detecting the failure, reconstruct, at a second storage region, data stored at the first storage region;receive a I/O request targeting the first storage region during reconstruction of the data; andin response to receiving the I/O request targeting the first storage region during reconstruction of the data, access both the first storage region and the second storage region in order to service the I/O request, andservice the received I/O request with the requested data from the storage region that is identified as responding first with the requested data.
  • 7. The apparatus of claim 6, wherein the computer memory includes computer program instructions that when executed by the computer processor cause the apparatus to: in response to determining an error occurs when accessing the given data storage location, use reconstructed data to both service the received I/O request and copy data in the given data storage location in the second storage region.
  • 8. The apparatus of claim 6, wherein the computer memory includes computer program instructions that when executed by the computer processor cause the apparatus to: in response to determining an error occurs when accessing the given data storage location, service the received I/O request using the second storage region.
  • 9. The apparatus of claim 6, wherein detecting the failure associated with a first storage region comprises determining a number of errors when accessing data in the first storage region exceeds a threshold.
  • 10. The apparatus of claim 6, wherein detecting the failure associated with a first storage region comprises determining an error rate when accessing data in the first storage region exceeds a threshold.
  • 11. A non-transitory computer readable storage medium storing program instructions that when executed by a computer processor cause the computer processor to: detect a failure associated with a first storage region;in response to detecting the failure, reconstruct, at a second storage region, data stored at the first storage region;receive a I/O request targeting the first storage region during reconstruction of the data; andin response to receiving the I/O request targeting the first storage region during reconstruction of the data, simultaneously access, in parallel, the first storage region and the second storage region in order to service the I/O request; andservice the received I/O request with the requested data from the storage region that is identified as responding first with the requested data.
  • 12. The computer readable storage medium of claim 11, wherein the computer memory includes computer program instructions when executed by a computer processor cause the computer processor to: in response to determining an error occurs when accessing the given data storage location, use reconstructed data to both service the received I/O request and copy data in the given data storage location in the second storage region.
  • 13. The computer readable storage medium of claim 11, wherein the computer memory includes computer program instructions when executed by a computer processor cause the computer processor to: in response to determining an error occurs when accessing the given data storage location, service the received I/O request using the second storage region.
  • 14. The computer readable storage medium of claim 11, wherein detecting the failure associated with a first storage region comprises determining a number of errors when accessing data in the first storage region exceeds a threshold.
  • 15. The computer readable storage medium of claim 11, wherein detecting the failure associated with a first storage region comprises determining an error rate when accessing data in the first storage region exceeds a threshold.
CROSS-REFERENCE TO RELATED APPLICATION

This application is a continuation application of and claims priority from U.S. patent application Ser. No. 15/416,151, filed Jan. 26, 2017, which is a continuation of U.S. Pat. No. 9,588,842, issued Mar. 7, 2017.

US Referenced Citations (135)
Number Name Date Kind
5208813 Stallmo May 1993 A
5403639 Belsan Apr 1995 A
5761411 Teague Jun 1998 A
5802264 Chen Sep 1998 A
5940838 Schmuck et al. Aug 1999 A
6263350 Wollrath et al. Jul 2001 B1
6412045 DeKoning et al. Jun 2002 B1
6718448 Ofer Apr 2004 B1
6757769 Ofer Jun 2004 B1
6799283 Tamai et al. Sep 2004 B1
6834298 Singer et al. Dec 2004 B1
6850938 Sadjadi Feb 2005 B1
6915434 Kuroda Jul 2005 B1
6973549 Testardi Dec 2005 B1
7028216 Aizawa et al. Apr 2006 B2
7028218 Schwarm et al. Apr 2006 B2
7039827 Meyer et al. May 2006 B2
7216164 Whitmore et al. May 2007 B1
7783682 Patterson Aug 2010 B1
7873619 Faibish et al. Jan 2011 B1
7913300 Flank et al. Mar 2011 B1
7933936 Aggarwal et al. Apr 2011 B2
7979613 Zohar et al. Jul 2011 B2
8086652 Bisson et al. Dec 2011 B1
8117464 Kogelnik Feb 2012 B1
8200887 Bennett Jun 2012 B2
8205065 Matze Jun 2012 B2
8352540 Anglin et al. Jan 2013 B2
8527544 Colgrove et al. Sep 2013 B1
8560747 Tan et al. Oct 2013 B1
8621241 Stephenson Dec 2013 B1
8700875 Barron et al. Apr 2014 B1
8751463 Chamness Jun 2014 B1
8806160 Colgrove et al. Aug 2014 B2
8874850 Goodson et al. Oct 2014 B1
8959305 Lecrone et al. Feb 2015 B1
9081713 Bennett Jul 2015 B1
9189334 Bennett Nov 2015 B2
9311182 Bennett Apr 2016 B2
9423967 Colgrove et al. Aug 2016 B2
9436396 Colgrove et al. Sep 2016 B2
9436720 Colgrove et al. Sep 2016 B2
9454476 Colgrove et al. Sep 2016 B2
9454477 Colgrove et al. Sep 2016 B2
9513820 Shalev Dec 2016 B1
9516016 Colgrove et al. Dec 2016 B2
9552248 Miller et al. Jan 2017 B2
9632870 Bennett Apr 2017 B2
10248516 Sanvido Apr 2019 B1
20020038436 Suzuki Mar 2002 A1
20020087544 Selkirk et al. Jul 2002 A1
20020178335 Selkirk et al. Nov 2002 A1
20030140209 Testardi Jul 2003 A1
20040049572 Yamamoto et al. Mar 2004 A1
20050066095 Mullick et al. Mar 2005 A1
20050216535 Saika et al. Sep 2005 A1
20050223154 Uemura Oct 2005 A1
20060074940 Craft et al. Apr 2006 A1
20060136365 Kedem et al. Jun 2006 A1
20060155946 Ji Jul 2006 A1
20070067585 Ueda et al. Mar 2007 A1
20070162954 Pela Jul 2007 A1
20070171562 Maejima et al. Jul 2007 A1
20070174673 Kawaguchi et al. Jul 2007 A1
20070220313 Katsuragi Sep 2007 A1
20070245090 King et al. Oct 2007 A1
20070266179 Chavan et al. Nov 2007 A1
20080059699 Kubo et al. Mar 2008 A1
20080065852 Moore et al. Mar 2008 A1
20080134174 Sheu et al. Jun 2008 A1
20080155191 Anderson et al. Jun 2008 A1
20080178040 Kobayashi Jul 2008 A1
20080209096 Lin et al. Aug 2008 A1
20080244205 Amano et al. Oct 2008 A1
20080275928 Shuster Nov 2008 A1
20080285083 Aonuma Nov 2008 A1
20080307270 Li Dec 2008 A1
20090006587 Richter Jan 2009 A1
20090037662 La Frese et al. Feb 2009 A1
20090204858 Kawaba Aug 2009 A1
20090228648 Wack Sep 2009 A1
20090300084 Whitehouse Dec 2009 A1
20100057673 Savov Mar 2010 A1
20100058026 Heil et al. Mar 2010 A1
20100067706 Anan et al. Mar 2010 A1
20100077205 Ekstrom et al. Mar 2010 A1
20100082879 McKean et al. Apr 2010 A1
20100106905 Kurashige et al. Apr 2010 A1
20100153620 McKean et al. Jun 2010 A1
20100153641 Jagadish et al. Jun 2010 A1
20100191897 Zhang et al. Jul 2010 A1
20100250802 Waugh et al. Sep 2010 A1
20100250882 Hutchison et al. Sep 2010 A1
20100281225 Chen et al. Nov 2010 A1
20100287327 Li et al. Nov 2010 A1
20110072300 Rousseau Mar 2011 A1
20110145598 Smith et al. Jun 2011 A1
20110161559 Yurzola et al. Jun 2011 A1
20110167221 Pangal et al. Jul 2011 A1
20110238634 Kobara Sep 2011 A1
20120023375 Dutta et al. Jan 2012 A1
20120036309 Dillow et al. Feb 2012 A1
20120117029 Gold May 2012 A1
20120198175 Atkisson Aug 2012 A1
20120330954 Sivasubramanian et al. Dec 2012 A1
20130042052 Colgrove et al. Feb 2013 A1
20130046995 Movshovitz Feb 2013 A1
20130047029 Ikeuchi Feb 2013 A1
20130091102 Nayak Apr 2013 A1
20130205110 Kettner Aug 2013 A1
20130227236 Flynn et al. Aug 2013 A1
20130275391 Batwara et al. Oct 2013 A1
20130275656 Talagala et al. Oct 2013 A1
20130283058 Fiske et al. Oct 2013 A1
20130290648 Shao et al. Oct 2013 A1
20130318314 Markus et al. Nov 2013 A1
20130339303 Potter et al. Dec 2013 A1
20140052946 Kimmel Feb 2014 A1
20140068791 Resch Mar 2014 A1
20140089730 Watanabe et al. Mar 2014 A1
20140101361 Gschwind Apr 2014 A1
20140143517 Jin et al. May 2014 A1
20140172929 Sedayao et al. Jun 2014 A1
20140201150 Kumarasamy et al. Jul 2014 A1
20140215129 Kuzmin et al. Jul 2014 A1
20140229131 Cohen et al. Aug 2014 A1
20140229452 Serita et al. Aug 2014 A1
20140281308 Lango et al. Sep 2014 A1
20140325115 Ramsundar et al. Oct 2014 A1
20150234709 Koarashi Aug 2015 A1
20150244775 Vibhor et al. Aug 2015 A1
20150278534 Thiyagarajan et al. Oct 2015 A1
20160019114 Han et al. Jan 2016 A1
20160098191 Golden et al. Apr 2016 A1
20160098199 Golden et al. Apr 2016 A1
Foreign Referenced Citations (14)
Number Date Country
103370685 Oct 2013 CN
103370686 Oct 2013 CN
104025010 Nov 2016 CN
3066610 Sep 2016 EP
3082047 Oct 2016 EP
3120235 Jan 2017 EP
2007-087036 Apr 2007 JP
2007-094472 Apr 2007 JP
2008-250667 Oct 2008 JP
2010-211681 Sep 2010 JP
WO-1995002349 Jan 1995 WO
WO 1999013403 Mar 1999 WO
WO-2008102347 Aug 2008 WO
WO-2010071655 Jun 2010 WO
Non-Patent Literature Citations (2)
Entry
Microsoft Corporation, “GCSettings.IsServerGC Property”, Retrieved Oct. 27,2013 via the WayBack Machine, 3 pages.
Microsoft Corporation, “Fundamentals of Garbage Collection”, Retrieved Aug. 30, 2013 via the WayBack Machine, 11 pages.
Continuations (2)
Number Date Country
Parent 15416151 Jan 2017 US
Child 16279174 US
Parent 14567960 Dec 2014 US
Child 15416151 US