The present embodiments relate generally to data storage systems and more particularly, but without limitation, to dynamically storing data in different storage formats within a generic storage capacity.
Data storage systems can comprise a central processing unit (CPU), one or more disc controllers and a plurality of disc drives. Computer program code operating on the CPU and/or disc controllers controls the manner in which user data is stored and accessed on the disc drives. User data can be stored in various data storage formats, such as redundant array of independent drive (RAID) formats, for example. Multiple RAID formats can be concurrently employed within the data storage system. The optimal RAID format employed for particular data files can reflect the importance of the data and the frequency of access or update. Data that is easily accessed from other sources or otherwise reacquired can be stored in a non fault-tolerant format such a RAID 0, for example. Data that is stored in a fault tolerant format that is accessed frequently can be stored in a RAID 1 format. Data stored in a fault tolerant format that is accessed less frequently can be stored in a RAID 5 format. Alternatively RAID 5 can be employed to provide greater storage capacity as compared to RAID 1. Data such as financial data or system configuration data can be stored in a RAID 6 or Raid 1×3 format that provides greater redundancy than RAID-1 or RAID-5. As such, there can exist the need for a plurality of data storage formats to be concurrently supported within the system storage capacity. Also, it can be advantageous to change dynamically from a first storage format to a second storage format. For example, RAID 1 can be advantageously employed to update daily billing records, and then converted to RAID 5 format for archival storage. The rate at which storage can be dynamically allocated and the efficiency of utilization of available storage capacity significantly affects the marketability and profit margin of a data storage system.
As embodied herein and as claimed below, the present embodiments are generally directed to a device and associated method for storing data in a plurality of data storage formats wherein each format employs a same predefined data storage structure.
In some embodiments a generic storage container system is provided for a grid-based storage architecture, comprising a storage container comprising a plurality of storage domains along one axis against a plurality of rows of stripes along another axis defining a preselected storage capacity, and configuration information allocating the stripes in response to a storage format specified by an allocation request.
In some embodiments a method is provided for storing data in a data storage system, comprising: providing a storage container comprising a plurality of storage domains along one axis against a plurality of rows of stripes along another axis defining a preselected storage capacity; providing configuration information adapted for selectively allocating the stripes in relation to a data storage format; specifying a desired storage format; and allocating the stripes in response to the desired format.
In some embodiments a data storage system is provided comprising a data storage grid defining data stripes across a plurality of domains, and means for configuring the data stripes in relation to a desired data storage format.
These and various other features and advantages which characterize the embodiments of the claimed invention will become apparent upon reading the following detailed description and upon reviewing the associated drawings.
Operating systems 100 each contain at least one CPU, a plurality of data storage devices 101 defining a data storage capacity, and metadata describing the configuration of the data storage capacity. By “configuration” it is meant that a description is provided to the system 100 regarding whether particular portions of the storage capacity are being used to store data, or “allocated” space, as opposed to the portions that are available for storing data, or “allocatable” space.
Other operating systems suited for use with various embodiments of the present invention include additional data storage devices 101, additional data storage arrays, additional data storage device controllers or network interface controllers, and the like, that are not depicted in
To illustrate an exemplary environment in which presently preferred embodiments of the present invention can be advantageously practiced,
The system 100C includes a number of host computers 102, respectively identified as hosts A, B, and C. The host computers 102 interact with each other as well as with a pair of data storage arrays 104 (denoted A and B, respectively) via a fabric 106. The fabric 106 is preferably characterized as fibre-channel based switching network, although other configurations can be utilized as well, including the Internet.
Each array 104 includes a pair of controllers 108 (denoted A1, A2 and B1, B2) and a set of data storage devices 101 preferably characterized as disc drives operated as a RAID. The controllers 108 and data storage devices 101 preferably utilize a fault tolerant arrangement so that the various controllers 108 utilize parallel, redundant links and at least some of the user data stored by the system 100C is stored in a redundant format within at least one set of the data storage devices 101.
It is further contemplated that the A host computer 102 and the A data storage array 104 can be physically located at a first site, the B host computer 102 and B storage array 104 can be physically located at a second site, and the C host computer 102 can be yet at a third site, although such is merely illustrative and not limiting.
A fabric interface (I/F) circuit 118 communicates with the other controllers 108 and the host computers 102 via the fabric 106, and a device I/F circuit 120 communicates with the storage devices 101. The I/F circuits 118, 120 and a path controller 122 form a communication path to pass commands and data between the storage array 104 and the host 102, such as by employing the cache memory 124. Although illustrated discretely, it will be understood that the path controller 122 and the I/F circuits 118, 120 can be unitarily constructed.
The data storage capacity of an array 104, defined by the extent of the data storage devices 101 in a given array 104, is organized into logical units (LUNs) that can be written to and read from the array 104. System configuration information defines the relationship between user data, including any associated parity and mirror data, with the respective storage locations. The system configuration information furthermore identifies the relationship between blocks of storage capacity allocated to user data and the memory storage locations, such as logical block addresses (LBA). The system configuration information can furthermore include virtualization by defining virtual block addresses that are mapped to LBAs.
System configuration information is changed in response to a configuration change request from the system 100C. In response, storage capacity can be allocated, such as when saving new files or enlarging existing files, or storage capacity can be deallocated, such as when deleting files or reducing the size of existing files. For purposes of this description and the appended claims, the term “allocation” means either an allocation or deallocation of storage capacity. System metadata is resultant data defining file allocation information states and other data structures that support allocation processes.
Embodiments of the present invention enable the dynamic capacity assignment, by providing a generic storage container defining a preselected capacity, but which is allocatable for use with multiple storage formats. The container is termed a grid, and a plurality of commonly allocated grids is termed a sheet of data storage capacity. Accordingly, by fixing the size of each grid, the specified storage capacity incident with an allocation request is determinative of the number of grids within a sheet of storage capacity needed to satisfy the request.
Spare capacity is distributed across sub-partitions of data storage devices 101. Accordingly, stripes 132 in some sub-partitions (101A-101H) are used to store data, and in other sub-partitions (101I, 101J) are used as spares. Stripes 132 containing data store such things as user data, minor data, or parity data. Minor and parity data is sometimes referred to herein as fault tolerance information. Spare stripes 132 can be employed to restore data from a failed drive. Spare stripes 132 can be configured to exist only on specific data storage devices 101. For example, the grid 160 of
Preferably, the designation of spare sub-partitions 132 can vary from one grid 160 to the next. This serves to distribute data across all data storage devices 101 and allows all data storage devices 101 to be accessing data, offering higher performance in multitasking environments. The designation of which data storage devices 101 provide spare stripes 132 can be provided, as shown in
In embodiments where two or more data storage devices 101 provide spare capacity for each sheet and sparing version, data storage devices 101 can be organized into groups and pairs. In sheet organization table 170, for example, data storage devices 101A and 101B, 101C and 101D, 101E and 101F, 101G and 101H, and 101I and 101J form pairs and sparing can be defined such that stripes 132 of the paired data storage devices 101 have the same sparing designation for each sheet organization version. Data storage devices 101 can also be organized into groups. In sheet organization table 170, for example, a first group 172 contains data storage devices 101A, 101C, 101E, 101G, and 101; a second group 174 contains drives 101B, 101D, 101F, 101H and 101J.
Pairing can be employed for a RAID 1 format such that one member of the pair contains user data (primary data) and the other member of the pair contains an identical copy in the form of mirror data (secondary data). Grouping can be employed for a RAID 5 format such that user data is stored in one group 172, 174 and parity data is stored in the other group 172, 174. As such, pairing and grouping of data storage devices can be advantageously employed when defining fault tolerant data storage formats. The pairing and grouping of
Sheet organization versions can change on grid or sheet boundaries or can change for a predetermined number of grids or sheets. The starting address of a sheet can be divided by the number of grids 160 per sheet and the remainder of the division operation can be used to determine a sheet organization version. Other calculation methods can be employed and the result may be offset. Sheet organization table 170 can be a circular table that repetitively cycles through all versions. For example, a sixth sheet (not shown) can employ the same sheet organization version as a first sheet in the sheet organization table 170. The sheet organization table 170 can be modified to accommodate other grid 160 organizations reflecting a different number of data storage devices 101 and/or sparing capacity. Other data structures, other data storage device 101 or sparing order, or software program code can be employed to achieve the distribution of sparing capacity. The sheet organization table 170 allows sparing to be implemented in a simple and regular manner, and the sparing used for any grid 160 to be easily determined. While sheet organization table 170 depicts distributed sparing, embodiments of the present invention can be employed with dedicated spares, or with no spare capacity.
Again, the data storage grid 180 can be allocated to store user data or fault tolerant data. For example,
It will be noted that if a data storage grid 180 contains only parity values, such as the row parity grid 192 and the column parity grid 194, then one parity grid accommodates four user data grids. That is, a RAID 5 format comprises user data and row parity such that there is one row parity grid for every four user data grids. A RAID-6 format comprises user data and row/column parities such that there is one row parity grid and one column parity grid for every four user data grids.
It will be noted that there is some commonality between different storage formats that can be leveraged in converting from one format to another. Also, all the maps illustrate a “packed parity” arrangement where user data is stored exclusively in one portion of a sheet and mirror or parity data (if any) is stored in another portion of the sheet. This provides increased sequential read performance. The user data grids (
Embodiments of the present invention contemplate providing a generic storage container by organizing the data storage capacity into predetermined size sheets. The sheets are adapted for being allocated in a manner depending upon the RAID format specified by an allocation request. The number and placement of the corresponding RSUs are then automatically allocated within the sheet.
The sheet architectures described above accommodate RAID 0, RAID 1, RAID 1×3, RAID 5 or RAID 6 formats. If fewer RAID formats need accommodation, simplified sheet and grid architectures can be employed to yield a generic data storage container.
The generic container 230 in
The generic container 230 in
For data storage formats contemplated within the embodiments of the present invention, data is allocated in a predetermined manner within the data storage grid 180 and/or within the sheet containing a number of data storage grids 180. System configuration information, such as the user/parity map 190 and the allocation map 200, 202, 204, 206, 208, 220, 222, 224, defines the contents of stripes 132 and/or data storage grids 180 for the format that is specified in an allocation request. Grid format refers to the size of stripes 132 and the number of rows and columns in a data storage grid 180. Sheet format refers to the number of data storage grids 180 in a sheet. Data storage grids 180 and/or sheets of different grid and sheet formats can be employed simultaneously in the data storage capacity. For example, a video server system might advantageously employ a large grid or sheet format to store streaming video data, while employing a relatively smaller grid or sheet format to store billing information. Billing information might be stored in a highly fault tolerant format, such as a RAID 6 or a RAID 1×3 format, for example, while streaming video data might be stored in a less fault tolerant format, such as a RAID 0 or RAID 5 format.
Embodiments of the present invention provide generic storage containers that can be configured at the time of allocation to store data in a specified RAID format. This provides simplified metadata schemes, yet provides the flexibility of dynamic capacity assignment without the associated overhead or fragmentation of individual data stripe 132 management.
While foregoing examples have employed data storage devices 101 generally, embodiments of the present invention are not limited to the type of data storage device employed. Data storage devices 101 can comprise any type of data storage device including electrical, magnetic, optical, or chemical data storage devices including but not limited to hard disc drives, optical drives, WORM drives, CDROM, DVD, EEPROM, RAM drives including solid state memory devices, and the like and can include combinations thereof and further can include combinations of volatile and non-volatile data storage devices.
While disc based storage systems often include spare storage capacity, embodiments of the present invention can be employed in systems with no spare storage capacity. For example, data can be archived or otherwise stored in one or more storage formats on write only media, such as CDROM or DVD. Embodiments of the present invention simplify allocation of storage capacity when storing files in a plurality of data storage formats.
In block 256 a generic storage container is selected from one or more predefined candidates that accommodate the specified storage format. Given the generic storage container, which specifies the number of storage grids within the sheet, and the size of storage capacity specified by the allocation request, then in block 258 the number of sheets of storage capacity can be selected. Finally, in block 260 the sheets are allocated for storing the data in accordance with the system configuration information, such as the user/parity mapping discussed above, which defines which stripes 132 within the generic storage container are allocated for storing user data and which, if any, are allocated to store fault tolerance data.
It is to be understood that even though numerous characteristics and advantages of various embodiments of the present invention have been set forth in the foregoing description, together with details of the structure and function of various embodiments of the invention, this detailed description is illustrative only, and changes may be made in detail, especially in matters of structure and arrangements of parts within the principles of the present invention to the full extent indicated by the broad general meaning of the terms in which the appended claims are expressed. For example, the particular elements may vary depending on the particular data container structure without departing from the spirit and scope of the present invention.
| Number | Name | Date | Kind |
|---|---|---|---|
| 4847807 | Yule | Jul 1989 | A |
| 5285451 | Henson et al. | Feb 1994 | A |
| 5325363 | Lui | Jun 1994 | A |
| 5412661 | Hao et al. | May 1995 | A |
| 5519844 | Stallmo | May 1996 | A |
| 5537567 | Galbraith et al. | Jul 1996 | A |
| 5544339 | Baba | Aug 1996 | A |
| 5568629 | Gentry et al. | Oct 1996 | A |
| 5632027 | Martin et al. | May 1997 | A |
| 5671349 | Hashemi et al. | Sep 1997 | A |
| 5671390 | Brady et al. | Sep 1997 | A |
| 5682509 | Kabenjian | Oct 1997 | A |
| 5708769 | Stallmo | Jan 1998 | A |
| 5729763 | Leshem | Mar 1998 | A |
| 5790774 | Sarkozy | Aug 1998 | A |
| 5805788 | Johnson | Sep 1998 | A |
| 5812754 | Lui et al. | Sep 1998 | A |
| 5822782 | Humlicek et al. | Oct 1998 | A |
| 5832198 | Lucht | Nov 1998 | A |
| 5845319 | Yorimitsu | Dec 1998 | A |
| 5889934 | Peterson | Mar 1999 | A |
| 5893919 | Sarkozy et al. | Apr 1999 | A |
| 5933834 | Aichelen | Aug 1999 | A |
| 5960169 | Styczinski | Sep 1999 | A |
| 6016552 | Lee et al. | Jan 2000 | A |
| 6041407 | Claar et al. | Mar 2000 | A |
| 6052759 | Stallmo et al. | Apr 2000 | A |
| 6101615 | Lyons | Aug 2000 | A |
| 6154853 | Kedem | Nov 2000 | A |
| 6154854 | Stallmo | Nov 2000 | A |
| 6192448 | Mansur | Feb 2001 | B1 |
| 6195695 | Cheston et al. | Feb 2001 | B1 |
| 6219753 | Richardson | Apr 2001 | B1 |
| 6247157 | Edirisooriya | Jun 2001 | B1 |
| 6275898 | DeKoning | Aug 2001 | B1 |
| 6327672 | Wilner | Dec 2001 | B1 |
| 6338126 | Ohran et al. | Jan 2002 | B1 |
| 6353895 | Stephenson | Mar 2002 | B1 |
| 6356979 | Sicola et al. | Mar 2002 | B1 |
| 6378038 | Richardson et al. | Apr 2002 | B1 |
| 6425052 | Hashemi | Jul 2002 | B1 |
| 6453428 | Stephenson | Sep 2002 | B1 |
| 6457140 | Lindberg et al. | Sep 2002 | B1 |
| 6476830 | Farmer et al. | Nov 2002 | B1 |
| 6487633 | Horst et al. | Nov 2002 | B1 |
| 6502166 | Cassidy | Dec 2002 | B1 |
| 6529994 | Bleidt et al. | Mar 2003 | B1 |
| 6557123 | Wiencko, Jr. et al. | Apr 2003 | B1 |
| 6574687 | Teachout et al. | Jun 2003 | B1 |
| 6675176 | Shinkai et al. | Jan 2004 | B1 |
| 6675318 | Lee | Jan 2004 | B1 |
| 6742137 | Frey, Jr. | May 2004 | B1 |
| 6745284 | Lee et al. | Jun 2004 | B1 |
| 6754785 | Chow et al. | Jun 2004 | B2 |
| 6792391 | Nanda | Sep 2004 | B1 |
| 20030188097 | Holland et al. | Oct 2003 | A1 |
| 20040205387 | Kleiman et al. | Oct 2004 | A1 |
| Number | Date | Country |
|---|---|---|
| 1168176 | Jan 2002 | EP |
| WO 2004046971 | Jun 2004 | WO |
| Number | Date | Country | |
|---|---|---|---|
| 20060085593 A1 | Apr 2006 | US |