Intelligent logical unit provisioning

Information

  • Patent Grant
  • 7721053
  • Patent Number
    7,721,053
  • Date Filed
    Monday, October 24, 2005
    19 years ago
  • Date Issued
    Tuesday, May 18, 2010
    14 years ago
Abstract
In one embodiment, a storage controller comprises a first I/O port that provides an interface to a host computer, a second I/O port that provides an interface to a storage device, a processor that receives I/O requests generated by the host computer and, in response to the I/O requests, generates and transmits I/O requests to the storage device, and a memory module communicatively connected to the processor. The memory module comprises logic instructions which, when executed by the processor, configure the processor to collect performance data and availability data for a plurality of logical devices (LDEVS) managed by the storage controller, and present the performance data and availability data to a reporting interface.
Description
BACKGROUND

The described subject matter relates to data storage in electronic computing, and more particularly to intelligent logical unit provisioning.


Effective collection, management, and control of information have become a central component of modern business processes. To this end, many businesses, both large and small, now implement computer-based information management systems.


Data management is an important component of computer-based information management systems. Many users implement storage networks to manage data operations in computer-based information management systems. Storage networks have evolved in computing power and complexity to provide highly reliable, managed storage solutions that may be distributed across a wide geographic area, and across physical storage devices that are under the management of a storage controller (i.e., internal) or outside the management of a storage controller (i.e., external).


Adroit management of storage network resources contributes to the effective management of storage networks. Existing management interfaces provide limited information for managing storage resources. Management interfaces that provide additional management information would be useful.


SUMMARY

In one embodiment, a storage controller comprises a first I/O port that provides an interface to a host computer, a second I/O port that provides an interface a storage device, a processor that receives I/O requests generated by the host computer and, in response to the I/O requests, generates and transmits I/O requests to the storage device, and a memory module communicatively connected to the processor. The memory module comprises logic instructions which, when executed by the processor, configure the processor to collect performance data and availability data for a plurality of logical devices (LDEVS) managed by the storage controller, and present the performance data and availability data to a reporting interface.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a schematic illustration of one embodiment of a storage network environment.



FIG. 2 is a schematic illustration of one embodiment of an array controller.



FIG. 3 is a flowchart illustrating operations in one embodiment of a method for intelligent logical unit provisioning



FIG. 4 is a schematic illustration of one embodiment of a user interface for intelligent logical unit provisioning.





DETAILED DESCRIPTION

Described herein are exemplary systems and methods for intelligent logical unit provisioning in a storage device, array, or network. The methods described herein may be embodied as logic instructions on a computer-readable medium. When executed on a processor such as, e.g., an array controller, the logic instructions cause the processor to be programmed as a special-purpose machine that implements the described methods. The processor, when configured by the logic instructions to execute the methods recited herein, constitutes structure for performing the described methods. The methods will be explained with reference to one or more logical volumes in a storage system, but the methods need not be limited to logical volumes. The methods are equally applicable to storage systems that map to physical storage, rather than logical storage.



FIG. 1 is a schematic illustration of an exemplary implementation of a networked computing environment 100. Referring to FIG. 1, computing environment 100 includes a storage pool 110 that provides data storage services to one or more computing devices. Storage pool 110 may be implemented in one or more networked storage cells 140A, 140B, 140C. Exemplary storage cells include the STORAGEWORKS line of storage devices commercially available from Hewlett-Packard Corporation of Palo Alto, Calif., USA. Storage cells 140A, 140B, 140C may be co-located or may be geographically distributed, and may be connected by a suitable communication network. The communication network may be embodied as a private, dedicated network such as, e.g., a Fibre Channel (FC) switching fabric. Alternatively, portions of communication network may be implemented using public communication networks pursuant to a suitable communication protocol such as, e.g., the Internet Small Computer Serial Interface (iSCSI) protocol. The number of storage cells 140A, 140B, 140C that can be included in any storage network is limited primarily by the connectivity implemented in the communication network. For example, a switching fabric comprising a single FC switch can interconnect 256 or more ports, providing a possibility of hundreds of storage cells in a single storage network.


Computing environment 100 further includes one or more host computing devices which utilize storage services provided by the storage pool 110 on their own behalf or on behalf of other client computing or data processing systems or devices. Client computing devices such as client 126 access storage the storage pool 110 embodied by storage cells 140A, 140B, 140C through a host computer. For example, client computer 126 may access storage pool 110 via a host such as server 124. Server 124 may provide file services to client 126, and may provide other services such as transaction processing services, email services, etc. Host computer 122 may also utilize storage services provided by storage pool 110 on its own behalf. Clients such as clients 132, 134 may be connected to host computer 128 directly, or via a network 130 such as a Local Area Network (LAN) or a Wide Area Network (WAN).



FIG. 2 is a schematic illustration of an exemplary embodiment of a storage cell 200. Storage cell 200 may correspond to one of the storage cells 140A, 140B, 140C depicted in FIG. 1. It will be appreciated that the storage cell 200 depicted in FIG. 2 is merely one exemplary embodiment, which is provided for purposes of explanation.


Referring to FIG. 2, storage cell 200 includes two Network Storage Controllers (NSCs), also referred to as “disk array controllers” or just “array controllers” 210a, 210b to manage operations and the transfer of data to and from one or more sets of disk drives 240, 242. Array controllers 210a, 210b may be implemented as plug-in cards having a microprocessor 216a, 216b, and memory 218a, 218b. Each array controller 210a, 210b includes dual host adapter ports 212a, 214a, 212b, 214b that provide an interface to a host, i.e., through a communication network such as a switching fabric. In a Fibre Channel implementation, host adapter ports 212a, 212b, 214a, 214b may be implemented as FC N_Ports. Each host adapter port 212a, 212b, 214a, 214b manages the login and interface with a switching fabric, and is assigned a fabric-unique port ID in the login process. The architecture illustrated in FIG. 2 provides a fully-redundant storage cell. This redundancy is entirely optional; only a single array controller is required to implement a storage cell.


Each array controller 210a, 210b further includes a communication port 228a, 228b that enables a communication connection 238 between the array controllers 210a, 210b. The communication connection 238 may be implemented as a FC point-to-point connection, or pursuant to any other suitable communication protocol.


In an exemplary implementation, array controllers 210a, 210b further include a plurality of Fiber Channel Arbitrated Loop (FCAL) ports 220a-226a, 220b-226b that implements an FCAL communication connection with a plurality of storage devices, e.g., sets of disk drives 240, 242. While the illustrated embodiment implement FCAL connections with the sets of disk drives 240, 242, it will be understood that the communication connection with sets of disk drives 240, 242 may be implemented using other communication protocols. For example, rather than an FCAL configuration, a FC switching fabric may be used.


In operation, the storage capacity provided by the sets of disk drives 240, 242 may be added to the storage pool 110. When an application requires storage capacity, logic instructions on a host computer such as host computer 128 establish a LUN from storage capacity available on the sets of disk drives 240, 242 available in one or more storage sites. It will be appreciated that, because a LUN is a logical unit, not a physical unit, the physical storage space that constitutes the LUN may be distributed across multiple storage cells. Data for the application may be stored on one or more LUNs in the storage network. An application that needs to access the data queries a host computer, which retrieves the data from the LUN and forwards the data to the application.


In operation, a user, administrator, or software module responsible for managing the storage pool 110 may periodically need to provision a new logical unit, such as logical unit 112a, 112b, in the storage pool 110. FIG. 3 is a flowchart illustrating operations in one embodiment of a method for intelligent logical unit provisioning, and FIG. 4 is a schematic illustration of one embodiment of a user interface for intelligent logical unit provisioning. The operations of FIG. 3 may be implemented in a storage controller such as one of the storage controllers 210a, 210b to enable the storage controller to collect performance data and availability data from internal logical units (i.e., logical units defined from physical storage media within the storage cell) and external logical units (i.e., logical units defined from physical storage media outside the storage cell. Once collected, the performance data and availability data may be presented to a reporting interface, which may organize the data and present the data in a suitable interface.


Referring to FIG. 3, at operation 310 a logical device is selected. In one embodiment, a logical device may correspond to a logical unit such as logical units 112a, 112b, managed by the storage controller. At operation 315 a performance test is initiated on the logical device selected in operation 310. In one embodiment, the storage controller initiates an online transaction performance (OLTP) test in which 8 KB blocks of data are written to and read from the logical unit for a predetermined time period such as, e.g., 250 ms. In one embodiment, the performance test implements a 60:40 ratio of read operations to write operations, although other ratios may be implemented. One technique for estimating the performance of an OLTP application may be accomplished by subjecting a storage unit to a workload including the following attributes: Block-size: 8 kB, access pattern: Random, read percentage 60%, write percentage: 40%, queue depth: 1 to n, where in causes an average response time of 30 ms. The general metric of concern is the maximum number of input/output operations per second (IO/sec) that the storage unit can support. This information permits proper matching of application users and storage resources to maintain an acceptable performance experience for the application. One technique for measuring such performance, may include supplying the storage with a predetermined workload for a defined amount of time. The performance rate may be calculated by dividing the number of I/O operations completed by the time (e.g., in seconds) to give a result in units of I/O operations per second.


At operation 320 a data warehouse test is initiated on the logical device selected in operation 310. In one embodiment, the storage controller initiates a simulated data warehouse workload test in which 64 KB blocks of data are read sequentially from the logical unit for a predetermined time period such as, e.g., 250 ms. In alternate embodiments, different data block sizes may be read. One technique for estimating the performance of a data warehouse application may be accomplished by subjecting a storage unit to a workload consisting of the following attributes: Block-size: 64 KB, access pattern: Sequential, read percentage 100%, queue depth: 1 to n (where in causes an limited increase in MB/sec throughput as compared to n-1). A point of interest is to determine a maximum MB/sec that a configuration can sustain. Knowing these limits allows for successful sizing of the storage resources to be able to meet the high throughput demands from the application. In one embodiment, a performance rate may be calculated by the following formula: MB/sec=(((C*B))/1000000)/s, where C is defined as the number of I/O operations completed, B is defined as the Block Size of each I.O in bytes, and s is defined as the duration of the measurement time in seconds.


If, at operation 325, the LDEV selected in operation 310 is an external LDEV, then control passes to operation 330, and the minimum number of paths to each external disk in the LDEV and to the array controller managing the LDEV is determined. This number is relevant in that a value of less than two represents storage which would not be considered as highly available. This number may be entered into the data table by the user at the time of external storage configuration. In one embodiment, a criteria for high availability is that no single point of failure causes data in the storage system to be inaccessible. So, a goal for a high availability configuration is to allow a user to have access to data stored in the storage product. Although a storage unit may be accessed through different paths, it is common that only a single path will be utilized at any particular point in time. One goal for high availability is not usually associated with general performance and has a different type of testing techniques which often includes such things as: 1) interface cable failures, 2) servers to be power-cycled during operation, and 3) disks failure. These types of failures are easily accomplished by physically removing an interface cable or disk and by turning off a server during a data integrity test. The testing philosophy for validating a high availability solution may focus on data integrity, where data is written and later read and checked to see if the retrieved data matches that which has been previously written. Storage performance, during this type of test, may not be related to a pass or fail criteria because the computer doing the test may be busy doing other tasks other than strict IO performance on the storage.


By contrast, if at operation 325 the LDEV selected in operation 310 is an internal LDEV, then control passes to operation 335 and the minimum number of paths to each internal disk and array controller is determined. For example, in the embodiment depicted in FIG. 2, there are two paths to each disk and to the array controllers that manage the LDEVs. The industry default for all internal disks would tend to be two paths, to provide redundancy. By contrast the number of paths to external storage can vary widely.


At operation 340 characteristics of the array controller managing and the disk array housing the LDEV selected in operation 310 are collected. In one embodiment, the storage controller determines the RAID level implemented in the LDEV, the RAID group with which the LDEV is associated, the size of the LDEV, and the size and type of the disk(s) housing the LDEV. In one embodiment, the storage controller maintains these characteristics in a data table established when LDEV is created. In this embodiment, the storage controller can retrieve these characteristics from the data table.


At operation 345 the results of the OLTP test initiated in operation 315 and the data warehouse test initiated in operation 320, the path information collected in operations 330, 335, and the characteristics collected in operation 345 are recorded in a suitable memory module. In one embodiment, the information collected in FIG. 3 may be stored in a memory table such as the memory table 400 depicted in FIG. 4. Referring to FIG. 4, the data table 400 includes a column 405 that includes an identifier for the LDEV, a column 410 that identifies the RAID type associated with the LDEV, a column 415 that identifies the RAID group with which the LDEV is associated, a column 420 that identifies the size of the LDEV, a column 425 that identifies the disk type and size, a column 430 that identifies whether the LDEV is internal or external, a column 435 that includes the results of the OLTP performance test, a column 440 that includes the results of the data warehouse test, a column 445 that includes the minimum number of physical paths to the disks, and a column 450 that includes the minimum number of physical paths to the controller.


Referring back to FIG. 3, at operation 350 the information in the memory table 400 may be forwarded to a reporting interface. In one embodiment, the reporting interface may include a user interface that presents the information to a user, e.g., on a suitable display. The user interface may further include logic instruction that permit a user to sort the data using one or more columns as a key. A user such as, e.g., a network administrator, may consult this information to make an informed judgment about which disk group(s) are good selections for provisioning a new LDEV as a host viewable logical unit (LU). Alternatively, the information in memory table 400 may be input to a software module that provisions LDEVs as host viewable LUs.


Reference in the specification to “one embodiment” or “an embodiment” means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least an implementation. The appearances of the phrase “in one embodiment” in various places in the specification are not necessarily all referring to the same embodiment.


Thus, although embodiments have been described in language specific to structural features and/or methodological acts, it is to be understood that claimed subject matter may not be limited to the specific features or acts described. Rather, the specific features and acts are disclosed as sample forms of implementing the claimed subject matter.

Claims
  • 1. A method, comprising: collecting, in a storage controller, performance data and availability data for a plurality of logical devices (LDEVS) managed by the storage controller, wherein the availability data is based, at least in part, on a number of communication paths to each of the plurality of logical devices; andpresenting the performance data and availability data to a reporting interface.
  • 2. The method of claim 1, further comprising using the performance data and the availability data to select a LDEV to provision as a host viewable LU.
  • 3. The method of claim 1, wherein performance data and availability data are collected on a periodic basis.
  • 4. The method of claim 1, wherein the performance data comprises at least one of results from an online transaction performance (OLTP) test or results from a data warehouse workload test.
  • 5. The method of claim 1, wherein the availability data comprises at least one of: a minimum number of physical paths to an external storage volume;a minimum number of paths per external controller;an array identifier; ora disk drive characteristic.
  • 6. The method of claim 1, wherein the availability data comprises at least one of: a disk drive characteristic; or a disk controller characteristic.
  • 7. The method of claim 1, further comprising using the performance data and the availability data as keys in a sorting routine implemented in the reporting interface.
  • 8. The method of claim 1, wherein the performance data comprises at least one of results from a performance test in which a predetermined number of blocks are written to and read from a logical device in a predetermined time period.
  • 9. A storage controller, comprising: a first I/O port that provides an interface to a host computer;a second I/O port that provides an interface a storage device; andmeans for collecting performance data and availability data for a plurality of logical devices (LDEVS) managed by the storage controller before assigning a logical unit number to a logical device, wherein the availability data is based, at least in part, on a number of communication paths to each of the plurality of logical devices.
  • 10. The storage controller of claim 9, further comprising means for presenting the performance data and availability data to a reporting interface.
  • 11. The storage controller of claim 9, further comprising means for using the performance data and the availability data to select a LDEV to provision as a host viewable LU.
  • 12. The storage controller of claim 9, wherein the performance data comprises at least one of results from an online transaction performance (OLTP) test or results from a data warehouse workload test.
  • 13. The storage controller of claim 9, wherein the availability data comprises at least one of: a minimum number of physical paths to an external storage volume;a minimum number of paths per external controller;an array identifier; ora disk drive characteristic.
  • 14. The storage controller of claim 9, wherein the availability data comprises at least one of: a disk drive characteristic; ora disk controller characteristic.
  • 15. The storage controller of claim 9, wherein the performance data comprises at least one of results from a performance test in which a predetermined number of blocks are written to and read from a logical device in a predetermined time period.
  • 16. A storage controller, comprising: a first I/O port that provides an interface to a host computer;a second I/O port that provides an interface a storage device;a processor that receives I/O requests generated by the host computer and, in response to the I/O requests, generates and transmits I/O requests to the storage device; anda memory module communicatively connected to the processor and comprising logic instructions which, when executed by the processor, configure the processor to: collect performance data and availability data for a plurality of logical devices (LDEVS) managed by the storage controller; andpresent the performance data and availability data to a reporting interface, wherein the availability data is based, at least in part, on a number of communication paths to each of the plurality of logical devices.
  • 17. The storage controller of claim 16, further comprising using the performance data and the availability data to select a LDEV to provision as a host viewable LU.
  • 18. The storage controller of claim 16, wherein performance data and availability data are collected on a periodic basis.
  • 19. The storage controller of claim 16, wherein the performance data comprises at least one of results from an online transaction performance (OLTP) test or results from a data warehouse workload test.
  • 20. The storage controller of claim 16, wherein the availability data comprises at least one of: a minimum number of physical paths to an external storage volume;a minimum number of paths per external controller;an array identifier; ora disk drive characteristic.
  • 21. The storage controller of claim 16, wherein the availability data comprises at least one of: a disk drive characteristic; ora disk controller characteristic.
  • 22. The storage controller of claim 16, further comprising logic instructions which, when executed by the processor, configure the processor to use the performance data and the availability data as keys in a sorting routine implemented in the reporting interface.
  • 23. The storage controller of claim 16, wherein the performance data comprises at least one of results from a performance test in which a predetermined number of blocks are written to and read from a logical device in a predetermined time period.
  • 24. A computer program product comprising logic instructions stored on a computer-readable medium which, when executed by a processor, configure the processor to: collect performance data and availability data for a plurality of logical devices (LDEVS) managed by a storage controller before assigning a logical unit number to a logical device, wherein the availability data is based, at least in part, on a number of communication paths to each of the plurality of logical devices; anduse the performance data and availability data to allocate a resources to provision a logical unit.
  • 25. The computer program product of claim 24, further comprising using the performance data and the availability data to select a LDEV to provision as a host viewable LU.
  • 26. The computer program product of claim 24, wherein performance data and availability data are collected on a periodic basis.
  • 27. The computer program product of claim 24, wherein the performance data comprises at least one of results from an online transaction performance (OLTP) test or results from a data warehouse workload test.
  • 28. The computer program product of claim 24, wherein the availability data comprises at least one of: a minimum number of physical paths to an external storage volume;a minimum number of paths per external controller; an array identifier; ora disk drive characteristic.
  • 29. The computer program product of claim 24, wherein the availability data comprises at least one of: a disk drive characteristic; or a disk controller characteristic.
  • 30. The computer program product of claim 24, further comprising logic instructions which, when executed by the processor, configure the processor to use the performance data and the availability data as keys in a sorting routine implemented in the reporting interface.
  • 31. The computer program product of claim 24, wherein the performance data comprises at least one of results from a performance test in which a predetermined number of blocks are written to and read from a logical device in a predetermined time period.
US Referenced Citations (103)
Number Name Date Kind
5923876 Teague Jul 1999 A
6161192 Lubbers Dec 2000 A
6170063 Golding Jan 2001 B1
6295578 Dimitroff Sep 2001 B1
6397293 Shrader May 2002 B2
6487636 Dolphin Nov 2002 B1
6490122 Holmquist et al. Dec 2002 B1
6493656 Houston Dec 2002 B1
6505268 Schultz Jan 2003 B1
6523749 Reasoner Feb 2003 B2
6546459 Rust Apr 2003 B2
6560673 Elliott May 2003 B2
6587962 Hepner Jul 2003 B1
6594745 Grover Jul 2003 B2
6601187 Sicola Jul 2003 B1
6606690 Padovano Aug 2003 B2
6609145 Thompson Aug 2003 B1
6629108 Frey Sep 2003 B2
6629273 Patterson Sep 2003 B1
6643795 Sicola Nov 2003 B1
6647514 Umberger Nov 2003 B1
6658590 Sicola Dec 2003 B1
6663003 Johnson Dec 2003 B2
6681308 Dallmann Jan 2004 B1
6708285 Oldfield Mar 2004 B2
6715101 Oldfield Mar 2004 B2
6718404 Reuter Apr 2004 B2
6718434 Veitch Apr 2004 B2
6721902 Cochran Apr 2004 B1
6725393 Pellegrino Apr 2004 B1
6742020 Dimitroff May 2004 B1
6745207 Reuter Jun 2004 B2
6763409 Elliott Jul 2004 B1
6772231 Reuter Aug 2004 B2
6775790 Reuter Aug 2004 B2
6795904 Kamvysselis Sep 2004 B1
6802023 Oldfield Oct 2004 B2
6807605 Umberger Oct 2004 B2
6817522 Brignone Nov 2004 B2
6823453 Hagerman Nov 2004 B1
6832299 Shimada et al. Dec 2004 B2
6839824 Camble Jan 2005 B2
6842833 Phillips Jan 2005 B1
6845403 Chadalapaka Jan 2005 B2
7272686 Yagisawa et al. Sep 2007 B2
20020019863 Reuter Feb 2002 A1
20020019908 Reuter Feb 2002 A1
20020019920 Reuter Feb 2002 A1
20020019922 Reuter Feb 2002 A1
20020019923 Reuter Feb 2002 A1
20020048284 Moulton Apr 2002 A1
20020188800 Tomaszewski Dec 2002 A1
20030051109 Cochran Mar 2003 A1
20030056038 Cochran Mar 2003 A1
20030063134 Lord Apr 2003 A1
20030074492 Cochran Apr 2003 A1
20030079014 Lubbers Apr 2003 A1
20030079074 Sicola Apr 2003 A1
20030079082 Sicola Apr 2003 A1
20030079083 Lubbers Apr 2003 A1
20030079102 Lubbers Apr 2003 A1
20030079156 Sicola Apr 2003 A1
20030084241 Lubbers May 2003 A1
20030101318 Kaga May 2003 A1
20030110237 Kitamura Jun 2003 A1
20030126315 Tan Jul 2003 A1
20030126347 Tan Jul 2003 A1
20030140191 McGowen Jul 2003 A1
20030145045 Pellegrino Jul 2003 A1
20030145130 Schultz Jul 2003 A1
20030170012 Cochran Sep 2003 A1
20030177323 Popp Sep 2003 A1
20030187847 Lubbers Oct 2003 A1
20030187947 Lubbers Oct 2003 A1
20030188085 Arakawa Oct 2003 A1
20030188114 Lubbers Oct 2003 A1
20030188119 Lubbers Oct 2003 A1
20030188153 Demoff Oct 2003 A1
20030188218 Lubbers Oct 2003 A1
20030188229 Lubbers Oct 2003 A1
20030188233 Lubbers Oct 2003 A1
20030191909 Asano Oct 2003 A1
20030191919 Sato Oct 2003 A1
20030196023 Dickson Oct 2003 A1
20030212781 Kaneda Nov 2003 A1
20030229651 Mizuno Dec 2003 A1
20030236953 Grieff Dec 2003 A1
20040019740 Nakayama Jan 2004 A1
20040022546 Cochran Feb 2004 A1
20040024838 Cochran Feb 2004 A1
20040024961 Cochran Feb 2004 A1
20040030727 Armangau Feb 2004 A1
20040030846 Armangau Feb 2004 A1
20040049634 Cochran Mar 2004 A1
20040078638 Cochran Apr 2004 A1
20040078641 Fleischmann Apr 2004 A1
20040128404 Cochran Jul 2004 A1
20040168034 Homma et al. Aug 2004 A1
20040215602 Cioccarelli Oct 2004 A1
20040230859 Cochran Nov 2004 A1
20040267959 Cochran Dec 2004 A1
20050044336 Shimada et al. Feb 2005 A1
20050071596 Aschoff et al. Mar 2005 A1
Foreign Referenced Citations (3)
Number Date Country
2001350707 Dec 2001 JP
2003296290 Oct 2003 JP
2003345514 Dec 2003 JP
Related Publications (1)
Number Date Country
20070094393 A1 Apr 2007 US