The field of the disclosure is data processing, or, more specifically, methods, apparatus, and products for compressing data in dependence upon characteristics of a storage system.
Enterprise storage systems frequently include a plurality of storage devices. Each of the storage devices may be capable of storing a particular amount of data, and as such, the storage system as a whole is characterized by the cumulative capacity of the storage devices that make up the storage system. In order to better utilize the capacity of the storage system, data reduction techniques are often applied to reduce the size of the data stored in the storage system. One such technique is data compression. Data compression, however, is frequently carried out in an unsophisticated, non-optimal manner.
Methods, apparatuses, and products for compressing data in dependence upon characteristics of a storage system, including: receiving an amount of processing resources available in the storage system; receiving an amount of space available in the storage system; and selecting, in dependence upon a priority for conserving the amount of processing resources and the amount of space, a data compression algorithm to utilize to compress the data.
The foregoing and other objects, features and advantages of the disclosure will be apparent from the following more particular descriptions of example embodiments of the disclosure as illustrated in the accompanying drawings wherein like reference numbers generally represent like parts of example embodiments of the disclosure.
Example methods, apparatuses, and products for compressing data in dependence upon characteristics of a storage system in accordance with the present disclosure are described with reference to the accompanying drawings, beginning with
The computing devices (164, 166, 168, 170) in the example of
The local area network (160) of
The example storage arrays (102, 104) of
Each storage array controller (106, 112) may be implemented in a variety of ways, including as a Field Programmable Gate Array (‘FPGA’), a Programmable Logic Chip (‘PLC’), an Application Specific Integrated Circuit (‘ASIC’), or computing device that includes discrete components such as a central processing unit, computer memory, and various adapters. Each storage array controller (106, 112) may include, for example, a data communications adapter configured to support communications via the SAN (158) and the LAN (160). Although only one of the storage array controllers (112) in the example of
Each write buffer device (148, 152) may be configured to receive, from the storage array controller (106, 112), data to be stored in the storage devices (146). Such data may originate from any one of the computing devices (164, 166, 168, 170). In the example of
A ‘storage device’ as the term is used in this specification refers to any device configured to record data persistently. The term ‘persistently’ as used here refers to a device's ability to maintain recorded data after loss of a power source. Examples of storage devices may include mechanical, spinning hard disk drives, Solid-state drives (e.g., “Flash drives”), and the like.
The storage array controllers (106, 112) of
The arrangement of computing devices, storage arrays, networks, and other devices making up the example system illustrated in
Compressing data in dependence upon characteristics of a storage system in accordance with embodiments of the present disclosure is generally implemented with computers. In the system of
The storage array controller (202) of
The storage array controller (202) of
Stored in RAM (214) is an operating system (246). Examples of operating systems useful in storage array controllers (202) configured for intelligently compressing data in a storage array according to embodiments of the present disclosure include UNIX™, Linux™, Microsoft Windows™, and others as will occur to those of skill in the art. Also stored in RAM (236) is a compression optimization module (248), a module that includes computer program instructions useful in compressing data in dependence upon characteristics of a storage system that includes a plurality of storage devices according to embodiments of the present disclosure.
The compression optimization module (248) may compress data in dependence upon characteristics of a storage system by: receiving an amount of processing resources available in the storage system, receiving an amount of space available in the storage system, and selecting, in dependence upon a priority for conserving the amount of processing resources and the amount of space, a data compression algorithm to utilize to compress the data, as will be described in greater detail below.
The compression optimization module (248) may further intelligently compress data in a storage array that includes a plurality of storage devices by: determining an expected amount of space in the storage system to be consumed within a predetermined period of time, determining an expected amount of processing resources to be consumed by compressing the data utilizing the data compression algorithm for each of a plurality of data compression algorithms, determining an expected amount of data reduction to be achieved by compressing the data utilizing the data compression algorithm for each of a plurality of data compression algorithms, determining, an expected decompression speed associated with decompressing the data for each of a plurality of data compression algorithms, determining an average decompression speed associated with decompressing a pool of data for each of a plurality of data compression algorithms, compressing at least a portion of the data utilizing a plurality of data compression algorithms, identifying a data reduction level achieved by each data compression algorithm, determining a decompression speed associated with each data compression algorithm, and selecting the data compression algorithm to utilize to compress the data in dependence upon the data reduction level achieved by each data compression algorithm and the decompression speed associated with each data compression algorithm, as will be described in greater detail below.
The storage array controller (202) of
The storage array controller (202) of
The storage array controller (202) of
The storage array controller (202) of
Readers will recognize that these components, protocols, adapters, and architectures are for illustration only, not limitation. Such a storage array controller may be implemented in a variety of different ways, each of which is well within the scope of the present disclosure.
For further explanation,
The example method depicted in
The example method depicted in
The example method depicted in
Readers will appreciate that the amount of space (324) available in the storage system (300) is distinct from the amount of available memory that may be included as part of the amount of processing resources (322) available in the storage system (300). The amount of space (324) available in the storage system (300) represents the amount of long-term, persistent storage that is available within the storage devices (302, 304, 306) that are included in the storage system (300). The amount of space (324) that is available in the storage system (300) may include not only space that is available in the traditional sense, but also space that is currently in use but may be reclaimed through the use of one or more data reduction techniques such as garbage collection. In contrast to the amount of space (324) available in the storage system (300), the amount of available memory that may be included as part of the amount of processing resources (322) available in the storage system (300) represents short-term, possibly volatile, memory that is available for use in performing data compression operations. The amount of available memory that may be included as part of the amount of processing resources (322) available in the storage system (300) may be embodied, for example, as DRAM within a storage array controller such as the storage array controllers described above with reference to
The example method depicted in
Such a priority may be dynamic as the value changes over time and such a priority may be calculated based on additional factors and could change at different times. For example, one factor that impacts the priority may be the current operating conditions of the system (e.g., processing resources currently available in the storage system, the amount of space currently available in the storage system, the rate at which consumption of each type of resource is accelerating or decelerating), another factor that impacts the priority may be historical operating conditions of the system (e.g., times and dates at which consumption of each type of resource spikes or decreases), another factor that impacts the priority could be system settings (e.g., a general preference to run out of processing resources rather than storage resources, or vice versa), and so on.
In the example method depicted in
Readers will appreciate that the predetermined formula may be configured to strike a balance between the amount of processing resources (322) available in the storage system (300) and the amount of space (324) available in the storage system. For example, when the amount of processing resources (322) available in the storage system (300) are relatively low, the compression prioritization module (314) may select (320) to compress the data (308, 310, 312) utilizing quick, lightweight compression algorithms that consume relatively small amounts of computing resources and also produce relatively small amounts of data reduction. Alternatively, when the amount of processing resources (322) available in the storage system (300) are relatively large, the compression prioritization module (314) may select (320) to compress the data (308, 310, 312) slower, heavier compression algorithms that consume relatively large amounts of computing resources and also produce relatively large amounts of data reduction. When the amount of space (324) available in the storage system is relatively low, however, the compression prioritization module (314) may select (320) to compress the data (308, 310, 312) using slower, heavier compression algorithms that consume relatively large amounts of computing resources and also produce relatively large amounts of data reduction, as a premium will be placed on data reduction rather than conservation of processing resources in such situations. Alternatively, when the amount of space (324) available in the storage system is relatively high, a premium will be placed conserving processing resources rather than data reduction in such situations.
For further explanation,
The example method depicted in
In the example method depicted in
In the example method depicted in
Readers will appreciate that when the expected amount of space (404) in the storage system (300) to be consumed within the predetermined period of time is relatively high, a data compression algorithm that can achieve higher levels of data reduction may be selected (406) even if such a data compression algorithm consumes relatively large amounts of processing resources, given that the available storage within the storage system (300) is being rapidly consumed. Alternatively, when the expected amount of space (404) in the storage system (300) to be consumed within the predetermined period of time is relatively low, a data compression algorithm that consumes relatively small amounts of processing resources may be selected (406) even if such a data compression algorithm only achieves relatively small levels of data reduction, given that the available storage within the storage system (300) is not being rapidly consumed. In such a way, data compression algorithms that achieve higher levels of data reduction may be selected (406) as the expected amount of space (404) in the storage system (300) to be consumed within the predetermined period of time increases, even if selecting such data compression algorithms requires a larger amount of processing resources to execute relative to data compression algorithms that achieve lower levels of data reduction.
For further explanation,
The example method depicted in
In the example method depicted in
In the example method depicted in
Readers will appreciate that when the expected amount of processing resources (504) to be consumed by compressing the data (308, 310, 312) utilizing a particular data compression algorithm is relatively high, such a data compression algorithm may only be selected (506) when the amount of processing resources (322) available in the storage system (300) is also relatively high, in order to avoid performing resource intensive compression when other consumers of processing resources are demanding a relatively high amount of processing resources. Alternatively, when the expected amount of processing resources (504) to be consumed by compressing the data (308, 310, 312) utilizing a particular data compression algorithm is relatively low, such a data algorithm may be selected (506) even when the amount of processing resources (322) available in the storage system (300) is relatively low, as performing such data compression is less likely to prevent other consumers of processing resources from accessing such processing resources. In such a way, data compression algorithms may be selected such that performing data compression does not cause other consumers of processing resources to be blocked from accessing such processing resources.
For further explanation,
The example method depicted in
In the example method depicted in
In the example method depicted in
Readers will appreciate that when the expected amount of data reduction (604) to be achieved by compressing the data (308, 310, 312) utilizing the data compression algorithm is relatively high, such a data compression algorithm may be selected (606) when the amount of free space within the storage system (300) is relatively low, even if executing such a data compression algorithm consumes relatively large amounts of processing resources. Alternatively, when the amount of free space within the storage system (300) is relatively high, less emphasis may be placed on selecting (606) a data compression algorithm whose expected amount of data reduction (604) is relatively high.
For further explanation,
The example method depicted in
In the example method depicted in
In the example method depicted in
Readers will appreciate that when the decompression speed (704) associated with decompressing the data (308, 310, 312) utilizing the data compression algorithm is relatively high, such a data compression algorithm may be selected (706) for data stored by applications that require relatively fast response times, even if utilizing such a data compression algorithm may result in relatively low levels of data reduction. Alternatively, data stored by applications that do not require relatively fast response times may be compressed utilizing data compression algorithms that deliver relatively high levels of data reduction, data compression algorithm that consume relatively low amounts of processing resources, and so on, even when the decompression speed (704) associated with decompressing the data (308, 310, 312) utilizing the data compression algorithm is relatively low.
For further explanation,
The example method depicted in
In the example method depicted in
In the example method depicted in
Readers will appreciate that when the average decompression speed (804) associated with decompressing a pool of data utilizing the data compression algorithm is relatively high, such a data compression algorithm may be selected (806) for data stored by applications that require relatively fast response times, even if utilizing such a data compression algorithm may result in relatively low levels of data reduction. Alternatively, data stored by applications that do not require relatively fast response times may be compressed utilizing data compression algorithms that deliver relatively high levels of data reduction, data compression algorithm that consume relatively low amounts of processing resources, and so on, even when average decompression speed (804) associated with decompressing a pool of data utilizing the data compression algorithm is relatively low.
For further explanation,
In the example method depicted in
In the example method depicted in
In the example method depicted in
In the example method depicted in
In the example method depicted in
In the example embodiments described above, a data compression algorithm (328) to utilize to compress data (308, 310, 312) is selected (320) based on factors such as (but not limited to): the priority for conserving the amount of processing resources (322) and the amount of space (324); the amount of processing resources available in the storage system; the amount of space available in the storage system; the expected amount of space in the storage system to be consumed within a predetermined period of time; the expected amount of processing resources to be consumed by compressing the data utilizing the data compression algorithm; the expected amount of data reduction to be achieved by compressing the data utilizing the data compression algorithm; the expected decompression speed associated with decompressing the data; and the average decompression speed associated with decompressing a pool of data. While the example embodiments described above typically describe selecting a data compression algorithm (328) to utilize to compress data (308, 310, 312) based on only a small subset of such factors, such a description is included only for ease of explanation. Readers will appreciate that embodiments are contemplated where a data compression algorithm (328) to utilize to compress data (308, 310, 312) is selected (320) based on combinations of factors not expressly identified above, including utilizing all of the factors described above to select a data compression algorithm (328) to utilize to compress data (308, 310, 312). In fact, such factors may be given equal or unequal consideration in various embodiments of the present disclosure. Furthermore, a data compression algorithm (328) to utilize to compress data (308, 310, 312) may be selected (320) in further dependence on additional factors as will occur to those of skill in the art in view of the present disclosure.
Example embodiments of the present disclosure are described largely in the context of a fully functional computer system. Readers of skill in the art will recognize, however, that the present disclosure also may be embodied in a computer program product disposed upon computer readable media for use with any suitable data processing system. Such computer readable storage media may be any transitory or non-transitory media. Examples of such media include storage media for machine-readable information, including magnetic media, optical media, or other suitable media. Examples of such media also include magnetic disks in hard drives or diskettes, compact disks for optical drives, magnetic tape, and others as will occur to those of skill in the art. Persons skilled in the art will immediately recognize that any computer system having suitable programming means will be capable of executing the steps of the method of the invention as embodied in a computer program product. Persons skilled in the art will recognize also that, although some of the example embodiments described in this specification are oriented to software installed and executing on computer hardware, nevertheless, alternative embodiments implemented as firmware, as hardware, or as an aggregation of hardware and software are well within the scope of embodiments of the present disclosure.
It will be understood from the foregoing description that modifications and changes may be made in various embodiments of the present disclosure without departing from its true spirit. The descriptions in this specification are for purposes of illustration only and are not to be construed in a limiting sense. The scope of the present disclosure is limited only by the language of the following claims.
This application is a continuation application of and claims priority from U.S. patent application Ser. No. 15/041,307, filed on Feb. 11, 2016, which is now U.S. Pat. No. 10,572,460 issued Feb. 25, 2020.
Number | Name | Date | Kind |
---|---|---|---|
5191548 | Balkanski et al. | Mar 1993 | A |
5686961 | Gasztonyi et al. | Nov 1997 | A |
5706210 | Kumano et al. | Jan 1998 | A |
5799200 | Brant et al. | Aug 1998 | A |
5918225 | White et al. | Jun 1999 | A |
5933598 | Scales et al. | Aug 1999 | A |
6012032 | Donovan et al. | Jan 2000 | A |
6085333 | DeKoning et al. | Jul 2000 | A |
6393584 | McLaren | May 2002 | B1 |
6643641 | Snyder | Nov 2003 | B1 |
6647514 | Umberger et al. | Nov 2003 | B1 |
6789162 | Talagala et al. | Sep 2004 | B1 |
7024414 | Sah | Apr 2006 | B2 |
7089272 | Garthwaite et al. | Aug 2006 | B1 |
7107389 | Inagaki et al. | Sep 2006 | B2 |
7123718 | Moskowitz et al. | Oct 2006 | B1 |
7146521 | Nguyen | Dec 2006 | B1 |
7334124 | Pham et al. | Feb 2008 | B2 |
7437530 | Rajan | Oct 2008 | B1 |
7493424 | Bali et al. | Feb 2009 | B1 |
7669029 | Mishra et al. | Feb 2010 | B1 |
7689609 | Lango et al. | Mar 2010 | B2 |
7743191 | Liao | Jun 2010 | B1 |
7899780 | Shmuylovich et al. | Mar 2011 | B1 |
8042163 | Karr et al. | Oct 2011 | B1 |
8086585 | Brashers et al. | Dec 2011 | B1 |
8140821 | Raizen et al. | Mar 2012 | B1 |
8200887 | Bennett | Jun 2012 | B2 |
8271700 | Annem et al. | Sep 2012 | B1 |
8369324 | Breight et al. | Feb 2013 | B1 |
8387136 | Lee et al. | Feb 2013 | B2 |
8422733 | Reisman | Apr 2013 | B2 |
8437189 | Montierth et al. | May 2013 | B1 |
8465332 | Hogan et al. | Jun 2013 | B2 |
8527544 | Colgrove et al. | Sep 2013 | B1 |
8566546 | Marshak et al. | Oct 2013 | B1 |
8578442 | Banerjee | Nov 2013 | B1 |
8613066 | Brezinski et al. | Dec 2013 | B1 |
8620970 | English et al. | Dec 2013 | B2 |
8751463 | Chamness | Jun 2014 | B1 |
8762642 | Bates et al. | Jun 2014 | B2 |
8769622 | Chang et al. | Jul 2014 | B2 |
8800009 | Beda et al. | Aug 2014 | B1 |
8812860 | Bray | Aug 2014 | B1 |
8850546 | Field et al. | Sep 2014 | B1 |
8898346 | Simmons | Nov 2014 | B1 |
8909854 | Yamagishi et al. | Dec 2014 | B2 |
8931041 | Banerjee | Jan 2015 | B1 |
8949863 | Coatney et al. | Feb 2015 | B1 |
8984602 | Bailey et al. | Mar 2015 | B1 |
8990905 | Bailey et al. | Mar 2015 | B1 |
9081713 | Bennett et al. | Jul 2015 | B1 |
9124569 | Hussain et al. | Sep 2015 | B2 |
9134922 | Rajagopal et al. | Sep 2015 | B2 |
9189334 | Bennett et al. | Nov 2015 | B2 |
9209973 | Aikas et al. | Dec 2015 | B2 |
9235590 | Wang | Jan 2016 | B1 |
9250823 | Kamat et al. | Feb 2016 | B1 |
9300660 | Borowiec et al. | Mar 2016 | B1 |
9311182 | Bennett | Apr 2016 | B2 |
9444822 | Borowiec et al. | Sep 2016 | B1 |
9507532 | Colgrove et al. | Nov 2016 | B1 |
9632870 | Bennett et al. | Apr 2017 | B2 |
10769016 | Leggette | Sep 2020 | B2 |
20020013802 | Mori et al. | Jan 2002 | A1 |
20030056146 | Freeman | Mar 2003 | A1 |
20030145172 | Galbraith et al. | Jul 2003 | A1 |
20030191783 | Wolczko et al. | Oct 2003 | A1 |
20030225961 | Chow et al. | Dec 2003 | A1 |
20040080985 | Chang et al. | Apr 2004 | A1 |
20040111573 | Garthwaite | Jun 2004 | A1 |
20040153844 | Ghose et al. | Aug 2004 | A1 |
20040193814 | Erickson et al. | Sep 2004 | A1 |
20040260967 | Guha et al. | Dec 2004 | A1 |
20050160416 | Jamison et al. | Jul 2005 | A1 |
20050188246 | Emberty et al. | Aug 2005 | A1 |
20050216800 | Bicknell et al. | Sep 2005 | A1 |
20050288836 | Glass et al. | Dec 2005 | A1 |
20060015771 | van Gundy et al. | Jan 2006 | A1 |
20060129817 | Borneman et al. | Jun 2006 | A1 |
20060161726 | Lasser | Jul 2006 | A1 |
20060230245 | Gounares et al. | Oct 2006 | A1 |
20060239075 | Williams et al. | Oct 2006 | A1 |
20070022227 | Miki | Jan 2007 | A1 |
20070028068 | Golding et al. | Feb 2007 | A1 |
20070038738 | Iyengar | Feb 2007 | A1 |
20070055702 | Fridella et al. | Mar 2007 | A1 |
20070109856 | Pellicone et al. | May 2007 | A1 |
20070150689 | Pandit et al. | Jun 2007 | A1 |
20070168321 | Saito et al. | Jul 2007 | A1 |
20070220227 | Long | Sep 2007 | A1 |
20070294563 | Bose | Dec 2007 | A1 |
20070294564 | Reddin et al. | Dec 2007 | A1 |
20080005587 | Ahlquist | Jan 2008 | A1 |
20080050029 | Bashyam et al. | Feb 2008 | A1 |
20080077825 | Bello et al. | Mar 2008 | A1 |
20080162674 | Dahiya | Jul 2008 | A1 |
20080195833 | Park | Aug 2008 | A1 |
20080253330 | Bartlett | Oct 2008 | A1 |
20080270678 | Cornwell et al. | Oct 2008 | A1 |
20080282045 | Biswas et al. | Nov 2008 | A1 |
20090077340 | Johnson et al. | Mar 2009 | A1 |
20090100115 | Park et al. | Apr 2009 | A1 |
20090198889 | Ito et al. | Aug 2009 | A1 |
20100052625 | Cagno et al. | Mar 2010 | A1 |
20100211723 | Mukaida | Aug 2010 | A1 |
20100246266 | Park et al. | Sep 2010 | A1 |
20100257142 | Murphy et al. | Oct 2010 | A1 |
20100262764 | Liu et al. | Oct 2010 | A1 |
20100325345 | Ohno et al. | Dec 2010 | A1 |
20100332754 | Lai et al. | Dec 2010 | A1 |
20110072290 | Davis et al. | Mar 2011 | A1 |
20110106769 | Baptist | May 2011 | A1 |
20110125955 | Chen | May 2011 | A1 |
20110131231 | Haas et al. | Jun 2011 | A1 |
20110167221 | Pangal et al. | Jul 2011 | A1 |
20120023144 | Rub | Jan 2012 | A1 |
20120054264 | Haugh et al. | Mar 2012 | A1 |
20120079318 | Colgrove et al. | Mar 2012 | A1 |
20120131253 | McKnight et al. | May 2012 | A1 |
20120254417 | Luna | Oct 2012 | A1 |
20120303919 | Hu et al. | Nov 2012 | A1 |
20120311000 | Post et al. | Dec 2012 | A1 |
20130007845 | Chang et al. | Jan 2013 | A1 |
20130031414 | Dhuse et al. | Jan 2013 | A1 |
20130036272 | Nelson | Feb 2013 | A1 |
20130071087 | Motiwala et al. | Mar 2013 | A1 |
20130145447 | Maron | Jun 2013 | A1 |
20130191555 | Liu | Jul 2013 | A1 |
20130198459 | Joshi et al. | Aug 2013 | A1 |
20130205067 | Kettner | Aug 2013 | A1 |
20130205173 | Yoneda | Aug 2013 | A1 |
20130219164 | Hamid | Aug 2013 | A1 |
20130227201 | Talagala et al. | Aug 2013 | A1 |
20130275396 | Condict | Oct 2013 | A1 |
20130290607 | Chang et al. | Oct 2013 | A1 |
20130311434 | Jones | Nov 2013 | A1 |
20130318297 | Jibbe et al. | Nov 2013 | A1 |
20130332614 | Brunk et al. | Dec 2013 | A1 |
20140020083 | Fetik | Jan 2014 | A1 |
20140074850 | Noel et al. | Mar 2014 | A1 |
20140082715 | Grajek et al. | Mar 2014 | A1 |
20140086146 | Kim et al. | Mar 2014 | A1 |
20140090009 | Li et al. | Mar 2014 | A1 |
20140096220 | Pinto et al. | Apr 2014 | A1 |
20140101434 | Senthurpandi et al. | Apr 2014 | A1 |
20140164774 | Nord et al. | Jun 2014 | A1 |
20140173232 | Reohr et al. | Jun 2014 | A1 |
20140195636 | Karve et al. | Jul 2014 | A1 |
20140201512 | Seethaler et al. | Jul 2014 | A1 |
20140201541 | Paul et al. | Jul 2014 | A1 |
20140208155 | Pan | Jul 2014 | A1 |
20140215590 | Brand | Jul 2014 | A1 |
20140229654 | Goss et al. | Aug 2014 | A1 |
20140230017 | Saib | Aug 2014 | A1 |
20140237201 | Swift | Aug 2014 | A1 |
20140258526 | Le Sant et al. | Sep 2014 | A1 |
20140282983 | Ju et al. | Sep 2014 | A1 |
20140285917 | Cudak et al. | Sep 2014 | A1 |
20140325262 | Cooper et al. | Oct 2014 | A1 |
20140347331 | Amit et al. | Nov 2014 | A1 |
20140351627 | Best et al. | Nov 2014 | A1 |
20140373104 | Gaddam et al. | Dec 2014 | A1 |
20140373126 | Hussain et al. | Dec 2014 | A1 |
20150026387 | Sheredy et al. | Jan 2015 | A1 |
20150074463 | Jacoby et al. | Mar 2015 | A1 |
20150089569 | Sondhi et al. | Mar 2015 | A1 |
20150095515 | Krithivas et al. | Apr 2015 | A1 |
20150113203 | Dancho et al. | Apr 2015 | A1 |
20150121137 | McKnight et al. | Apr 2015 | A1 |
20150134920 | Anderson et al. | May 2015 | A1 |
20150149822 | Coronado et al. | May 2015 | A1 |
20150193169 | Sundaram et al. | Jul 2015 | A1 |
20150293896 | Runkis et al. | Oct 2015 | A1 |
20150339059 | Kang et al. | Nov 2015 | A1 |
20150347443 | Reid et al. | Dec 2015 | A1 |
20150358537 | Mirza | Dec 2015 | A1 |
20150378888 | Zhang et al. | Dec 2015 | A1 |
20160098323 | Mutha et al. | Apr 2016 | A1 |
20160350009 | Cerreta et al. | Dec 2016 | A1 |
20160352720 | Hu et al. | Dec 2016 | A1 |
20160352830 | Borowiec et al. | Dec 2016 | A1 |
20160352834 | Borowiec et al. | Dec 2016 | A1 |
20190073306 | Kowles | Mar 2019 | A1 |
Number | Date | Country |
---|---|---|
0725324 | Aug 1996 | EP |
3414652 | Dec 2018 | EP |
2012087648 | Jun 2012 | WO |
2013071087 | May 2013 | WO |
2014110137 | Jul 2014 | WO |
2016015008 | Jan 2016 | WO |
2016190938 | Dec 2016 | WO |
2016195759 | Dec 2016 | WO |
2016195958 | Dec 2016 | WO |
2016195961 | Dec 2016 | WO |
2017138990 | Aug 2017 | WO |
Entry |
---|
Bellamy-McIntyre et al., “OpenID and the Enterprise: A Model-based Analysis of Single Sign-On Authentication”, 15th IEEE International Enterprise Distributed Object Computing Conference (EDOC), Aug. 29, 2011, pp. 129-138, IEEE Computer Society, USA, DOI: 10.1109/EDOC.2011.26, ISBN: 978-1-4577-0362-1. |
ETSI, “Network Function Virtualisation (NFV); Resiliency Requirements”, ETSI GS NFCV-REL 001, V1.1.1, Jan. 2015, 82 pages, etsi.org (online), URL: www.etsi.org/deliver/etsi_gs/NFV-REL/001 099/001/01.01.01_60/gs_NFV-REL001v010101p.pdf. |
Faith, “dietzip file format”, GitHub.com (online), accessed Jul. 28, 2015, 1 page, URL: github.com/fidlej/idzip. |
Google Search of “storage array define” performed by the Examiner on Nov. 4, 2015 for U.S. Appl. No. 14/725,278, Results limited to entries dated before 2012, 1 page. |
Hota et al., “Capability-based Cryptographic Data Access Control in Cloud Computing”, International Journal of Advanced Networking and Applications, col. 1, Issue 1, Aug. 2011, 10 pages, Eswar Publications, India. |
Hu et al., “Container Marking: Combining Data Placement, Garbage Collection and Wear Levelling for Flash”, 19th Annual IEEE International Symposium on Modelling, Analysis, and Simulation of Computer and Telecommunications Systems, Jul. 25-27, 2011, 11 pages, ISBN: 978-0-7695-4430-4, DOI: 10.1109/MASCOTS.2011.50. |
International Search Report and Written Opinion, PCT/US2016/015006, dated Apr. 29, 2016, 12 pages. |
International Search Report and Written Opinion, PCT/US2016/015008, dated May 4, 2016, 12 pages. |
International Search Report and Written Opinion, PCT/US2016/016333, dated Jun. 8, 2016, 12 pages. |
International Search Report and Written Opinion, PCT/US2016/020410, dated Jul. 8, 2016, 12 pages. |
International Search Report and Written Opinion, PCT/US2016/032052, dated Aug. 30, 2016, 17 pages. |
International Search Report and Written Opinion, PCT/US2016/032084, dated Jul. 18, 2016, 12 pages. |
International Search Report and Written Opinion, PCT/US2016/035492, dated Aug. 17, 2016, 10 pages. |
International Search Report and Written Opinion, PCT/US2016/036693, dated Aug. 29, 2016, 10 pages. |
International Search Report and Written Opinion, PCT/US2016/038758, dated Oct. 7, 2016, 10 pages. |
International Search Report and Written Opinion, PCT/US2016/040393, dated Sep. 22, 2016, 10 pages. |
International Search Report and Written Opinion, PCT/US2016/044020, dated Sep. 30, 2016, 11 pages. |
International Search Report and Written Opinion, PCT/US2016/044874, dated Oct. 7, 2016, 11 pages. |
International Search Report and Written Opinion, PCT/US2016/044875, dated Oct. 5, 2016, 13 pages. |
International Search Report and Written Opinion, PCT/US2016/044876, dated Oct. 21, 2016, 12 pages. |
International Search Report and Written Opinion, PCT/US2016/044877, dated Sep. 29, 2016, 13 pages. |
International Search Report and Written Opinion, PCT/US2016/059787, dated Jan. 17, 2017, 10 pages. |
Kong, “Using PCI Express As The Primary System Interconnect In Multiroot Compute, Storage, Communications And Embedded Systems”, White Paper, IDT.com (online), Aug. 28, 2008, 12 pages, URL: www.idt.com/document/whp/idt-pcie-multi-root-white-paper. |
Li et al., “Access Control for the Services Oriented Architecture”, Proceedings of the 2007 ACM Workshop on Secure Web Services (SWS '07), Nov. 2007, pp. 9-17, ACM New York, NY. |
Microsoft, “Hybrid for SharePoint Server 2013—Security Reference Architecture”, Microsoft (online), Oct. 2014, 53 pages, URL: hybrid.office.com/img/Security_Reference_Architecture.pdf. |
Microsoft, “Hybrid Identity Management”, Microsoft (online), Apr. 2014, 2 pages, URL: download.microsoft.com/download/E/A/E/EAE57CD1-A80B-423C-96BB-142FAAC630B9/Hybrd_Identity_Datasheet.pdf. |
Microsoft, “Hybrid Identity”, Microsoft (online), Apr. 2014, 36 pages, URL: www.aka.ms/HybridIdentityWp. |
Paul Sweere, “Creating Storage Class Persistent Memory with NVDIMM”, Published in Aug. 2013, Flash Memory Summit 2013, 22 pages. |
PCMAG, “Storage Array Definition”, Published May 10, 2013, 2 pages. |
Storer et al., “Secure Data Deduplication”, Proceedings of the 4th ACM International Workshop on Storage Security and Survivability (StorageSS'08), Oct. 2008, 10 pages, ACM New York, NY. USA, DOI 10.1145/1456469.1456471. |
Techopedia, “What is a disk array”, techopedia.com (online), Jan. 13, 2012, 1 page, URL: web.archive.org/web/20120113053358/http://www.techopedia.com/definition/1009/disk-array. |
Webopedia, “What is a disk array”, webopedia.com (online), May 26, 2011, 2 pages, URL: web/archive.org/web/20110526081214/http://www.webopedia.com/TERM/D/disk_array html. |
Wikipedia, “Convergent Encryption”, Wikipedia.org (online), accessed Sep. 8, 2015, 2 pages, URL: en. wikipedia.org/wiki/Convergent_encryption. |
Number | Date | Country | |
---|---|---|---|
Parent | 15041307 | Feb 2016 | US |
Child | 16743024 | US |