The present invention relates to modeling temperature distributions in an indoor environment, such as a data center, and more particularly, to techniques for real-time modeling temperature distributions based on streaming sensor data.
The energy consumption of data centers has dramatically increased in recent years, primarily because of the massive computing demands driven essentially by every sector of the economy, ranging from accelerating online sales in the retail business to banking services in the financial industry. A study estimated the total U.S. DC energy consumption in the year 2005 to be approximately 1.2% of the total U.S. consumption (up by 15% from the year 2000). See, for example, “EPA Report to Congress on Server and Data Center Energy Efficiency” Public Law 109-431, United States Code, Aug. 2, 2007.
In order to improve data center energy efficiency, it is important to be able to accurately assess the temperature distributions with the data center. That way cooling systems can be effectively implemented with greater efficiency. Further, changing conditions within the data center make it desirable to be able to determine the temperature distributions in a timely manner, in order to continually maintain an efficient cooling operation within the data center.
Thus, techniques for real-time modeling temperature distributions would be desirable.
The present invention provides techniques for real-time modeling temperature distributions based on streaming sensor data. In one aspect of the invention, a method for creating a three-dimensional temperature distribution model for a room having a floor and a ceiling is provided. The method includes the following steps. A ceiling temperature distribution in the room is determined. A floor temperature distribution in the room is determined. An interpolation between the ceiling temperature distribution and the floor temperature distribution is used to obtain the three-dimensional temperature distribution model for the room.
A more complete understanding of the present invention, as well as further features and advantages of the present invention, will be obtained by reference to the following detailed description and drawings.
with w=10 (1.67 feet) according to an embodiment of the present invention;
with w=5 (0.83 feet) according to an embodiment of the present invention;
Provided herein are techniques for real-time modeling temperature distributions based on streaming sensor data. Also described herein is how this method can be benchmarked against “detailed” temperature measurements to improve the model accuracy. It is notable that while the present techniques are described in the context of data center temperature modeling, the fundamental concepts can be applied to any indoor space, including but not limited to buildings, rooms, residential and commercial dwellings, etc. in the same manner as described.
Traditionally, computational fluid dynamics (CFD) methods are used to model such temperature distributions solving a complex system of coupled, partial differential equations (PDEs). Although CFD methods have been successfully deployed to data centers or buildings they come with long computation times. In fact, it is not often clear whether the “quality” of the input data required to specify the boundaries of the PDEs warrants such detailed computation. Because of these large computation times, some other methods have been explored where temperature distributions are simply obtained by “interpolation” from real-time sensor data. Such techniques may involve inverse distance interpolation, kriging methods and/or proper orthogonal decomposition. Although such techniques are much faster than CFD methods they lack the physical insights and accuracy as well as a way to benchmark and improve their accuracy.
In a first aspect of the present techniques a much faster method for modeling temperature distributions is described.
In
The ACUs typically receive chilled water from a refrigeration chiller plant (not shown), also referred to herein simply as a “chiller.” Each ACU typically includes a blower motor to circulate air through the ACU and to blow cooled air, e.g., into the sub-floor plenum. As such, in most data centers, the ACUs are simple heat exchangers mainly consuming power needed to blow the cooled air into the sub-floor plenum. Coolant distribution units (CDUs) (not shown) can be employed at the interface between the chiller and the ACUs. In general, a CDU includes a heat exchanger and one or more circulating pumps to circulate the chilled water through the ACUs. Thus, as will be described in detail below, the CDUs contribute to the overall power consumption in the data center.
Typically, one or more power distribution units (PDUs) (not shown) are present that distribute power to the IT equipment racks 101. As will be described in detail below, power consumption by the PDUs can be an important consideration in the present techniques. In general, since the PDUs supply electrical power required by the IT equipment in a data center, a total electrical power intake of the PDUs represents an important parameter in determining the energy efficiency of a data center. According to an exemplary embodiment, each of the PDUs is outfitted with commercially available power and current sensors which measure the electric power drawn by each of the PDUs.
Uninterruptable power supplies or UPS (not shown) are also typically implemented in a data center to protect the IT equipment in the event of a power disruption so as to prevent data loss (i.e., UPS provides short term power when the power source fails). As is known in the art, the UPSs might also correct common utility power issues, such as voltage spikes.
The pressure differential between the “pressurized” sub-floor plenum and the raised floor is measured with a pressure sensor(s) (see sensor labeled “pressure sensor” in
As is apparent from the description of
In contrast to other techniques, in the present techniques the interpolation solution is bound to obey energy balance. The steps of the present techniques will be described in detail in conjunction with the description of
While so far only the mean ceiling temperature has been determined, a more detailed ceiling temperature distribution is next calculated by leveraging the inlet thermal sensors distributed at one or more of the inlets of each rack. In order to estimate the exhaust temperatures for each rack, virtual sensors may optionally be placed at the “exhaust” side of the rack opposite to the inlet thermal sensors. With virtual sensors a knowledge base is used to “estimate” the difference between inlet and exhaust temperatures—see
Using the real and virtual thermal sensors, a 2D inverse distance interpolation is then used to calculate the ceiling temperature distribution. In order to account for the fact that the different sensors are at different heights in the data center (based, for example, on the placement of the sensor on the rack) and thus influence the ceiling temperature to a greater or lesser extent, each sensor is weighted by its distance to the ceiling. The dependence is accomplished with a “damping” factor as described in further detail below. Once the ceiling temperature distribution has been estimated, the mean value of that distribution is adjusted so that it obeys the principle of energy balance as previously established.
Next, the floor temperature distribution is estimated. In order to do so, the air flow through each perforated tile is first calculated using the tile impedance of each tile and the pressure differential at the tile location. The plenum pressure distribution is obtained by 2D inverse distance interpolation of the real-time pressure sensors sensing the pressure differential between the raised floor and the plenum at various locations in the data center. In the example shown in
Finally, from the ceiling and floor temperature distributions, a three dimensional temperature distribution for the data center room can be obtained by interpolating between these values using different approaches. One approach includes using an s-curve. S-curves are described, for example, in U.S. patent application Ser. No. 12/540,213 filed by Hamann et al., entitled “Knowledge-Based Models for Data Centers,” the contents of which are incorporated by reference herein. In the instant description, a simple linear interpolation between the floor and the ceiling temperature is employed.
In a second aspect of the present techniques, a method is described for improving the model described in the first aspect above by using detailed temperature distribution measurements from the MMT tool/cart. MMT is a technology for optimizing data center infrastructures for improved energy and space efficiency which involves a combination of advanced metrology techniques for rapid measuring/surveying data centers (see, for example, U.S. Pat. No. 7,366,632, issued to Hamann et al., entitled “Method and Apparatus for Three-Dimensional Measurements,” the contents of which are incorporated by reference herein) and physics-based modeling techniques for optimizing a data center facility within a given thermal envelope for optimum space and most-efficient energy utilization (see, for example, U.S. Application Publication Number 2008/0288193 A1, filed by Claassen et al., entitled “Techniques for Analyzing Data Center Energy Utilization Practices,” the contents of which are incorporated by reference herein. The MMT cart data can be used to fit parameters as described immediately below. Basically, the goal is to optimize/fine-tune the process by finding the parameters (see below) that give the best fit to the data obtained from the MMT cart.
The parameters to fit include 1) the width of the Lorentzian fields (which can be a function of the perforated air flow), 2) the temperature increase across the rack inlet and exhaust (which can be a function of the height in the data center, equipment type, etc.) and 3) a damping parameter, which determines the weights of the thermal sensors as a function of height. These parameters can be adjusted to minimize the error between the interpolated three dimensional temperature distribution for the data center room and the detailed temperature distribution measurements from the MMT tool/cart.
The present techniques are now described in detail by way of reference to the description of
To begin the process, information is gathered about the data center (DC) and the elements (ACUs, PDUs, CDUs, sensors, perforated tiles, etc.) within the data center. See, for example, steps 302-330 which are now described. According to an exemplary embodiment, in step 302, the area of the data center floor, i.e., floorArea: ADC, is determined. The data center floor area can be determined by one of skill in the art given the physical dimensions of the data center. For instance, given the exemplary data center shown in
Next, in step 304, power consumption data is obtained from the data center. According to an exemplary embodiment, the following power consumption data is obtained:
In step 306, the following data is obtained regarding the ACUs and PDUs:
In step 308, real-time data is obtained from each of the sensors deployed throughout the data center. As provided above, the sensors employed in the data center can include, but are not limited to, inlet and discharge thermal sensors, flow sensors, pressure sensors, and power/current sensors. As will be described in detail below, a weight (i.e., wn weight of nth sensor) may be applied to each sensor to account for the fact that different sensors are at different heights in the data center. Thus, each sensor is weighted by its distance to the ceiling.
Air flow rates through the perforated tiles in the raised floor depend on the flow resistance (or flow impedance) of the tiles. Thus, the flow impedance for each tile is obtained:
where At is the area of the tile and K is a loss coefficient. The loss coefficient can be determined by:
wherein o is the fractional opening or perforation of the tile. See, for example, Yogendra Joshi and Pramod Kumar (eds.), “Energy Efficient Thermal Management of Data Centers,” 2012, DOI: 10.1007/978-1-4419-7124-1, the contents of which are incorporated by reference herein. The perforated tiles in the data center might be fitted with dampers which can be adjusted to regulate air flow. Dampers increase the impedance. Thus, in that case, a very large impedance (e.g., 10 times more impedance) is applied.
As will be described in detail below, a floor temperature distribution in the data center will be determined based on convolution of a plenum temperature distribution and a perforated tile air flow distribution. As is known in the art, convolution is an operation performed on two functions that results in a third function that is a modified version of one of the original functions. With convolution, masks (convolution masks) are applied to the function that ‘translate’ the function. Convolution masks are commonly used in image filtering to enhance image quality. The convolution masks typically include a matrix of values or weights, which represents the second function.
Thus, in step 310, plenum, raised floor (RF) and ceiling masks are created. The information needed to create the mask is the data center dimensions and the location/dimensions of all the structures (e.g., pillars), equipment (e.g., racks), furniture etc. in the data center. An exemplary plenum mask is shown in
As described above, optionally, virtual sensors may be employed in accordance with the present techniques at the “exhaust”/outlet side of the rack opposite to the inlet thermal sensors. These sensors will also be referred to herein as “virtual (outlet) sensors.” This enables one to estimate the exhaust temperatures for each rack. With virtual sensors a knowledge base can be used to “estimate” the difference between inlet and exhaust temperatures. An exemplary virtual sensor framework that may be implemented in accordance with the present techniques is described, for example, in W. Minker et al. (eds.) “Advanced Intelligent Environments,” Springer Dordrecht (2009) (hereinafter “Minker”), the contents of which are incorporated by reference herein. For example, in section 3.2, FIG. 2 of Minker a virtual sensor framework architecture is shown which includes a knowledge base and a framework controller. The knowledge base manages information related to the virtual sensors and the framework controller determines which virtual sensors need to be created based on data from the knowledge base.
According to the present techniques, in step 312, the placement of the virtual (outlet) sensors (creating virtual sensor locations) will be based on the location of the inlet thermal sensors, see above. Namely, it is preferable to create a virtual (outlet) sensor at the exhaust side of a rack opposite a physical inlet sensor. That way the exhaust temperatures for each rack can be determined. Thus, according to an exemplary embodiment, the first step in creating the virtual (outlet) sensor locations is to create an equipment index. The equipment index is basically a mapping of all of the equipment (e.g., racks) in the data center. See for example
The real-time data collected from the physical sensors (see description of step 308, above) may or may not be consistent with the model description of a sensor. For example, MMT has a data model which has data for every sensor like its location, its type and other description. It is possible that someone physically puts in a sensor but forgets to update the data model. In the case of inconsistencies, sensor filtering may be performed. For instance, some sensors may be unbound meaning that a real-time value is obtained for a sensor but a corresponding sensor description is not found in the data model. For example, it may be unknown what type of sensor it is and where it is located.
Another inconsistency that might occur is when a sensor(s), for which there is an entry in the data model, may not transmit a real-time value. By way of example only, this might occur because the sensor may be faulty, someone might have removed it or disconnected its power/battery, a network/IT issue may have occurred. Such sensors are classified as “missing.” A further distinction may be made herein between those “missing” sensors which are permanently missing (i.e., those sensors which provide no value at all) and those that are temporarily missing (i.e., sensors which provide data but after a specified time out period has ended). These sensors which provide data but after a specified time out period has ended are also referred to herein as “out of time sensors,” or “timed out” sensors and as indicated immediately above are considered temporarily missing.
Yet another inconsistency is with regard to sensors which are out of range. This can be caused by “bit errors,” which can yield non-physical values. These “out of range” sensors can be identified by specifying an expected range, and their values can be removed from consideration.
It is notable that, as described above, power consumption by the PDUs can be determined by monitoring power values from each PDU. Alternatively, current values may be monitored for each of three inputs (I) of a PDU (i.e., three phase units will have three values for current). These current values can then be used to determine the power consumption associated with this particular PDU using:
PPDUi=(I1i+I2i+I3i)/3·208V·√{square root over (3)}·PF,
where PF is the power factor (PF˜0.9).
Taking the above-described sensor inconsistencies into account, the number (#) of real-time sensors (RTS) (i.e., sensors for which real-time values are present) is specified based on the number of model sensors (MS), the number of unbound sensors (see above) and the number of missing sensors (see above) as follows:
#RTS=#MS+#unbound−#missing.
The model sensors (MS) are those sensors which are specified in the model. See for example
Given the data collected from the sensors, in step 314, the mean plenum temperature Tpmean is calculated. According to an exemplary embodiment, the mean plenum temperature Tpmean is calculated by first using inverse weighted distance (IWD) to determine a plenum temperature distribution Tp(x,y) based on the plenum temperature sensor readings. The plenum temperature readings can be taken from the thermal sensors (discharge) and/or from the plenum thermal sensors, if available (see description of
The plenum temperature distribution (See
Another factor needed to calculate the ceiling temperature is the heat load in the data center. Thus, in step 316, the heat load in the data center (also referred to herein as the raised floor or RF) PRF is calculated. This heat load PRF calculation takes into account the power consumption of the PDUs PPDUi, the data center floor area ADC, the blower power Pblowerj and blower settings θblowerj, miscellaneous power consumption Pmisc and power consumption by the CDUs PCDU (all of which were obtained as described above) as follows:
Yet another factor needed to calculate the ceiling temperature is the ACU air flow in the data center (i.e., the air flow in the data center attributable to the ACUs/blowers). Thus, in step 318, the ACU air flow φACUtotal in the data center is calculated. This ACU air flow φACUtotal calculation takes into account the flow capacity of the ACUs φACU,0j and the blower settings θblowerj as follows:
From the mean plenum temperature Tpmean (calculated in step 314), heat load in the data center PRF (calculated in step 316) and the ACU air flow φACUtotal in the data center (calculated in step 318), the mean ceiling temperature Tcmean can be calculated in step 320 as follows:
Tcmean=3140 [cfm·F/kW]·PRF/φACUtotal+Tpmean.
While so far only the mean ceiling temperature has been determined, a more detailed ceiling temperature distribution is next calculated by leveraging the inlet thermal sensors distributed at one or more of the inlets of each rack measured using the physical and virtual sensors. The mean ceiling temperature is calculated based on energy balance and it is important to know this number because the methodology aims to preserve energy balance. The detailed ceiling temperature distribution is “scaled” such that its mean is the same as the calculated mean ceiling temperature. Specifically, in step 322, a weight of each sensor is calculated. As described above, in order to account for the fact that the different sensors are at different heights in the data center (based, for example, on the placement of the sensor on the rack) and thus influence the ceiling temperature to a greater or lesser extent, each thermal sensor (inlet) and virtual (outlet) sensor is weighted by its distance from the ceiling. The dependence is accomplished with a damping constant k. According to an exemplary embodiment, the weight of each sensor n is calculated as follows:
wRFn=exp(−(dz−zRFn)/k),
with k being the damping constant. Exemplary ceiling temperature distributions are shown in
Based on the inlet thermal sensor readings, IWD is used to calculate a ceiling temperature distribution Tc(x,y). Sensor weights are used in IWD. In this case the weight is w_rf/distance. IWD was described above. Exemplary ceiling temperature distribution/array are shown in
The ceiling temperature distribution (see
Next, the floor temperature distribution is determined. In order to do so, the air flow through each perforated tile is first calculated using the tile impedance of each tile and the pressure differential at the tile location. According to an exemplary embodiment, in step 324, based on the plenum pressure sensor readings (obtained, e.g., in step 308, see above), IWD is used to calculate the plenum pressure distribution PP(x,y). As provided above, IWD is a standard interpolation technique. An exemplary plenum pressure distribution/array is shown in
The plenum pressure distribution (See
Next, the plenum pressure Pp is used to calculate the perforated air flow from each tile φPERFq as follows:
φPERFq=√{square root over (Pp(xPERFq,yPERFq)/RPERFq)}.
Using the perforated air flow from each tile φPERFq, in step 326, the total perforated tile air flow φPERFtotal can be calculated as follows:
Next, in step 328, the floor temperature Tf is calculated. In order to calculate the floor temperature Tf, a normalized Lorentzian field (resp) is applied to each tile location as follows:
zq=√{square root over ((x−xPERFq)2+(y−yPERFq)2)}
which takes into account the air flow through each perforated tile, which was calculated as described above. The Lorentzian field is chosen to yield 0 at the center of the tile and converge to 1 at large distances from the tile. By way of example only, as described in conjunction with the description of
An average temperature increase distribution in the data center is then calculated, which is given by the difference between ceiling Tc and plenum temperature Tp distributions, which were determined in steps 320 and 314, respectively, described above. This average temperature increase distribution is multiplied by the product of the Lorentzian fields and the plenum temperature distribution is added to it to yield the floor temperature distribution Tf as follows:
The process yields floor temperatures very near to the plenum temperature at locations close to tiles and average temperatures at locations far away from tiles.
An exemplary matrix of
with w=10 (1.67 feet) is shown in
with w=5 (0.83 feet) is shown in
The floor temperature distribution Tf (see
The corresponding resulting floor temperature distributions after the RF mask has been applied are shown in
Finally, in step 330, from the ceiling and floor temperature distributions, a three dimensional temperature distribution for the (e.g., data center) room (also referred to herein as a raised floor temperature distribution) is obtained by interpolating between these values using different approaches. One approach includes using an s-curve. S-curves are described, for example, in U.S. patent application Ser. No. 12/540,213 filed by Hamann et al., entitled “Knowledge-Based Models for Data Centers,” the contents of which are incorporated by reference herein. Alternatively, a simple linear interpolation between the floor and the ceiling temperature may be employed. Linear interpolation is a standard method known to those of skill in the art and thus is not described further herein.
Exemplary resulting raised floor temperature distributions are shown in
Turning now to
Apparatus 1800 includes a computer system 1810 and removable media 1850. Computer system 1810 includes a processor device 1820, a network interface 1825, a memory 1830, a media interface 1835 and an optional display 1840. Network interface 1825 allows computer system 1810 to connect to a network, while media interface 1835 allows computer system 1810 to interact with media, such as a hard drive or removable media 1850.
As is known in the art, the methods and apparatus discussed herein may be distributed as an article of manufacture that itself includes a machine-readable medium containing one or more programs which when executed implement embodiments of the present invention. For instance, the machine-readable medium may contain a program configured to determine a ceiling temperature distribution in the room; determine a floor temperature distribution in the room; and interpolate between the ceiling temperature distribution and the floor temperature distribution to obtain the three-dimensional temperature distribution model for the room.
The machine-readable medium may be a recordable medium (e.g., floppy disks, hard drive, optical disks such as removable media 1850, or memory cards) or may be a transmission medium (e.g., a network comprising fiber-optics, the world-wide web, cables, or a wireless channel using time-division multiple access, code-division multiple access, or other radio-frequency channel). Any medium known or developed that can store information suitable for use with a computer system may be used.
Processor device 1820 can be configured to implement the methods, steps, and functions disclosed herein. The memory 1830 could be distributed or local and the processor device 1820 could be distributed or singular. The memory 1830 could be implemented as an electrical, magnetic or optical memory, or any combination of these or other types of storage devices. Moreover, the term “memory” should be construed broadly enough to encompass any information able to be read from, or written to, an address in the addressable space accessed by processor device 1820. With this definition, information on a network, accessible through network interface 1825, is still within memory 1830 because the processor device 1820 can retrieve the information from the network. It should be noted that each distributed processor that makes up processor device 1820 generally contains its own addressable memory space. It should also be noted that some or all of computer system 1810 can be incorporated into an application-specific or general-use integrated circuit.
Optional video display 1840 is any type of video display suitable for interacting with a human user of apparatus 1800. Generally, video display 1840 is a computer monitor or other similar video display.
Although illustrative embodiments of the present invention have been described herein, it is to be understood that the invention is not limited to those precise embodiments, and that various other changes and modifications may be made by one skilled in the art without departing from the scope of the invention.
This invention was made with Government support under Contract number DE-EE00002897 awarded by Department of Energy. The Government has certain rights in this invention.
Number | Name | Date | Kind |
---|---|---|---|
7366632 | Hamann et al. | Apr 2008 | B2 |
8244502 | Hamann et al. | Aug 2012 | B2 |
8306794 | Hamann et al. | Nov 2012 | B2 |
20080288193 | Claassen et al. | Nov 2008 | A1 |
20090164811 | Sharma et al. | Jun 2009 | A1 |
20090326879 | Hamann et al. | Dec 2009 | A1 |
20090326884 | Amemiya et al. | Dec 2009 | A1 |
20100312498 | Hamann | Dec 2010 | A1 |
20110040532 | Hamann et al. | Feb 2011 | A1 |
20110060561 | Lugo et al. | Mar 2011 | A1 |
Entry |
---|
Joshi et al., “Energy Efficient Thermal Management” 2012, Chapters 1, 2, 3, Springer. |
Webster et al., “Modeling of Underfloor Air Distribution (UFAD) Systems”, 2008, 3rd Natl. Conf. of IBPSA pp. 214-221. |
Sarkar et al., “CFD Modeling and Validation of Temperature and Flow Distribution in Air-Conditioned Space”, 2008, International Refrigeration and Air Conditioning Conference. Paper 972. |
Joshi et al., “Energy Efficient Thermal Management” 2012, Chapters 1 & 11, Springer. |
Kumar et al., “Effect of Supply Air Temperature on Rack Cooling in a High Density Raised Floor Data Center Facility”, 2011 Proceedings of the ASME 2011 International Mechanical Engineering Congress & Exposition IMECE2011 Nov. 11-17, 2011, Denver, Colorado, USA. |
Energy Efficient Thermal Management of Data Centers 2012, pp. 273-334 Hendrik F. Hamann, Vanessa Lopez, Data Center Metrology and Measurement-Based Modeling Methods. |
Hamann et al., “Uncovering energy-efficiency opportunities in data centers,” IBM J. Res. & Dev., vol. 53, No. 3, pp. 10:1-10:12, May 2009. |
EPA Report to Congress on Server and Data Center Energy Efficiency Public Law 109-431, United States Code, Aug. 2, 2007. |
Li et al., “ThermoCast: A Cyber-Physical Forecasting Model for Data Centers,” KDD 2011: 17th ACM SIGKDD, Conference on Knowledge Discovery and Data Mining, Aug. 21-24, 2011 San Diego, CA. |
Lopez et al., “Measurement-Based Modeling for Data Centers,” 12th IEEE Intersociety Conference on Thermal and Thermomechanical Phenomena in Electronic Systems (Itherm), pp. 1-8, Jun. 2-5, 2010. |
Tang et al., “Sensor-Based Fast Thermal Evaluation Model for Energy Efficient High-Performance Datacenters,” 4th International Conference on Intelligent Sensing and Information Processing—ICISIP 2006 Oct. 15, 2006-Dec. 18, 2006, pp. 203-208. |
A. Terzis, “Active Cyber-Thermal Management of Data Centers for Energy Efficiency,” Computer Science Department Johns Hopkins University, Hopkins Internetworking Research Group HiNRG, Mar. 17, 2011. |
Number | Date | Country | |
---|---|---|---|
20140257740 A1 | Sep 2014 | US |