US20090138313A1 - Methods and systems for managing facility power and cooling - Google Patents

Methods and systems for managing facility power and cooling Download PDF

Info

Publication number
US20090138313A1
US20090138313A1 US12/121,113 US12111308A US2009138313A1 US 20090138313 A1 US20090138313 A1 US 20090138313A1 US 12111308 A US12111308 A US 12111308A US 2009138313 A1 US2009138313 A1 US 2009138313A1
Authority
US
United States
Prior art keywords
data center
capacity
equipment
determining
rack
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/121,113
Inventor
Jane E. Morgan
Neil Rasmussen
James W. VanGilder
Mikkel Dalgas
Ted Ives
Morten Bock
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Schneider Electric IT Corp
Original Assignee
American Power Conversion Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by American Power Conversion Corp filed Critical American Power Conversion Corp
Priority to US12/121,113 priority Critical patent/US20090138313A1/en
Assigned to AMERICAN POWER CONVERSION CORPORATION reassignment AMERICAN POWER CONVERSION CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: IVES, TED, MORGAN, JANE E., BOCK, MORTEN, DALGAS, MIKKEL, RASMUSSEN, NEIL, VANGILDER, JAMES W.
Publication of US20090138313A1 publication Critical patent/US20090138313A1/en
Assigned to SCHNEIDER ELECTRIC IT CORPORATION reassignment SCHNEIDER ELECTRIC IT CORPORATION CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: AMERICAN POWER CONVERSION CORPORATION
Priority to US16/019,052 priority patent/US11076507B2/en
Priority to US17/351,636 priority patent/US11503744B2/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H05ELECTRIC TECHNIQUES NOT OTHERWISE PROVIDED FOR
    • H05KPRINTED CIRCUITS; CASINGS OR CONSTRUCTIONAL DETAILS OF ELECTRIC APPARATUS; MANUFACTURE OF ASSEMBLAGES OF ELECTRICAL COMPONENTS
    • H05K7/00Constructional details common to different types of electric apparatus
    • H05K7/20Modifications to facilitate cooling, ventilating, or heating
    • H05K7/20709Modifications to facilitate cooling, ventilating, or heating for server racks or cabinets; for data centers, e.g. 19-inch computer racks
    • H05K7/20763Liquid cooling without phase change
    • H05K7/2079Liquid cooling without phase change within rooms for removing heat from cabinets
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/06Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/06Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
    • G06Q10/063Operations research, analysis or management
    • G06Q10/0631Resource planning, allocation, distributing or scheduling for enterprises or organisations
    • G06Q10/06313Resource planning in a project environment
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01HELECTRIC SWITCHES; RELAYS; SELECTORS; EMERGENCY PROTECTIVE DEVICES
    • H01H21/00Switches operated by an operating part in the form of a pivotable member acted upon directly by a solid body, e.g. by a hand
    • H01H21/02Details
    • H01H21/18Movable parts; Contacts mounted thereon
    • H01H21/22Operating parts, e.g. handle
    • H01H21/24Operating parts, e.g. handle biased to return to normal position upon removal of operating force
    • HELECTRICITY
    • H05ELECTRIC TECHNIQUES NOT OTHERWISE PROVIDED FOR
    • H05KPRINTED CIRCUITS; CASINGS OR CONSTRUCTIONAL DETAILS OF ELECTRIC APPARATUS; MANUFACTURE OF ASSEMBLAGES OF ELECTRICAL COMPONENTS
    • H05K7/00Constructional details common to different types of electric apparatus
    • H05K7/20Modifications to facilitate cooling, ventilating, or heating
    • H05K7/20709Modifications to facilitate cooling, ventilating, or heating for server racks or cabinets; for data centers, e.g. 19-inch computer racks
    • H05K7/20836Thermal management, e.g. server temperature control
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02PCLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
    • Y02P90/00Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
    • Y02P90/80Management or planning
    • Y02P90/82Energy audits or management systems therefor

Definitions

  • Embodiments of the invention relate generally to entity management, and more specifically to methods and systems for managing facility power and cooling.
  • Centralized data centers for computer, communications and other electronic equipment have been in use for a number of years, and more recently with the increasing use of the Internet, large scale data centers that provide hosting services for Internet Service Providers (ISPs), Application Service Providers (ASPs) and Internet content providers are becoming more prevalent.
  • ISPs Internet Service Providers
  • ASPs Application Service Providers
  • Typical centralized data centers contain numerous racks of equipment that require power, cooling and connections to external communications facilities.
  • the increased density of computing equipment used in these facilities has put strains on the cooling and power systems of the facilities.
  • typical power consumption for each equipment enclosure in a data facility was on the order of 1 kW.
  • server blades and other high power density equipment in equipment racks it is not uncommon for an equipment rack to have a power draw of 10 kW or even as high as 25 kW.
  • the power consumed by computer equipment is converted to heat and typically, the cooling requirements of a facility are determined based on the power requirements of the facility.
  • Typical data centers utilize air plenums under raised floors to distribute cooling air through a data center.
  • One or more computer room air conditioners (CRACs) or computer room air handlers (CRAHs) are typically distributed along the periphery of the data room, and these units draw return air from the room or a ceiling plenum and distribute cooling air beneath the raised floor.
  • Perforated tiles may be placed in front or beneath racks of equipment to be cooled to allow the cooling air from beneath the floor to cool equipment within the racks.
  • the system uses built-in cell to cell interaction rules to predict overall thermal performance and to optimize equipment layout. While this system may offer some improvements in speed over a full CFD analysis, it is limited to the use of pre-characterized cells, and does not provide analysis below the cell level. Also, the cells must be characterized using, for example, a CFD analysis or physical testing.
  • Programs and systems such as those described above provide idealized results for the cooling performance of a facility and often fail to account for situations which often occur in real life installations, which can dramatically affect the cooling performance of a data center.
  • a data center For example, in a facility using a raised floor, the absence of one or more floor panels, or the misplacement of one or more perforated floor panels can greatly affect the cooling performance of the data center and cause the actual performance to vary greatly from a calculated idealized performance.
  • the degradation in performance of one or more air conditioning units can drastically change airflow and cooling characteristics of a facility.
  • the inability to properly analyze the cooling performance of a facility typically causes a data center designer to over design the facility from a cooling perspective, which results in the facility to be more expensive and less efficient.
  • a data center manager may know, or can determine, if the total cooling capacity of the data center is sufficient for the total power draw.
  • hot spots in a facility may develop, particularly where high power density equipment is used, and it may not be enough to merely analyze cooling capacity at the facility level.
  • a data center manager may resort to manual temperature measurements throughout a facility and try to implement fixes to correct the hot spots. Such fixes may involve a rearrangement or replacement of perforated floor panels, a rearrangement of enclosures, and/or adding additional cooling capacity.
  • these fixes are typically done on a trial and error basis, and while some hot spots may be eliminated, the fixes may cause other hot spots to arise due to a redirection of the cooling air in the facility.
  • This trial and error approach can lead to unexpected failures of equipment, which is unacceptable in critical data centers.
  • data center managers typically over design facilities and fail to use facilities to their full capacity.
  • Embodiments of the invention relate generally to management of data center entities and their associated resources.
  • Embodiments of the invention provide systems and methods for determining data center cooling and power requirements and for monitoring performance of cooling and power systems in data centers.
  • At least one embodiment provides a system and method that enables a data center operator to determine available power and cooling at specific areas and enclosures in a data center to assist in locating new equipment in the data center.
  • One aspect is directed to a method that includes determining cooling capacity of a number of equipment enclosures in a data center, determining cooling requirements of each of the number of equipment enclosures, and providing an indication of remaining cooling capacity for each of the number of equipment enclosures.
  • the method may further include developing a floor plan model of the data center, wherein the floor plan model includes a floor plan that indicates location of each of the number of equipment enclosures in the data center, and for each of the number of equipment enclosures, displaying on the floor plan, the indication of remaining cooling capacity.
  • the indication of remaining cooling capacity may include an indication of additional power that can be drawn by each of the number of equipment enclosures based on the remaining cooling capacity.
  • Determining cooling capacity may include calculating a predicted cooling capacity based on the floor plan model. Determining cooling capacity may include measuring airflow at a first plurality of locations in the facility to obtain a measured cooling capacity. Determining cooling capacity may include measuring air temperature at a second plurality of locations in the facility.
  • At least one of the first plurality of locations and the second plurality of locations includes at least one air vent of a raised floor.
  • the method may further include comparing predicted cooling capacity with measured cooling capacity to obtain a comparison result and providing an indication when the comparison result is greater than a threshold.
  • the method may further include adjusting the predicted cooling capacity based on measured airflow.
  • the method may further include determining placement of new equipment in an equipment enclosure in the data center by comparing power ratings of the new equipment with cooling capacity of the equipment enclosure.
  • the method may further include, for each of the number of equipment enclosures, determining electrical power capacity and remaining electrical power availability, and displaying on the floor plan remaining electrical power availability.
  • determining remaining electrical power availability may include measuring at least one parameter of electrical power provided to at least one of the number of equipment enclosures.
  • determining cooling capacity of an equipment enclosure may include estimating available cooling air at the equipment enclosure using a weighted summation of available airflows from a plurality of airflow sources, the weights used in the weighted summation may decrease with distance from the equipment enclosure to each of the airflow sources, and the weights may be based on mechanical characteristics of the plurality of airflow sources.
  • the method may further include determining available airflow of at least one of the plurality of airflow devices using at least one of specifications of the at least one of the plurality of airflow devices and measured data for the at least one of the plurality of airflow devices in the data center.
  • the method may further include determining available airflow of at least a second one of the plurality of airflow devices based on the measured data for the at least one of the plurality of airflow devices.
  • determining cooling capacity may include using superposition to combine airflows.
  • determining airflow into and out of each of a plurality of sides of each control volume may include computing airflows using equations based on at least one of conservation of mass and conservation of momentum. Further, determining airflow into and out of each of a plurality of sides of each control volume may include determining airflows using empirical rules derived from CFD, physical measurement, or any other means.
  • the system includes at least one input to receive data related to equipment and equipment enclosures and to receive data related to cooling characteristics of the data center, a controller operatively coupled to the input and configured to determine cooling capacity of each equipment enclosure, and at least one output operatively coupled to the controller that provides an indication of remaining cooling capacity for each of the equipment enclosures.
  • the system may further include an output device coupled to the at least one output, wherein the system is configured to display a floor plan of at least a portion of the data center indicating location of at least one of the equipment enclosures in the data center and indicating the remaining cooling capacity for the at least one of the equipment enclosures.
  • the output device may be configured to include an indication of additional power that can be drawn by the at least one of the equipment enclosures.
  • the system may further include at least one airflow monitor operatively coupled to the controller to provide data related to at least one airflow in the data center.
  • the system may further include at least one air monitor operatively coupled to the controller to provide data related to air temperature at a location in the data center.
  • the controller may be configured to compare a predicted cooling capacity with a measured cooling capacity to obtain a comparison result and to provide an indication when the comparison result is greater than a threshold.
  • the system may further include at least one power monitor operatively coupled to the controller, and at least one airflow controller operatively coupled to the controller and responsive to signals from the controller to alter cooling airflow in the data center.
  • the system may further include at least one power controller operatively coupled to the controller and responsive to signals from the controller to alter at least one characteristic of power in the data center.
  • Another aspect of the invention is directed to a system for managing a data center.
  • the system includes at least one input to receive data related to equipment and equipment enclosures and to receive data related to cooling characteristics of the data center, and means, coupled to the at least one input, for determining remaining cooling capacity for each of the number of equipment enclosures and providing an indication of remaining cooling capacity.
  • the system may further include means for providing an indication of additional power that can be drawn by each of the equipment enclosures and means for updating the indication of remaining cooling capacity based on measured airflows in the data center.
  • the system may still further include means for determining placement of equipment in the data center based on remaining cooling capacity, and means for estimating available cooling air of at least one of the equipment enclosures using a weighted summation of available airflows from a plurality of airflow sources.
  • Another aspect of the invention is directed to a computer-implemented method of managing power and cooling capacity of a data center.
  • the method includes receiving data center parameters, determining an equipment layout in the data center, the equipment layout identifying a location of each of a plurality of devices in the data center, based on the location, determining available power and available cooling in the data center for each of the plurality of devices, comparing the available power and available cooling with power requirements and cooling requirements of each of the plurality of devices to obtain a cooling comparison result and a power comparison result for each of the plurality of devices.
  • the method may further include comparing each of the comparison cooling results and the power cooling results with thresholds and providing at least one recommendation for correcting an out of tolerance condition.
  • the at least one recommendation may include adding an in-row cooling unit in a row of equipment of the data center along with a proposed location for the in-row cooling unit.
  • the method may further include determining at least one of stranded cooling capacity and stranded power capacity in the data center and providing recommendations for reducing at least one of the stranded power capacity and the stranded cooling capacity in the data center.
  • the method may further include displaying a floor plan model of the data center, wherein the floor plan model includes a floor plan that indicates a location of each of the plurality of devices in the data center, and for each of the plurality of devices, displaying on the floor plan an indication of remaining cooling capacity.
  • the method may include displaying on the floor plan an indication of remaining power capacity for each of the plurality of devices.
  • the method may still further include setting a redundancy level for at least some of the plurality of devices, and wherein the available power and available cooling are determined to meet the redundancy level.
  • the act of determining the equipment layout may include arranging the plurality of devices in at least two substantially parallel rows with a hot aisle formed between the two rows, and the method may further include conducting a cooling analysis by analyzing airflows in the hot aisle.
  • the method may include selecting at least one in-row cooling unit to be placed in the layout in one of the at least two substantially parallel rows.
  • At least one of the plurality of devices is an equipment rack, and the method may include determining a capture index for the at least one in-row cooling unit and the equipment rack.
  • the method may further include on a display screen, simultaneously displaying a first floor plan model of the data center and a second floor plan model of the data center, wherein the first floor plan model includes at least a partial view of the data center and the second floor plan model includes a full view of the data center.
  • the second floor plan model may include an indication of a portion of the data center that is shown in the first floor plan model.
  • the method may further include displaying a three dimensional view of at least a portion of the data center.
  • the method may include selecting a camera location for use in the data center and displaying a three dimensional view of a viewing area of the camera.
  • the method may further include selecting a subset of the plurality of devices and defining a power zone for each device of the subset of the plurality of devices, wherein each power zone includes at least one UPS.
  • each of the plurality of devices may be an equipment rack, and the method may further include selecting components for each of the plurality of devices from a displayed list of components.
  • the method may further include determining operational power costs and operational cooling costs for a subset of the plurality of devices, and the operational power costs and operational cooling costs may be determined in terms of kilowatts.
  • the method may further include transferring an electronic file containing data for the equipment layout from a design system to a management system.
  • the method may also include displaying measured data for the data center on a display of a layout of the data center.
  • Another aspect of the invention is directed to a computer-implemented method for designing a layout of a data center.
  • the method includes receiving information from a user regarding parameters of the data center, determining an equipment layout for the data center, the equipment layout identifying a location of each of a plurality of devices in the data center, and on a display screen, simultaneously displaying a first floor plan model of the data center and a second floor plan model of the data center, wherein the first floor plan model includes at least a partial view of the data center and the second floor plan model includes a full view of the data center.
  • the second floor plan model may include an indication of a portion of the data center that is shown in the first floor plan model.
  • the method may further include determining available power and available cooling in the data center for each of the plurality of devices, and comparing the available power and available cooling with power requirements and cooling requirements of each of the plurality of devices to obtain a cooling comparison result and a power comparison result for each of the plurality of devices.
  • the method may include, for each of the plurality of devices, displaying on the first floor plan model an indication of remaining cooling capacity, and displaying on the first floor plan model an indication of remaining power capacity for each of the plurality of devices.
  • the system includes an input to receive instructions from a user, an output to provide output data to a display device, and a controller coupled to the input and to the output and configured to determine an equipment layout of a data center, the equipment layout identifying a location of each of a plurality of devices in the data center.
  • the controller is further configured to, based on the location, determine available power and available cooling in the data center for each of the plurality of devices, and compare the available power and available cooling with power requirements and cooling requirements of each of the plurality of devices to obtain a cooling comparison result and a power comparison result for each of the plurality of devices.
  • the controller may be configured to compare each of the comparison cooling results and the power cooling results with thresholds, and based on at least one of the cooling comparison result and the power comparison result, provide at least one recommendation for correcting an out of tolerance condition.
  • the at least one recommendation may include adding an in-row cooling unit in a row of equipment of the data center, and the at least one recommendation may include a proposed location for the in-row cooling unit.
  • the controller may be configured to determine at least one of stranded cooling capacity and stranded power capacity in the data center, and to provide recommendations for reducing at least one of the stranded power capacity and the stranded cooling capacity in the data center.
  • the controller may be further configured to provide data to the output for displaying a floor plan model of the data center, wherein the floor plan model includes a floor plan that indicates a location of each of the plurality of devices in the data center, and provide data to the output for displaying on the floor plan an indication of remaining cooling capacity.
  • the controller may be further configured to provide data to the output for displaying on the floor plan an indication of remaining power capacity for each of the plurality of devices, and the controller may be configured to determine the available power and available cooling based on a user-selected redundancy level.
  • the controller may be configured to arrange the plurality of devices in at least two substantially parallel rows with a hot aisle formed between the two rows, and to conduct a cooling analysis by analyzing airflows in the hot aisle.
  • the controller may be configured to select at least one in-row cooling unit to be placed in the layout in one of the at least two substantially parallel rows.
  • At least one of the plurality of devices may be an equipment rack, and the controller may be configured to determine a capture index for the at least one in-row cooling unit and the equipment rack.
  • the controller may be configured to provide data to the output to simultaneously display a first floor plan model of the data center and a second floor plan model of the data center, wherein the first floor plan model includes at least a partial view of the data center and the second floor plan model includes a full view of the data center.
  • the second floor plan model may also include an indication of a portion of the data center that is shown in the first floor plan model.
  • the controller may be configured to provide data to the output to display a three dimensional view of at least a portion of the data center, and to provide data to the output to display a three dimensional view of a viewing area of a camera to be located in the data center.
  • the controller may be further configured to select a subset of the plurality of devices and define a power zone for each device of the subset of the plurality of devices, wherein each power zone includes at least one UPS.
  • the system may further include a database module containing data for components to be loaded into the plurality of devices, and the controller may be configured to provide data to the output for displaying a list of the components.
  • the controller may be configured to determine operational power costs and operational cooling costs for a subset of the plurality of devices, and the operational power costs and operational cooling costs may be determined in terms of kilowatts.
  • the controller may also be configured to provide an output electronic file containing data for the equipment layout.
  • the controller may also be configured to, based on at least one of the cooling comparison result and the power comparison result, provide at least one recommendation for placement of rack mount equipment.
  • Another aspect of the invention is directed to a computer-implemented method for designing a layout of a data center.
  • the method includes receiving information from a user regarding parameters of the data center, determining an equipment layout for the data center, the equipment layout identifying a location of each of a plurality of devices in the data center, including identifying a row location for each of a plurality of equipment racks with a first subset of the plurality of equipment racks being included in a first row; and on a display screen, displaying a rack view of the data center showing a front view of each of the plurality of equipment racks of the first subset in the first row.
  • displaying a rack view may include displaying a front view of a second subset of the plurality of equipment racks of a second row along with the front view of the first subset of the plurality of equipment racks.
  • displaying a rack view may include displaying a front view of a second subset of the plurality of equipment racks of a second row along with the front view of the first subset of the plurality of equipment racks.
  • the first row may include additional equipment, with the additional equipment not included in the rack view.
  • the method may further include simultaneously displaying on the display screen the rack view and a full room view of the equipment layout for the data center.
  • the method may also include, in response to selection of a selected equipment rack in the full room view by a user, displaying the selected equipment rack in the rack view, and displaying the selected equipment rack in the rack view may include displaying a front view of the selected equipment rack.
  • Another aspect of the invention is directed to a computer-implemented method for evaluating the cooling performance of a cluster of equipment racks in a data center, wherein the cluster of equipment racks includes at least a first row of racks and a second row of racks separated by a cool aisle, with each of the equipment racks being configured to draw cooling air from the cool aisle.
  • the method includes obtaining at least one of power data and airflow data for each of the equipment racks, obtaining cool airflow data for cool air supplied to the cool aisle from a source of cool air, and conducting an analysis of airflows in the cool aisle to determine a recirculation index for at least one of the equipment racks, wherein the recirculation index is indicative of a quantity of recirculated air included in an input airflow of the at least one equipment rack.
  • the recirculation index may be equal to a ratio of recirculated air to total air in the input airflow of the at least one equipment rack, and the method may further include determining a recirculation index for each of the equipment racks.
  • the act of obtaining cool airflow data may include obtaining cool airflow data for an in-row cooling unit included in the cluster of racks.
  • the act of obtaining cool airflow data may include obtaining cool airflow data of at least one perforated tile included in the cool aisle.
  • the act of conducting an analysis may include defining a plurality of control volumes in the cool aisle, and the method may further include determining airflows in the cool aisle by determining airflow into and out of at least one of the control volumes.
  • the method may further include comparing the recirculation index for each of the plurality of equipment enclosures with a threshold.
  • the method may further include determining a cooling capacity for each of the equipment enclosures based on the recirculation index for each of the equipment enclosures, and displaying the cooling capacity for each of the equipment enclosures along with a representation of a data center containing the cluster.
  • the act of conducting an analysis may include assigning different chemical concentration identifiers to the airflows for at least two of the plurality of equipment racks.
  • the act of conducting an analysis may include importing empirical data and determining end of aisle airflows using the empirical data.
  • the act of conducting an analysis may include determining cool aisle airflows in isolation from the data center to obtain isolated results, and combining the isolated results with the empirical data.
  • Another aspect of the invention is directed to a computer-implemented method for evaluating the cooling performance of a cluster of equipment racks in a data center, wherein the cluster of equipment racks includes at least a first row of racks and a second row of racks separated by a hot aisle, with each of the equipment racks being configured to exhaust air into the hot aisle.
  • the method includes obtaining at least one of power data and airflow data for each of the equipment racks, obtaining airflow data for at least one air removal unit contained in one of the first row of equipment racks and the second row of equipment racks, and conducting an analysis of airflows in the hot aisle to determine a capture index for at least one of the equipment racks, wherein the capture index is indicative of a fraction of air that is exhausted by the at least one of the equipment racks and captured by the at least one air removal unit.
  • the at least one air removal unit may include an in-row cooling unit, and the capture index may be equal to a ratio of captured air to total air exhausted by the at least one equipment rack.
  • the method may further include determining a capture index for each of the equipment racks.
  • the act of conducting an analysis may include defining a plurality of control volumes in the hot aisle, and the method may further include determining airflows in the hot aisle by determining airflow into and out of at least one of the control volumes.
  • the method may further include comparing the capture index for each of the plurality of equipment enclosures with a threshold.
  • the method may include determining a cooling capacity for each of the equipment enclosures based on the capture index for each of the equipment enclosures, and displaying the cooling capacity for each of the equipment enclosures along with a representation of a data center containing the cluster.
  • the act of conducting an analysis may include assigning different chemical concentration identifiers to the airflows for at least two of the plurality of equipment racks.
  • the act of conducting an analysis may include importing empirical data and determining end of aisle airflows using the empirical data.
  • the act of conducting an analysis may include determining hot aisle airflows in isolation from the data center to obtain isolated results, and combining the isolated results with the empirical data.
  • the act of conducting an analysis may include importing empirical rules, and determining the capture index using the empirical rules, and the empirical rules may include coefficients for use in determining at least one capture index.
  • Another aspect of the invention is directed to a computer-readable medium encoded with instructions for execution on a computer system.
  • the instructions when executed, perform a method comprising acts of obtaining at least one of power data and airflow data for a plurality of equipment racks arranged in a cluster, wherein the cluster of equipment racks includes at least a first row of racks and a second row of racks separated by a cool aisle, with each of the equipment racks being configured to draw cooling air from the cool aisle, obtaining cool airflow data for cool air supplied to the cool aisle from a source of cool air, and conducting an analysis of airflows in the cool aisle to determine a recirculation index for at least one of the equipment racks, wherein the recirculation index is indicative of a quantity of recirculated air included in an input airflow of the at least one equipment rack.
  • the recirculation index may equal to a ratio of recirculated air to total air in the input airflow of the at least one equipment rack, and the acts may further include determining a recirculation index for each of the equipment racks.
  • the act of obtaining cool airflow data may include obtaining cool airflow data for an in-row cooling unit included in the cluster of racks.
  • the act of obtaining cool airflow data may include obtaining cool airflow data of at least one perforated tile included in the cool aisle.
  • the act of conducting an analysis ma include defining a plurality of control volumes in the cool aisle, and wherein the method further includes determining airflows in the cool aisle by determining airflow into and out of at least one of the control volumes.
  • the acts may further include comparing the recirculation index for each of the plurality of equipment enclosures with a threshold, and determining a cooling capacity for each of the equipment enclosures based on the recirculation index for each of the equipment enclosures.
  • the acts may further include displaying the cooling capacity for each of the equipment enclosures along with a representation of a data center containing the cluster.
  • the act of conducting an analysis may include assigning different chemical concentration identifiers to the airflows for at least two of the plurality of equipment racks.
  • the act of conducting an analysis may include importing empirical data and determining end of aisle airflows using the empirical data.
  • the act of conducting an analysis may include determining cool aisle airflows in isolation from the data center to obtain isolated results, and combining the isolated results with the empirical data.
  • Another aspect of the invention is directed to a computer-readable medium encoded with instructions for execution on a computer system.
  • the instructions when executed, perform a method comprising acts of obtaining at least one of power data and airflow data for a plurality of equipment racks arranged in a cluster, wherein the cluster of equipment racks includes at least a first row of racks and a second row of racks separated by a hot aisle, with each of the equipment racks being configured to exhaust air into the hot aisle, obtaining airflow data for at least one air removal unit contained in one of the first row of equipment racks and the second row of equipment racks, and conducting an analysis of airflows in the hot aisle to determine a capture index for at least one of the equipment racks, wherein the capture index is indicative of a fraction of air that is exhausted by the at least one of the equipment racks and captured by the at least one air removal unit.
  • the at least one air removal unit may be an in-row cooling unit, and the capture index may be equal to a ratio of captured air to total air exhausted by the at least one equipment rack.
  • the acts may further include determining a capture index for each of the equipment racks.
  • the act of conducting an analysis may include defining a plurality of control volumes, and wherein the method further includes determining airflows in the hot aisle by determining airflow into and out of at least one of the control volumes.
  • the acts may further include comparing the capture index for each of the plurality of equipment enclosures with a threshold.
  • the acts may further include determining a cooling capacity for each of the equipment enclosures based on the capture index for each of the equipment enclosures, and displaying the cooling capacity for each of the equipment enclosures along with a representation of a data center containing the cluster.
  • the act of conducting an analysis may include assigning different chemical concentration identifiers to the airflows for at least two of the plurality of equipment racks.
  • the act of conducting an analysis may include importing empirical data and determining end of aisle airflows using the empirical data.
  • the act of conducting an analysis may include determining hot aisle airflows in isolation from the data center to obtain isolated results, and combining the isolated results with the empirical data.
  • the act of conducting an analysis may include importing empirical rules, and determining the capture index using the empirical rules.
  • the empirical rules include coefficients for use in determining at least one capture index.
  • a computer-implemented method for providing a representation of capacity of a data center resource comprises acts of determining a first indication of excess capacity of a first data center resource that is unusable due to insufficient capacity of at least one other data center resource associated with the first data center resource, and providing the first indication of excess capacity of the first data center resource to an external entity.
  • the act of determining the first indication of excess capacity includes an act of determining an amount of excess capacity of the first data center resource that is unusable due to insufficient capacity of the at least one other data center resource associated with the first data center resource.
  • the act of determining the first indication of excess capacity includes an act of determining an indication of excess capacity at a location in a rack.
  • the act of determining the first indication of excess capacity at the location in the rack includes an act of determining an indication of excess capacity at a U space position.
  • the act of determining the first indication of excess capacity includes an act of determining an indication of excess capacity of rack space that is unusable due to insufficient capacity of power.
  • the act of determining the first indication of excess capacity includes an act of determining an indication of excess capacity of power that is unusable due to insufficient capacity of cooling.
  • the act of determining the first indication of excess capacity includes an act of determining an indication of excess capacity of power distribution that is unusable due to insufficient capacity of power available for distribution.
  • the act of determining the first indication of excess capacity includes an act of determining an indication of excess capacity of physical space that is unusable due to insufficient capacity of cooling.
  • the act of determining the first indication of excess capacity includes an act of determining an indication of excess capacity of power distribution that is unusable due to insufficient capacity of network connectivity.
  • the act of determining the first indication of excess capacity includes an act of determining an indication of excess capacity of U space that is unusable due to insufficient capacity of weight support.
  • the act of providing the first indication of excess capacity includes an act of providing the indication to another system element.
  • the act of providing the first indication of excess capacity includes an act of presenting the first indication to a user of a computer system.
  • the act of presenting the first indication to the user of the computer system includes an act of presenting, to the user of the computer system, an amount of excess capacity of the first data center resource that is unusable due to insufficient capacity of the at least one other data center resource associated with the first data center resource.
  • the act of presenting, to the user of the computer system, the amount of excess capacity includes acts of presenting, to the user of the computer system, an identifier that identifies the first data center resource and presenting, to the user of the computer system, an identifier that identifies the at least one other data center resource.
  • the method further comprises determining a configuration of data center equipment that minimizes, relative to at least one other configuration, the excess capacity of the first data center resource, and providing the configuration of data center equipment to an external entity.
  • determining the configuration of data center equipment includes determining a location for at least one element of data center equipment.
  • the method further comprises determining a second indication of excess capacity of a second data center resource that is unusable due to insufficient capacity of at least one other data center resource associated with the second data center resource determining a configuration of data center equipment that minimizes, relative to at least one other configuration, an aggregate of the first indication and the second indication, and providing the configuration of data center equipment to an external entity.
  • a computer-implemented method for providing a representation of a data center resource comprises acts of determining a status of capacity for the data center resource provided to either a rack or a location in a rack, and providing the status to an external entity.
  • determining the status of capacity for the data center resource includes determining a status for the data center resource provided to a U space position.
  • the act of determining the status of capacity for the data center resource includes an act of determining a status of capacity of power.
  • the act of determining the status of capacity for the data center resource includes an act of determining a status of capacity of cooling.
  • the act of determining the status of capacity for the data center resource includes an act of determining a status of capacity of physical space.
  • the act of determining the status of capacity for the data center resource includes an act of determining a status of capacity of weight support.
  • the act of determining the status of capacity for the data center resource includes an act of determining a status of available capacity.
  • the act of determining the status of capacity for the data center resource includes an act of determining a status of utilized capacity.
  • a computer-implemented method for providing a representation of a data center resource comprises acts of determining a rate of change of a utilized capacity of the data center resource based at least in part on changes to a data center configuration, and providing a predicted utilization of the data center resource based at least in part on the rate of change.
  • determining the rate of change of utilized capacity of the data center resource includes determining the rate of change using linear regression.
  • the data center resource includes power.
  • the data center resource includes cooling.
  • a system for providing a representation of a capacity of data center resource comprises an input configured to receive a capacity of a first data center resource and a capacity of at least one other data center resource, an output configured to provide an amount of excess capacity of the first data center resource, and a controller coupled to the input and the output and configured to determine an amount of excess capacity of the first data center resource that is unusable due to insufficient capacity of the at least one other data center resource.
  • a system for providing a representation of a data center resource comprises an input configured to receive capacity information for a data center resource, an output configured to provide a status of capacity for the data center resource, and a controller coupled to the input and the output and configured to determine the status of capacity for the data center resource at either a rack or a location in a rack.
  • a system for providing an evaluation of a data center resource comprises an input configured to receive changes in a data center configuration, an output configured to provide a predicted utilization of the data center resource based on a rate of change of utilized capacity of the data center resource, and a controller configured to determine the rate of change of the utilized capacity of the data center resource based on the changes in a data center configuration.
  • a computer-implemented method for distributing data describing models of data center equipment comprises acts of indicating a first group of the data describing models of data center equipment, the first group identifying standard models of data center equipment, indicating a second group of the data describing models of data center equipment, the second group identifying models approved for use in a first data center, indicating a third group of the data describing models of data center equipment, the third group identifying models preferred for use in the first data center, providing the second group to a first remote computing device located in the first data center, and providing the third group to the first remote computing device located in the first data center.
  • the act of providing the second group comprises an act of providing data including at least one of the group comprising data center equipment manufacturer, data center equipment model, data center equipment power plug type, data center equipment number of plugs required, data center equipment power required, data center equipment airflow required, data center equipment network connectivity required, data center equipment weight support required, data center equipment cooling required, and data center equipment physical space required.
  • the method further comprises an act of providing the third group from the first remote computing device to a second remote computing device.
  • the act of providing the third group from the first remote computing device includes an act of providing the third group to a mobile computing device.
  • the act of indicating a first group includes storing an indicator of the first group in a database hosted by a system provider.
  • the method further comprises an act of receiving first non-standard data describing models of data center equipment from the first remote computing device.
  • the method further comprises an act of receiving non-standard data describing models of data center equipment from an external entity.
  • the act of receiving non-standard data includes receiving additional data from a data center equipment manufacturer.
  • the method further comprises acts of indicating a fourth group of the data describing models of data center equipment, the fourth group identifying models approved for use in a second data center, indicating a fifth group of the data describing models of data center equipment, the fifth group identifying models preferred for use in the second data center, providing the fourth group to a third remote computing device located in the second data center, and providing the fifth group to the third remote computing device located in the second data center.
  • the method further comprises an act of receiving second non-standard data describing models of data center equipment from the third remote computing device.
  • a computer-implemented method for maintaining a data center equipment database comprises acts of receiving a set of configuration management data from each of a plurality of data centers, each set of configuration management data identifying one of the plurality data centers, a capacity for power of the identified data center, a capacity for cooling of the identified data center and at least one element of data center equipment disposed within the identified data center.
  • the acts of receiving include acts of receiving configuration management data including at least one of the group comprising data center physical dimensions, data center equipment location within the data center physical dimensions, data center power distribution provided to the data center equipment location, data center cooling distribution provided to the data center equipment location, data center network connectivity provided to the data center equipment location, data center equipment manufacturer, and data center equipment model.
  • the method further comprises acts of receiving a set of data center resource consumption and production data from each of the plurality of data centers, each set of data center resource consumption and production data identifying one of the plurality of data centers and being associated with at least one element of data center equipment disposed within the identified data center.
  • the acts of receiving a set of data center resource consumption and production data from each of the plurality of data centers include acts of receiving data center resource consumption and production data including at least one from the group comprising data center equipment identity, data center equipment manufacturer, data center equipment model, data center equipment power plug type used, data center equipment number of plugs used, data center equipment power used, data center equipment airflow used, data center equipment network connectivity used, data center equipment weight support used, data center equipment cooling used, and data center equipment physical space used.
  • the acts of receiving a set of data center resource consumption and production data from each of the plurality of data centers include acts of receiving data center resource consumption and production data including at least one from the group comprising data center equipment identity, data center equipment manufacturer, data center equipment model, data center equipment power plug type provided, data center equipment number of plugs provided, data center equipment power provided, data center equipment airflow provided, data center equipment network connectivity provided, data center equipment weight support provided, data center equipment cooling provided, and data center equipment physical space provided.
  • a computer-implemented method for distributing data describing models of data center equipment comprises acts of receiving, from a first remote computing device, a first group of the data describing models of data center equipment, the first group identifying models approved for use in a data center, receiving, from the first remote computing device, a second group of the data describing models of data center equipment, the second group identifying models preferred for use in the data center, and providing the second group to a second remote computing device.
  • the method further comprises acts of receiving, from an external entity, non-standard data describing models of data center equipment that are not approved for use in the data center, and providing the non-standard data to the first remote computing device.
  • a system for managing equipment data describing models of data center equipment comprises an input configured to receive data associated with the equipment data, an output configured to provide data associated with the equipment data, and a controller coupled to the input and the output and configured to allow an external entity to indicate that a first group of the equipment data describes standard models of data center equipment; to indicate that a second group of the equipment data describes models of data center equipment approved for use in a data center, to indicate that a third group of the equipment data describes models of data center equipment preferred for use in the data center and further configured to provide, via the output, the second group to a remote computing device located in the data center and to provide, via the output, the third group to the remote computing device located in the data center.
  • a system for managing equipment data describing models of data center equipment comprises a computer readable medium, an input configured to receive data associated with the equipment data, an output configured to provide data associated with the equipment data, and a controller coupled to the input, the output and the computer readable medium and configured to store, in the computer readable medium, an first indication that a first group of the equipment data describes models of data center equipment approved for use in a data center and to store a second indication that a second group of equipment data describes models of data center equipment preferred for use in the data center and further configured to provide, via the output, the second group to a remote computing device located in the data center.
  • a system for maintaining a data center equipment database comprises an input configured to receive a set of configuration management data from each of a plurality of data centers, each set of configuration management data identifying one of the plurality data centers, a capacity for power of the identified data center, a capacity for cooling of the identified data center and at least one element of data center equipment disposed within the identified data center, and a controller coupled to the input and configured to store, in the data center equipment database, each set of configuration management data.
  • a computer-implemented method for managing a data center configuration comprises acts of receiving, from a user, a proposed change to the data center configuration though an interactive representation of a layout of an identified space in a data center, the proposed change including data identifying at least one element of data center equipment to be provisioned in the data center, the at least one element of data center equipment having a plurality of data center resource requirements, determining at least one location in the data center having a plurality of data center resources that satisfy the plurality of data center resource requirements of the at least one element of data center equipment, and storing the at least one location on a computer readable medium.
  • the act of receiving, from the user, the proposed change to the data center configuration includes an act of receiving data identifying at least one element of floor mounted data center equipment to be provisioned in the data center.
  • the act of receiving, from the user, the proposed change includes an act of receiving, from the user, the proposed change to the data center configuration though an interactive representation of a layout of a data center room.
  • the act of receiving, from the user, the proposed change includes an act of receiving, from the user, the proposed change to the data center configuration though an interactive representation of a layout of a rack.
  • the act of receiving, from the user, the proposed change to the data center configuration includes an act of receiving information regarding at least one element of data center equipment represented in the interactive representation.
  • the act of receiving, from the user, the proposed change to the data center configuration includes an act of scanning a barcode.
  • the act of determining the at least one location includes an act of determining at least one location in near real-time.
  • the act of determining the at least one location includes an act of receiving, from an external entity, data identifying at least one location in the data center for the at least one element of data center equipment having a plurality of data center resources.
  • the act of determining the at least one location includes an act of determining a location that minimizes, relative to at least one other location, an amount of excess data center resource capacity that is unusable due to insufficient associated capacity of at least one other data center resource, and the method further comprises providing the at least one location to an external entity.
  • the act of providing the at least one location further includes an act of presenting the at least one location to a user of a computer system.
  • the act of determining the at least one location includes an act of determining a series of locations arranged consecutively according to an amount of excess data center resource capacity that is unusable due to insufficient associated capacity of at least one other data center resource and the method further comprises providing the series of locations to an external entity.
  • the act of providing the at least one location includes an act of presenting the series to a user of a computer system.
  • the method further comprises an act of presenting, to the user, information regarding the proposed change to the data center configuration, the information including a representation of at least one element of data center equipment that is disposed in a rack.
  • the act of presenting, to the user, the information regarding the proposed change to the data center configuration includes an act of presenting, to the user, a representation of the at least one element of data center equipment as placed in inactive storage, and the method further includes an act of recording the at least one element of data center equipment as placed in inactive storage.
  • the act of presenting, to the user, the information regarding the proposed change to the data center configuration includes an act of presenting, to the user, an interactive representation that is updated based on the proposed change.
  • the act of presenting, to the user, the interactive representation that is updated based on the proposed change includes an act of presenting, to the user, the interactive representation updated in near real-time.
  • the method further comprises generating a work order identifying a set of tasks required to complete the proposed change, and presenting the set of tasks via an output of a computer system.
  • the method further comprises acts of detecting a change in the power demanded at power distribution device associated with the work order, and recording the work order as completed based on the change in power demanded.
  • the method is performed using a remote computing device.
  • a computer-implemented method for managing a data center configuration comprises acts of receiving at least one policy identifying at least one characteristic of a data center resource as provided to at least one element of data center equipment, receiving a proposed change to the data center configuration, and determining a compliance result that indicates whether the data center configuration, including the proposed change, complies with the at least one policy.
  • the act of receiving the at least one policy includes an act of receiving at least one policy identifying a redundancy level.
  • the act of receiving the at least one policy includes an act of receiving at least one policy identifying a runtime requirement.
  • the act of receiving the at least one policy includes an act of receiving at least one policy identifying a security requirement.
  • the act of receiving the at least one policy includes an act of receiving at least one policy identifying a requirement of cooling.
  • the act of receiving the at least one policy includes an act of receiving at least one policy identifying a requirement of network connectivity.
  • the act of receiving the at least one policy includes an act of receiving at least one policy identifying a requirement of power.
  • the act of receiving the at least one policy identifying a requirement of power includes an act of receiving at least one policy identifying a requirement of power distribution.
  • a computer-implemented method for designing a data center comprises acts of receiving at least one policy identifying at least one characteristic of a data center resource as provided to at least one element of data center equipment, receiving data identifying at least one location in the data center for at least one element of data center equipment, the at least one location having a plurality of data center resources and the at least one element of data center equipment having a plurality of data center resource requirements, and determining at least one element of data center equipment that satisfies the policy when located at the at least one location, wherein the plurality of data center resources of the at least one location satisfy the plurality of data center resource requirements of the at least one element of data center equipment.
  • a system for managing a data center configuration comprising an input configured to receive, from a user, a proposed change to the data center configuration though an interactive representation of a layout of an identified space in a data center, an output configured to provide a set of tasks to a user, and a controller coupled to the input and the output and configured to generate a work order identifying the set of tasks required to complete the proposed change and to provide the set of tasks to the output.
  • the system further comprises a computer readable medium coupled to the controller and wherein the input is further configured to receive status information associated with the work order and the controller is further configured to record the status information in the computer readable medium.
  • the system further comprises an input configured to receive, from a user, status information associated with the work order and wherein the controller is further configured to provide the status information to the output.
  • a system for managing a data center configuration comprises a first input configured to receive at least one policy identifying at least one characteristic of a data center resource as provided to at least one element of data center equipment, a second input configured to receive a proposed change to the data center configuration, and a controller coupled to the first and second inputs and configured to determine a compliance result that indicates whether the data center configuration, including the proposed change, complies with the at least one policy.
  • a system for designing a data center comprises an first input configured to receive at least one policy identifying at least one characteristic of a data center resource as provided to at least one element of data center equipment, a second input configured to receive data identifying at least one location in the data center for at least one element of data center equipment, the at least one location having a plurality of data center resources and the at least one element of data center equipment having a plurality of data center resource requirements, and a controller coupled to the first and second inputs and configured to determine at least one element of data center equipment that satisfies the policy when located at the at least one location, wherein the plurality of data center resources of the at least one location satisfy the plurality of data center resource requirements of the at least one element of data center equipment.
  • a remote computing device for navigating a representation of a data center.
  • the remote computing device comprises an interface configured to present, to a user, an interactive representation of a layout of an identified space in a data center and to receive, from the user, an identifier of at least one element of data center equipment located in the data center, a memory configured to store information regarding data center equipment represented by the interface, and a controller coupled to the interface and memory and configured to retrieve information regarding the identified at least one element of data center equipment from the memory and provide the information to the interface.
  • a computer-implemented method for managing data center resource redundancy comprises acts of determining a redundancy level of a data center resource at a location in a rack, and providing the redundancy level of the data center resource.
  • the act of determining the redundancy level of the data center resource includes an act of determining a redundancy level of cooling.
  • the act of determining the redundancy level of the data center resource includes an act of determining a redundancy level of network connectivity.
  • the act of determining the redundancy level of the data center resource includes an act of determining a redundancy level of power.
  • the act of determining the redundancy level of the data center resource includes an act of determining a redundancy level of power distribution.
  • the act of determining the redundancy level of the data center resource at a location in a rack includes an act of determining a redundancy level of a data center resource at a U space position.
  • the act of providing the redundancy level includes an act of presenting, to a user of a computer system, a representation of the at least one location in a rack.
  • the act of presenting, to the user of the computer system, the representation of the at least one location in the rack includes an act of presenting, to a user of a computer system, a representation at least one U space position in a rack.
  • the act of presenting, to the user of the computer system, the representation of the at least one location in the rack includes an act of presenting, to a user of a computer system, a representation of a redundancy relationship between elements disposed in a data center.
  • the act of presenting, to the user of the computer system, a representation of a redundancy relationship between elements disposed in a data center includes an act of presenting, to a user of a computer system, a redundancy relationship between elements disposed in a rack.
  • a system for managing data center resource redundancy comprises a controller configured to determine the redundancy level of a data center resource at a location in a rack, and an output coupled to the controller and configured to provide a representation of the redundancy level of the data center resource.
  • the location in the rack includes a U space position in a rack.
  • the system includes a remote computing device.
  • FIG. 1 is a top view of a data center of the type with which embodiments of the present invention may be used;
  • FIG. 2 is a side view of the data center of FIG. 1 .
  • FIG. 3 is a functional block diagram of a system in accordance with one embodiment of the present invention.
  • FIG. 4 is a flowchart of a process that may be implemented using the system of FIG. 3 in accordance with one embodiment of the invention
  • FIG. 5 is a diagram showing facility information that can be displayed using at least one embodiment of the invention.
  • FIGS. 5A and 5B are diagrams showing additional information that can be displayed using embodiments of the invention.
  • FIGS. 5C and 5D show graphical user interface screens that exhibit various aspects of the present invention.
  • FIG. 6 is a functional block diagram of a management system in accordance with one embodiment of the invention.
  • FIG. 7 is a flow chart of a management process in accordance with one embodiment of the invention.
  • FIG. 8 shows a perspective view of a cluster of racks whose cooling performance can be analyzed using embodiments of the invention
  • FIG. 9 shows a top view of a cluster of racks whose cooling performance can be analyzed using a control volume analysis technique of at least one embodiment
  • FIG. 9A shows the cluster of racks of FIG. 9 along with staggered control volumes that may be used in the control volume analysis technique
  • FIG. 10 is a flow chart of a control volume analysis technique in accordance with one embodiment of the invention.
  • FIG. 11 is a diagram demonstrating a principle of superposition used in one embodiment
  • FIG. 12 is a graph used in determining airflows in one embodiment
  • FIG. 13 is a diagram identifying airflows used with one analysis method of one embodiment
  • FIG. 14 is a flow chart of a process for determining a recirculation index in one embodiment
  • FIG. 15 is a schematic diagram showing a layout of equipment racks used in an analysis in one embodiment to determine a capture index
  • FIG. 16 is a flowchart of a process for determining a capture index in accordance with one embodiment
  • FIG. 17 is a functional block diagram of a computer system that may be used in embodiments of the invention.
  • FIG. 18 is a functional block diagram of a storage system that may be used with the computer system of FIG. 17 ;
  • FIG. 19 is a flowchart of a process that may be implemented using the system of FIG. 3 in accordance with one embodiment of the invention.
  • FIG. 20 shows graphical user interface screen that exhibits various aspects of the present invention
  • FIG. 21 depicts another graphical user interface screens that exhibits various aspects of the present invention.
  • FIG. 22 illustrates another graphical user interface screen that exhibits various aspects of the present invention.
  • FIG. 23 shows a representation used to analyze stranded capacity according to an aspect of the present invention.
  • Embodiments of the present invention may be used to design, manage and retrofit a data center, such as data center 100 which is shown in FIGS. 1 and 2 with FIG. 1 showing a top view of the data center 100 , and FIG. 2 showing a side view of the data center 100 .
  • a data center resource may include, for example, any characteristic of a data center that supports and/or ensures data center equipment functionality. Examples of data center resources include power, cooling, physical space, weight support, remote equipment control capability, physical and logical security and physical and logical network connectivity.
  • Power data center resources may include power distribution resources, such as transformers, PDUs and outlets, and power available for distribution, such as utility power supplied to the data center, power generated by an onsite generator and power supplied by UPSs.
  • Physical space resources in a data center may include data center floor space and rack U space. Cooling resources in a data center may include cooling distribution capacity and cooling generation capacity.
  • Physical security resources in a data center may include security cameras and door locks.
  • Logical network connectivity resources in a data center may include Virtual Local Area Networks, Domain Name Services, and Dynamic Host Configuration Protocol Services.
  • Physical network connectivity resources may include network cabling and patch panels.
  • Remote equipment control capability resources in a data center may include Keyboard Video Mouse services.
  • Embodiments of the invention are not limited for use with data centers like that shown in FIGS. 1 and 2 and may be used with other facilities that do not include raised floors and may be used with facilities that house equipment other than computing equipment, including telecommunications facilities and other facilities. Further, embodiments of the invention may be used with raised floor and equipment layouts that are not neatly arranged in the manner shown in FIGS. 1 and 2 . Embodiments of the present invention may use systems, devices and methods described in U.S. patent application Ser. No. 10/038,106, filed Jan. 2, 2002, titled “Rack Power System and Method,” incorporated herein in its entirety by reference.
  • the data center 100 includes rows of racks 102 A, 102 B, 102 C and 102 D, cooling units 104 A and 104 B, and a raised floor 106 .
  • Each of the rows includes racks 108 , at least a number of which draw cool air from the front of the rack and return warm air to the rear or top or rear and top of the rack.
  • Each rack may contain U space positions designed to house rack mounted data center equipment, such as, for example, servers, cooling equipment and network connectivity equipment.
  • rows of racks are often positioned to create alternating cold aisles and hot aisles.
  • aisles 110 A, 110 B and 110 C are hot aisles and aisles 112 A and 112 B are cold aisles.
  • perforated floor tiles 114 are used to provide cooling air from under the raised floor.
  • the raised floor may include solid floor tiles.
  • the cooling units 104 A and 104 B are designed to provide cool air to the area under the raised floor and to receive return warm air from the space adjacent the ceiling of the data center.
  • in-row cooling units such as those available from American Power Conversion Corporation, may be used.
  • half-rack in-row cooling units may be used, as described in co-pending U.S. patent application Ser. No. 11/335,901, entitled COOLING SYSTEM AND METHOD, Attorney Docket No. A2000-704819, by Neil Rasmussen et al., filed on Jan. 19, 2006, and incorporated herein by reference.
  • the half-rack, in-row units have a width of twelve inches, which is approximately half of that of a standard data center rack.
  • FIG. 3 shows a functional block diagram of a design and management system 200 .
  • Embodiments of the invention are not limited to the functions provided by the functional blocks or the particular arrangement of the blocks.
  • the functions provided by the system 200 need not be implemented on one computer system, but rather may be implemented using a number of networked devices as described further below that provide the functions described.
  • particular embodiments may have more or less functions and functional modules than those described below with reference to FIG. 3 .
  • the functions described with reference to FIG. 3 may be performed on one processor or controller or may be distributed across a number of different devices.
  • the system 200 includes an input module 202 , a display module 204 , a builder module 206 , a facility management module 208 , an integration module 210 , a database module 212 , and a work order module 214 .
  • the input module 202 provides an interface to allow users to enter data into the system 200 .
  • the input module may include, for example, one of a number of known user input devices for computer systems, and in addition, in at least one embodiment, electronic data regarding a facility and/or equipment to be loaded into a facility may be entered into the system through a network interface or using an electronic media storage reader. Information may flow between these modules using any technique known in the art. Such techniques include passing the information over the network via TCP/IP, passing the information between modules in memory and passing the information by writing to a file, database, or other storage entity, such as a storage device, disk or other type of storage entity.
  • the display module includes a display interface and may include a graphical display to display output data to a user.
  • the display module may include an interface for one or more printers that provide a hard copy of output data.
  • the builder module 206 includes routines for designing optimal layout of equipment in a facility, determining data center resource requirements, such as power requirements and cooling requirements, for electronic enclosures and/or equipment racks, ensuring that the placement of equipment, cooling units and power distribution branches in the facility allow the data center resource requirements, such as power and cooling requirements, to be met, and calculating for each electronic enclosure, and/or equipment rack, the remaining data center resource capacity, such as power capacity and cooling capacity, available based on the layout of equipment in the facility.
  • data center resource requirements such as power requirements and cooling requirements
  • the builder module 206 includes routines for designing optimal layout of equipment in a facility, determining data center resource requirements, such as power requirements and cooling requirements, for electronic enclosures and/or equipment racks, ensuring that the placement of equipment, cooling units and power distribution branches in the facility allow the data center resource requirements, such as power and cooling requirements, to be met, and calculating for each electronic enclosure, and/or equipment rack, the remaining data center resource capacity, such as power capacity and cooling capacity, available based on the layout of
  • builder module 206 exposes interfaces that allow for entry of data center resource supply policies. These policies may detail additional desired characteristics of the data center resources as provided to data center equipment. For example, data center resource supply policies may specify desired data center resource redundancy levels and runtime requirements. Additionally, data center resource supply policies may specify security needs for data center equipment, such as, for example, the need to locate servers with financial data in racks that made be locked and/or under surveillance. As is discussed further below, both system and method embodiments may compare particular data center configurations to data center supply polices to determine whether the particular data center configuration complies with the data center policies. The result of this comparison may be stored on a computer readable medium for later use.
  • the facility management module 208 is used by the system 200 after equipment is installed in the facility.
  • the management module includes routines to monitor data center resource characteristics, such as power and cooling characteristics, of equipment in a facility.
  • the management module may be coupled, either directly or through one or more networks, to measurement devices and control devices throughout the facility and may record their history for analysis, summarization and exportation. Based at least in part on methods discussed further below, the facility management module 208 may ascertain data center resources provided to a particular space or location in the data center.
  • the integration module 210 is the main module in the system and coordinates flow of data in the system to perform methods of embodiments of the present invention.
  • the database module is used to store data regarding various devices that may be used in a data center, such as servers, uninterruptible power supplies, power strips, network connectivity equipment (such as network cabling, hubs, routers, wireless routers, switches, patch panels, etc.), automatic transfer switches, power distribution units, air conditioning units, racks and any other data center equipment.
  • the data stored may include data center resource supply policies for data center equipment.
  • the data stored may also include data center resource consumption and production benchmarks for data center equipment such as physical parameters (e.g., dimensions/rack space requirements, power plug types, network cabling specifications, etc.) as well as power and cooling consumption data.
  • data center resource production benchmarks may include network connectivity, cooling and power output characteristics (e.g. total output capacity, number and type of connections/plugs available, etc.). It should be appreciated that benchmarks may indicate that certain data center equipment produce, and/or consume, data center resources. Examples of data center resource providing equipment include electrical generators and CRACs, among others.
  • the database module may be used in embodiments of the invention to provide a complete bill of materials (BOM) for a completed design.
  • a centralized web-accessible database server may be used to store equipment information and warnings and error messages, allowing easy access to the information for editing.
  • a user may maintain data center equipment information, such as the information handled by database module 212 , with the assistance of system 300 .
  • System 300 is also depicted in functional block diagram FIG. 3 .
  • System 300 which may be hosted by the system provider, includes data center equipment database 304 and data center equipment database interface 302 .
  • interface 302 may be an interface to database 304 that may receive or provide any data appropriate for storage in database 304 including data center configuration, equipment or resource information.
  • Database 304 may receive and store from interface 302 or retrieve and provide to interface 302 data center information including data center equipment resource requirements, data center configurations and data center resource redundancy and runtime requirements.
  • Database 304 may serve as a master database for the system provider and thus may include several types of information related to data centers.
  • database 304 includes a data center configuration management database (CMDB), which may encompass various configuration management data including one or more specific data center physical and logical configurations, and a data center equipment characteristics database, which may include theoretical and actual data center resource production and consumption information for data center equipment.
  • CMDB data center configuration management database
  • database 304 may accept and store specific CMDB information for the system provider's installed base of data centers or a subset thereof. This information may include the entire physical layout of a data center, e.g. its physical dimensions, the location and identity of data center equipment, and data center resource capacity, redundancy and runtime requirements, among other information that may have relevance to the design and performance of a data center.
  • the source of this information may include the initial design on the data center as discussed with regard to method 300 below.
  • database 304 may include the information discussed above with regard to database module 212 , e.g. nameplate values as referred to in the art. Additionally, database 304 may also store data center resource consumption and production history for data center equipment and may use this information to maintain a set of benchmarks that are specific to the manufacturer and model of data center equipment. These specific historical measurements may be summarized into various forms to establish a benchmark that is specific to the manufacturer and model of data center equipment and that is based on actual, practical usage of the data center equipment rather than theoretical values (e.g. nameplate values). These data center resource consumption and production summaries may include, among others, minimum, maximum and average data center resource consumption or production, data center resource consumption or production as a function of time, e.g.
  • the source of this information may include the initial design on the data center, including customer or system provider entered benchmarks, and ongoing parameter measurement as discussed with regard to method 300 below. According to one aspect, it is appreciated that these benchmarks, in turn, may be more accurate than the nameplate values and may be used for simulation purposes during the design and retrofitting processes discussed below.
  • Interface 302 may expose both user interfaces (UIs) and system interfaces to exchange database 304 information with external entities. These external entities may include systems and/or users. Interface 302 may both restrict input to a predefined information domain and validate any information entered prior to using the information or supplying the information to other modules.
  • interface 302 may include a Structured Query Language (SQL) interface to enable a user or application program to interrogate database 304 .
  • SQL Structured Query Language
  • This SQL interface may include graphical elements that a user may actuate to build SQL statements and may also include elements that enable a user to simply input SQL statement directly.
  • interface 302 may include more complex grouping, translation, validation and/or restriction logic. For instance, interface 302 may validate that a SQL statement entered by a user conforms to proper SQL syntax prior to executing it against database 304 . In one embodiment, interface 302 may expose a UI with characteristics similar to those of builder module 206 to enable users to create database information representing various data center configurations. In another example, interface 302 may expose a user interface allowing a user to enter new data center equipment resource information, including manufacturer, model and data center resource consumption and production benchmarks. Interface 302 may restrict entry of resource consumption and production benchmarks to predefined data center resources, such as power, cooling, physical space, etc.
  • a user may establish groups of data center equipment through interface 302 and approve these equipment groups for use within a set of data centers or upon certain types of computing devices, such as remote or mobile computing devices.
  • interface 302 may enable a user to designate a hierarchical relationship between groups of equipment. As discussed further in regard to method 300 below, arrangement of this information into a hierarchy of groups may ease database maintenance and distribution.
  • interface 302 may expose a system interface that imports data center configuration and measured parameter information from a system provider's installed base of data centers or data center equipment providers.
  • Interface 302 may utilize a standard protocol, such as, for example, SQL or SQL wrapped in SOAP, to implement such a system interface and may include non-standard protocol elements that are parsed and resolved to commands to be executed by database 304 .
  • Interface 302 may validate database commands prior to execution on database 304 . Any customer specific data center equipment, e.g. data center equipment entered by a customer that is not present in the database 304 , may be imported into database 304 by interface 302 if any applicable validation is successful. Similarly, equipment not present in database 304 that is made available by data center equipment providers may be imported into database 304 by interface 302 if any applicable validation is successful.
  • importation of measure parameters may trigger a parameter summarization process that updates data center equipment resource consumption and production benchmarks to reflect actual usage history.
  • the summarization process may include updating, among others, minimum, maximum and average power consumed or produced benchmarks, benchmarks for power consumption or production as a function of time, e.g. power consumed or produced by day of week, week of year, etc., benchmarks for power actually consumed or produced when a change in either is requested, and/or benchmarks for power consumed or produced as a function of data center equipment utilization.
  • interface 302 may expose a system interface that exports data center equipment information, or catalogs, to external storage.
  • This external storage may reside at various locations on various computing devices. Furthermore, as discussed below, the particular information that is exported to these locations and computing devices may be relegated to specific groups of data center equipment.
  • system 300 may be included in database module 212 of system 200 .
  • the work order module 214 may enable management of changes to the configuration of the data center, such as those resulting from a data center retrofit process as discussed below with regard to process 600 .
  • work order module 214 analyzes any changes made to the data center configuration by the other modules in system 200 . Then work order module 214 generates a set of tasks that, once complete, will implement those changes. Next, work order module 214 translates the set of tasks to a set of work orders in human-readable form. Lastly, work order module 214 facilitates tracking of those work orders to completion. To accomplish these objectives, work order module 214 may expose various UIs and system interfaces to enable communication and interoperation with external entities.
  • work order module 214 exposes a system interface through which integration module 210 may supply any changes made to the data center configuration.
  • Work order module 214 may also expose a user interface including elements that enable a user, such as a data center manager, to modify and dispatch work orders to other users, such as technicians, for implementation.
  • Work order module 214 may also expose a user interface to enable a user to configure a set of automatic dispatching rules.
  • work order module 214 may expose other user interfaces that enable users, such as technicians, to modify the details, including status information, of work orders. These user interfaces may reside on various computing devices, including remote or mobile computing devices.
  • Implementing such a user interface on a remote or mobile computing device may allow users, such as technicians, to update the CMDB of the data center as configuration changes are made. This may produce several benefits including increased accuracy of the CMDB, due to the increased compliance with, and promptness of, configuration change updates. Another benefit may be increased productivity for users, such as technicians, because configuration changes may be entered while working on data center equipment, rather than entering changes at a workstation as a separate activity.
  • Work order module 214 may also implement modification and assignment system interfaces to interoperate with other systems. For example, work order module 214 may, through a system interface, utilize an external email system to notify users of a work order assignments. Lastly, work order module 214 may utilize a system interface that enables it to detect changes in data center resource demands and may use this information, when appropriate, to modify the status information of a work order.
  • a user may create groups of data describing data center equipment. These groups may enable management of multiple types and/or models of data center equipment as a single collective entity. For example, inclusion of data describing data center equipment in a group may designate such equipment as approved for use within all of the data centers owned by a customer or specific subsets thereof. Any data center equipment approved for use within a data center may be used during the design or retrofit of the data center.
  • a first group is created that includes all of the data center equipment supported by the system
  • a second group is created that is a subset of the first group and that designates equipment that is approved for use within the data centers of a particular customer
  • a third group is created that is a subset of the second group and that designates equipment that is preferred for use within a particular customer data center.
  • a user may target specific groups for exportation to specific customers, customer data centers or specific computing devices within a customer data center, such as remote or mobile computing devices.
  • Fashioning groups of equipment data may ease maintenance and distribution of electronic data center equipment catalogs, which may include data describing characteristics of data center equipment. For example, specific inclusion of data center equipment data in a group that is lower in the hierarchy, e.g. a group designated for distribution to a remote or mobile computing device within a specific data center, may require that the data be present in ancestor groups. This inclusion may be performed automatically by the system. These groups may also be employed to enforce standardization of data center equipment within the customer organizational hierarchy by requiring that equipment be approved for use by the customer before it is allowed to be used to retrofit a specific customer data center.
  • the system provider may create a group of data describing standard data center equipment.
  • a group of the standard data center equipment data may be approved for use in customer data centers.
  • An equipment catalog including data associated with the standard and/or approved data center equipment may be delivered with the initial system installation in a data center.
  • the standard and/or approved equipment data may then be supplemented with customer specific, non-standard equipment data used in a customer data center, as discussed with regard to stage 304 below.
  • Approved equipment may be further grouped into a preferred set for use with one or more customer data centers.
  • the creation of these preferred groups may be explicitly performed by the user or may be implicitly performed by the system based on usage of specific equipment by a customer or within a customer data center during data center design or retrofit.
  • stage 302 of the method 300 information regarding the facility is loaded into the system.
  • the information includes, for example, dimensions of the facility, number of rooms, locations of doors, support columns, other obstructions, parameters of data center resource capabilities, such as available power, cooling capabilities of the facility, whether a raised floor or drop ceiling is in use, and characteristics of any such floor and roof.
  • Data center resource supply policies may also be entered in this stage.
  • capability information may be loaded by receiving data center equipment information from interface 302 of system 300 .
  • this stage of method 300 collects information helpful in determining the data center resources provided to particular data center locations and spaces. In at least one embodiment, as discussed further below with regard to cooling, this information may be further processed to determine the data center resources provided to particular data center locations and spaces.
  • stage 304 of the method information regarding equipment to be installed in the facility is entered.
  • the information includes, for example, the number of racks of equipment, maximum power draw for each of the racks, dimensions of the racks, and cooling requirements for the racks.
  • the need for backup power sources and multiple power sources for equipment and or racks may also be entered at stage 304 .
  • data center resource consumption and production characteristics of individual pieces of equipment that are to be loaded into racks may also be entered.
  • the weight of equipment (including equipment loaded into racks) may be used to ensure that the weight of the installed equipment is within any facility constraints. These characteristics may include, in addition to network connectivity, power and cooling requirements, the amount of rack space that the equipment needs to occupy and the type and/or number of electrical plugs that the equipment requires.
  • the database module 212 contains information regarding a number of devices, such as uninterruptible power supplies, equipment racks, cooling units, generator systems, power distribution units, automatic transfer switches, electrical routing devices, including cables, and servers and other computer equipment.
  • devices such as uninterruptible power supplies, equipment racks, cooling units, generator systems, power distribution units, automatic transfer switches, electrical routing devices, including cables, and servers and other computer equipment.
  • interface 302 of system 300 provides these device/equipment characteristics to database module 212 .
  • Equipment related to fire protection and security may also be included in the design.
  • all equipment and components within equipment racks may include RFID tags, which can be used by systems of the invention to track location of equipment and racks.
  • any data center equipment characteristics that are added or changed by a user at this or the previous stage may be transmitted to interface 302 of system 300 for importation into database 304 .
  • the system determines a layout for the equipment in the facility, taking into account the data center resource requirements, such as power and cooling requirements, of the equipment as well as other characteristics of the equipment that were entered at stage 304 or retrieved from the database module.
  • the user may create the layout graphically, adding racks and other equipment where desired, and in this embodiment, the system will provide feedback during the layout process, disallowing some choices and making intelligent suggestions.
  • These rules may include, for example: a standard alternating hot aisle/cold aisle layout must be specified, the plenum must be greater than some minimum value, the total room cooling capacity must exceed total room cooling load, aisles must be wide enough for access purposes and to meet building codes, distance between PDU and IT racks served by the PDU must not exceed some maximum value, PDU must be located immediately adjacent to a UPS, where a cable ladder spans an aisle, the aisle cannot exceed a maximum width, one or more data center resource capacities must be at a level sufficient to support target data center resource redundancy and/or runtime requirements, etc.
  • a cooling analysis is conducted to determine if the design provides adequate cooling for each of the racks and the equipment installed in the racks.
  • one of a number of different methods may be used to conduct the cooling analysis.
  • the procedure may return to stage 306 to change the layout of the equipment based on feedback provided from the analysis conducted at stage 308 .
  • a room model is displayed showing the locations of the equipment in one or more rooms of the facility.
  • the room model may include, for each equipment rack, or for individual data center equipment, information regarding the total data center resources, such as power and cooling, being consumed or produced as well as an indication of total available data center resources, such as power and cooling, to the rack or data center equipment.
  • actual data center resource data such as power and cooling data
  • colors may be used, either alone or in combination with data, to display different levels of data center resource availability, such as power and cooling availability.
  • the rack may be indicated in green on the display, if the cooling air availability is closer to the threshold, the rack may be indicated in yellow, and if the rack does not have sufficient cooling air it may be indicated in red. Still further, the results of the analysis may indicate that adequate data center resources, such as power and/or cooling, are being provided for equipment, but that specified redundancy levels and/or runtime margins are not being met, either at the room level, a row level, a rack level, or at a specific piece/element of data center equipment.
  • adequate data center resources such as power and/or cooling
  • the system may display multiple room models and may allow the user to find one or a series of alternate satisfactory locations for the equipment based on the data center resource requirements of the equipment and any applicable redundancy and/or runtime requirements. Another embodiment may allow the user to find an optimum location followed by a series of alternate satisfactory locations for the equipment. Each series of locations may be arranged consecutively according to an order of preference, such as in order of decreasing excess or stranded capacity. As discussed further below, stranded capacity includes excess capacity that is nominally available, but unusable, due to insufficient associated capacity of another data center resource required by data center equipment.
  • Yet another embodiment may allow the user to specify both the data center equipment and location and may validate whether the location provides sufficient data center resources to satisfy the requirements of the equipment and any applicable data center supply policies. This validation may be performed by comparing the data center resources provided to the location, the data center requirements of the equipment and any applicable data center resource supply policies. The result of this comparison, a compliance result, may be stored for later use.
  • the system may suggest data center equipment to be placed at a user specified location. In this case, the system may ensure that applicable data center resource supply policies, and data center resource requirements of the suggested equipment, are met by comparing the polices, the data center resources provided to the location and the data center resource requirements of the equipment prior to suggesting the equipment based on a compliance result. Specific example details regarding the room model are described further below with reference to FIGS. 5 through 5D .
  • a determination may be made by, for example, a facility designer as to whether the layout generated in stage 310 is satisfactory. The determination may be based on additional criteria of importance to the designer that was not included during the design of the original layout. For example, it may be desirable to have certain racks near each other or to have certain racks isolated from one another.
  • additional criteria or other feedback can be provided and the process then returns to stages 306 and 308 where the room model can be refined. Stages 306 to 312 may be repeated until a satisfactory model is achieved at stage 312 .
  • a bill of materials is generated and may be used to provide the cost of the equipment to be installed in the facility and may also be used to generate a sales order for the equipment, providing a simple solution for ordering all equipment associated with a new data center.
  • CAD drawings and electronic files that capture the designed layout may also be generated.
  • this data center configuration is transmitted to interface 302 of system 300 for storage in database 304 in the form of a CMDB for the installed data center.
  • the equipment is installed in the facility according to the layout generated at stages 306 to 314 .
  • measurement equipment to measure cooling characteristics and power characteristics may be installed with the equipment.
  • the measurement equipment is described further below, and may include, for example, devices for measuring power, airflow, humidity and temperature at various locations in the facility and within equipment racks located in the facility.
  • power and cooling parameters are measured using the measurement equipment. Additional temperature measurements may also be provided by devices, such as servers, that have the capability to detect internal temperatures. The parameters measured may be used continuously by the management module of the system 200 to detect error conditions and to monitor trends that may lead to an error condition. Further, in the process 300 , the measured parameters can be compared with predicted parameters calculated during the design process in stages 306 and 308 . For example, in one embodiment, the airflow through a perforated floor tile of a raised floor is used to determine the available cooling air of a rack located adjacent the floor tile.
  • the airflow through the perforated tile may be determined in stage 308 using one of a number of computational methods that are described further below, or the airflow may be determined using data from related physical measurements or simulations.
  • the perforated floor tile may be instrumented to measure the actual airflow through the tile. The actual measured value may then be compared with the calculated value at stage 320 . If the two differ by more than a predetermined threshold, then an indication or warning may be provided and the calculations conducted in stage 308 may be conducted once again at stage 322 using measured values in place of calculated values as appropriate to obtain updated parameters.
  • measured parameters are transmitted to interface 302 of system 300 for storage in database 304 . As discussed above, the storage of these measure parameters by interface 302 may trigger further analysis and summarization of the measure parameters into data center equipment consumption and production benchmarks.
  • stage 322 the model of the facility described above with reference to stage 310 may be displayed with values of power and cooling availability and consumption updated to reflect any differences between measured parameters and calculated parameters. Any out of tolerance conditions (for either cooling or power) may be indicated on the display using, for example, a color coded scheme as described above.
  • a user may be provided with a number of available options to correct an out of tolerance condition. The options may include upgrading or adding facility equipment (i.e., an air conditioning unit or an uninterruptible power supply) or may include moving equipment and/or racks. Stages 318 to 322 of the process may be performed continuously as part of a management system of the data facility.
  • stages 302 to 314 of the process 300 are implemented using a build-out system accessible by a user over the Internet.
  • the user provides the requested information
  • the build-out system provides the processing described above, provides outputs to the user over the Internet, and stores results locally.
  • the management system 500 (described below) may access the build-out system to download information related to the equipment.
  • the management system may contact the build-out system to coordinate the design of the retrofit.
  • electronic files may be imported/exported between the systems to provide a complete transfer of all information related to a data center's design.
  • FIG. 5 shows an example of a display of a room model that may be generated using the system 200 and the process 300 and shown on a computer display. It should be appreciated that this computer display may be any be coupled to, or included in, any sort of computing device including a remote or mobile computing device.
  • the room model shown in FIG. 5 is essentially the data center 100 previously discussed above with reference to FIGS. 1 and 2 . However, in FIG. 5 , additional data related to the power and cooling consumption and capacity of each rack, U space positions, and/or elements of data center equipment housed within each rack may be included in an informational block, such as informational blocks 120 A and 120 B shown on two of the racks 108 A and 108 B in FIG. 5 . Informational blocks may be included on data center equipment, each rack, or on less than all racks, for example, by row, zone, or cluster.
  • FIGS. 5A and 5B show enlarged views of respectively racks 108 A and 108 B that may also be shown on a computer display of systems of embodiments of the invention.
  • specific information regarding the racks and U space positions is included in the informational block.
  • the information in the block includes a rack identifier 122 , a rack type 124 , power capacity 126 , power usage 128 , cooling capacity 130 , cooling usage 132 , contents of the rack 134 , power redundancy 136 , cooling redundancy 138 , and UPS runtime 140 .
  • the information in the block includes a rack identifier 122 , a rack type 124 , power capacity 126 , power usage 128 , cooling capacity 130 , cooling usage 132 , contents of the rack 134 , power redundancy 136 , cooling redundancy 138 , and UPS runtime 140 .
  • UPS runtime 140 UPS runtime
  • information for each rack may be included in tabular form, or in the form of columnar bars, on a graphical display showing the room layout.
  • Data center information that may be displayed includes runtime measurements, data center resource redundancy measurements, including relationships between data center equipment working in combination to providing redundancy, and capacity measurements, such as available capacity, utilized capacity, and stranded capacity.
  • the rack identifier 122 includes a row number and a rack number, however, in other embodiments, the rack identifier may also include an indicator of the type of rack, membership of the rack to a particular row, zone, group, or cluster, manufacturer of the rack, as well as other information.
  • the rack type 124 identifies the particular type of rack, i.e., server rack, router rack or telecommunications rack.
  • the power capacity 126 indicates the maximum power capacity of the rack, and the power usage indicator 128 indicates the percentage of maximum capacity at which the rack is expected to operate. In different embodiments, the power usage indicator may be calculated based on manufacturer supplied data for equipment contained in the rack and/or based on actual power measurements of equipment.
  • the power capacity for a rack is determined based on limitations of devices and/or power cables that supply power to the rack, such as circuit breakers, UPSs or any other devices.
  • the contents of the rack 134 includes a list of the equipment contained in the rack and may include an indication of remaining space in the rack displayed, for example, in terms of rack units, which are typically referred to as “U” with 1 U equal to 1.75 inches. Details regarding the equipment in the rack, including operational status and network addresses, such as an IP address for a device may also be included.
  • the cooling capacity indicator 130 and cooling usage indicator 132 identify respectively the quantity of cooling air available to the rack and the percentage of that cooling air that is being used by equipment in the rack.
  • power and cooling usage may be indicated using various types of graphical gauges, such as a bar graph, that indicates power and cooling usage and capacity.
  • the cooling capacity is shown in terms of kilowatts (kW).
  • kW kilowatts
  • many equipment racks typically require approximately 160 cfm (cubic feet per minute) of cooling air per kilowatt of power used by the rack.
  • the cooling usage indicator is equal to the power consumed by the rack.
  • the cooling required by a rack may not be equal to that consumed by the rack and may be calculated based on manufacturer's data for the equipment, based on test results, or in any other manner.
  • the cooling capacity of an equipment rack is determined based on a number of different factors. For example, for a raised-floor environment, these factors may include: location of the rack in the facility, proximity of perforated tiles to the rack, the amount and temperature of cooling air provided through any such perforated tile, the physical or geometric layout of the racks and building, and the cooling requirements of other equipment in the facility located near the rack.
  • the cooling capacity of one rack in a facility may be affected by the cooling usage of nearby racks, and accordingly, in one embodiment, the cooling capacity of a rack is adjusted when the cooling usage of a nearby rack is changed.
  • calculations for determining cooling capacity of a rack are based in part on the ability of one rack to borrow cooling air available to adjacent racks. Particular methods for determining cooling capacity for racks in embodiments of the present invention are described further below. In one embodiment, when the cooling usage of one rack is changed, the cooling capacity of that rack, and all racks located near the changed rack is recalculated.
  • the true available capacity of a data center can be determined at U space position level, the rack level, at the row level and at the room level.
  • available capacity including unused capacity
  • data center resources including both cooling and power capacity are used, and the true available capacity can be determined using the lowest data center resource capacity.
  • the excess capacity can be considered wasted or stranded capacity that can not be used in the present design.
  • the stranded capacity can be determined at the U space position level or rack level and can be totaled to determine stranded capacity at the row level and at the room level.
  • this data resource information may be displayed to the user in the form of graphs.
  • FIG. 21 depicts total capacity 2102 , space capacity 2104 , cooling capacity 2106 and power capacity 2108 , each as a function of time.
  • Each of these representations may include total data center capacity specific to the data center resource shown.
  • the representations may include data center capacity specific to a specific location in the data center, e.g. high density or medium density racks. Still further, these capacities may be projected into the future to enable a user, such as a data center manager, to predict when capacity may need expansion.
  • FIG. 23 illustrates an embodiment that a user, such as a data center manager, may reference to assess stranded data center capacity and ascertain the root cause of the stranded capacity. More specifically, FIG. 23 depicts report 2302 which includes rows and columns under various data center resource headings. The intersection of these rows and columns may indicate how much excess capacity of the data resource listed in column 2304 is unusable due to insufficient capacity of the data resource listed in row 2306 . For example, intersection 2308 may indicate that 20% of the weight support capacity of the data center is unusable due to insufficient power capacity.
  • the actual amount of stranded capacity may be used to articulate these relationships and the meaning ascribed to the rows and columns may be transposed. As will be apparent to one of ordinary skill in the art, other embodiments illustrating this relationship are possible without departing from the present invention.
  • the recommendations may include reducing capacity of data center resources, such as power and cooling resources, (thereby reducing operational costs) or adding data center resource capacity, such as cooling capacity or power capacity, appropriately to reduce the stranded capacity. Warnings may be generated when the stranded capacity is greater than preset thresholds, and in addition, recommendations for more optimal locations of equipment, including power and cooling equipment, may also be provided to minimize the amount of stranded capacity. Further, costs of the stranded capacity may be calculated.
  • data center resource capacity and availability such as power and cooling capacity and availability
  • changes to the availability rate are monitored and based on these changes, the growth rate of a data center may be determined, and predictions of dates when additional capacity will be required can be provided.
  • the forecasting method used to make these determinations is linear regression. Other forecasting methods are possible without departing from the scope of the present invention.
  • the ability to monitor capacity and predict future capacity needs allows data center operators to control costs and plan for upcoming expenses. Further, determinations may be made as to the additional expenses that will be incurred if new equipment is added. The total cost (for example per kilowatt) can also be determined during the design phase or during operation.
  • data center layouts may be designed to provide specific redundancy levels (i.e., n, n+1 or 2 n) for data center resource design, such as both the power design and the cooling design.
  • additional room cooling units are typically provided to include some redundancy in a data center, such that an overall cooling capacity of the data center can be maintained, even when one or more of the room cooling units fails or must be powered down to conduct maintenance.
  • the cooling redundancy is designed at the room level and not the rack level, and while overall cooling capacity may meet redundancy requirements, cooling at the rack level may not meet the desired redundancy requirements.
  • the ability to provide accurate airflow data at the rack level and U space position level allows true cooling redundancy to be designed into a solution.
  • graphical user interfaces may be used with embodiments of the present invention to assist in the design and management of data centers.
  • these graphical user interfaces may be used to present an interactive representation of a layout of a various spaces in the data center.
  • interactive representations may be tailored to particular spaces within the data center, such as the entire data center, data center floors, rooms, racks and U spaces.
  • these interfaces may be displayed on a local or remote computing device.
  • Remote computing devices may include general computer systems, mobile computing devices and computing devices that are included with, embedded in, or affixed to, another physical structure or device, such as a wall, ceiling, another computer system or an enclosure, e.g. a rack enclosure, etc.
  • FIG. 5C shows a floor editor interface 402 used in one embodiment to layout equipment in a data center
  • FIG. 5D shows a rack editor interface 404 used in one embodiment to provide further details of the contents of equipment in the data center.
  • tutorials are provided for a user to assist the user by providing best practice design guidelines. The tutorials may be accessed by a user as desired or may be configured to be displayed as a user is taking a particular action.
  • the floor editor interface includes a main menu 403 , a tool bar 406 , a configuration box 408 , a generic components box 410 , a floor layout box 412 , a status box 414 a full-image viewing box 416 , and an unplaced equipment box 418 .
  • the main menu 403 provides a drop-down menu in a format known to those skilled in the art, and allows a user to perform various functions, including the ability to “undo” and/or “redo” changes that are made to the layout.
  • the tool bar 406 provides short hand access to functions of the design system, and in one embodiment includes a floor editor button 406 A and a rack editor button 406 B. Activation of the floor editor button results in the display of the screen shown in FIG. 5C , while activation of the rack editor button results in display of the screen shown in FIG. 5D .
  • the floor editor box 412 shows the layout of equipment in a data center being designed and provides text that identifies the equipment contained in the layout.
  • a room perimeter 412 A shows the exterior walls of the room along with dimensions of the room that can be set by a user.
  • the user is presented with a screen showing a number of basic room configurations that can be selected.
  • the walls of the room can be moved by a user by selecting one of buttons 412 B, and additional buttons can be added where needed to expand or shrink an area of the room.
  • the room size may be changed as equipment is added to the layout.
  • Three rows 412 C, 412 D and 412 E are outlined in the room shown in FIG. 5C .
  • Row 412 D includes three equipment racks (identified by “R”), two half-rack cooling units (identified by “C”) a UPS (“U”) and a power distribution unit (“P”).
  • Row 412 E includes one rack, and row 412 C as presently configured does not include any equipment. During the design phase additional equipment may be added to each of the rows.
  • the room also includes an automatic transfer switch (ATS) 412 G and a cooling distribution unit (CDU) 412 F. Hatched areas are shown on the display around the ATS and CDU to indicate that these areas should be kept clear of equipment.
  • ATS automatic transfer switch
  • CDU cooling distribution unit
  • Each piece of equipment in the room may include identifiers that indicate the type of rack as well as the rack's location in the room and the power source for the rack. Further, as discussed above, each rack may include information regarding data center resource use and availability, such as power and cooling use and availability. Still further, text may be provided on each row to indicate total data center resource information, such as power and cooling information for each row.
  • the configuration box 408 includes eight configuration options for designing a data center.
  • a room properties configuration option when selected, allows a user to identify data center resource values, such as physical, power, and cooling values, that affect the data center design as a whole including dimensions of the room, aisle widths and total anticipated power density for the data center.
  • Data center resource redundancy and/or runtime requirements such as power redundancy requirements (i.e. N, N+1, 2N), cooling redundancy requirements and runtime requirements for UPS systems may also be set.
  • the number of data troughs that will be used and location of power distribution and cooling line distribution i.e. overhead or under a raised floor) can also be configured.
  • only in-row cooling is provided, however, in other embodiments other types of cooling solutions may be used as well.
  • individual rows may be rotated to different angles in the data center.
  • only one room is shown in FIG. 5C , at least one embodiment allows a data center to include multiple rooms. These rooms may be active rooms, which hold active data center equipment, and inactive rooms for storage of spare or decommissioned equipment.
  • An add rack configuration option in the configurations box 408 is used to add equipment racks to the data center design. When this option is selected, a user is presented with choices of various types of racks to add to the data center. When racks are selected, an indicator is provided in the unplaced equipment box 418 , indicating that the racks still need to be placed into the room layout.
  • An add in-row cooling option in the configuration box is used to add in-row cooling units to the data center design.
  • this option is selected, a user is presented with various types of cooling units that can be added in the rows.
  • an indicator is provided in the unplaced equipment box 418 , indicating that the cooling unit still needs to be placed in the room layout.
  • a power zone option in the configuration box is used to identify and select PDUs and UPSs and to indicate which equipment will be powered from the UPSs and PDUs. Characteristics of the PDUs and UPSs may also be selected. Once selected, an indicator is provided in the unplaced equipment box 418 for the UPSs and PDUs. In one embodiment, multiple racks may be included in a selection on the layout to identify the equipment that belongs to a particular power group, also known as power zone.
  • an automatic power zone option may be implemented in which the system matches equipment power requirements (i.e., redundancy levels, runtime durations, voltages, phasing) to those of the UPSs and PDUs and assigns power zones automatically and determines lengths of power cables that are needed to power equipment from the assigned PDU.
  • equipment power requirements i.e., redundancy levels, runtime durations, voltages, phasing
  • the system may also identify the need for additional UPSs and PDUs.
  • a power generation option in the configuration box 408 is used to identify and select an automatic transfer switch (ATS) and generator. Again, once these are selected, an indicator is provided in the unplaced equipment box 418 .
  • ATS automatic transfer switch
  • An emergency power off option in the configuration box is used to select an emergency power off (EPO) solution for the data center design, and once selected, an indicator for the EPO solution will be added in the unplaced equipment box.
  • EPO emergency power off
  • a management option in the configuration box 408 allows a data center manager, such as the InfrastruXure® Manager and/or InfrastruXure® Central discussed above, to be added.
  • a data center manager such as the InfrastruXure® Manager and/or InfrastruXure® Central discussed above.
  • a rack location for the manager is also selected.
  • a service option in the configuration box 408 allows a user to select a level of service to be provided to the data center by a data center services organization.
  • Other configuration options may include a row planning configurator that allows a user to plan how many racks a row can support by defining the power and cooling settings for the row, prior to placing equipment in a row.
  • the row planning configurator may provide an estimate on the number of racks that can be supported based on the power components and in-row cooling units contained in the row.
  • the row planning configurator may provide a complete layout based on design best practices.
  • the generic components box 410 includes a number of icons to designate pre-existing equipment in a data center.
  • the components may be selected and “dragged” into position in the layout.
  • the generic components include blocks and gaps.
  • the gaps can be used to identify areas over which cables and conduits can be routed (i.e. a walkway), while the blocks are used to identify areas over which cables and conduits can not be routed (i.e. a column). Once dragged onto the layout, the blocks and gaps can be sized appropriately.
  • an icon appears in the unplaced equipment box 418 .
  • the icon is selected and dragged into the appropriate location in the layout.
  • Existing equipment may be repositioned using this same method. For example, existing equipment may be dragged from an active room and dropped in an inactive storage room, thus enabling the system to track unutilized equipment available for provisioning elsewhere.
  • the icon for the cooling unit can be placed between two adjacent racks and released, and the racks will then move in the row to allow the cooling unit to be inserted between the racks.
  • a snap-to feature is employed to align equipment properly in rows and along walls, and in addition, rows and equipment may be aligned along and “snapped to” floor tiles when, for example, a raised floor is in use. Using this feature, a user does not need to precisely align objects.
  • the full-image viewing box 416 provides a “bird's eye” view of the layout contained in the floor layout box 412 .
  • the zoom button on the tool bar can be used to enlarge the view of the data center layout in the floor layout box 412 .
  • the entire layout may not appear in the floor layout box.
  • the full-image box 416 still displays the full image of the layout for the user.
  • an overlay is used in the full-image box to indicate on the full-image, the portion of the layout that is displayed in the floor layout box.
  • the overlay may be selected and dragged within the full-image viewing box to select which part of the layout is displayed in the floor layout box.
  • the status box 414 is used to display warnings, errors and other conditions to the user.
  • the warnings may vary in severity and may include indications that design guidelines are being violated and may also include more severe warnings indicating that data center resource capacities, such as power and cooling capacities, have been exceeded, or that redundancy and/or runtime requirements are no longer met.
  • the status box indicates that there is an error or warning associated with a particular piece of equipment in the layout, the piece of equipment may be highlighted with a color such as red or yellow.
  • guidelines for correcting the error or warning are provided by either selecting a highlighted piece of equipment or the error or warning message directly.
  • the rack editor interface 404 will now be described further with reference to FIG. 5D .
  • the rack editor interface includes the tool bar 406 , the status box 414 and the full-image viewing box 416 discussed above. Further, the rack editor interface 404 also includes a rack editor box, a product catalog box 422 and a rack content section 424 .
  • the rack editor box 420 shows the front face of each of the equipment racks in the data center layout with the racks being arranged by row. In FIG. 5 , two rows of racks 420 A and 420 B are shown. As shown in FIG. 5 , in one embodiment, only the equipment racks are shown in the rack editor box. When a particular rack is selected in the rack editor box, then the contents of the rack appear in the rack content box 424 , and components may be added to the selected rack. Racks may be selected in the rack editor box or may also be selected in the full-image view box 416 . When a rack is selected in the full-image view box, then the image in the rack editor box will shift, if necessary, to provide a view that includes the selected rack.
  • the product catalog box 422 provides a comprehensive listing of components that may be used in equipment racks in data centers. The user may select equipment to be included in each rack, and as equipment is selected, it is included in the rack content box 424 . The list may include only equipment of a particular manufacturer or may include equipment from several manufacturers. In one embodiment, all necessary hardware and cabling associated with rack equipment may be selected from the product catalog box.
  • FIG. 20 a user may review and manage data center resource capacity and available supplied to data center equipment. Many of the features of this embodiment are similar to those described in reference to FIG. 5C above.
  • FIG. 20 includes capacity group explorer 2002 , which presents a logical grouping of data center equipment, such as racks, based on common capacity requirements. This grouping enables a user, such as a data center designer, to manage sets of data center equipment as one collective entity for capacity planning purposes. Representations of individual elements of data center equipment, such as UPS 2004 , Rack 2006 , and PDU 2008 may present the user with columnar bars depicting various data center resource redundancy and/or runtime measurements and capacity measurements, such as available capacity, utilized capacity, stranded capacity.
  • Rack 2006 has an as-configured potential power and cooling demand of 28.8 kW and a current actual power and cooling demand of 7.92 kW.
  • these demand or consumption measurements may be matched against as-configured potential power and cooling supply capacity and actual power and cooling capacity to ensure all capacity, consumption, redundancy and runtime requirements are met.
  • rack editor interface 404 may be displayed on a remote or mobile computing device.
  • Rack editor interface 404 includes user interface elements that allow for adding equipment 2202 , deleting equipment 2204 , editing equipment 2206 and moving equipment 2208 .
  • This embodiment further includes data center explorer 2210 and rack editor box 2212 . Because rack editor interface 404 may be provided on a remote or mobile computing device, users, such as technicians, may document as-built conditions, verify rack installations, troubleshoot installation in a more efficient manner and perform other rack-oriented activities.
  • equipment racks are shown in data center explorer 2210 .
  • the rack editor box 2212 shows the front face of the equipment rack currently selected in data center explorer 2210 .
  • rack editor box 2212 may display the rear face of the rack and/or the front face of the rack.
  • rack Front is selected in data center explorer 2210 and houses data center equipment sss, PowerEdge 2650 , PowerEdge 6650 , etc.
  • the equipment it houses may be modified using user interface elements 2202 through 2208 .
  • the user may add data center equipment to the currently selected rack by actuating user element 2202 and selecting the desired equipment and position with the rack.
  • the user may delete or edit data center equipment associated with the currently selected rack by selecting the desired equipment from data center explorer 2210 and actuating user element 2204 or 2206 , respectively.
  • the user may move data center equipment associated with the currently selected rack by selecting the desired equipment from data center explorer 2210 , actuating user element 2208 and selecting the desired position within the rack.
  • the system may recommend a satisfactory or optimum position. It should be appreciated that the user may select particular data center equipment by drilling down in data center explorer 2210 or by searching the data center using a unique identifier, such as a barcode scanned by the remote or mobile computing device. These methods of searching may allow a user, such as a technician, to quickly obtain information concerning specific data center equipment or modify the CMDB while physically located near the equipment.
  • an interactive representation of an identified space within a data center may be implemented using other interface technology.
  • a three-dimensional option is available allowing a user to view the design of a data center in 3D.
  • a design system includes software code programmed in Java that is used to generate 3D models that are rendered via OpenGL to allow for hardware acceleration.
  • 3D models may be exported from the design system to CAD tools such as AutoCAD, available from AutoDesk of San Rafael, Calif.
  • security cameras can be implemented into data centers designed using embodiments of the present invention.
  • security cameras may be included in the design and the 3D view may be used to view a simulation of a camera's view after installation.
  • networked cameras and other security monitoring devices available from Netbotz Corporation of Austin, Tex. may be used.
  • the system 200 may be used as part of a data center management system.
  • the management system may include the system 200 described above with the management module containing routines to perform management functions, or in other embodiments, the management functions may be performed by a designated manager controller contained in the data center and implemented, for example, in a computer server located in one of the racks of equipment and accessible by a user using a management console.
  • FIG. 6 shows a block diagram of a management system 500 that may be used in embodiments of the present invention.
  • the management system includes the manager controller 502 , the manager console 504 , power measurement devices 506 , and airflow, humidity and temperature measurement devices 508 .
  • manager console 504 may be implemented as any computing device, including a remote or mobile computing device. Implementing manager console 504 on a remote or mobile computing device may allow a user, such as a data center manager, to, for example, dispatch work orders to technicians while physically inspecting, installing, moving, and/or changing data center equipment.
  • the management system may include power control devices 520 to control application of power to one or more individual devices or racks contained within a data center, and the system may include airflow controllers 521 to control the airflow or supply temperature of an air conditioning unit or to control, for example, perforated tile dampers.
  • the management system may also include one or more security devices 523 , including security cameras.
  • the devices of the management system 500 may be directly coupled to the manager controller or may be coupled to the manager controller using a network 522 that may be a dedicated network, may include the Internet, or may include a LAN or WAN contained in the data center.
  • the manager controller may communicate with one or more servers 524 to obtain information from and control operation of the servers.
  • the management controller 502 may be implemented at least in part using an InfrastruXure® Manager and/or InfrastruXure® Central available from American Power Conversion Corporation (APC) of West Kingston, R.I., and devices may be coupled to the manager using, for example a controller area network (CAN) bus or an Ethernet network.
  • the power controllers and airflow controllers may be implemented using available known devices that monitor and/or control power and airflow in facilities.
  • the management system 500 may include systems and methods for monitoring and controlling power as described in U.S. Pat. No. 6,721,672 to Spitaels et al, which is incorporated by reference herein.
  • the management controller may communicate with the cooling units to control the units to ensure that adequate cooling at specified redundancy levels is being met. Further details regarding the control of in-row cooling units that can be used with embodiments of the invention are discussed in copending U.S. patent application Ser. No. 11/335,901, discussed above and filed on Jan. 19, 2006.
  • the retrofit system may be a standalone computer system configured to perform processes described herein, or in one embodiment, the retrofit system is implemented using the system 200 described above.
  • the builder module 206 of the system 200 may include routines to assist in retrofitting a data center.
  • FIG. 7 shows a flow chart of the process 600 .
  • data related to a present configuration of a data center to be retrofitted is provided to the builder module.
  • the data related to the present configuration may include the data displayed in the room model of FIG. 5 along with additional data that was generated during design of the data center.
  • the data related to the present configuration may include data generated during an initial design as updated by actual measurements conducted in a facility. For example, the cooling capacity of individual racks may be calculated in an initial design and then updated by the management module once the system is installed and operating. Cooling capacity data may be updated based on actual measurements of airflow from, for example, perforated floor tiles, while the original data may have been calculated based on predicted airflows.
  • interface 302 of system 300 provides this data center configuration and measured parameter information from database 304 .
  • Information related to the retrofit is then provided in stage 604 .
  • the information related to the retrofit may include information similar to that input at stage 304 of process 300 described above, such as type of equipment, characteristics of equipment, number of racks, as well as other information.
  • a user may designate one or more desired locations in the data center for the installation of new equipment. For example, a user may desire to add five additional servers to the data center, where the servers are similar to and have a related function to existing servers in the data center. The user may choose one or more preferred locations based on power specifications, cooling specifications and physical dimensions of the servers, and based on power capacity, power plug type and/or number, cooling capacity and contents of existing racks displayed on a floor model of the data center.
  • the user may indicate whether it is acceptable to move existing equipment to accommodate the installation of new equipment.
  • a user may desire to replace three servers in the data center. The user may choose the servers targeted for replacement and may make other data center modifications with these replacement servers in mind.
  • the user may indicate whether the replaced equipment should be removed from the data center or moved into an inactive storage room. Tracking inactive equipment may allow a user, such as a data center manager, to quickly ascertain equipment available for provisioning within the data center.
  • an updated layout for the data center is generated and cooling and power calculations are performed at stage 608 on the updated layout in the manner discussed above at stage 308 of process 300 .
  • the layout may first be determined using these locations, and if problems arise as a result of the desired layout (i.e., lack of cooling for a rack based on equipment or supply policy requirements), then the user will be able to note any such problems once the layout is displayed and can then choose to change the layout.
  • the system may suggest one or more elements of data center equipment to be placed at one or more locations. In this case, the system may ensure that applicable data center resource supply policies, and data center resource requirements of the suggested equipment, are met.
  • the system may provide the user with one or a series of satisfactory locations for the equipment based on the data center resource requirements of the equipment and any applicable data center resource redundancy and/or runtime requirements. In another embodiment, the system may provide the user with an optimum location, followed by other satisfactory locations in a decreasing order of preference, for the new equipment based on the data center resource requirements of the equipment and any applicable data center resource redundancy and/or runtime requirements. If a particular layout is not designated by a user, then the system 200 will determine the layout in the manner discussed above with respect to stage 306 of process 300 .
  • an updated floor model is displayed (for example, in the manner shown in FIGS. 5C and 5D ), and at stage 612 , a user can review the model and either provide feedback (stage 614 ) or indicate that the design is satisfactory.
  • the retrofit design process is complete, and the equipment may be installed, relocated, or removed and parameters of the data center may be measured and updated in the manner described above in stages 318 to 322 of process 300 using for example a data center management system.
  • changes to the configuration of the data center such as those designed by retrofit process 600 may be implemented by work order process 1900 as illustrated by the flowchart in FIG. 19 .
  • a data center configuration change may be made using a graphical user interface such as those discussed with regard to FIGS. 5C and 5D above. Details of the change may be used to construct a set of tasks to carry out the change. These tasks may be grouped together into work orders. Work orders may be human-readable and may include textual instructions as well as pictorial representations. It should be appreciated that a single configuration change, initiated using a visual editor, may be translated into multiple work orders.
  • a work order is dispatched to a user for implementation.
  • this dispatching may be automatic or manual.
  • Automatic dispatching may occur based on a set of preconfigured rules. These rules may consider characteristics of both the work to be performed and the user to whom the work order may be dispatched. Characteristics of the user that may be considered include the user's physical location, physical area of responsibility, recent and current work load, remaining schedule availability and area of expertise.
  • a work order may be automatically dispatched to one or more users in a “round robin” fashion.
  • a user such as a data center manager, may access the dispatch UI exposed by work order module 214 , discussed in relation to FIG. 3 above, to modify and/or dispatch work orders to users.
  • a user such as a technician, may access the dispatch UI exposed by work order module 214 to modify and/or dispatch work orders to himself.
  • a user performs the work dictated by the work order and modifies the work order appropriately.
  • Modification of the work order may include, for example, changing the details of the work performed such as recording installation of an alternative model of equipment, rearrangement of the order of equipment within the rack, reconfiguration of the flow of a data center's resources, such as power, to the equipment, etc. Modification may also include changing the status of the work order.
  • a user modifies a work order using a remote or mobile computing device certain information required to complete the modification may be entered using any interface supported by the remote or mobile computing device, including barcode scanning.
  • work order module 214 may monitor a data center resource provider, such as power outlet, that is targeted as part of the work order for a change in demand.
  • a change in demand is detected, such as an increase in the power demanded at a power outlet
  • the work order may be marked as complete.
  • the work order is not marked as complete until the data center resource consumption or production benchmark, such as the power consumption benchmark, of the data center equipment specified on the work order is recognized. For example, if a work order calls for the installation of a particular model of blade server, the system may not mark the work order as complete after initial activation of the blade server, but instead may wait until the power consumption history of the blade server matches a known power consumption benchmark for the model of blade server called for in the work order.
  • design and analysis stages are performed after all data is entered as part of an initial design or a retrofit of a facility.
  • analysis is performed real-time, and user displays are updated as the user enters data into the system.
  • data center operators are able to determine, in essentially real-time, whether additional equipment may be added to a data center and may also determine locations for the equipment, where data center resource requirements, such as both power and cooling requirements, of the equipment may be met. Further, reports may be generated that indicate to a user or data center manager how much capacity, redundancy and/or runtime is available for each row, for each rack, for each U space position, for each piece of data center equipment and for the facility in its entirety. Still further, as discussed above, in determining overall capacity, systems and methods are able to identify and display stranded capacity, and provide suggestions for reducing the stranded capacity.
  • cooling calculations for a data center and for equipment in the data center are performed as part of the design process for the data center, during operation of the data center, and during an upgrade or retrofit of the data center.
  • initial information on characteristics of the facility itself are identified to determine if there is sufficient cooling at the facility level. These characteristics include, for example, whether a raised floor or drop ceiling is used as an air plenum, the location and characteristics of air conditioning units (including in-row cooling units), dimensions of the room that are to house the data center, and total power draw of the data center.
  • an initial determination may be made as to whether there is sufficient cooling provided by the air conditioning units for the expected power draw in the room, and if not, a recommendation may be made for additional air conditioning units. For some facilities, desired redundancy and operating margins may be included in this determination.
  • the cooling capacity of a rack may be determined by increasing the power level of the rack to determine at what additional power level the airflow to the rack becomes inadequate. This can be performed individually for each of the racks in a data center. In different embodiments of the present invention, one or more of a number of different methods may be used to perform the cooling calculations.
  • CFD computational fluid dynamics
  • CGCFD Coarse-Grid CFD analysis
  • CVA control volume analysis
  • an analysis based on empirical rules and/or borrowing concepts a combination of two or more of the above methods may be used to conduct portions of an overall analysis.
  • the principle of superposition is used to combine results of portions of an analysis.
  • the airflows may be considered to be ideal, where an ideal airflow is inviscid, incompressible, irrotational without any other forces, such as buoyancy.
  • a complex application can be reduced to a number of less complex applications, analysis of the less complex applications can be performed using one of the methods described herein, and superposition can be used to combine the results of each of the less complex applications to obtain analysis results for the complex application.
  • a computational fluid dynamics analysis can be used in one embodiment in association with the design and retrofit of a data center to provide detailed results of the cooling performance of a facility, including determining the availability of adequate cooling air at racks and individual pieces of equipment in the facility and determining cooling capacity for each rack.
  • the techniques for implementing a CFD analysis of a data center are known.
  • a CFD analysis must typically be performed by someone particularly skilled in the art, typically requires detailed information regarding the facility and the layout of equipment in the facility, and depending on the complexity of the analysis conducted, and the computing equipment used to conduct the analysis, may take hours or days to run one iteration of the analysis.
  • an improved technique for conducting the cooling analysis is used.
  • the improved technique has been developed based on computational fluid dynamics techniques.
  • the improved technique is referred to herein as a Coarse-Grid CFD or simply CGCFD.
  • a data center to be analyzed is typically divided into non-uniform cells in the range of one to eight inches on a side.
  • a Cartesian grid system of cells that are one foot cubes are used. The use of uniform one foot cells typically reduces the number of cells used in the calculations from a traditional CFD analysis by at least an order of magnitude. Further, uniform grid cells generally make the CFD analysis faster and more reliable relative to a comparable non-uniform cell analysis.
  • CGCFD analysis uses other techniques to improve the computational efficiency of the analysis. These techniques include: the use of simple turbulence models, initializing the analysis with data obtained from the results from a prior similar solution, using two dimensional or partial two dimensional representations when possible to simplify calculations, and tailoring a CGCFD routine for a specific application.
  • the use of two dimensional representations may be used, for example, in a raised floor or ceiling plenum, where pressure gradients in the depth direction can be neglected in the computations.
  • a CGCFD routine can be used in embodiments of the present invention to significantly improve computational efficiency and improve robustness (for example, so the tool can be made to work reliably in an autonomous way) of the total analysis, and multiple tailored routines may be used in combination to produce a complete analysis.
  • a first CGCFD routine may be tailored for use with different raised floor configurations to determine the output airflow at each perforated floor tile of a raised floor in a data center
  • a second CGCFD routine may be tailored for use with a cluster of racks that include two rows of racks with a cold aisle between the rows.
  • the first CGCFD routine may be run to determine the output air at perforated tiles in the cold aisle of the rack cluster, and the second CGCFD routine may use the results of the first routine to determine the airflows and temperatures at inlets and outlets of the racks.
  • the second routine may be run a number of times to account for all of the clusters of racks located in a data center. As equipment is moved, and as different configurations are established within a cluster to optimize cooling performance, the second routine may be run to obtain new cooling data without the need to repeat the first routine, as the airflows from the perforated tiles generally would not change. In some cases, for perforated floor tiles having a large percentage of open area (for example, greater than 50%), it may be desirable to repeat the first routine as air flows may change based on room configuration.
  • Embodiments of the invention that utilize the CGCFD approach to conduct the analyses of a data center provide advantages over embodiments that utilize a traditional CFD approach. These advantages include computational efficiency and simplification of use. Iterations of cooling calculations using the CGCFD approach may be conducted in a matter of seconds or minutes versus hours or days with a full CFD analysis. Further, the CGCFD routines may be designed to operate with a limited set of input variables, allowing a less-skilled user to conduct analyses using the CGCFD approach.
  • the input variables may be limited to the height of the floor, location and type of perforated tiles, length and width of the floor, and the locations and characteristics of air conditioning units that provide cooling air to the raised floor.
  • the input data may be limited to airflow per tile (could be obtained automatically from the output of a separate CGCFD routine or using other methods), the number of racks in the cluster, the power draw of each of the racks, and room environmental details including the temperature of the surrounding environment, ceiling height, the presence of nearby walls, etc.
  • the output data for a cluster of racks may include the input temperature at each server, or other piece of equipment in a rack.
  • the output data may simply be a measure of the amount of warm recirculated air drawn into each rack.
  • the data may be expressed as an absolute number (e.g. in terms of cfm) or expressed as a fraction of the total air consumed by the rack.
  • a system such as system 200 described above, may use the output data to determine if the cooling performance of the cluster is satisfactory.
  • the improved technique is referred to herein as a control volume analysis or simply CVA.
  • the control volume analysis may be used in conjunction with a CFD analysis and/or a CGCFD analysis, or may be used as a stand alone process.
  • the CVA technique is similar in some aspects to the CGCFD technique described above, however, further simplification of the analysis process is provided.
  • the CVA technique is a computationally efficient technique that is particularly effective for computing the three-dimensional airflow, pressure and temperature distributions in the cold aisle of a raised floor data center.
  • the CVA technique is not limited in its use to this application and may be used for other applications as well.
  • the CVA technique can provide output data in essentially real-time, allowing a user to try various locations for equipment as part of an initial design or a retrofit and obtain cooling data for the different locations in real-time.
  • FIG. 8 shows a subsection 700 of a data center.
  • the subsection of the data center includes a cluster of racks that includes a first row of racks 702 , and a second row of racks 704 , which are located on a raised floor and separated by two rows of perforated tiles 706 , 708 .
  • a high-power rack may borrow underutilized air from an adjacent rack, or in some cases from a rack separated by several tiles.
  • the airflow patterns and temperature distribution in the cold aisle are complex functions.
  • the CVA technique can be used to simplify the solutions to these complex functions.
  • the airflow and temperature analysis is conducted on the volume of air contained in the cold aisle, between the racks, from the perforated tiles up to a height equal to the top height of the racks.
  • the volume of air is divided into a number of control volumes equal to the number of racks in the cluster.
  • Each control volume is defined as the volume above one of the perforated tiles extending from the perforated tile to the top of the racks.
  • the control volume analysis includes determining for each control volume, the airflow through each of the six faces of the control volume. Once the airflows are known, temperatures and mass species concentrations can be determined for each of the control volumes.
  • the temperature analysis can be decoupled from the airflow analysis because, as discussed above, buoyancy forces in the control volumes can be ignored.
  • mass species concentrations are not coupled to the airflow solution and may also be computed separately if desired in order to compute the fraction of recirculated air ingested by each rack.
  • the first assumption is that airflow across each face of a control volume (and therefore into the front face of a rack) is considered uniform. Resulting airflow and temperature values effectively represent an average of the airflow and temperature at each face.
  • the second assumption is that buoyancy forces within each control volume are negligible. Unless a significant hot spot develops, then there is insufficient heating of the air in a cold aisle to substantially affect airflow patterns, and even if some heating occurs, any buoyancy effects are small compared to the momentum of the airflow from typical perforated tiles.
  • the third initial assumption is that viscosity and turbulence within each control volume are negligible.
  • air is introduced through the perforated tiles and is pulled into the racks. The air is not required to change direction rapidly and there is no flow of air parallel to solid objects. Accordingly, viscosity and turbulence may be ignored and the competition of forces driving the airflow reduces to an interplay between pressure and momentum.
  • the CVA analysis may be further simplified by dividing a cluster of racks into slices of two racks separated by two perforated tiles 718 A, 718 B.
  • FIG. 9 shows a cluster of six racks 710 that can be divided into three two-rack slices 712 , 714 , 716 .
  • the nomenclature used for identifying racks, and airflows in FIG. 9 is defined in Table 1 below, along with other variables that will be used herein in the description of the CVA technique.
  • the rack airflows QA i , QB i and the tile airflows are known.
  • the tile airflows can be estimated based on the mean perforated tile airflow for the entire facility or determined using a CFD analysis, a CGCFD analysis, physical measurement, or using one of a number of known programs.
  • the rack airflows can be determined based on characteristics of equipment installed in the rack. In one embodiment, each rack airflow is determined on the basis of the power usage of the rack and using the relationship of 160 cfm per kilowatt of power as discussed above.
  • all airflows QAx i , QBx i , Qz i , QAtop i , and QBtop i , and pressures PA i , and PB i are computed based on the principle of conservation of mass and momentum.
  • a total of 7 n ⁇ 2 unknowns (5 n ⁇ 2 internal airflows plus 2 n pressures) may be determined using a total of 7 n ⁇ 2 equations, where n is the number of 2-rack slices (or length of cluster expressed in tile or rack widths).
  • an energy balance or mass species balance can then be used to compute the 2 n temperatures or 2 n species concentrations based on the airflows.
  • a semi-simultaneous approach is taken.
  • the five unknowns for each two-rack slices of a rack cluster, airflows Qz i , QAtop i , and QBtop i , and pressures PA i , and PB i are first determined simultaneously.
  • each two-rack slice is considered in isolation, which is the equivalent of having the ends of the slices blocked, such that QAx i and QBx i are equal to zero.
  • the side-to-side airflows (QAx i , QBx i ,) can be computed based on the calculated pressures within each control volume.
  • the side-to-side airflows affect the pressures, and after computing the side to side airflows, a second computation of the airflows and pressures is conducted for each of the two-rack slices. This process is repeated until there are no significant changes in the computed variables. Once all airflows are known, all temperatures or mass species concentrations can be computed directly without the need to do multiple iterations.
  • staggered control volumes are used to formulate the z-momentum equations.
  • Three staggered control volumes 730 A, 730 B and 730 C are shown in FIG. 9A .
  • the number of staggered control volumes is equal to the number of 2-rack slices.
  • the staggered control volumes are the same size as the main control volumes, but are shifted so that they are located midway between opposing racks.
  • the staggered control volumes allow pressure to be considered more easily for each face which is normal to the z-direction. If the original control volumes are used, then each control volume would have one face coplanar with a rack inlet, which is a face over which the pressure is not known and need not be known in the calculations.
  • Applying the conservation of momentum equation (M) in the z-direction to the staggered control volume in slice i results in the following equation:
  • PA i ⁇ PB i ( ⁇ /(4 A s 2 )) ⁇ ( Qz i +QB i ) 2 ⁇ ( QA i +Qz i ) 2 ⁇ +M L +M R +M T (3)
  • equation (3) the first term on the right hand side of equation (3) is generally dominant, as it accounts for the effect of rack airflow rates on control volume pressures.
  • M L , M R and M T account for losses or gains in z-momentum through the sides and the top of the control volume.
  • M L , M R and M T are determined using the equations of Table 2 below
  • equations (1) through (5) are solved simultaneously for each 2-rack slice of a cluster sequentially using the process 800 shown in FIG. 10 .
  • the user defines Q T (the airflow through the perforated tiles), the number of 2-rack slices in the cluster, and the power draw of each of the racks.
  • Q T may be estimated as the mean perforated tile airflow rate for the entire facility or determined separately using, for example, a CFD or CGCFD analysis or other analysis or physical measurement.
  • all airflow variables (except Q T and the rack inlet airflows) are initialized to zero.
  • equations (1) through (5) are solved simultaneously for each slice.
  • stage 806 a determination is made as to whether the equations have been solved for all slices, and if not, stage 806 is repeated. Once the equations have been solved for all slices, then at stage 810 , the x-direction airflow variables are updated based on the computed pressures in the control volumes, PA i and PB i as discussed below. At stage 812 , a determination is made as to whether the computed pressures have changed by more than a predetermined threshold since the previous iteration and if so, stages 806 to 812 are repeated. Once there is no significant change in the computed variables, the process 800 stops at stage 814 , at which point the pressures and airflows for all of the control spaces have been determined.
  • new x-direction airflow values (QA xi and QB xi ) are determined based on the assumption that the pressure drop between adjacent cells is proportional to the square of the airflow rate using the equations in Table 3.
  • adjustments to the x-direction airflow values at stage 810 are achieved gradually by introducing damping into the iterative process and updating the values of QA xi and QB xi using the following equations (6) and (7).
  • QAx i ⁇ QAx i new +(1 ⁇ ) QAx i old (6)
  • is a linear relaxation factor. If ⁇ is set to zero, then no changes will occur from iteration to iteration. If ⁇ is set to 1, then there will be no damping introduced. For smaller values of ⁇ , more iterations will be required, however, the chances of obtaining a stable solution increase.
  • the particular optimum choice of ⁇ is problem specific, however, it has been found that values of ⁇ around 0.05 work well in the process described above.
  • the CVA technique described above can be conducted separately, one for each cluster of racks in a facility to obtain a complete cooling analysis of the facility.
  • the control volume analysis may be done for all clusters, or only for those in the vicinity of any changes to the facility.
  • empirical rules are used either alone or in combination with one of the methods described above to determine proper placement of equipment and the adequacy of cooling air.
  • the empirical rules that are used may take a number of different forms, and programs incorporating the empirical rules may be updated as more data is generated to support the empirical rules.
  • empirical rules are based, at least in part, on the ability of equipment racks to borrow unused capacity from surrounding neighbors.
  • the amount that may be borrowed may be limited to an allowable fraction (or weight) of the unused capacity and the particular allowable fraction may differ depending on a number of variables such as borrower-donor separation distance, tile flow rate, and the total power draw of both the borrower and the donor.
  • the cooling air available to a given rack is computed based on a weighted summation of the available airflows from airflow sources (i.e., supply devices, including in-row cooling units, or vents), net of airflows computed to be used by other racks, where the weights associated with the available airflows for a given rack decrease with distance between the rack and the air supply devices or vents.
  • airflow sources i.e., supply devices, including in-row cooling units, or vents
  • the cooling air available to each rack may initially be set equal to the cooling air supplied by the perforated tile in front of the rack, or to reflect possible losses, and provide safety margin, the cooling air available may be set equal to some amount (i.e. 90%) of the total air from the perforated tile.
  • the cooling load for each rack is then subtracted from the available air to provide a net available cooling air figure for each perforated tile and to provide an initial indication of a lack of cooling air for any equipment rack.
  • the available cooling air is then increased by assigning to each rack, a percentage of the net available cooling air from nearby perforated tiles.
  • the cooling air available may include 10% of the net available cooling air from a perforated tile associated with either an adjacent rack or a rack across an aisle, and 5% of the net available cooling air from a perforated tile of a diagonal rack or a rack two positions over in a row.
  • the particular percentages or weights used may be changed based on actual results or as a result of analyses conducted.
  • the loads of each rack may then be compared with the total available cooling air to determine remaining cooling capacity and to identify any potential problem racks.
  • empirical rules may be used in combination with superposition to analyze data centers and provide recommended equipment layouts. Using superposition, complex problems may be broken down into simpler problems that can then be solved using empirical rules.
  • empirical rules are established by initially performing a number of CFD analyses on typical rack layouts, and the results of these analyses are used to produce simple equations or look-up tables that can be used in real-time to design layouts of equipment.
  • the side-to-side airflows such as those shown in FIG. 9 may be determined for each rack one at a time with one rack turned “on” and all other racks turned “off”.
  • the airflows at the ends of a cluster for a number of different configurations may also be determined using CFD.
  • the airflows may be determined for a number of different air intake values for each rack and a number of different values of air flow from the perforated tiles.
  • the total air flows for different configurations can then be determined in real-time using superposition and the stored results.
  • the airflows through the top (in or out) of the volume in front of each rack may then be determined based on conservation of mass. In one embodiment, when the airflow into the top of one of the volumes exceeds some percentage (i.e., 20%) of the total air flow into the rack associated with the volume, then an overheating problem may exist requiring a design around.
  • mass species concentration analyses may be used in combination with empirical rules to determine what percentage of the total air entering a control volume is recirculated air to determine when an overheating problem may exist.
  • symmetry of the clusters can be used to reduce the number of CFD analyses that need to be performed, and the control volumes discussed above with respect to FIG. 9 may be used to establish a reference grid for the analysis.
  • CFD analyses need only be performed for Rack A i ⁇ 1 and Rack A i , and the results for each other rack may be determined based on the resulting airflows and the relative position of the racks.
  • the airflows in the cluster associated with Rack B i+1 are the same as those associated with Rack A i ⁇ 1 with the direction of some of the airflows changed for corresponding Rack A and Rack B inlet airflow and tile airflow rates.
  • RI recirculation index
  • one cooling objective in a data center is to manage the equipment rack inlet air temperatures.
  • the rack inlet air temperatures are dominated by the airflow patterns within the cold aisle and the temperatures within and around the cold aisle. Air drawn in from outside the cold aisle is generally heated to some degree by the rack exhaust and will be hereafter referred to as “recirculated air”.
  • a cluster of racks which receives its required cooling air exclusively from the perforated tiles within the cluster, represents an autonomous scalable unit from which a larger facility with predictable cooling performance may be constructed. A reasonable requirement is therefore to ensure that racks are adequately cooled by air originating from the racks own cold aisle. Conversely, it is acceptable for the rack to ingest no more than a small fraction of recirculated air.
  • the recirculation index (ri) is defined as the fraction of recirculated air ingested by the rack.
  • An ri of 0% implies that all of the rack inlet air was drawn directly from the perforated tiles while an ri of 100% implies that all of the rack inlet air was drawn from outside of the cold aisle. Note that a low ri is sufficient to guarantee cool inlet temperatures; however, a high ri does not guarantee excessively high inlet temperatures.
  • control volumes, symmetry and superposition are used in the present example to determine air flows and ultimately RI for a cold aisle.
  • superposition a sum of velocity potentials (or actual velocity components or total airflows over a consistent area) of simpler, elemental flow solutions is used to obtain a new, composite flow solutions. For example, assume we know the airflow pattern associated with only Rack A 1 “on” subject to a particular tile airflow rate and we also know the airflow pattern with Rack B 3 “on” subject to the same perforated tile airflow rate. The relevant horizontal airflow components can be added to obtain a solution, which approximates the airflow pattern resulting from Racks A 1 and B 3 both “on” simultaneously.
  • the airflow pattern resulting from the superposition of the two separate airflow patterns is not exactly the same as the full solution—even for an ideal flow.
  • two solutions are added together which individually (and when added together) satisfy conservation of mass criteria.
  • the use of superposition does not guarantee that the combined solution will be the unique solution and the difference is in the boundary conditions.
  • the top airflow condition floats freely as a constant-pressure boundary condition in all cases.
  • the airflow pattern constructed from superposition may not provide a perfect match to air velocity over the top surface of the cold aisle.
  • an assumption is made that the face (inlet) of the rack is a symmetry boundary condition (which is consistent with an inviscid analysis). This result creates the opportunity for some flow parallel to the face of the rack, which would probably not exist when the rack is actually drawing air.
  • superposition is used to establish only the 3 n ⁇ 2 internal horizontal airflows (n being equal to the length of the row in terms of racks) while end-of-row horizontal airflows are computed based on separate empirical correlations.
  • Vertical airflow components are computed from a mass balance performed on each control volume.
  • the horizontal airflows clearly depend on tile airflow. For example, a rack of a given airflow rate may be able to draw cooling air from a distance of many tile-widths when the perforated tile airflow rate is very low. However, this range of influence is much less as the tile flow rate is substantially increased.
  • the tile airflow rate is included in the analysis; the floor tiles should be “turned on” in the CFD analysis used to correlate airflow patterns.
  • the floor tiles are left “on” and the effect of each rack is considered individually, when the flows for each rack are summed, the sum would have more airflow leaving the top of the cold aisle than in reality.
  • the answer is to correlate only the horizontal airflows and then simply compute the correct airflow into or out of the top of each control volume based on conservation of mass.
  • FIG. 11 illustrates which airflow boundary conditions are turned on and off in each of the elemental airflow solution to yield the total solution with all airflow boundary conditions “on”.
  • Each of arrows 1004 a to 1004 f represents one of the airflows. It is the airflow components internal to the cold aisle that are actually being combined. There are, in general, a total of 2 n+4 elemental solutions for any layout, which makes up a complete airflow solution. Obviously, fewer elemental solutions are required if some racks have zero airflow and the ends of the cold aisle are sealed (e.g. with doors).
  • the elemental airflows used with superposition may be determined in any manner including physical testing.
  • CFD modeling for the cold aisle is performed using the following boundary conditions:
  • each elemental solution covers a range of dimensionless flow rates so that any arbitrary, but practical, rack or end airflow rate can be considered. So, the task is reduced to determining and storing the 3 n ⁇ 2 internal horizontal control volume airflows over an appropriate range of dimensionless airflow rates.
  • the flow in the cold aisle is considered to be ideal with no viscosity or turbulence. To verify this assumption, sample CFD cases were run with turbulence and viscosity included, and little difference was detected between models that included viscosity and turbulence and those that did not.
  • a method includes predicting end airflow as a function of rack power and airflow distribution for any row length and perforated tile flow rate while assuming a fixed room environment.
  • the example environment is large and free of other racks or objects.
  • Air is supplied at 60° F. and is exhausted uniformly over a 14 ft-high ceiling.
  • air velocity at points near the racks to scale with the dimensionless rack inlet velocities.
  • these “response” velocities vary nearly linearly with dimensionless rack flow rate (or velocity). It is, therefore, reasonable to estimate the dimensionless end airflows based on the following expression:
  • QAx 0 * a 0 +a A1 QRA 1 *+a A2 QRA 2 *+ . . . +a An QRA n *+a B1 QRB 1 *+a B2 QRB 2 + . . . +a Bn QRB n * (9)
  • QAx 0 * is one of four dimensionless end airflows for a cluster and the coefficients a Ai and a Bi effectively weight the relative importance of each rack on the end airflow.
  • the weighting coefficients associated with racks located near the end of the row will be much larger than those associated with more interior racks. Further, empirically it is found that only racks in the first four or five positions nearest the end of the row need be retained in Equation 9. For the fixed conditions considered, the constant a 0 is negative, implying that the flow is “out” (driven by buoyancy) when there is zero rack airflow.
  • CFD simulations may be performed at a number of different perforated tile flow rates.
  • a large pool of rack power values may be created from which the many CFD simulations draw rack power and airflow data from, either randomly or systematically.
  • the rack power values may be based on the frequency distribution of actual data center racks as determined from a survey.
  • the rack power and airflow values used in the CFD simulations may be scaled as necessary to achieve practical total supply-to-equipment airflow ratios in the range of, for example, 0.9 to 3 for each perforated tile flow rate considered.
  • the CFD data is then used to determine a least-squares fit of the coefficients in Equation 9 for each tile flow rate considered.
  • a simple end airflow model has been described which accurately accounts for a non-uniform distribution of rack airflow and power for a fixed set of room conditions.
  • the model is generalized to include the effects of geometric environment, the thermal environment, and supply airflow rate. The effects of the end airflow penetrate only a few rack distances down the row; for longer row lengths predictions for the majority of the racks in the cluster will be good even if the end airflow model is not as accurate as desired.
  • QA top i Q T ⁇ QRA i +QAx i ⁇ 1 ⁇ Qz i ⁇ QAx i (10a)
  • equations 9a and 9b represent a total of 2 n equations. At this stage, there is only one unknown per equation (QAtop i and QBtop i ) so they may be solved sequentially.
  • RI is the fraction of recirculated air ingested by a rack.
  • the recirculated air can enter the cold aisle at any point where there is inflow at the ends of the rows or along the top of the cold aisle.
  • the warm recirculated air need not directly enter the cold aisle via the control volume immediately adjacent to a rack of interest; it may enter anywhere, travel anywhere the airflow patterns take it, and end up at the inlet of any rack.
  • Equation 11 To compute RI for each rack the cool supply air is distinguished from the warm recirculated air at all points in the cold aisle. Mathematically, this is accomplished by defining the concentration of recirculated air at any point in the cold aisle using Equation 11.
  • the recirculated air can be assumed to have the same physical properties as the cool supply air so that it has no effect, e.g. due to a density difference, on the airflow patterns in the cold aisle.
  • Equation 11 applied to this volume represents the average C recirc over this volume. Dividing the numerator and denominator by a small time increment ⁇ t and taking the limit as ⁇ t ⁇ 0, demonstrates that the average C recirc over a rack inlet is precisely the rack recirculation index. Thus, to determine the RR's for each rack the average C recirc over each rack inlet is determined. Referring back to FIG. 8 , we can estimate the RR for each rack as the average C recirc of the control volume immediately adjacent to the rack of interest. C recirc over all 2 n control volumes can be computed from the conservation of mass of the recirculated air using Equation 12.
  • Q is the total airflow rate through each control volume face and is a known value at this stage of the calculation.
  • FIG. 13 shows control volumes 1008 and 1010 of a transverse section of a cold aisle 1006 . Equation 12 is applied to the control volumes 1008 and 1010 .
  • Equation 12 is applied to the control volumes 1008 and 1010 .
  • Equations 13a and 13b are not solved directly because the number of C recirc values exceeds the number of equations. Estimating each C recirc as the average C recirc from the “upwind” control volume, results in a proper balance of 2 n unknown C recirc 's and 2 n equations. Based on this “upwind” approach, the appropriate C recirc values are inserted into Equations 13a and 13b after the airflow patterns in the cold aisle have been computed thereby establishing the direction of airflow crossing each face of each control volume.
  • the 2 n Equations represented by 13a and 13b may be solved simultaneously for the 2 n CA i and CB i values. These simple linear equations can be solved without iteration virtually instantaneously for any practical row length using common computing hardware. Finally, as discussed above, the computed CA i and CB i values may be directly interpreted as the recirculation index of the adjacent “A” and “B” racks respectively. In other embodiments, because of the similarity between the energy and concentration equations, bulk average temperature could be determined over each control volume instead of RI following a very similar procedure.
  • stage 1050 the row length, tile airflow, rack airflow and rack power are defined for a cold aisle to be analyzed.
  • stage 1054 empirical data used for computing airflows is imported from a CFD analysis as described above.
  • stage 1056 The end airflows are then determined at stage 1056 based on details of the cluster and details of the room environment. All horizontal airflows are then determined at stage 1058 .
  • stage 1060 horizontal airflows induced by the 4 end airflows are computed, and at stage 1062 , complete horizontal airflows are computed by adding the airflows from stages 1058 and 1060 .
  • Vertical airflows are computed at stage 1064 , and then at stage 1066 , the recirculation index may be determined for each rack by solving a set of conservation equations for the recirculated air as described above.
  • a threshold recirculation index is first established, below which a design is considered unsatisfactory. For each rack, after a satisfactory design is achieved, the power of the rack is increased until the recirculation index of the that rack (or any other rack) reaches the threshold level, and the power at which that occurs represents the maximum cooling capacity for the rack.
  • a similar method for determining cooling capacity can be used with other analyses described herein, including the analysis using capture index values described below.
  • control volume and superposition methods described above may be modified. These modifications may include the use of more complex statistical methods (e.g., the use of neural networks) to determine end airflow conditions from large pools of CFD data. Further, the number of control volumes may be substantially increased to improve accuracy and resolution of computed variables. In particular, the latter improvement would allow airflow variations at various rack elevations (e.g., due to a variety of equipment installed in a rack) to be considered.
  • the basic methodology could further be modified to include layouts beyond the scope discussed above including layouts involving an arbitrary number of perforated tiles of arbitrary flow rate, an arbitrary cold aisle width, arbitrary rack dimensions or other such variations from examples discussed above.
  • cooling analyses of a data center have focused primarily on determining airflows in the cool aisle for a cluster of racks located in a data center having a raised floor.
  • Embodiments described above are not limited for use in data centers having raised floors, and aspects of the embodiments are also applicable to data centers that do not include raised floor cooling systems.
  • At least one embodiment described above provides a decoupling of the cold aisle from the remainder of the data center to compute airflows in the cold aisle. The effect of the room environment is then built back into the analysis using end-of-row airflows that are computed, for example, in separate CFD calculations that may be computed offline and made available through look-up tables or empirical correlations.
  • a hot aisle in a data center can be analyzed by decoupling the aisle from the remainder of the room and later building the effects of the room environment back into the analysis.
  • a raised floor data center is not used, but rather, cooling is provided using in-row cooling units as described above.
  • a capture index (CI) is calculated and used to analyze a cluster of racks in a data center.
  • the capture index is used in one embodiment with a row or cluster of racks having one or more in-row cooling units, and the capture index is defined as the percentage of air released by a rack into a hot aisle, which is captured by cooling units bounding the hot aisle.
  • the CI may be considered as a complementary metric to the RI described above for use with the hot aisle.
  • CI techniques may also be used with cold aisle analysis as discussed further in U.S. Provisional Application 60/897,076, filed on Jan. 24, 2007 which is hereby incorporated herein for all purposes.
  • the CI is useful when the focus of a design is to keep the hot air within the hot aisle.
  • rack inlet temperatures are typically the ultimate cooling metric, however, if all of the hot air is captured in the hot aisle, the rest of the data center (including rack inlets) can be designed and controlled to remain at “room temperature.”
  • the cold-aisle capture index is defined as the fraction of air ingested by the rack which originates from local cooling resources (e.g., perforated floor tiles or local coolers).
  • the hot-aisle capture index is defined as the fraction of air exhausted by a rack which is captured by local extracts (e.g., local coolers or return vents).
  • CI therefore varies between 0 and 100% with better cooling performance generally indicated by greater CI values.
  • high CI's ensure that the bulk of the air ingested by a rack comes from local cooling resources rather than being drawn from the room environment or from air which may have already been heated by electronics equipment.
  • rack inlet temperatures will closely track the perforated-tile airflow temperatures and, assuming these temperatures are within the desired range, acceptable cooling will be achieved.
  • high CI's ensure that rack exhaust is captured locally and there is little heating of the surrounding room environment.
  • CI has been defined above as a rack-level quantity relative to local cooling resources
  • the definition of CI could be extended to any grouping of inlets and outlets, for example, a single server with reference to a single perforated tile.
  • the metric would indicate the fraction of airflow ingested by the server which originated from the single perforated tile.
  • the use of chemical concentrations with, for example a CFD analysis can be used to quantitatively determine CI for both a hot aisle and a cold aisle.
  • the exhaust of each rack is identified in such an analysis as a separate species having the same properties as air, so as not to change the physics of airflow.
  • the fraction of hot air released from rack i (identified as C i ) which is captured by an in-row cooler identified as cooler j may be computed using Equation 14 below.
  • C ji is the concentration of C i at the inlet of cooler j
  • Q coolerj is the airflow rate (e.g. in cfm) through coolerj
  • the capture index (CI) is the sum of all of the f ij 's over all N coolers and can be expressed using Equation 15 below.
  • a set of empirical rules can be used to determine the CI for each rack 1082 of the cluster.
  • the resulting CI values may be displayed on a display with the associated racks.
  • racks having a CI less than 60% are identified in red indicating a warning
  • racks having a CI between 60% and 80% are indicated in yellow as a caution
  • racks having a CI greater than 80% are indicated in green indicating that the CI is satisfactory.
  • the cluster 1080 includes two rows (row A and row B) of parallel racks that exhaust air to a common hot aisle 1084 .
  • Each Rack is labeled A 1 -A 6 and B 1 -B 7 , identifying the row and position in the row of the rack, and for the example shown each rack has a power draw of 2 kW.
  • the cluster also includes in-row cooling units 1086 .
  • FIG. 15 a number of half-rack cooling units 1086 are shown, but embodiments of the invention may also be used with full-width rack cooling units or other devices.
  • the half-rack cooling units used in the example associated with FIG. 15 have a nominal cooling capacity of 17 kW.
  • the CI in terms of percentage for each rack. The CI is shown as a percentage and indicates for each rack, the percentage of its exhaust air that is captured by one of the cooling units.
  • the CI is determined based on the concept that all rack-cooling unit interactions depend only on the magnitude of airflow associated with the racks and cooling units and their relative geometric positions.
  • Each rack location can be thought of as having a certain potential to supply airflow to other rack locations. This potential varies inversely with the distance of separation.
  • rack A 1 in FIG. 15 could potentially supply a large portion of its airflow to the area near rack A 2 . However, much less of rack A 1 's airflow could make its way to position A 6 .
  • the amount of airflow a rack can supply to other locations is in direct proportion to its own total airflow.
  • Equation 16 The net airflow which can be supplied to a particular rack location A 1 can be represented using Equation 16 below.
  • the net maximum airflow that can be supplied to various locations in the B row is computed using a similar expression. Finally, the same expression is used to compute the net maximum airflow, which can be captured at any rack location—with the sum over all coolers instead of racks.
  • the CI is then estimated as the ratio of net airflow captured and net airflow supplied at any location expressed as a percentage and with values capped at 100%.
  • the constants A, B, and C are selected to provide the best statistical fit to benchmark CFD data. Different values of the constants may be used to account for alternative configurations including different cooler types, different average rack power or peak-to-average power ratios and alternative hot-aisle spacing, row lengths, or room environments. As an example, consider a cluster of average power racks with modest rack-to-rack power variations.
  • the CI is calculated for a cluster of racks having uniform depth and width. In other embodiments, the same processes may be used for racks of non-uniform depth and width. In one embodiment, the CI calculations described above are programmed into a Microsoft Excel Spreadsheet program that allows a user to add and move cooling units to see the effect of different numbers of cooling units and their placements. In other embodiments, the process for determining CI described above may be incorporated into data center design and management systems, such as system 200 discussed above.
  • an exponential expression is used for modeling the rack and cooler interactions.
  • other expressions may be used, such as a polynomial or any other mathematical expression which contains a number of parameters which may be tuned to provide the best fit to benchmark performance data.
  • different curves and/or different coefficients may be used for the portion of the calculation associated with determining air supplied by racks than used in the portion of the calculation used for determining air captured by cooling units.
  • the rules may be further refined to address specific situations. For example, a Rack A may have no effect on another Rack B where a third rack C is located between Rack A and Rack B and has greater airflow than either Rack A or Rack B.
  • the effects of the ends of rows may be accounted for explicitly in hot aisle and cold aisle CI analysis.
  • Separate CFD simulations may be conducted to determine the net inflow or outflow of air at each end of a row for given layouts of racks and cooling units.
  • the results of the CFD simulations may be incorporated into the empirical methods described above to determine CI for racks in a cluster.
  • the results of the CFD simulations can be used to provide correct airflow estimates at the ends of the row, while one of the algorithms discussed above may be used to determine CI at more interior portions of the row.
  • the effects of a missing rack or racks may be simulated using CFD with the results incorporated into the empirical methods.
  • the capture index method of analysis discussed above provides a rack-by-rack, or local, cooling performance metric for equipment racks in a data center.
  • cluster-wide performance metrics are determined, and the overall cooling performance of the cluster may be determined based on both the CI metric and the global metric.
  • the cluster-wide performance metrics indicate whether the cluster as a whole will have adequate cooling performance.
  • the CI identifies which racks are not having their exhaust airflow adequately captured. If a rack having a low CI is a low power rack, this may not result in a problem.
  • a rack may have a relatively high CI, yet still cause problems if it is a high power rack with a high outlet temperature.
  • the global metric that is used is a determination of the net power which escapes the hot aisle. The net escaped power can be determined using equation 17.
  • the net escaped power determined by equation 17 may be correlated to maximum cluster inlet temperatures (e.g., a net escaped power of 25 kW may imply a maximum cluster rack inlet temperature of seventy-nine degrees F.) for clusters of particular geometric layouts (e.g., hot aisle width, row length, etc.), room environments and rack and cooler details (e.g., cooling unit flow rate and rack airflow/power (cfm/kW)). Accordingly, the net escaped power may be used to determine the highest rack inlet temperature.
  • maximum cluster inlet temperatures e.g., a net escaped power of 25 kW may imply a maximum cluster rack inlet temperature of seventy-nine degrees F.
  • the net escaped power may be used to determine the highest rack inlet temperature.
  • a process 1100 is provided for determining CI in a hot aisle using a CFD analysis.
  • the CFD analysis is performed only on the hot aisle itself and end of row airflows are determined separately and may be patched into the CFD solution.
  • Common computer hardware can be used to conduct such a CFD analysis in 10 to 20 seconds.
  • the process 1100 is shown in flowchart form in FIG. 16 .
  • data related to the layout is loaded either manually, read from a database, or in any other manner.
  • the data related to the layout may include row length, power draw and airflow for each rack, rack dimensions, cooling unit types, locations and flow rates, hot aisle widths and room environment details (i.e., size, temperatures).
  • hot-aisle end of row airflows are determined using best-fit curves, or look-up tables based on prior CFD studies. Hot aisle end of row calculations may be performed using the same techniques used to determine cold aisle end of row airflows with the input rack airflow provided as a positive value and the cooling units airflow input as a negative value.
  • a CFD analysis of just the hot aisle is performed using results of the end airflow analysis of stage 1104 and with the top of the hot aisle taken as a constant-pressure boundary. In the analysis, the exhaust air for each rack is “tagged” with a particular concentration.
  • the CI for each rack is determined based on the ratio of captured and supplied airflows computed by equation 16 and the results of stage 1106 .
  • the global cluster cooling metric may also be determined at this point using equation 17.
  • the rack CI's and the global cluster cooling metric can be used to determine if cooling for the cluster is adequate.
  • additional cooling analysis methods may be used in embodiments of the invention. These additional methods include a panel method, a potential analysis method, flow network/zonal modeling, principal component analysis or a combination of any of these and the methods discussed above. These additional cooling analysis methods are generally known to those of skill in the art.
  • Various embodiments according to the invention may be implemented on one or more computer systems as discussed above. These computer systems may include cellular phones, personal digital assistants and/or other types of mobile computing devices. Moreover, a computer system may include any sort of computing device such as a personal computer, a workstation, a mainframe, a networked client, a server, media servers and application servers. For example, system 200 may be implemented in a single computer system or in multiple computer systems. These computer systems may be, for example, general-purpose computers such as those based on Intel PENTIUM-type processor, Motorola PowerPC, Sun UltraSPARC, Hewlett-Packard PA-RISC processors, or any other type of processor.
  • the computer system 900 may include a processor 903 connected to one or more memory devices 904 , such as a disk drive, memory, or other device for storing data. Memory 904 is typically used for storing programs and data during operation of the computer system 900 .
  • the computer system 900 may also include a storage system 906 that provides additional storage capacity.
  • Components of computer system 900 may be coupled by an interconnection mechanism 905 , which may include one or more busses (e.g., between components that are integrated within a same machine) and/or a network (e.g., between components that reside on separate discrete machines).
  • the interconnection mechanism 905 enables communications (e.g., data, instructions) to be exchanged between system components of system 900 .
  • Computer system 900 also includes one or more input devices 902 , for example, a keyboard, mouse, trackball, microphone, touch screen, and one or more output devices 907 , for example, a printing device, display screen, speaker.
  • input devices 902 for example, a keyboard, mouse, trackball, microphone, touch screen
  • output devices 907 for example, a printing device, display screen, speaker.
  • computer system 900 may contain one or more interfaces (not shown) that connect computer system 900 to a communication network (in addition or as an alternative to the interconnection mechanism 905 ).
  • the storage system 906 typically includes a computer readable and writeable nonvolatile recording medium 911 in which signals are stored that define a program to be executed by the processor or information stored on or in the medium 911 to be processed by the program to perform one or more functions associated with embodiments described herein.
  • the medium may, for example, be a disk or flash memory.
  • the processor causes data to be read from the nonvolatile recording medium 911 into another memory 912 that allows for faster access to the information by the processor than does the medium 911 .
  • This memory 912 is typically a volatile, random access memory such as a dynamic random access memory (DRAM) or static memory (SRAM).
  • DRAM dynamic random access memory
  • SRAM static memory
  • the processor 903 generally manipulates the data within the integrated circuit memory 904 , 912 and then copies the data to the medium 911 after processing is completed.
  • a variety of mechanisms are known for managing data movement between the medium 911 and the integrated circuit memory element 904 , 912 , and the invention is not limited thereto. The invention is not limited to a particular memory system 904 or storage system 906 .
  • the computer system may include specially-programmed, special-purpose hardware, for example, an application-specific integrated circuit (ASIC).
  • ASIC application-specific integrated circuit
  • computer system 900 is shown by way of example as one type of computer system upon which various aspects of the invention may be practiced, it should be appreciated that aspects of the invention are not limited to being implemented on the computer system as shown in FIG. 17 . Various aspects of the invention may be practiced on one or more computers having a different architecture or components shown in FIG. 17 . Further, where functions or processes of embodiments of the invention are described herein (or in the claims) as being performed on a processor or controller, such description is intended to include systems that use more than one processor or controller to perform the functions.
  • Computer system 900 may be a general-purpose computer system that is programmable using a high-level computer programming language. Computer system 900 may be also implemented using specially programmed, special purpose hardware.
  • processor 903 is typically a commercially available processor such as the well-known Pentium class processor available from the Intel Corporation. Many other processors are available.
  • processor usually executes an operating system which may be, for example, the Windows 95, Windows 98, Windows NT, Windows 2000 (Windows ME) or Windows XP operating systems available from the Microsoft Corporation, MAC OS System X operating system available from Apple Computer, the Solaris operating system available from Sun Microsystems, or UNIX operating systems available from various sources. Many other operating systems may be used.
  • the processor and operating system together define a computer platform for which application programs in high-level programming languages are written. It should be understood that embodiments of the invention are not limited to a particular computer system platform, processor, operating system, or network. Also, it should be apparent to those skilled in the art that the present invention is not limited to a specific programming language or computer system. Further, it should be appreciated that other appropriate programming languages and other appropriate computer systems could also be used.
  • One or more portions of the computer system may be distributed across one or more computer systems coupled to a communications network.
  • a computer system that performs build-out functions may be located remotely from a system manager.
  • These remote computer systems also may include remote general-purpose computer systems and/or remote computing devices.
  • various aspects of the invention may be distributed among one or more computer systems configured to provide a service (e.g., servers) to one or more client computers, or to perform an overall task as part of a distributed system.
  • a service e.g., servers
  • various aspects of the invention may be performed on a client-server or multi-tier system that includes components distributed among one or more server systems that perform various functions according to various embodiments of the invention.
  • These components may be executable, intermediate (e.g., IL) or interpreted (e.g., Java) code which communicate over a communication network (e.g., the Internet) using a communication protocol (e.g., TCP/IP).
  • a communication network e.g., the Internet
  • a communication protocol e.g., TCP/IP
  • one or more database servers may be used to store device data that is used in designing layouts, and one or more servers may be used to efficiently perform cooling calculations associated with embodiments of the present invention.
  • Various embodiments of the present invention may be programmed using an object-oriented programming language, such as SmallTalk, Java, C++, Ada, or C# (C-Sharp). Other object-oriented programming languages may also be used. Alternatively, functional, scripting, and/or logical programming languages may be used.
  • object-oriented programming languages such as SmallTalk, Java, C++, Ada, or C# (C-Sharp).
  • Other object-oriented programming languages may also be used.
  • functional, scripting, and/or logical programming languages may be used.
  • Various aspects of the invention may be implemented in a non-programmed environment (e.g., documents created in HTML, XML or other format that, when viewed in a window of a browser program, render aspects of a graphical-user interface (GUI) or perform other functions).
  • GUI graphical-user interface
  • Various aspects of the invention may be implemented as programmed or non-programmed elements, or any combination thereof.
  • a variety of system elements including inputs, outputs and interfaces may exchange information with various external entities that may be information providers and/or information consumers. These external entities may include users and/or systems. Each of these system elements may both restrict exchanged information to a predefined set of values and validate any information exchanged prior to using the information or providing the information to other components.
  • the indication of remaining cooling capacity may be a direct indication of remaining cooling in terms of, for example, kilowatts or BTU per hour, or the indication may be indirect such as providing the total capacity of cooling available to an enclosure along with an indication of how much cooling is being used, for example, in terms of percentage.
  • calculated values, including the capture index and the recirculation index may be used to determine the sufficiency of a particular design and to determine additional cooling capacity before a warning or error condition will result.
  • Embodiments of a systems and methods described above are generally described for use in relatively large data centers having numerous equipment racks, however, embodiments of the invention may also be used with smaller data centers and with facilities other than data centers. Further, as discussed above, embodiments of the present invention may be used with facilities having raised floors as well as with facilities that do not have a raised floor.
  • results of analyses are described as being provided in real-time.
  • real-time is not meant to suggest that the results are available immediately, but rather, are available quickly giving a designer the ability to try a number of different designs over a short period of time, such as a matter of minutes.

Abstract

Systems and methods are provided for determining data center resource requirements, such as cooling and power requirements, and for monitoring performance of data center resource systems, such as cooling and power systems, in data centers. At least one aspect provides a system and method that enables a data center operator to determine available data center resources, such as power and cooling, at specific areas and enclosures in a data center to assist in locating new equipment in the data center.

Description

    RELATED APPLICATIONS
  • This application claims priority under 35 U.S.C. § 119(e) to U.S. Provisional Application Ser. No. 60/938,034, entitled “METHODS AND SYSTEMS FOR MANAGING FACILITY POWER AND COOLING,” filed on May 15, 2007, which is herein incorporated by reference in its entirety.
  • BACKGROUND OF INVENTION
  • 1. Field of Invention
  • Embodiments of the invention relate generally to entity management, and more specifically to methods and systems for managing facility power and cooling.
  • 2. Discussion of Related Art
  • Centralized data centers for computer, communications and other electronic equipment have been in use for a number of years, and more recently with the increasing use of the Internet, large scale data centers that provide hosting services for Internet Service Providers (ISPs), Application Service Providers (ASPs) and Internet content providers are becoming more prevalent. Typical centralized data centers contain numerous racks of equipment that require power, cooling and connections to external communications facilities. In modern data centers and network rooms, the increased density of computing equipment used in these facilities has put strains on the cooling and power systems of the facilities. In the past, typical power consumption for each equipment enclosure in a data facility was on the order of 1 kW. With the use of server blades and other high power density equipment in equipment racks, it is not uncommon for an equipment rack to have a power draw of 10 kW or even as high as 25 kW.
  • Typically, the power consumed by computer equipment is converted to heat and typically, the cooling requirements of a facility are determined based on the power requirements of the facility. Typical data centers utilize air plenums under raised floors to distribute cooling air through a data center. One or more computer room air conditioners (CRACs) or computer room air handlers (CRAHs) are typically distributed along the periphery of the data room, and these units draw return air from the room or a ceiling plenum and distribute cooling air beneath the raised floor. Perforated tiles may be placed in front or beneath racks of equipment to be cooled to allow the cooling air from beneath the floor to cool equipment within the racks.
  • Several tools are available to assist a data center designer in configuring a layout of a data center to provide necessary power and cooling to equipment to be located in the data center. These tools typically assist a designer in determining total power requirements and accordingly overall cooling requirements for a data center. In addition, these tools may assist a designer in determining optimum equipment layout and proper sizing of power cabling and circuit breakers.
  • While existing tools provide a designer with detailed layout information regarding power distribution, these tools typically provide far less help in determining cooling needs for a facility. Advanced programs that use computational fluid dynamics (CFD) may be used to model the cooling design of a facility, but the use of such programs is extremely limited due to the complexity of the programs, which results in their use being prohibitively expensive and time consuming. U.S. Patent Application US2003/0158718 A1 to Nakagawa et al describes an automated system for designing a cooling system for a facility. In the system of Nakagawa, the facility is divided into a number of pre-characterized cells (such as a cluster of racks) over which the response of various parameters, such as maximum temperature, are known based on key parameters. The system uses built-in cell to cell interaction rules to predict overall thermal performance and to optimize equipment layout. While this system may offer some improvements in speed over a full CFD analysis, it is limited to the use of pre-characterized cells, and does not provide analysis below the cell level. Also, the cells must be characterized using, for example, a CFD analysis or physical testing.
  • Programs and systems such as those described above provide idealized results for the cooling performance of a facility and often fail to account for situations which often occur in real life installations, which can dramatically affect the cooling performance of a data center. For example, in a facility using a raised floor, the absence of one or more floor panels, or the misplacement of one or more perforated floor panels can greatly affect the cooling performance of the data center and cause the actual performance to vary greatly from a calculated idealized performance. Further, the degradation in performance of one or more air conditioning units can drastically change airflow and cooling characteristics of a facility.
  • The inability to properly analyze the cooling performance of a facility typically causes a data center designer to over design the facility from a cooling perspective, which results in the facility to be more expensive and less efficient.
  • In existing data centers, it is often desirable to replace equipment with upgraded equipment and/or add new equipment to existing enclosures in the facility. Several tools exist which enable a manager of a data center to monitor power usage in a facility. These tools include the InfrastruXure® Manager and/or InfrastruXure® Central product available from American Power Conversion Corporation of West Kingston, R.I.
  • With the increasing cooling and power requirements of computer equipment, it is desirable for a data center manager to determine if there is adequate power and cooling available in the facility before new or replacement equipment may be added. Typically, a data center manager may know, or can determine, if the total cooling capacity of the data center is sufficient for the total power draw. However, hot spots in a facility may develop, particularly where high power density equipment is used, and it may not be enough to merely analyze cooling capacity at the facility level. To attempt to identify hot spots, a data center manager may resort to manual temperature measurements throughout a facility and try to implement fixes to correct the hot spots. Such fixes may involve a rearrangement or replacement of perforated floor panels, a rearrangement of enclosures, and/or adding additional cooling capacity. In any case, these fixes are typically done on a trial and error basis, and while some hot spots may be eliminated, the fixes may cause other hot spots to arise due to a redirection of the cooling air in the facility. This trial and error approach can lead to unexpected failures of equipment, which is unacceptable in critical data centers. To avoid such failures, data center managers typically over design facilities and fail to use facilities to their full capacity.
  • SUMMARY OF INVENTION
  • Aspects of the present invention relate generally to management of data center entities and their associated resources. Embodiments of the invention provide systems and methods for determining data center cooling and power requirements and for monitoring performance of cooling and power systems in data centers. At least one embodiment provides a system and method that enables a data center operator to determine available power and cooling at specific areas and enclosures in a data center to assist in locating new equipment in the data center.
  • One aspect is directed to a method that includes determining cooling capacity of a number of equipment enclosures in a data center, determining cooling requirements of each of the number of equipment enclosures, and providing an indication of remaining cooling capacity for each of the number of equipment enclosures.
  • The method may further include developing a floor plan model of the data center, wherein the floor plan model includes a floor plan that indicates location of each of the number of equipment enclosures in the data center, and for each of the number of equipment enclosures, displaying on the floor plan, the indication of remaining cooling capacity. The indication of remaining cooling capacity may include an indication of additional power that can be drawn by each of the number of equipment enclosures based on the remaining cooling capacity. Determining cooling capacity may include calculating a predicted cooling capacity based on the floor plan model. Determining cooling capacity may include measuring airflow at a first plurality of locations in the facility to obtain a measured cooling capacity. Determining cooling capacity may include measuring air temperature at a second plurality of locations in the facility. At least one of the first plurality of locations and the second plurality of locations includes at least one air vent of a raised floor. The method may further include comparing predicted cooling capacity with measured cooling capacity to obtain a comparison result and providing an indication when the comparison result is greater than a threshold. The method may further include adjusting the predicted cooling capacity based on measured airflow. The method may further include determining placement of new equipment in an equipment enclosure in the data center by comparing power ratings of the new equipment with cooling capacity of the equipment enclosure. The method may further include, for each of the number of equipment enclosures, determining electrical power capacity and remaining electrical power availability, and displaying on the floor plan remaining electrical power availability. In the method, determining remaining electrical power availability may include measuring at least one parameter of electrical power provided to at least one of the number of equipment enclosures. In the method, determining cooling capacity of an equipment enclosure may include estimating available cooling air at the equipment enclosure using a weighted summation of available airflows from a plurality of airflow sources, the weights used in the weighted summation may decrease with distance from the equipment enclosure to each of the airflow sources, and the weights may be based on mechanical characteristics of the plurality of airflow sources. The method may further include determining available airflow of at least one of the plurality of airflow devices using at least one of specifications of the at least one of the plurality of airflow devices and measured data for the at least one of the plurality of airflow devices in the data center. The method may further include determining available airflow of at least a second one of the plurality of airflow devices based on the measured data for the at least one of the plurality of airflow devices. In the method, determining cooling capacity may include using superposition to combine airflows. In the method, determining airflow into and out of each of a plurality of sides of each control volume may include computing airflows using equations based on at least one of conservation of mass and conservation of momentum. Further, determining airflow into and out of each of a plurality of sides of each control volume may include determining airflows using empirical rules derived from CFD, physical measurement, or any other means.
  • Another aspect of the invention is directed to a system for managing a data center. The system includes at least one input to receive data related to equipment and equipment enclosures and to receive data related to cooling characteristics of the data center, a controller operatively coupled to the input and configured to determine cooling capacity of each equipment enclosure, and at least one output operatively coupled to the controller that provides an indication of remaining cooling capacity for each of the equipment enclosures.
  • The system may further include an output device coupled to the at least one output, wherein the system is configured to display a floor plan of at least a portion of the data center indicating location of at least one of the equipment enclosures in the data center and indicating the remaining cooling capacity for the at least one of the equipment enclosures. The output device may be configured to include an indication of additional power that can be drawn by the at least one of the equipment enclosures. The system may further include at least one airflow monitor operatively coupled to the controller to provide data related to at least one airflow in the data center. The system may further include at least one air monitor operatively coupled to the controller to provide data related to air temperature at a location in the data center. The controller may be configured to compare a predicted cooling capacity with a measured cooling capacity to obtain a comparison result and to provide an indication when the comparison result is greater than a threshold. The system may further include at least one power monitor operatively coupled to the controller, and at least one airflow controller operatively coupled to the controller and responsive to signals from the controller to alter cooling airflow in the data center. The system may further include at least one power controller operatively coupled to the controller and responsive to signals from the controller to alter at least one characteristic of power in the data center.
  • Another aspect of the invention is directed to a system for managing a data center. The system includes at least one input to receive data related to equipment and equipment enclosures and to receive data related to cooling characteristics of the data center, and means, coupled to the at least one input, for determining remaining cooling capacity for each of the number of equipment enclosures and providing an indication of remaining cooling capacity.
  • The system may further include means for providing an indication of additional power that can be drawn by each of the equipment enclosures and means for updating the indication of remaining cooling capacity based on measured airflows in the data center. The system may still further include means for determining placement of equipment in the data center based on remaining cooling capacity, and means for estimating available cooling air of at least one of the equipment enclosures using a weighted summation of available airflows from a plurality of airflow sources.
  • Another aspect of the invention is directed to a computer-implemented method of managing power and cooling capacity of a data center. The method includes receiving data center parameters, determining an equipment layout in the data center, the equipment layout identifying a location of each of a plurality of devices in the data center, based on the location, determining available power and available cooling in the data center for each of the plurality of devices, comparing the available power and available cooling with power requirements and cooling requirements of each of the plurality of devices to obtain a cooling comparison result and a power comparison result for each of the plurality of devices.
  • The method may further include comparing each of the comparison cooling results and the power cooling results with thresholds and providing at least one recommendation for correcting an out of tolerance condition. The at least one recommendation may include adding an in-row cooling unit in a row of equipment of the data center along with a proposed location for the in-row cooling unit. The method may further include determining at least one of stranded cooling capacity and stranded power capacity in the data center and providing recommendations for reducing at least one of the stranded power capacity and the stranded cooling capacity in the data center. The method may further include displaying a floor plan model of the data center, wherein the floor plan model includes a floor plan that indicates a location of each of the plurality of devices in the data center, and for each of the plurality of devices, displaying on the floor plan an indication of remaining cooling capacity. The method may include displaying on the floor plan an indication of remaining power capacity for each of the plurality of devices. The method may still further include setting a redundancy level for at least some of the plurality of devices, and wherein the available power and available cooling are determined to meet the redundancy level. In the method, the act of determining the equipment layout may include arranging the plurality of devices in at least two substantially parallel rows with a hot aisle formed between the two rows, and the method may further include conducting a cooling analysis by analyzing airflows in the hot aisle. The method may include selecting at least one in-row cooling unit to be placed in the layout in one of the at least two substantially parallel rows. At least one of the plurality of devices is an equipment rack, and the method may include determining a capture index for the at least one in-row cooling unit and the equipment rack. The method may further include on a display screen, simultaneously displaying a first floor plan model of the data center and a second floor plan model of the data center, wherein the first floor plan model includes at least a partial view of the data center and the second floor plan model includes a full view of the data center. The second floor plan model may include an indication of a portion of the data center that is shown in the first floor plan model. The method may further include displaying a three dimensional view of at least a portion of the data center. The method may include selecting a camera location for use in the data center and displaying a three dimensional view of a viewing area of the camera. The method may further include selecting a subset of the plurality of devices and defining a power zone for each device of the subset of the plurality of devices, wherein each power zone includes at least one UPS. In the method, each of the plurality of devices may be an equipment rack, and the method may further include selecting components for each of the plurality of devices from a displayed list of components. The method may further include determining operational power costs and operational cooling costs for a subset of the plurality of devices, and the operational power costs and operational cooling costs may be determined in terms of kilowatts. The method may further include transferring an electronic file containing data for the equipment layout from a design system to a management system. The method may also include displaying measured data for the data center on a display of a layout of the data center.
  • Another aspect of the invention is directed to a computer-implemented method for designing a layout of a data center. The method includes receiving information from a user regarding parameters of the data center, determining an equipment layout for the data center, the equipment layout identifying a location of each of a plurality of devices in the data center, and on a display screen, simultaneously displaying a first floor plan model of the data center and a second floor plan model of the data center, wherein the first floor plan model includes at least a partial view of the data center and the second floor plan model includes a full view of the data center.
  • In the method, the second floor plan model may include an indication of a portion of the data center that is shown in the first floor plan model. The method may further include determining available power and available cooling in the data center for each of the plurality of devices, and comparing the available power and available cooling with power requirements and cooling requirements of each of the plurality of devices to obtain a cooling comparison result and a power comparison result for each of the plurality of devices. The method may include, for each of the plurality of devices, displaying on the first floor plan model an indication of remaining cooling capacity, and displaying on the first floor plan model an indication of remaining power capacity for each of the plurality of devices.
  • Another aspect of the invention is directed to a system for use with a data center. The system includes an input to receive instructions from a user, an output to provide output data to a display device, and a controller coupled to the input and to the output and configured to determine an equipment layout of a data center, the equipment layout identifying a location of each of a plurality of devices in the data center. The controller is further configured to, based on the location, determine available power and available cooling in the data center for each of the plurality of devices, and compare the available power and available cooling with power requirements and cooling requirements of each of the plurality of devices to obtain a cooling comparison result and a power comparison result for each of the plurality of devices.
  • In the system, the controller may be configured to compare each of the comparison cooling results and the power cooling results with thresholds, and based on at least one of the cooling comparison result and the power comparison result, provide at least one recommendation for correcting an out of tolerance condition. The at least one recommendation may include adding an in-row cooling unit in a row of equipment of the data center, and the at least one recommendation may include a proposed location for the in-row cooling unit. The controller may be configured to determine at least one of stranded cooling capacity and stranded power capacity in the data center, and to provide recommendations for reducing at least one of the stranded power capacity and the stranded cooling capacity in the data center. The controller may be further configured to provide data to the output for displaying a floor plan model of the data center, wherein the floor plan model includes a floor plan that indicates a location of each of the plurality of devices in the data center, and provide data to the output for displaying on the floor plan an indication of remaining cooling capacity. The controller may be further configured to provide data to the output for displaying on the floor plan an indication of remaining power capacity for each of the plurality of devices, and the controller may be configured to determine the available power and available cooling based on a user-selected redundancy level. The controller may be configured to arrange the plurality of devices in at least two substantially parallel rows with a hot aisle formed between the two rows, and to conduct a cooling analysis by analyzing airflows in the hot aisle. The controller may be configured to select at least one in-row cooling unit to be placed in the layout in one of the at least two substantially parallel rows. At least one of the plurality of devices may be an equipment rack, and the controller may be configured to determine a capture index for the at least one in-row cooling unit and the equipment rack. The controller may be configured to provide data to the output to simultaneously display a first floor plan model of the data center and a second floor plan model of the data center, wherein the first floor plan model includes at least a partial view of the data center and the second floor plan model includes a full view of the data center. The second floor plan model may also include an indication of a portion of the data center that is shown in the first floor plan model. The controller may be configured to provide data to the output to display a three dimensional view of at least a portion of the data center, and to provide data to the output to display a three dimensional view of a viewing area of a camera to be located in the data center. The controller may be further configured to select a subset of the plurality of devices and define a power zone for each device of the subset of the plurality of devices, wherein each power zone includes at least one UPS. The system may further include a database module containing data for components to be loaded into the plurality of devices, and the controller may be configured to provide data to the output for displaying a list of the components. The controller may be configured to determine operational power costs and operational cooling costs for a subset of the plurality of devices, and the operational power costs and operational cooling costs may be determined in terms of kilowatts. The controller may also be configured to provide an output electronic file containing data for the equipment layout. The controller may also be configured to, based on at least one of the cooling comparison result and the power comparison result, provide at least one recommendation for placement of rack mount equipment.
  • Another aspect of the invention is directed to a computer-implemented method for designing a layout of a data center. The method includes receiving information from a user regarding parameters of the data center, determining an equipment layout for the data center, the equipment layout identifying a location of each of a plurality of devices in the data center, including identifying a row location for each of a plurality of equipment racks with a first subset of the plurality of equipment racks being included in a first row; and on a display screen, displaying a rack view of the data center showing a front view of each of the plurality of equipment racks of the first subset in the first row.
  • In the method, displaying a rack view may include displaying a front view of a second subset of the plurality of equipment racks of a second row along with the front view of the first subset of the plurality of equipment racks. In the method, displaying a rack view may include displaying a front view of a second subset of the plurality of equipment racks of a second row along with the front view of the first subset of the plurality of equipment racks. The first row may include additional equipment, with the additional equipment not included in the rack view. The method may further include simultaneously displaying on the display screen the rack view and a full room view of the equipment layout for the data center. The method may also include, in response to selection of a selected equipment rack in the full room view by a user, displaying the selected equipment rack in the rack view, and displaying the selected equipment rack in the rack view may include displaying a front view of the selected equipment rack.
  • Another aspect of the invention is directed to a computer-implemented method for evaluating the cooling performance of a cluster of equipment racks in a data center, wherein the cluster of equipment racks includes at least a first row of racks and a second row of racks separated by a cool aisle, with each of the equipment racks being configured to draw cooling air from the cool aisle. The method includes obtaining at least one of power data and airflow data for each of the equipment racks, obtaining cool airflow data for cool air supplied to the cool aisle from a source of cool air, and conducting an analysis of airflows in the cool aisle to determine a recirculation index for at least one of the equipment racks, wherein the recirculation index is indicative of a quantity of recirculated air included in an input airflow of the at least one equipment rack.
  • In the method, the recirculation index may be equal to a ratio of recirculated air to total air in the input airflow of the at least one equipment rack, and the method may further include determining a recirculation index for each of the equipment racks. In the method, the act of obtaining cool airflow data may include obtaining cool airflow data for an in-row cooling unit included in the cluster of racks. In the method, the act of obtaining cool airflow data may include obtaining cool airflow data of at least one perforated tile included in the cool aisle. In the method, the act of conducting an analysis may include defining a plurality of control volumes in the cool aisle, and the method may further include determining airflows in the cool aisle by determining airflow into and out of at least one of the control volumes. The method may further include comparing the recirculation index for each of the plurality of equipment enclosures with a threshold. The method may further include determining a cooling capacity for each of the equipment enclosures based on the recirculation index for each of the equipment enclosures, and displaying the cooling capacity for each of the equipment enclosures along with a representation of a data center containing the cluster. In the method, the act of conducting an analysis may include assigning different chemical concentration identifiers to the airflows for at least two of the plurality of equipment racks. In the method, the act of conducting an analysis may include importing empirical data and determining end of aisle airflows using the empirical data. In the method, the act of conducting an analysis may include determining cool aisle airflows in isolation from the data center to obtain isolated results, and combining the isolated results with the empirical data.
  • Another aspect of the invention is directed to a computer-implemented method for evaluating the cooling performance of a cluster of equipment racks in a data center, wherein the cluster of equipment racks includes at least a first row of racks and a second row of racks separated by a hot aisle, with each of the equipment racks being configured to exhaust air into the hot aisle. The method includes obtaining at least one of power data and airflow data for each of the equipment racks, obtaining airflow data for at least one air removal unit contained in one of the first row of equipment racks and the second row of equipment racks, and conducting an analysis of airflows in the hot aisle to determine a capture index for at least one of the equipment racks, wherein the capture index is indicative of a fraction of air that is exhausted by the at least one of the equipment racks and captured by the at least one air removal unit.
  • In the method, the at least one air removal unit may include an in-row cooling unit, and the capture index may be equal to a ratio of captured air to total air exhausted by the at least one equipment rack. The method may further include determining a capture index for each of the equipment racks. In the method, the act of conducting an analysis may include defining a plurality of control volumes in the hot aisle, and the method may further include determining airflows in the hot aisle by determining airflow into and out of at least one of the control volumes. The method may further include comparing the capture index for each of the plurality of equipment enclosures with a threshold. The method may include determining a cooling capacity for each of the equipment enclosures based on the capture index for each of the equipment enclosures, and displaying the cooling capacity for each of the equipment enclosures along with a representation of a data center containing the cluster. In the method, the act of conducting an analysis may include assigning different chemical concentration identifiers to the airflows for at least two of the plurality of equipment racks. In the method, the act of conducting an analysis may include importing empirical data and determining end of aisle airflows using the empirical data. The act of conducting an analysis may include determining hot aisle airflows in isolation from the data center to obtain isolated results, and combining the isolated results with the empirical data. In the method, the act of conducting an analysis may include importing empirical rules, and determining the capture index using the empirical rules, and the empirical rules may include coefficients for use in determining at least one capture index.
  • Another aspect of the invention is directed to a computer-readable medium encoded with instructions for execution on a computer system. The instructions, when executed, perform a method comprising acts of obtaining at least one of power data and airflow data for a plurality of equipment racks arranged in a cluster, wherein the cluster of equipment racks includes at least a first row of racks and a second row of racks separated by a cool aisle, with each of the equipment racks being configured to draw cooling air from the cool aisle, obtaining cool airflow data for cool air supplied to the cool aisle from a source of cool air, and conducting an analysis of airflows in the cool aisle to determine a recirculation index for at least one of the equipment racks, wherein the recirculation index is indicative of a quantity of recirculated air included in an input airflow of the at least one equipment rack.
  • The recirculation index may equal to a ratio of recirculated air to total air in the input airflow of the at least one equipment rack, and the acts may further include determining a recirculation index for each of the equipment racks. The act of obtaining cool airflow data may include obtaining cool airflow data for an in-row cooling unit included in the cluster of racks. The act of obtaining cool airflow data may include obtaining cool airflow data of at least one perforated tile included in the cool aisle. The act of conducting an analysis ma include defining a plurality of control volumes in the cool aisle, and wherein the method further includes determining airflows in the cool aisle by determining airflow into and out of at least one of the control volumes. The acts may further include comparing the recirculation index for each of the plurality of equipment enclosures with a threshold, and determining a cooling capacity for each of the equipment enclosures based on the recirculation index for each of the equipment enclosures. The acts may further include displaying the cooling capacity for each of the equipment enclosures along with a representation of a data center containing the cluster. The act of conducting an analysis may include assigning different chemical concentration identifiers to the airflows for at least two of the plurality of equipment racks. The act of conducting an analysis may include importing empirical data and determining end of aisle airflows using the empirical data. The act of conducting an analysis may include determining cool aisle airflows in isolation from the data center to obtain isolated results, and combining the isolated results with the empirical data.
  • Another aspect of the invention is directed to a computer-readable medium encoded with instructions for execution on a computer system. The instructions when executed, perform a method comprising acts of obtaining at least one of power data and airflow data for a plurality of equipment racks arranged in a cluster, wherein the cluster of equipment racks includes at least a first row of racks and a second row of racks separated by a hot aisle, with each of the equipment racks being configured to exhaust air into the hot aisle, obtaining airflow data for at least one air removal unit contained in one of the first row of equipment racks and the second row of equipment racks, and conducting an analysis of airflows in the hot aisle to determine a capture index for at least one of the equipment racks, wherein the capture index is indicative of a fraction of air that is exhausted by the at least one of the equipment racks and captured by the at least one air removal unit.
  • In the method, the at least one air removal unit may be an in-row cooling unit, and the capture index may be equal to a ratio of captured air to total air exhausted by the at least one equipment rack. The acts may further include determining a capture index for each of the equipment racks. The act of conducting an analysis may include defining a plurality of control volumes, and wherein the method further includes determining airflows in the hot aisle by determining airflow into and out of at least one of the control volumes. The acts may further include comparing the capture index for each of the plurality of equipment enclosures with a threshold. The acts may further include determining a cooling capacity for each of the equipment enclosures based on the capture index for each of the equipment enclosures, and displaying the cooling capacity for each of the equipment enclosures along with a representation of a data center containing the cluster. The act of conducting an analysis may include assigning different chemical concentration identifiers to the airflows for at least two of the plurality of equipment racks. The act of conducting an analysis may include importing empirical data and determining end of aisle airflows using the empirical data. The act of conducting an analysis may include determining hot aisle airflows in isolation from the data center to obtain isolated results, and combining the isolated results with the empirical data. The act of conducting an analysis may include importing empirical rules, and determining the capture index using the empirical rules. The empirical rules include coefficients for use in determining at least one capture index.
  • According to one aspect of the present invention, a computer-implemented method for providing a representation of capacity of a data center resource is provided. The method comprises acts of determining a first indication of excess capacity of a first data center resource that is unusable due to insufficient capacity of at least one other data center resource associated with the first data center resource, and providing the first indication of excess capacity of the first data center resource to an external entity. According to one embodiment of the present invention, the act of determining the first indication of excess capacity includes an act of determining an amount of excess capacity of the first data center resource that is unusable due to insufficient capacity of the at least one other data center resource associated with the first data center resource. According to another embodiment of the invention, the act of determining the first indication of excess capacity includes an act of determining an indication of excess capacity at a location in a rack. According to another embodiment of the invention, the act of determining the first indication of excess capacity at the location in the rack includes an act of determining an indication of excess capacity at a U space position. According to another embodiment of the invention, the act of determining the first indication of excess capacity includes an act of determining an indication of excess capacity of rack space that is unusable due to insufficient capacity of power.
  • According to one embodiment of the present invention, the act of determining the first indication of excess capacity includes an act of determining an indication of excess capacity of power that is unusable due to insufficient capacity of cooling. According to another embodiment of the invention, the act of determining the first indication of excess capacity includes an act of determining an indication of excess capacity of power distribution that is unusable due to insufficient capacity of power available for distribution. According to another embodiment of the invention, the act of determining the first indication of excess capacity includes an act of determining an indication of excess capacity of physical space that is unusable due to insufficient capacity of cooling. According to another embodiment of the invention, the act of determining the first indication of excess capacity includes an act of determining an indication of excess capacity of power distribution that is unusable due to insufficient capacity of network connectivity. According to another embodiment of the invention, the act of determining the first indication of excess capacity includes an act of determining an indication of excess capacity of U space that is unusable due to insufficient capacity of weight support.
  • According to one embodiment of the present invention, the act of providing the first indication of excess capacity includes an act of providing the indication to another system element. According to another embodiment of the invention, the act of providing the first indication of excess capacity includes an act of presenting the first indication to a user of a computer system. According to another embodiment of the invention, the act of presenting the first indication to the user of the computer system includes an act of presenting, to the user of the computer system, an amount of excess capacity of the first data center resource that is unusable due to insufficient capacity of the at least one other data center resource associated with the first data center resource. According to another embodiment of the invention, the act of presenting, to the user of the computer system, the amount of excess capacity includes acts of presenting, to the user of the computer system, an identifier that identifies the first data center resource and presenting, to the user of the computer system, an identifier that identifies the at least one other data center resource. According to another embodiment of the invention, the method further comprises determining a configuration of data center equipment that minimizes, relative to at least one other configuration, the excess capacity of the first data center resource, and providing the configuration of data center equipment to an external entity.
  • According to one embodiment of the present invention, determining the configuration of data center equipment includes determining a location for at least one element of data center equipment. According to another embodiment of the invention, the method further comprises determining a second indication of excess capacity of a second data center resource that is unusable due to insufficient capacity of at least one other data center resource associated with the second data center resource determining a configuration of data center equipment that minimizes, relative to at least one other configuration, an aggregate of the first indication and the second indication, and providing the configuration of data center equipment to an external entity.
  • According to one aspect of the present invention, a computer-implemented method for providing a representation of a data center resource is provided. The method comprises acts of determining a status of capacity for the data center resource provided to either a rack or a location in a rack, and providing the status to an external entity. According to another embodiment of the invention, determining the status of capacity for the data center resource includes determining a status for the data center resource provided to a U space position. According to one embodiment of the present invention, the act of determining the status of capacity for the data center resource includes an act of determining a status of capacity of power. According to another embodiment of the invention, the act of determining the status of capacity for the data center resource includes an act of determining a status of capacity of cooling. According to another embodiment of the invention, the act of determining the status of capacity for the data center resource includes an act of determining a status of capacity of physical space. According to another embodiment of the invention, the act of determining the status of capacity for the data center resource includes an act of determining a status of capacity of weight support. According to another embodiment of the invention, the act of determining the status of capacity for the data center resource includes an act of determining a status of available capacity. According to another embodiment of the invention, the act of determining the status of capacity for the data center resource includes an act of determining a status of utilized capacity.
  • According to one aspect of the present invention, a computer-implemented method for providing a representation of a data center resource is provided. The method comprises acts of determining a rate of change of a utilized capacity of the data center resource based at least in part on changes to a data center configuration, and providing a predicted utilization of the data center resource based at least in part on the rate of change. According to another embodiment of the invention, determining the rate of change of utilized capacity of the data center resource includes determining the rate of change using linear regression. According to another embodiment of the invention, the data center resource includes power. According to another embodiment of the invention, the data center resource includes cooling.
  • According to one aspect of the present invention, a system for providing a representation of a capacity of data center resource is provided. The system comprises an input configured to receive a capacity of a first data center resource and a capacity of at least one other data center resource, an output configured to provide an amount of excess capacity of the first data center resource, and a controller coupled to the input and the output and configured to determine an amount of excess capacity of the first data center resource that is unusable due to insufficient capacity of the at least one other data center resource.
  • According to one aspect of the present invention, a system for providing a representation of a data center resource is provided. The system comprises an input configured to receive capacity information for a data center resource, an output configured to provide a status of capacity for the data center resource, and a controller coupled to the input and the output and configured to determine the status of capacity for the data center resource at either a rack or a location in a rack.
  • According to one aspect of the present invention, a system for providing an evaluation of a data center resource is provided. The system comprises an input configured to receive changes in a data center configuration, an output configured to provide a predicted utilization of the data center resource based on a rate of change of utilized capacity of the data center resource, and a controller configured to determine the rate of change of the utilized capacity of the data center resource based on the changes in a data center configuration.
  • According to one aspect of the present invention, a computer-implemented method for distributing data describing models of data center equipment is provided. The method comprises acts of indicating a first group of the data describing models of data center equipment, the first group identifying standard models of data center equipment, indicating a second group of the data describing models of data center equipment, the second group identifying models approved for use in a first data center, indicating a third group of the data describing models of data center equipment, the third group identifying models preferred for use in the first data center, providing the second group to a first remote computing device located in the first data center, and providing the third group to the first remote computing device located in the first data center. According to one embodiment of the present invention, the act of providing the second group comprises an act of providing data including at least one of the group comprising data center equipment manufacturer, data center equipment model, data center equipment power plug type, data center equipment number of plugs required, data center equipment power required, data center equipment airflow required, data center equipment network connectivity required, data center equipment weight support required, data center equipment cooling required, and data center equipment physical space required. According to another embodiment of the invention, the method further comprises an act of providing the third group from the first remote computing device to a second remote computing device.
  • According to one embodiment of the present invention, the act of providing the third group from the first remote computing device includes an act of providing the third group to a mobile computing device. According to another embodiment of the invention, the act of indicating a first group includes storing an indicator of the first group in a database hosted by a system provider. According to another embodiment of the invention, the method further comprises an act of receiving first non-standard data describing models of data center equipment from the first remote computing device. According to another embodiment of the invention, the method further comprises an act of receiving non-standard data describing models of data center equipment from an external entity. According to another embodiment of the invention, the act of receiving non-standard data includes receiving additional data from a data center equipment manufacturer. According to another embodiment of the invention, the method further comprises acts of indicating a fourth group of the data describing models of data center equipment, the fourth group identifying models approved for use in a second data center, indicating a fifth group of the data describing models of data center equipment, the fifth group identifying models preferred for use in the second data center, providing the fourth group to a third remote computing device located in the second data center, and providing the fifth group to the third remote computing device located in the second data center. According to another embodiment of the invention, the method further comprises an act of receiving second non-standard data describing models of data center equipment from the third remote computing device.
  • According to one aspect of the present invention, a computer-implemented method for maintaining a data center equipment database is provided. The method comprises acts of receiving a set of configuration management data from each of a plurality of data centers, each set of configuration management data identifying one of the plurality data centers, a capacity for power of the identified data center, a capacity for cooling of the identified data center and at least one element of data center equipment disposed within the identified data center. According to one embodiment of the present invention, the acts of receiving include acts of receiving configuration management data including at least one of the group comprising data center physical dimensions, data center equipment location within the data center physical dimensions, data center power distribution provided to the data center equipment location, data center cooling distribution provided to the data center equipment location, data center network connectivity provided to the data center equipment location, data center equipment manufacturer, and data center equipment model. According to another embodiment of the invention, the method further comprises acts of receiving a set of data center resource consumption and production data from each of the plurality of data centers, each set of data center resource consumption and production data identifying one of the plurality of data centers and being associated with at least one element of data center equipment disposed within the identified data center.
  • According to one embodiment of the present invention, the acts of receiving a set of data center resource consumption and production data from each of the plurality of data centers include acts of receiving data center resource consumption and production data including at least one from the group comprising data center equipment identity, data center equipment manufacturer, data center equipment model, data center equipment power plug type used, data center equipment number of plugs used, data center equipment power used, data center equipment airflow used, data center equipment network connectivity used, data center equipment weight support used, data center equipment cooling used, and data center equipment physical space used. According to another embodiment of the invention, the acts of receiving a set of data center resource consumption and production data from each of the plurality of data centers include acts of receiving data center resource consumption and production data including at least one from the group comprising data center equipment identity, data center equipment manufacturer, data center equipment model, data center equipment power plug type provided, data center equipment number of plugs provided, data center equipment power provided, data center equipment airflow provided, data center equipment network connectivity provided, data center equipment weight support provided, data center equipment cooling provided, and data center equipment physical space provided.
  • According to one aspect of the present invention, a computer-implemented method for distributing data describing models of data center equipment is provided. The method comprises acts of receiving, from a first remote computing device, a first group of the data describing models of data center equipment, the first group identifying models approved for use in a data center, receiving, from the first remote computing device, a second group of the data describing models of data center equipment, the second group identifying models preferred for use in the data center, and providing the second group to a second remote computing device. According to one embodiment of the present invention, the method further comprises acts of receiving, from an external entity, non-standard data describing models of data center equipment that are not approved for use in the data center, and providing the non-standard data to the first remote computing device.
  • According to one aspect of the present invention, a system for managing equipment data describing models of data center equipment is provided. The system comprises an input configured to receive data associated with the equipment data, an output configured to provide data associated with the equipment data, and a controller coupled to the input and the output and configured to allow an external entity to indicate that a first group of the equipment data describes standard models of data center equipment; to indicate that a second group of the equipment data describes models of data center equipment approved for use in a data center, to indicate that a third group of the equipment data describes models of data center equipment preferred for use in the data center and further configured to provide, via the output, the second group to a remote computing device located in the data center and to provide, via the output, the third group to the remote computing device located in the data center.
  • According to one aspect of the present invention, a system for managing equipment data describing models of data center equipment is provided. The system comprises a computer readable medium, an input configured to receive data associated with the equipment data, an output configured to provide data associated with the equipment data, and a controller coupled to the input, the output and the computer readable medium and configured to store, in the computer readable medium, an first indication that a first group of the equipment data describes models of data center equipment approved for use in a data center and to store a second indication that a second group of equipment data describes models of data center equipment preferred for use in the data center and further configured to provide, via the output, the second group to a remote computing device located in the data center.
  • According to one aspect of the present invention, a system for maintaining a data center equipment database is provided. The system comprises an input configured to receive a set of configuration management data from each of a plurality of data centers, each set of configuration management data identifying one of the plurality data centers, a capacity for power of the identified data center, a capacity for cooling of the identified data center and at least one element of data center equipment disposed within the identified data center, and a controller coupled to the input and configured to store, in the data center equipment database, each set of configuration management data.
  • According to one aspect of the present invention, a computer-implemented method for managing a data center configuration is provided. The method comprises acts of receiving, from a user, a proposed change to the data center configuration though an interactive representation of a layout of an identified space in a data center, the proposed change including data identifying at least one element of data center equipment to be provisioned in the data center, the at least one element of data center equipment having a plurality of data center resource requirements, determining at least one location in the data center having a plurality of data center resources that satisfy the plurality of data center resource requirements of the at least one element of data center equipment, and storing the at least one location on a computer readable medium. According to one embodiment of the present invention, the act of receiving, from the user, the proposed change to the data center configuration includes an act of receiving data identifying at least one element of floor mounted data center equipment to be provisioned in the data center. According to another embodiment of the invention, the act of receiving, from the user, the proposed change includes an act of receiving, from the user, the proposed change to the data center configuration though an interactive representation of a layout of a data center room. According to another embodiment of the invention, the act of receiving, from the user, the proposed change includes an act of receiving, from the user, the proposed change to the data center configuration though an interactive representation of a layout of a rack. According to another embodiment of the invention, the act of receiving, from the user, the proposed change to the data center configuration includes an act of receiving information regarding at least one element of data center equipment represented in the interactive representation.
  • According to one embodiment of the present invention, the act of receiving, from the user, the proposed change to the data center configuration includes an act of scanning a barcode. According to another embodiment of the invention, the act of determining the at least one location includes an act of determining at least one location in near real-time. According to another embodiment of the invention, the act of determining the at least one location includes an act of receiving, from an external entity, data identifying at least one location in the data center for the at least one element of data center equipment having a plurality of data center resources. According to another embodiment of the invention, the act of determining the at least one location includes an act of determining a location that minimizes, relative to at least one other location, an amount of excess data center resource capacity that is unusable due to insufficient associated capacity of at least one other data center resource, and the method further comprises providing the at least one location to an external entity. According to another embodiment of the invention, the act of providing the at least one location further includes an act of presenting the at least one location to a user of a computer system. According to another embodiment of the invention, the act of determining the at least one location includes an act of determining a series of locations arranged consecutively according to an amount of excess data center resource capacity that is unusable due to insufficient associated capacity of at least one other data center resource and the method further comprises providing the series of locations to an external entity.
  • According to one embodiment of the present invention, the act of providing the at least one location includes an act of presenting the series to a user of a computer system. According to another embodiment of the invention, the method further comprises an act of presenting, to the user, information regarding the proposed change to the data center configuration, the information including a representation of at least one element of data center equipment that is disposed in a rack. According to another embodiment of the invention, the act of presenting, to the user, the information regarding the proposed change to the data center configuration includes an act of presenting, to the user, a representation of the at least one element of data center equipment as placed in inactive storage, and the method further includes an act of recording the at least one element of data center equipment as placed in inactive storage. According to another embodiment of the invention, the act of presenting, to the user, the information regarding the proposed change to the data center configuration includes an act of presenting, to the user, an interactive representation that is updated based on the proposed change. According to another embodiment of the invention, the act of presenting, to the user, the interactive representation that is updated based on the proposed change includes an act of presenting, to the user, the interactive representation updated in near real-time. According to another embodiment of the invention, the method further comprises generating a work order identifying a set of tasks required to complete the proposed change, and presenting the set of tasks via an output of a computer system. According to another embodiment of the invention, the method further comprises acts of detecting a change in the power demanded at power distribution device associated with the work order, and recording the work order as completed based on the change in power demanded. According to another embodiment of the invention, the method is performed using a remote computing device.
  • According to one aspect of the present invention, a computer-implemented method for managing a data center configuration is provided. The method comprises acts of receiving at least one policy identifying at least one characteristic of a data center resource as provided to at least one element of data center equipment, receiving a proposed change to the data center configuration, and determining a compliance result that indicates whether the data center configuration, including the proposed change, complies with the at least one policy. According to one embodiment of the present invention, the act of receiving the at least one policy includes an act of receiving at least one policy identifying a redundancy level. According to another embodiment of the invention, the act of receiving the at least one policy includes an act of receiving at least one policy identifying a runtime requirement. According to another embodiment of the invention, the act of receiving the at least one policy includes an act of receiving at least one policy identifying a security requirement. According to another embodiment of the invention, the act of receiving the at least one policy includes an act of receiving at least one policy identifying a requirement of cooling. According to another embodiment of the invention, the act of receiving the at least one policy includes an act of receiving at least one policy identifying a requirement of network connectivity. According to another embodiment of the invention, the act of receiving the at least one policy includes an act of receiving at least one policy identifying a requirement of power. According to another embodiment of the invention, the act of receiving the at least one policy identifying a requirement of power includes an act of receiving at least one policy identifying a requirement of power distribution.
  • According to one aspect of the present invention, a computer-implemented method for designing a data center is provided. The method comprises acts of receiving at least one policy identifying at least one characteristic of a data center resource as provided to at least one element of data center equipment, receiving data identifying at least one location in the data center for at least one element of data center equipment, the at least one location having a plurality of data center resources and the at least one element of data center equipment having a plurality of data center resource requirements, and determining at least one element of data center equipment that satisfies the policy when located at the at least one location, wherein the plurality of data center resources of the at least one location satisfy the plurality of data center resource requirements of the at least one element of data center equipment.
  • According to one aspect of the present invention, a system for managing a data center configuration comprising an input configured to receive, from a user, a proposed change to the data center configuration though an interactive representation of a layout of an identified space in a data center, an output configured to provide a set of tasks to a user, and a controller coupled to the input and the output and configured to generate a work order identifying the set of tasks required to complete the proposed change and to provide the set of tasks to the output. According to one embodiment of the present invention, the system further comprises a computer readable medium coupled to the controller and wherein the input is further configured to receive status information associated with the work order and the controller is further configured to record the status information in the computer readable medium. According to another embodiment of the invention, the system further comprises an input configured to receive, from a user, status information associated with the work order and wherein the controller is further configured to provide the status information to the output.
  • According to one aspect of the present invention, a system for managing a data center configuration comprises a first input configured to receive at least one policy identifying at least one characteristic of a data center resource as provided to at least one element of data center equipment, a second input configured to receive a proposed change to the data center configuration, and a controller coupled to the first and second inputs and configured to determine a compliance result that indicates whether the data center configuration, including the proposed change, complies with the at least one policy.
  • According to one aspect of the present invention, a system for designing a data center is provided. The system comprises an first input configured to receive at least one policy identifying at least one characteristic of a data center resource as provided to at least one element of data center equipment, a second input configured to receive data identifying at least one location in the data center for at least one element of data center equipment, the at least one location having a plurality of data center resources and the at least one element of data center equipment having a plurality of data center resource requirements, and a controller coupled to the first and second inputs and configured to determine at least one element of data center equipment that satisfies the policy when located at the at least one location, wherein the plurality of data center resources of the at least one location satisfy the plurality of data center resource requirements of the at least one element of data center equipment.
  • According to one aspect of the present invention, a remote computing device for navigating a representation of a data center is provided. The remote computing device comprises an interface configured to present, to a user, an interactive representation of a layout of an identified space in a data center and to receive, from the user, an identifier of at least one element of data center equipment located in the data center, a memory configured to store information regarding data center equipment represented by the interface, and a controller coupled to the interface and memory and configured to retrieve information regarding the identified at least one element of data center equipment from the memory and provide the information to the interface.
  • According to one aspect of the present invention, a computer-implemented method for managing data center resource redundancy is provided. The method comprises acts of determining a redundancy level of a data center resource at a location in a rack, and providing the redundancy level of the data center resource. According to one embodiment of the present invention, the act of determining the redundancy level of the data center resource includes an act of determining a redundancy level of cooling. According to another embodiment of the invention, the act of determining the redundancy level of the data center resource includes an act of determining a redundancy level of network connectivity. According to another embodiment of the invention, the act of determining the redundancy level of the data center resource includes an act of determining a redundancy level of power. According to another embodiment of the invention, the act of determining the redundancy level of the data center resource includes an act of determining a redundancy level of power distribution. According to another embodiment of the invention, the act of determining the redundancy level of the data center resource at a location in a rack includes an act of determining a redundancy level of a data center resource at a U space position. According to another embodiment of the invention, the act of providing the redundancy level includes an act of presenting, to a user of a computer system, a representation of the at least one location in a rack. According to another embodiment of the invention, the act of presenting, to the user of the computer system, the representation of the at least one location in the rack includes an act of presenting, to a user of a computer system, a representation at least one U space position in a rack. According to another embodiment of the invention, the act of presenting, to the user of the computer system, the representation of the at least one location in the rack includes an act of presenting, to a user of a computer system, a representation of a redundancy relationship between elements disposed in a data center. According to another embodiment of the invention, the act of presenting, to the user of the computer system, a representation of a redundancy relationship between elements disposed in a data center includes an act of presenting, to a user of a computer system, a redundancy relationship between elements disposed in a rack.
  • According to one aspect of the present invention, a system for managing data center resource redundancy is provided. The system comprises a controller configured to determine the redundancy level of a data center resource at a location in a rack, and an output coupled to the controller and configured to provide a representation of the redundancy level of the data center resource. According to one embodiment of the present invention, the location in the rack includes a U space position in a rack. According to another embodiment of the invention, the system includes a remote computing device.
  • Further features and advantages of the present invention as well as the structure and operation of various embodiments of the present invention are described in detail below with reference to the accompanying drawings. In the drawings, like reference numerals indicate like or functionally similar elements. Additionally, the left-most one or two digits of a reference numeral identifies the drawing in which the reference numeral first appears.
  • BRIEF DESCRIPTION OF DRAWINGS
  • The accompanying drawings, are not intended to be drawn to scale. In the drawings, each identical or nearly identical component that is illustrated in various figures is represented by a like numeral. For purposes of clarity, not every component may be labeled in every drawing. In the drawings:
  • BRIEF DESCRIPTION OF DRAWINGS
  • The accompanying drawings are not intended to be drawn to scale. In the drawings, each identical or nearly identical component that is illustrated in various figures is represented by a like numeral. For purposes of clarity, not every component may be labeled in every drawing. In the drawings:
  • FIG. 1 is a top view of a data center of the type with which embodiments of the present invention may be used;
  • FIG. 2 is a side view of the data center of FIG. 1.
  • FIG. 3 is a functional block diagram of a system in accordance with one embodiment of the present invention;
  • FIG. 4 is a flowchart of a process that may be implemented using the system of FIG. 3 in accordance with one embodiment of the invention;
  • FIG. 5 is a diagram showing facility information that can be displayed using at least one embodiment of the invention;
  • FIGS. 5A and 5B are diagrams showing additional information that can be displayed using embodiments of the invention;
  • FIGS. 5C and 5D show graphical user interface screens that exhibit various aspects of the present invention;
  • FIG. 6 is a functional block diagram of a management system in accordance with one embodiment of the invention;
  • FIG. 7 is a flow chart of a management process in accordance with one embodiment of the invention;
  • FIG. 8 shows a perspective view of a cluster of racks whose cooling performance can be analyzed using embodiments of the invention;
  • FIG. 9 shows a top view of a cluster of racks whose cooling performance can be analyzed using a control volume analysis technique of at least one embodiment;
  • FIG. 9A shows the cluster of racks of FIG. 9 along with staggered control volumes that may be used in the control volume analysis technique;
  • FIG. 10 is a flow chart of a control volume analysis technique in accordance with one embodiment of the invention;
  • FIG. 11 is a diagram demonstrating a principle of superposition used in one embodiment;
  • FIG. 12 is a graph used in determining airflows in one embodiment;
  • FIG. 13 is a diagram identifying airflows used with one analysis method of one embodiment;
  • FIG. 14 is a flow chart of a process for determining a recirculation index in one embodiment;
  • FIG. 15 is a schematic diagram showing a layout of equipment racks used in an analysis in one embodiment to determine a capture index;
  • FIG. 16 is a flowchart of a process for determining a capture index in accordance with one embodiment;
  • FIG. 17 is a functional block diagram of a computer system that may be used in embodiments of the invention;
  • FIG. 18 is a functional block diagram of a storage system that may be used with the computer system of FIG. 17;
  • FIG. 19 is a flowchart of a process that may be implemented using the system of FIG. 3 in accordance with one embodiment of the invention;
  • FIG. 20 shows graphical user interface screen that exhibits various aspects of the present invention;
  • FIG. 21 depicts another graphical user interface screens that exhibits various aspects of the present invention;
  • FIG. 22 illustrates another graphical user interface screen that exhibits various aspects of the present invention; and
  • FIG. 23 shows a representation used to analyze stranded capacity according to an aspect of the present invention.
  • DETAILED DESCRIPTION
  • This invention is not limited in its application to the details of construction and the arrangement of components set forth in the following description or illustrated in the drawings. The invention is capable of other embodiments and of being practiced or of being carried out in various ways. Also, the phraseology and terminology used herein is for the purpose of description and should not be regarded as limiting. The use of “including,” “comprising,” or “having,” “containing,” “involving,” and variations thereof herein, is meant to encompass the items listed thereafter and equivalents thereof as well as additional items.
  • Embodiments of the present invention may be used to design, manage and retrofit a data center, such as data center 100 which is shown in FIGS. 1 and 2 with FIG. 1 showing a top view of the data center 100, and FIG. 2 showing a side view of the data center 100. As discussed further below, the design of the layout of the data center 100, including various data center resources such as power and cooling considerations may be performed using systems and processes of embodiments of the present invention. A data center resource may include, for example, any characteristic of a data center that supports and/or ensures data center equipment functionality. Examples of data center resources include power, cooling, physical space, weight support, remote equipment control capability, physical and logical security and physical and logical network connectivity. Power data center resources may include power distribution resources, such as transformers, PDUs and outlets, and power available for distribution, such as utility power supplied to the data center, power generated by an onsite generator and power supplied by UPSs. Physical space resources in a data center may include data center floor space and rack U space. Cooling resources in a data center may include cooling distribution capacity and cooling generation capacity. Physical security resources in a data center may include security cameras and door locks. Logical network connectivity resources in a data center may include Virtual Local Area Networks, Domain Name Services, and Dynamic Host Configuration Protocol Services. Physical network connectivity resources may include network cabling and patch panels. Remote equipment control capability resources in a data center may include Keyboard Video Mouse services.
  • Embodiments of the invention, however, are not limited for use with data centers like that shown in FIGS. 1 and 2 and may be used with other facilities that do not include raised floors and may be used with facilities that house equipment other than computing equipment, including telecommunications facilities and other facilities. Further, embodiments of the invention may be used with raised floor and equipment layouts that are not neatly arranged in the manner shown in FIGS. 1 and 2. Embodiments of the present invention may use systems, devices and methods described in U.S. patent application Ser. No. 10/038,106, filed Jan. 2, 2002, titled “Rack Power System and Method,” incorporated herein in its entirety by reference.
  • The data center 100 includes rows of racks 102A, 102B, 102C and 102D, cooling units 104A and 104B, and a raised floor 106. Each of the rows includes racks 108, at least a number of which draw cool air from the front of the rack and return warm air to the rear or top or rear and top of the rack. Each rack may contain U space positions designed to house rack mounted data center equipment, such as, for example, servers, cooling equipment and network connectivity equipment.
  • As understood by those skilled in the art, to optimize cooling performance in a data center, rows of racks are often positioned to create alternating cold aisles and hot aisles. In the configuration shown in FIGS. 1 and 2, aisles 110A, 110B and 110C are hot aisles and aisles 112A and 112B are cold aisles. To provide cooling to the racks, in front of each of the racks in the cold aisle, perforated floor tiles 114 are used to provide cooling air from under the raised floor. In the data center 100, in addition to the perforated floor tiles shown, the raised floor may include solid floor tiles. The cooling units 104A and 104B are designed to provide cool air to the area under the raised floor and to receive return warm air from the space adjacent the ceiling of the data center. In other embodiments, in addition to or in place of the cooling units 104A and 104B, in-row cooling units, such as those available from American Power Conversion Corporation, may be used. Further, in at least one embodiment, half-rack in-row cooling units may be used, as described in co-pending U.S. patent application Ser. No. 11/335,901, entitled COOLING SYSTEM AND METHOD, Attorney Docket No. A2000-704819, by Neil Rasmussen et al., filed on Jan. 19, 2006, and incorporated herein by reference. As described in the referenced application, the half-rack, in-row units have a width of twelve inches, which is approximately half of that of a standard data center rack.
  • One embodiment of the invention, directed to a system and a method for designing, monitoring, and upgrading the equipment installed in a data center, such as data center 100, will now be described with reference to FIG. 3. This equipment may include rack mounted equipment, such as servers, storage devices, and network connectivity equipment, and floor mounted equipment, such as three phase power distribution units and CRACs. FIG. 3 shows a functional block diagram of a design and management system 200. Embodiments of the invention are not limited to the functions provided by the functional blocks or the particular arrangement of the blocks. In addition, the functions provided by the system 200 need not be implemented on one computer system, but rather may be implemented using a number of networked devices as described further below that provide the functions described. Further, particular embodiments may have more or less functions and functional modules than those described below with reference to FIG. 3. In different embodiments, the functions described with reference to FIG. 3 may be performed on one processor or controller or may be distributed across a number of different devices.
  • The system 200 includes an input module 202, a display module 204, a builder module 206, a facility management module 208, an integration module 210, a database module 212, and a work order module 214. The input module 202 provides an interface to allow users to enter data into the system 200. The input module may include, for example, one of a number of known user input devices for computer systems, and in addition, in at least one embodiment, electronic data regarding a facility and/or equipment to be loaded into a facility may be entered into the system through a network interface or using an electronic media storage reader. Information may flow between these modules using any technique known in the art. Such techniques include passing the information over the network via TCP/IP, passing the information between modules in memory and passing the information by writing to a file, database, or other storage entity, such as a storage device, disk or other type of storage entity.
  • The display module includes a display interface and may include a graphical display to display output data to a user. In addition, the display module may include an interface for one or more printers that provide a hard copy of output data.
  • The builder module 206 includes routines for designing optimal layout of equipment in a facility, determining data center resource requirements, such as power requirements and cooling requirements, for electronic enclosures and/or equipment racks, ensuring that the placement of equipment, cooling units and power distribution branches in the facility allow the data center resource requirements, such as power and cooling requirements, to be met, and calculating for each electronic enclosure, and/or equipment rack, the remaining data center resource capacity, such as power capacity and cooling capacity, available based on the layout of equipment in the facility.
  • In another embodiment, builder module 206 exposes interfaces that allow for entry of data center resource supply policies. These policies may detail additional desired characteristics of the data center resources as provided to data center equipment. For example, data center resource supply policies may specify desired data center resource redundancy levels and runtime requirements. Additionally, data center resource supply policies may specify security needs for data center equipment, such as, for example, the need to locate servers with financial data in racks that made be locked and/or under surveillance. As is discussed further below, both system and method embodiments may compare particular data center configurations to data center supply polices to determine whether the particular data center configuration complies with the data center policies. The result of this comparison may be stored on a computer readable medium for later use.
  • The facility management module 208 is used by the system 200 after equipment is installed in the facility. The management module includes routines to monitor data center resource characteristics, such as power and cooling characteristics, of equipment in a facility. The management module may be coupled, either directly or through one or more networks, to measurement devices and control devices throughout the facility and may record their history for analysis, summarization and exportation. Based at least in part on methods discussed further below, the facility management module 208 may ascertain data center resources provided to a particular space or location in the data center.
  • According to one embodiment, the integration module 210 is the main module in the system and coordinates flow of data in the system to perform methods of embodiments of the present invention.
  • The database module is used to store data regarding various devices that may be used in a data center, such as servers, uninterruptible power supplies, power strips, network connectivity equipment (such as network cabling, hubs, routers, wireless routers, switches, patch panels, etc.), automatic transfer switches, power distribution units, air conditioning units, racks and any other data center equipment. The data stored may include data center resource supply policies for data center equipment. The data stored may also include data center resource consumption and production benchmarks for data center equipment such as physical parameters (e.g., dimensions/rack space requirements, power plug types, network cabling specifications, etc.) as well as power and cooling consumption data. In the case of data center equipment that provide data center resources such as network connectivity equipment, power supplies and air conditioning units, data center resource production benchmarks may include network connectivity, cooling and power output characteristics (e.g. total output capacity, number and type of connections/plugs available, etc.). It should be appreciated that benchmarks may indicate that certain data center equipment produce, and/or consume, data center resources. Examples of data center resource providing equipment include electrical generators and CRACs, among others. As described below, the database module may be used in embodiments of the invention to provide a complete bill of materials (BOM) for a completed design. In one embodiment, a centralized web-accessible database server may be used to store equipment information and warnings and error messages, allowing easy access to the information for editing.
  • In another embodiment, a user may maintain data center equipment information, such as the information handled by database module 212, with the assistance of system 300. System 300 is also depicted in functional block diagram FIG. 3. System 300, which may be hosted by the system provider, includes data center equipment database 304 and data center equipment database interface 302. In general, interface 302 may be an interface to database 304 that may receive or provide any data appropriate for storage in database 304 including data center configuration, equipment or resource information. Database 304, in turn, may receive and store from interface 302 or retrieve and provide to interface 302 data center information including data center equipment resource requirements, data center configurations and data center resource redundancy and runtime requirements.
  • Database 304 may serve as a master database for the system provider and thus may include several types of information related to data centers. In one embodiment, database 304 includes a data center configuration management database (CMDB), which may encompass various configuration management data including one or more specific data center physical and logical configurations, and a data center equipment characteristics database, which may include theoretical and actual data center resource production and consumption information for data center equipment. More particularly, database 304 may accept and store specific CMDB information for the system provider's installed base of data centers or a subset thereof. This information may include the entire physical layout of a data center, e.g. its physical dimensions, the location and identity of data center equipment, and data center resource capacity, redundancy and runtime requirements, among other information that may have relevance to the design and performance of a data center. The source of this information may include the initial design on the data center as discussed with regard to method 300 below.
  • The type of information stored by database 304 for data center equipment characteristics may include the information discussed above with regard to database module 212, e.g. nameplate values as referred to in the art. Additionally, database 304 may also store data center resource consumption and production history for data center equipment and may use this information to maintain a set of benchmarks that are specific to the manufacturer and model of data center equipment. These specific historical measurements may be summarized into various forms to establish a benchmark that is specific to the manufacturer and model of data center equipment and that is based on actual, practical usage of the data center equipment rather than theoretical values (e.g. nameplate values). These data center resource consumption and production summaries may include, among others, minimum, maximum and average data center resource consumption or production, data center resource consumption or production as a function of time, e.g. power or cooling consumption or production by day of week, week of year, etc., actual data center resource consumption or production when a change in either is requested, and data center resource consumption or production as a function of data center equipment utilization. The source of this information may include the initial design on the data center, including customer or system provider entered benchmarks, and ongoing parameter measurement as discussed with regard to method 300 below. According to one aspect, it is appreciated that these benchmarks, in turn, may be more accurate than the nameplate values and may be used for simulation purposes during the design and retrofitting processes discussed below.
  • Interface 302 may expose both user interfaces (UIs) and system interfaces to exchange database 304 information with external entities. These external entities may include systems and/or users. Interface 302 may both restrict input to a predefined information domain and validate any information entered prior to using the information or supplying the information to other modules. For instance, in one embodiment, interface 302 may include a Structured Query Language (SQL) interface to enable a user or application program to interrogate database 304. This SQL interface may include graphical elements that a user may actuate to build SQL statements and may also include elements that enable a user to simply input SQL statement directly.
  • In other embodiments, interface 302 may include more complex grouping, translation, validation and/or restriction logic. For instance, interface 302 may validate that a SQL statement entered by a user conforms to proper SQL syntax prior to executing it against database 304. In one embodiment, interface 302 may expose a UI with characteristics similar to those of builder module 206 to enable users to create database information representing various data center configurations. In another example, interface 302 may expose a user interface allowing a user to enter new data center equipment resource information, including manufacturer, model and data center resource consumption and production benchmarks. Interface 302 may restrict entry of resource consumption and production benchmarks to predefined data center resources, such as power, cooling, physical space, etc. In still another embodiment, a user may establish groups of data center equipment through interface 302 and approve these equipment groups for use within a set of data centers or upon certain types of computing devices, such as remote or mobile computing devices. Furthermore, interface 302 may enable a user to designate a hierarchical relationship between groups of equipment. As discussed further in regard to method 300 below, arrangement of this information into a hierarchy of groups may ease database maintenance and distribution.
  • In one embodiment, interface 302 may expose a system interface that imports data center configuration and measured parameter information from a system provider's installed base of data centers or data center equipment providers. Interface 302 may utilize a standard protocol, such as, for example, SQL or SQL wrapped in SOAP, to implement such a system interface and may include non-standard protocol elements that are parsed and resolved to commands to be executed by database 304. Interface 302 may validate database commands prior to execution on database 304. Any customer specific data center equipment, e.g. data center equipment entered by a customer that is not present in the database 304, may be imported into database 304 by interface 302 if any applicable validation is successful. Similarly, equipment not present in database 304 that is made available by data center equipment providers may be imported into database 304 by interface 302 if any applicable validation is successful. These features enable system 300 to easily add new data center equipment to its set of managed elements.
  • Furthermore, importation of measure parameters may trigger a parameter summarization process that updates data center equipment resource consumption and production benchmarks to reflect actual usage history. The summarization process may include updating, among others, minimum, maximum and average power consumed or produced benchmarks, benchmarks for power consumption or production as a function of time, e.g. power consumed or produced by day of week, week of year, etc., benchmarks for power actually consumed or produced when a change in either is requested, and/or benchmarks for power consumed or produced as a function of data center equipment utilization.
  • In still another embodiment, interface 302 may expose a system interface that exports data center equipment information, or catalogs, to external storage. This external storage may reside at various locations on various computing devices. Furthermore, as discussed below, the particular information that is exported to these locations and computing devices may be relegated to specific groups of data center equipment.
  • It should be appreciated that in at least one embodiment, the functionality of system 300 may be included in database module 212 of system 200.
  • In yet another embodiment, the work order module 214 may enable management of changes to the configuration of the data center, such as those resulting from a data center retrofit process as discussed below with regard to process 600. In general, work order module 214 analyzes any changes made to the data center configuration by the other modules in system 200. Then work order module 214 generates a set of tasks that, once complete, will implement those changes. Next, work order module 214 translates the set of tasks to a set of work orders in human-readable form. Lastly, work order module 214 facilitates tracking of those work orders to completion. To accomplish these objectives, work order module 214 may expose various UIs and system interfaces to enable communication and interoperation with external entities.
  • In one embodiment, work order module 214 exposes a system interface through which integration module 210 may supply any changes made to the data center configuration. Work order module 214 may also expose a user interface including elements that enable a user, such as a data center manager, to modify and dispatch work orders to other users, such as technicians, for implementation. Work order module 214 may also expose a user interface to enable a user to configure a set of automatic dispatching rules. Further, work order module 214 may expose other user interfaces that enable users, such as technicians, to modify the details, including status information, of work orders. These user interfaces may reside on various computing devices, including remote or mobile computing devices.
  • Implementing such a user interface on a remote or mobile computing device may allow users, such as technicians, to update the CMDB of the data center as configuration changes are made. This may produce several benefits including increased accuracy of the CMDB, due to the increased compliance with, and promptness of, configuration change updates. Another benefit may be increased productivity for users, such as technicians, because configuration changes may be entered while working on data center equipment, rather than entering changes at a workstation as a separate activity.
  • Work order module 214 may also implement modification and assignment system interfaces to interoperate with other systems. For example, work order module 214 may, through a system interface, utilize an external email system to notify users of a work order assignments. Lastly, work order module 214 may utilize a system interface that enables it to detect changes in data center resource demands and may use this information, when appropriate, to modify the status information of a work order.
  • A flow chart of a method 300 in accordance with one embodiment that may be performed using the system 200 will now be described with reference to FIG. 4. Initially, and optionally, at stage 301 of the method 300, a user may create groups of data describing data center equipment. These groups may enable management of multiple types and/or models of data center equipment as a single collective entity. For example, inclusion of data describing data center equipment in a group may designate such equipment as approved for use within all of the data centers owned by a customer or specific subsets thereof. Any data center equipment approved for use within a data center may be used during the design or retrofit of the data center.
  • Moreover, the user may arrange these groups in a hierarchical fashion. In one embodiment, a first group is created that includes all of the data center equipment supported by the system, a second group is created that is a subset of the first group and that designates equipment that is approved for use within the data centers of a particular customer, and a third group is created that is a subset of the second group and that designates equipment that is preferred for use within a particular customer data center. In one embodiment, a user may target specific groups for exportation to specific customers, customer data centers or specific computing devices within a customer data center, such as remote or mobile computing devices.
  • Fashioning groups of equipment data may ease maintenance and distribution of electronic data center equipment catalogs, which may include data describing characteristics of data center equipment. For example, specific inclusion of data center equipment data in a group that is lower in the hierarchy, e.g. a group designated for distribution to a remote or mobile computing device within a specific data center, may require that the data be present in ancestor groups. This inclusion may be performed automatically by the system. These groups may also be employed to enforce standardization of data center equipment within the customer organizational hierarchy by requiring that equipment be approved for use by the customer before it is allowed to be used to retrofit a specific customer data center.
  • In another embodiment, the system provider may create a group of data describing standard data center equipment. A group of the standard data center equipment data may be approved for use in customer data centers. An equipment catalog including data associated with the standard and/or approved data center equipment may be delivered with the initial system installation in a data center. The standard and/or approved equipment data may then be supplemented with customer specific, non-standard equipment data used in a customer data center, as discussed with regard to stage 304 below.
  • Approved equipment may be further grouped into a preferred set for use with one or more customer data centers. The creation of these preferred groups may be explicitly performed by the user or may be implicitly performed by the system based on usage of specific equipment by a customer or within a customer data center during data center design or retrofit.
  • In stage 302 of the method 300, information regarding the facility is loaded into the system. The information includes, for example, dimensions of the facility, number of rooms, locations of doors, support columns, other obstructions, parameters of data center resource capabilities, such as available power, cooling capabilities of the facility, whether a raised floor or drop ceiling is in use, and characteristics of any such floor and roof. Data center resource supply policies may also be entered in this stage. For data center resource providing equipment, such as electrical generators or CRACs, capability information may be loaded by receiving data center equipment information from interface 302 of system 300. Thus, this stage of method 300 collects information helpful in determining the data center resources provided to particular data center locations and spaces. In at least one embodiment, as discussed further below with regard to cooling, this information may be further processed to determine the data center resources provided to particular data center locations and spaces.
  • In stage 304 of the method, information regarding equipment to be installed in the facility is entered. The information includes, for example, the number of racks of equipment, maximum power draw for each of the racks, dimensions of the racks, and cooling requirements for the racks. The need for backup power sources and multiple power sources for equipment and or racks may also be entered at stage 304. In one embodiment, data center resource consumption and production characteristics of individual pieces of equipment that are to be loaded into racks may also be entered. Also, the weight of equipment (including equipment loaded into racks) may be used to ensure that the weight of the installed equipment is within any facility constraints. These characteristics may include, in addition to network connectivity, power and cooling requirements, the amount of rack space that the equipment needs to occupy and the type and/or number of electrical plugs that the equipment requires. In one embodiment, the database module 212 contains information regarding a number of devices, such as uninterruptible power supplies, equipment racks, cooling units, generator systems, power distribution units, automatic transfer switches, electrical routing devices, including cables, and servers and other computer equipment. In this embodiment, when a particular model number of a device is entered, characteristics of the device are retrieved from the database module. In one embodiment, interface 302 of system 300 provides these device/equipment characteristics to database module 212. Equipment related to fire protection and security may also be included in the design. Further, in at least one version, all equipment and components within equipment racks may include RFID tags, which can be used by systems of the invention to track location of equipment and racks. In another embodiment, any data center equipment characteristics that are added or changed by a user at this or the previous stage may be transmitted to interface 302 of system 300 for importation into database 304.
  • Once all of the information is entered into the system, at stage 306, the system in one embodiment determines a layout for the equipment in the facility, taking into account the data center resource requirements, such as power and cooling requirements, of the equipment as well as other characteristics of the equipment that were entered at stage 304 or retrieved from the database module. In another embodiment, the user may create the layout graphically, adding racks and other equipment where desired, and in this embodiment, the system will provide feedback during the layout process, disallowing some choices and making intelligent suggestions. These rules may include, for example: a standard alternating hot aisle/cold aisle layout must be specified, the plenum must be greater than some minimum value, the total room cooling capacity must exceed total room cooling load, aisles must be wide enough for access purposes and to meet building codes, distance between PDU and IT racks served by the PDU must not exceed some maximum value, PDU must be located immediately adjacent to a UPS, where a cable ladder spans an aisle, the aisle cannot exceed a maximum width, one or more data center resource capacities must be at a level sufficient to support target data center resource redundancy and/or runtime requirements, etc.
  • Next, at stage 308, a cooling analysis is conducted to determine if the design provides adequate cooling for each of the racks and the equipment installed in the racks. As described further below, in different embodiments of the present invention one of a number of different methods may be used to conduct the cooling analysis. In one embodiment, if the results of the cooling analysis indicate that one or more devices and/or racks are not receiving adequate cool air, then the procedure may return to stage 306 to change the layout of the equipment based on feedback provided from the analysis conducted at stage 308.
  • At the completion of the cooling analysis, at stage 310, a room model is displayed showing the locations of the equipment in one or more rooms of the facility. The room model may include, for each equipment rack, or for individual data center equipment, information regarding the total data center resources, such as power and cooling, being consumed or produced as well as an indication of total available data center resources, such as power and cooling, to the rack or data center equipment. In one embodiment actual data center resource data, such as power and cooling data, may be displayed, while in other embodiments colors may be used, either alone or in combination with data, to display different levels of data center resource availability, such as power and cooling availability. For example, if a rack is operating with sufficient cooling air with a margin above a threshold, the rack may be indicated in green on the display, if the cooling air availability is closer to the threshold, the rack may be indicated in yellow, and if the rack does not have sufficient cooling air it may be indicated in red. Still further, the results of the analysis may indicate that adequate data center resources, such as power and/or cooling, are being provided for equipment, but that specified redundancy levels and/or runtime margins are not being met, either at the room level, a row level, a rack level, or at a specific piece/element of data center equipment.
  • In one embodiment, the system may display multiple room models and may allow the user to find one or a series of alternate satisfactory locations for the equipment based on the data center resource requirements of the equipment and any applicable redundancy and/or runtime requirements. Another embodiment may allow the user to find an optimum location followed by a series of alternate satisfactory locations for the equipment. Each series of locations may be arranged consecutively according to an order of preference, such as in order of decreasing excess or stranded capacity. As discussed further below, stranded capacity includes excess capacity that is nominally available, but unusable, due to insufficient associated capacity of another data center resource required by data center equipment.
  • Yet another embodiment may allow the user to specify both the data center equipment and location and may validate whether the location provides sufficient data center resources to satisfy the requirements of the equipment and any applicable data center supply policies. This validation may be performed by comparing the data center resources provided to the location, the data center requirements of the equipment and any applicable data center resource supply policies. The result of this comparison, a compliance result, may be stored for later use. In still another embodiment, the system may suggest data center equipment to be placed at a user specified location. In this case, the system may ensure that applicable data center resource supply policies, and data center resource requirements of the suggested equipment, are met by comparing the polices, the data center resources provided to the location and the data center resource requirements of the equipment prior to suggesting the equipment based on a compliance result. Specific example details regarding the room model are described further below with reference to FIGS. 5 through 5D.
  • At decision block 312, a determination may be made by, for example, a facility designer as to whether the layout generated in stage 310 is satisfactory. The determination may be based on additional criteria of importance to the designer that was not included during the design of the original layout. For example, it may be desirable to have certain racks near each other or to have certain racks isolated from one another. At stage 314, additional criteria or other feedback can be provided and the process then returns to stages 306 and 308 where the room model can be refined. Stages 306 to 312 may be repeated until a satisfactory model is achieved at stage 312. In at least one embodiment, at the completion of the design stage, a bill of materials is generated and may be used to provide the cost of the equipment to be installed in the facility and may also be used to generate a sales order for the equipment, providing a simple solution for ordering all equipment associated with a new data center. Further, CAD drawings and electronic files that capture the designed layout may also be generated. In another embodiment, this data center configuration is transmitted to interface 302 of system 300 for storage in database 304 in the form of a CMDB for the installed data center.
  • At stage 316, the equipment is installed in the facility according to the layout generated at stages 306 to 314. In one embodiment, measurement equipment to measure cooling characteristics and power characteristics, may be installed with the equipment. The measurement equipment is described further below, and may include, for example, devices for measuring power, airflow, humidity and temperature at various locations in the facility and within equipment racks located in the facility.
  • At stage 318 of the process 300, power and cooling parameters are measured using the measurement equipment. Additional temperature measurements may also be provided by devices, such as servers, that have the capability to detect internal temperatures. The parameters measured may be used continuously by the management module of the system 200 to detect error conditions and to monitor trends that may lead to an error condition. Further, in the process 300, the measured parameters can be compared with predicted parameters calculated during the design process in stages 306 and 308. For example, in one embodiment, the airflow through a perforated floor tile of a raised floor is used to determine the available cooling air of a rack located adjacent the floor tile. The airflow through the perforated tile may be determined in stage 308 using one of a number of computational methods that are described further below, or the airflow may be determined using data from related physical measurements or simulations. Once the equipment is installed in the facility, the perforated floor tile may be instrumented to measure the actual airflow through the tile. The actual measured value may then be compared with the calculated value at stage 320. If the two differ by more than a predetermined threshold, then an indication or warning may be provided and the calculations conducted in stage 308 may be conducted once again at stage 322 using measured values in place of calculated values as appropriate to obtain updated parameters. In another embodiment, measured parameters are transmitted to interface 302 of system 300 for storage in database 304. As discussed above, the storage of these measure parameters by interface 302 may trigger further analysis and summarization of the measure parameters into data center equipment consumption and production benchmarks.
  • After stage 322, the model of the facility described above with reference to stage 310 may be displayed with values of power and cooling availability and consumption updated to reflect any differences between measured parameters and calculated parameters. Any out of tolerance conditions (for either cooling or power) may be indicated on the display using, for example, a color coded scheme as described above. In one embodiment, a user may be provided with a number of available options to correct an out of tolerance condition. The options may include upgrading or adding facility equipment (i.e., an air conditioning unit or an uninterruptible power supply) or may include moving equipment and/or racks. Stages 318 to 322 of the process may be performed continuously as part of a management system of the data facility.
  • In one embodiment of the invention, stages 302 to 314 of the process 300 are implemented using a build-out system accessible by a user over the Internet. In this embodiment, the user provides the requested information, and the build-out system provides the processing described above, provides outputs to the user over the Internet, and stores results locally. After the equipment has been installed in the facility, the management system 500 (described below) may access the build-out system to download information related to the equipment. In addition, when a retrofit of the facility is to occur, the management system may contact the build-out system to coordinate the design of the retrofit. In at least one embodiment, electronic files may be imported/exported between the systems to provide a complete transfer of all information related to a data center's design.
  • FIG. 5 shows an example of a display of a room model that may be generated using the system 200 and the process 300 and shown on a computer display. It should be appreciated that this computer display may be any be coupled to, or included in, any sort of computing device including a remote or mobile computing device. The room model shown in FIG. 5 is essentially the data center 100 previously discussed above with reference to FIGS. 1 and 2. However, in FIG. 5, additional data related to the power and cooling consumption and capacity of each rack, U space positions, and/or elements of data center equipment housed within each rack may be included in an informational block, such as informational blocks 120A and 120B shown on two of the racks 108A and 108B in FIG. 5. Informational blocks may be included on data center equipment, each rack, or on less than all racks, for example, by row, zone, or cluster.
  • FIGS. 5A and 5B show enlarged views of respectively racks 108A and 108B that may also be shown on a computer display of systems of embodiments of the invention. In the views of FIGS. 5A and 5B, specific information regarding the racks and U space positions is included in the informational block. In the embodiment shown, the information in the block includes a rack identifier 122, a rack type 124, power capacity 126, power usage 128, cooling capacity 130, cooling usage 132, contents of the rack 134, power redundancy 136, cooling redundancy 138, and UPS runtime 140. In other embodiments, such as the embodiment depicted in FIG. 20 discussed below, information for each rack may be included in tabular form, or in the form of columnar bars, on a graphical display showing the room layout. Data center information that may be displayed includes runtime measurements, data center resource redundancy measurements, including relationships between data center equipment working in combination to providing redundancy, and capacity measurements, such as available capacity, utilized capacity, and stranded capacity.
  • The rack identifier 122 includes a row number and a rack number, however, in other embodiments, the rack identifier may also include an indicator of the type of rack, membership of the rack to a particular row, zone, group, or cluster, manufacturer of the rack, as well as other information. The rack type 124 identifies the particular type of rack, i.e., server rack, router rack or telecommunications rack. The power capacity 126 indicates the maximum power capacity of the rack, and the power usage indicator 128 indicates the percentage of maximum capacity at which the rack is expected to operate. In different embodiments, the power usage indicator may be calculated based on manufacturer supplied data for equipment contained in the rack and/or based on actual power measurements of equipment. The power capacity for a rack, in at least one embodiment, is determined based on limitations of devices and/or power cables that supply power to the rack, such as circuit breakers, UPSs or any other devices. The contents of the rack 134 includes a list of the equipment contained in the rack and may include an indication of remaining space in the rack displayed, for example, in terms of rack units, which are typically referred to as “U” with 1 U equal to 1.75 inches. Details regarding the equipment in the rack, including operational status and network addresses, such as an IP address for a device may also be included.
  • The cooling capacity indicator 130 and cooling usage indicator 132 identify respectively the quantity of cooling air available to the rack and the percentage of that cooling air that is being used by equipment in the rack. In other embodiments power and cooling usage may be indicated using various types of graphical gauges, such as a bar graph, that indicates power and cooling usage and capacity. In the embodiment shown in FIGS. 5A and 5B, the cooling capacity is shown in terms of kilowatts (kW). As known to those skilled in the art, for typical data center applications, many equipment racks typically require approximately 160 cfm (cubic feet per minute) of cooling air per kilowatt of power used by the rack. All the power consumed by computing type devices is typically converted to heat, such that the required cooling (in terms of kW) for a rack can be assumed to be equal to the power consumption of the rack. Accordingly, in one embodiment, the cooling usage indicator is equal to the power consumed by the rack. In other embodiments, depending on the type of equipment that is installed in the racks, the cooling required by a rack may not be equal to that consumed by the rack and may be calculated based on manufacturer's data for the equipment, based on test results, or in any other manner.
  • The cooling capacity of an equipment rack is determined based on a number of different factors. For example, for a raised-floor environment, these factors may include: location of the rack in the facility, proximity of perforated tiles to the rack, the amount and temperature of cooling air provided through any such perforated tile, the physical or geometric layout of the racks and building, and the cooling requirements of other equipment in the facility located near the rack. The cooling capacity of one rack in a facility may be affected by the cooling usage of nearby racks, and accordingly, in one embodiment, the cooling capacity of a rack is adjusted when the cooling usage of a nearby rack is changed. In at least one embodiment of the present invention, calculations for determining cooling capacity of a rack are based in part on the ability of one rack to borrow cooling air available to adjacent racks. Particular methods for determining cooling capacity for racks in embodiments of the present invention are described further below. In one embodiment, when the cooling usage of one rack is changed, the cooling capacity of that rack, and all racks located near the changed rack is recalculated.
  • In embodiments of the present invention, during the design as well as during the management of a data center, the true available capacity of a data center can be determined at U space position level, the rack level, at the row level and at the room level. In determining available capacity (including unused capacity), data center resources including both cooling and power capacity are used, and the true available capacity can be determined using the lowest data center resource capacity. In situations where data center resource capacities are not equal, the excess capacity can be considered wasted or stranded capacity that can not be used in the present design. In other words, while stranded capacity is nominally available, it is unusable due to insufficient associated capacity. In embodiments of the present invention, the stranded capacity can be determined at the U space position level or rack level and can be totaled to determine stranded capacity at the row level and at the room level.
  • In one embodiment, as depicted in FIG. 21, this data resource information may be displayed to the user in the form of graphs. For example, FIG. 21 depicts total capacity 2102, space capacity 2104, cooling capacity 2106 and power capacity 2108, each as a function of time. Each of these representations may include total data center capacity specific to the data center resource shown. In addition, the representations may include data center capacity specific to a specific location in the data center, e.g. high density or medium density racks. Still further, these capacities may be projected into the future to enable a user, such as a data center manager, to predict when capacity may need expansion.
  • FIG. 23 illustrates an embodiment that a user, such as a data center manager, may reference to assess stranded data center capacity and ascertain the root cause of the stranded capacity. More specifically, FIG. 23 depicts report 2302 which includes rows and columns under various data center resource headings. The intersection of these rows and columns may indicate how much excess capacity of the data resource listed in column 2304 is unusable due to insufficient capacity of the data resource listed in row 2306. For example, intersection 2308 may indicate that 20% of the weight support capacity of the data center is unusable due to insufficient power capacity. In another embodiment, the actual amount of stranded capacity may be used to articulate these relationships and the meaning ascribed to the rows and columns may be transposed. As will be apparent to one of ordinary skill in the art, other embodiments illustrating this relationship are possible without departing from the present invention.
  • Recommendations are provided for reducing stranded capacity during the design phase as well as during the management phase. The recommendations may include reducing capacity of data center resources, such as power and cooling resources, (thereby reducing operational costs) or adding data center resource capacity, such as cooling capacity or power capacity, appropriately to reduce the stranded capacity. Warnings may be generated when the stranded capacity is greater than preset thresholds, and in addition, recommendations for more optimal locations of equipment, including power and cooling equipment, may also be provided to minimize the amount of stranded capacity. Further, costs of the stranded capacity may be calculated.
  • In management systems and methods of embodiments of the invention, as described above, data center resource capacity and availability, such as power and cooling capacity and availability, may be monitored in real time. In one version, changes to the availability rate (or the utilization rate) are monitored and based on these changes, the growth rate of a data center may be determined, and predictions of dates when additional capacity will be required can be provided. In one embodiment the forecasting method used to make these determinations is linear regression. Other forecasting methods are possible without departing from the scope of the present invention. The ability to monitor capacity and predict future capacity needs allows data center operators to control costs and plan for upcoming expenses. Further, determinations may be made as to the additional expenses that will be incurred if new equipment is added. The total cost (for example per kilowatt) can also be determined during the design phase or during operation.
  • In embodiments of the invention described herein, data center layouts may be designed to provide specific redundancy levels (i.e., n, n+1 or 2 n) for data center resource design, such as both the power design and the cooling design. In data centers in the past, additional room cooling units are typically provided to include some redundancy in a data center, such that an overall cooling capacity of the data center can be maintained, even when one or more of the room cooling units fails or must be powered down to conduct maintenance. One problem with these past solutions is that the cooling redundancy is designed at the room level and not the rack level, and while overall cooling capacity may meet redundancy requirements, cooling at the rack level may not meet the desired redundancy requirements. In embodiments described herein, the ability to provide accurate airflow data at the rack level and U space position level allows true cooling redundancy to be designed into a solution.
  • As discussed above, graphical user interfaces may be used with embodiments of the present invention to assist in the design and management of data centers. In one embodiment, these graphical user interfaces may be used to present an interactive representation of a layout of a various spaces in the data center. In another embodiment, interactive representations may be tailored to particular spaces within the data center, such as the entire data center, data center floors, rooms, racks and U spaces. Further these interfaces may be displayed on a local or remote computing device. Remote computing devices may include general computer systems, mobile computing devices and computing devices that are included with, embedded in, or affixed to, another physical structure or device, such as a wall, ceiling, another computer system or an enclosure, e.g. a rack enclosure, etc.
  • For example, particular user interface screens used in one embodiment to design a layout in a data center will now be described further with reference to FIGS. 5C and 5D. FIG. 5C shows a floor editor interface 402 used in one embodiment to layout equipment in a data center, while FIG. 5D shows a rack editor interface 404 used in one embodiment to provide further details of the contents of equipment in the data center. In one embodiment of a data center design system, tutorials are provided for a user to assist the user by providing best practice design guidelines. The tutorials may be accessed by a user as desired or may be configured to be displayed as a user is taking a particular action.
  • The floor editor interface includes a main menu 403, a tool bar 406, a configuration box 408, a generic components box 410, a floor layout box 412, a status box 414 a full-image viewing box 416, and an unplaced equipment box 418. The main menu 403 provides a drop-down menu in a format known to those skilled in the art, and allows a user to perform various functions, including the ability to “undo” and/or “redo” changes that are made to the layout. The tool bar 406 provides short hand access to functions of the design system, and in one embodiment includes a floor editor button 406A and a rack editor button 406B. Activation of the floor editor button results in the display of the screen shown in FIG. 5C, while activation of the rack editor button results in display of the screen shown in FIG. 5D.
  • The floor editor box 412 shows the layout of equipment in a data center being designed and provides text that identifies the equipment contained in the layout. A room perimeter 412A shows the exterior walls of the room along with dimensions of the room that can be set by a user. In one embodiment, when a new design is started, the user is presented with a screen showing a number of basic room configurations that can be selected. Further, the walls of the room can be moved by a user by selecting one of buttons 412B, and additional buttons can be added where needed to expand or shrink an area of the room. In one embodiment, the room size may be changed as equipment is added to the layout. Three rows 412C, 412D and 412E are outlined in the room shown in FIG. 5C. In other embodiments, more or less rows may be included. As shown in FIG. 5C, the rows are configured in a manner to provide alternating hot and cold aisles. Row 412D includes three equipment racks (identified by “R”), two half-rack cooling units (identified by “C”) a UPS (“U”) and a power distribution unit (“P”). Row 412E includes one rack, and row 412C as presently configured does not include any equipment. During the design phase additional equipment may be added to each of the rows. The room also includes an automatic transfer switch (ATS) 412G and a cooling distribution unit (CDU) 412F. Hatched areas are shown on the display around the ATS and CDU to indicate that these areas should be kept clear of equipment. Each piece of equipment in the room may include identifiers that indicate the type of rack as well as the rack's location in the room and the power source for the rack. Further, as discussed above, each rack may include information regarding data center resource use and availability, such as power and cooling use and availability. Still further, text may be provided on each row to indicate total data center resource information, such as power and cooling information for each row.
  • The configuration box 408 includes eight configuration options for designing a data center. A room properties configuration option, when selected, allows a user to identify data center resource values, such as physical, power, and cooling values, that affect the data center design as a whole including dimensions of the room, aisle widths and total anticipated power density for the data center. Data center resource redundancy and/or runtime requirements, such as power redundancy requirements (i.e. N, N+1, 2N), cooling redundancy requirements and runtime requirements for UPS systems may also be set. The number of data troughs that will be used and location of power distribution and cooling line distribution (i.e. overhead or under a raised floor) can also be configured. In one embodiment, only in-row cooling is provided, however, in other embodiments other types of cooling solutions may be used as well. In at least one embodiment, individual rows may be rotated to different angles in the data center. Further, while only one room is shown in FIG. 5C, at least one embodiment allows a data center to include multiple rooms. These rooms may be active rooms, which hold active data center equipment, and inactive rooms for storage of spare or decommissioned equipment.
  • An add rack configuration option in the configurations box 408 is used to add equipment racks to the data center design. When this option is selected, a user is presented with choices of various types of racks to add to the data center. When racks are selected, an indicator is provided in the unplaced equipment box 418, indicating that the racks still need to be placed into the room layout.
  • An add in-row cooling option in the configuration box is used to add in-row cooling units to the data center design. When this option is selected, a user is presented with various types of cooling units that can be added in the rows. As with equipment racks, when a cooling unit is selected, an indicator is provided in the unplaced equipment box 418, indicating that the cooling unit still needs to be placed in the room layout.
  • A power zone option in the configuration box is used to identify and select PDUs and UPSs and to indicate which equipment will be powered from the UPSs and PDUs. Characteristics of the PDUs and UPSs may also be selected. Once selected, an indicator is provided in the unplaced equipment box 418 for the UPSs and PDUs. In one embodiment, multiple racks may be included in a selection on the layout to identify the equipment that belongs to a particular power group, also known as power zone. In still another embodiment, after selection of equipment and UPSs and PDUs, an automatic power zone option may be implemented in which the system matches equipment power requirements (i.e., redundancy levels, runtime durations, voltages, phasing) to those of the UPSs and PDUs and assigns power zones automatically and determines lengths of power cables that are needed to power equipment from the assigned PDU. In automatically determining power zones, the system may also identify the need for additional UPSs and PDUs.
  • A power generation option in the configuration box 408 is used to identify and select an automatic transfer switch (ATS) and generator. Again, once these are selected, an indicator is provided in the unplaced equipment box 418.
  • An emergency power off option in the configuration box is used to select an emergency power off (EPO) solution for the data center design, and once selected, an indicator for the EPO solution will be added in the unplaced equipment box.
  • A management option in the configuration box 408 allows a data center manager, such as the InfrastruXure® Manager and/or InfrastruXure® Central discussed above, to be added. In one embodiment, when selecting the manager, a rack location for the manager is also selected.
  • A service option in the configuration box 408 allows a user to select a level of service to be provided to the data center by a data center services organization.
  • Other configuration options may include a row planning configurator that allows a user to plan how many racks a row can support by defining the power and cooling settings for the row, prior to placing equipment in a row. In one embodiment, the row planning configurator may provide an estimate on the number of racks that can be supported based on the power components and in-row cooling units contained in the row. In one embodiment, the row planning configurator may provide a complete layout based on design best practices.
  • The generic components box 410 includes a number of icons to designate pre-existing equipment in a data center. The components may be selected and “dragged” into position in the layout. In one embodiment, the generic components include blocks and gaps. The gaps can be used to identify areas over which cables and conduits can be routed (i.e. a walkway), while the blocks are used to identify areas over which cables and conduits can not be routed (i.e. a column). Once dragged onto the layout, the blocks and gaps can be sized appropriately.
  • As discussed above, when equipment is selected for use in the data center, an icon appears in the unplaced equipment box 418. To place the equipment in the layout, the icon is selected and dragged into the appropriate location in the layout. Existing equipment may be repositioned using this same method. For example, existing equipment may be dragged from an active room and dropped in an inactive storage room, thus enabling the system to track unutilized equipment available for provisioning elsewhere. In one embodiment, when adding an in-row cooling unit, the icon for the cooling unit can be placed between two adjacent racks and released, and the racks will then move in the row to allow the cooling unit to be inserted between the racks. Further, in one embodiment, a snap-to feature is employed to align equipment properly in rows and along walls, and in addition, rows and equipment may be aligned along and “snapped to” floor tiles when, for example, a raised floor is in use. Using this feature, a user does not need to precisely align objects.
  • The full-image viewing box 416 provides a “bird's eye” view of the layout contained in the floor layout box 412. In one embodiment, the zoom button on the tool bar can be used to enlarge the view of the data center layout in the floor layout box 412. When the view is enlarged, the entire layout may not appear in the floor layout box. The full-image box 416 still displays the full image of the layout for the user. In one embodiment, when the full layout does not appear in the floor layout box, an overlay is used in the full-image box to indicate on the full-image, the portion of the layout that is displayed in the floor layout box. In one embodiment, when the full image is not displayed in the floor layout box 412, the overlay may be selected and dragged within the full-image viewing box to select which part of the layout is displayed in the floor layout box.
  • The status box 414 is used to display warnings, errors and other conditions to the user. The warnings may vary in severity and may include indications that design guidelines are being violated and may also include more severe warnings indicating that data center resource capacities, such as power and cooling capacities, have been exceeded, or that redundancy and/or runtime requirements are no longer met. In one embodiment, when the status box indicates that there is an error or warning associated with a particular piece of equipment in the layout, the piece of equipment may be highlighted with a color such as red or yellow. In at least one embodiment, when an error or warning occurs, guidelines for correcting the error or warning are provided by either selecting a highlighted piece of equipment or the error or warning message directly.
  • The rack editor interface 404 will now be described further with reference to FIG. 5D. The rack editor interface includes the tool bar 406, the status box 414 and the full-image viewing box 416 discussed above. Further, the rack editor interface 404 also includes a rack editor box, a product catalog box 422 and a rack content section 424.
  • The rack editor box 420 shows the front face of each of the equipment racks in the data center layout with the racks being arranged by row. In FIG. 5, two rows of racks 420A and 420B are shown. As shown in FIG. 5, in one embodiment, only the equipment racks are shown in the rack editor box. When a particular rack is selected in the rack editor box, then the contents of the rack appear in the rack content box 424, and components may be added to the selected rack. Racks may be selected in the rack editor box or may also be selected in the full-image view box 416. When a rack is selected in the full-image view box, then the image in the rack editor box will shift, if necessary, to provide a view that includes the selected rack.
  • The product catalog box 422 provides a comprehensive listing of components that may be used in equipment racks in data centers. The user may select equipment to be included in each rack, and as equipment is selected, it is included in the rack content box 424. The list may include only equipment of a particular manufacturer or may include equipment from several manufacturers. In one embodiment, all necessary hardware and cabling associated with rack equipment may be selected from the product catalog box.
  • In one embodiment depicted in FIG. 20, a user may review and manage data center resource capacity and available supplied to data center equipment. Many of the features of this embodiment are similar to those described in reference to FIG. 5C above. In addition, FIG. 20 includes capacity group explorer 2002, which presents a logical grouping of data center equipment, such as racks, based on common capacity requirements. This grouping enables a user, such as a data center designer, to manage sets of data center equipment as one collective entity for capacity planning purposes. Representations of individual elements of data center equipment, such as UPS 2004, Rack 2006, and PDU 2008 may present the user with columnar bars depicting various data center resource redundancy and/or runtime measurements and capacity measurements, such as available capacity, utilized capacity, stranded capacity. For example, Rack 2006 has an as-configured potential power and cooling demand of 28.8 kW and a current actual power and cooling demand of 7.92 kW. In other embodiments, these demand or consumption measurements may be matched against as-configured potential power and cooling supply capacity and actual power and cooling capacity to ensure all capacity, consumption, redundancy and runtime requirements are met.
  • In another embodiment depicted in FIG. 22, rack editor interface 404 may be displayed on a remote or mobile computing device. Rack editor interface 404 includes user interface elements that allow for adding equipment 2202, deleting equipment 2204, editing equipment 2206 and moving equipment 2208. This embodiment further includes data center explorer 2210 and rack editor box 2212. Because rack editor interface 404 may be provided on a remote or mobile computing device, users, such as technicians, may document as-built conditions, verify rack installations, troubleshoot installation in a more efficient manner and perform other rack-oriented activities.
  • In the depicted embodiment, equipment racks are shown in data center explorer 2210. The rack editor box 2212 shows the front face of the equipment rack currently selected in data center explorer 2210. In another embodiment, rack editor box 2212 may display the rear face of the rack and/or the front face of the rack. In FIG. 22, rack Front is selected in data center explorer 2210 and houses data center equipment sss, PowerEdge 2650, PowerEdge 6650, etc. When a particular rack is selected in data center explorer 2210, the equipment it houses may be modified using user interface elements 2202 through 2208.
  • The user may add data center equipment to the currently selected rack by actuating user element 2202 and selecting the desired equipment and position with the rack. The user may delete or edit data center equipment associated with the currently selected rack by selecting the desired equipment from data center explorer 2210 and actuating user element 2204 or 2206, respectively. The user may move data center equipment associated with the currently selected rack by selecting the desired equipment from data center explorer 2210, actuating user element 2208 and selecting the desired position within the rack. In another embodiment, the system may recommend a satisfactory or optimum position. It should be appreciated that the user may select particular data center equipment by drilling down in data center explorer 2210 or by searching the data center using a unique identifier, such as a barcode scanned by the remote or mobile computing device. These methods of searching may allow a user, such as a technician, to quickly obtain information concerning specific data center equipment or modify the CMDB while physically located near the equipment.
  • According to other embodiments, an interactive representation of an identified space within a data center may be implemented using other interface technology. For example, in another embodiment, in addition to the graphical user interface screens shown above, a three-dimensional option is available allowing a user to view the design of a data center in 3D. In one embodiment, a design system includes software code programmed in Java that is used to generate 3D models that are rendered via OpenGL to allow for hardware acceleration. Further, 3D models may be exported from the design system to CAD tools such as AutoCAD, available from AutoDesk of San Rafael, Calif. As described above, security cameras can be implemented into data centers designed using embodiments of the present invention. In one version that has 3D viewing capabilities, security cameras may be included in the design and the 3D view may be used to view a simulation of a camera's view after installation. In one embodiment, networked cameras and other security monitoring devices available from Netbotz Corporation of Austin, Tex. may be used.
  • As discussed above, with reference to the process shown in FIG. 4, the system 200, and other systems of the present invention, may be used as part of a data center management system. The management system may include the system 200 described above with the management module containing routines to perform management functions, or in other embodiments, the management functions may be performed by a designated manager controller contained in the data center and implemented, for example, in a computer server located in one of the racks of equipment and accessible by a user using a management console.
  • FIG. 6 shows a block diagram of a management system 500 that may be used in embodiments of the present invention. The management system includes the manager controller 502, the manager console 504, power measurement devices 506, and airflow, humidity and temperature measurement devices 508. It should be appreciated that manager console 504 may be implemented as any computing device, including a remote or mobile computing device. Implementing manager console 504 on a remote or mobile computing device may allow a user, such as a data center manager, to, for example, dispatch work orders to technicians while physically inspecting, installing, moving, and/or changing data center equipment. In addition, in one embodiment, the management system may include power control devices 520 to control application of power to one or more individual devices or racks contained within a data center, and the system may include airflow controllers 521 to control the airflow or supply temperature of an air conditioning unit or to control, for example, perforated tile dampers. As discussed above, the management system may also include one or more security devices 523, including security cameras. The devices of the management system 500 may be directly coupled to the manager controller or may be coupled to the manager controller using a network 522 that may be a dedicated network, may include the Internet, or may include a LAN or WAN contained in the data center. The manager controller may communicate with one or more servers 524 to obtain information from and control operation of the servers.
  • In one embodiment, the management controller 502 may be implemented at least in part using an InfrastruXure® Manager and/or InfrastruXure® Central available from American Power Conversion Corporation (APC) of West Kingston, R.I., and devices may be coupled to the manager using, for example a controller area network (CAN) bus or an Ethernet network. The power controllers and airflow controllers may be implemented using available known devices that monitor and/or control power and airflow in facilities. Further, in at least one embodiment, the management system 500 may include systems and methods for monitoring and controlling power as described in U.S. Pat. No. 6,721,672 to Spitaels et al, which is incorporated by reference herein. Further, in at least one embodiment that uses in-row cooling devices, the management controller may communicate with the cooling units to control the units to ensure that adequate cooling at specified redundancy levels is being met. Further details regarding the control of in-row cooling units that can be used with embodiments of the invention are discussed in copending U.S. patent application Ser. No. 11/335,901, discussed above and filed on Jan. 19, 2006.
  • One aspect of the present invention, which will now be described, is directed to a retrofit system and method that is particularly useful for adding new equipment to a facility. The addition of new equipment may include adding equipment to existing racks or may include adding other data center equipment, such as racks or other floor mounted equipment, to a facility. The retrofit system may be a standalone computer system configured to perform processes described herein, or in one embodiment, the retrofit system is implemented using the system 200 described above. Specifically, the builder module 206 of the system 200 may include routines to assist in retrofitting a data center. A process 600 for using the system 200 (or some other system) to retrofit or upgrade a data center will now be described with reference to FIG. 7, which shows a flow chart of the process 600.
  • In a first stage 602 of the process 600, data related to a present configuration of a data center to be retrofitted is provided to the builder module. The data related to the present configuration may include the data displayed in the room model of FIG. 5 along with additional data that was generated during design of the data center. Further, in one embodiment, the data related to the present configuration may include data generated during an initial design as updated by actual measurements conducted in a facility. For example, the cooling capacity of individual racks may be calculated in an initial design and then updated by the management module once the system is installed and operating. Cooling capacity data may be updated based on actual measurements of airflow from, for example, perforated floor tiles, while the original data may have been calculated based on predicted airflows. In one embodiment, interface 302 of system 300 provides this data center configuration and measured parameter information from database 304.
  • Information related to the retrofit is then provided in stage 604. The information related to the retrofit may include information similar to that input at stage 304 of process 300 described above, such as type of equipment, characteristics of equipment, number of racks, as well as other information. In addition, a user may designate one or more desired locations in the data center for the installation of new equipment. For example, a user may desire to add five additional servers to the data center, where the servers are similar to and have a related function to existing servers in the data center. The user may choose one or more preferred locations based on power specifications, cooling specifications and physical dimensions of the servers, and based on power capacity, power plug type and/or number, cooling capacity and contents of existing racks displayed on a floor model of the data center. In addition, the user may indicate whether it is acceptable to move existing equipment to accommodate the installation of new equipment. In another example, a user may desire to replace three servers in the data center. The user may choose the servers targeted for replacement and may make other data center modifications with these replacement servers in mind. In addition, the user may indicate whether the replaced equipment should be removed from the data center or moved into an inactive storage room. Tracking inactive equipment may allow a user, such as a data center manager, to quickly ascertain equipment available for provisioning within the data center.
  • At stage 606, an updated layout for the data center is generated and cooling and power calculations are performed at stage 608 on the updated layout in the manner discussed above at stage 308 of process 300. If the user has designated specific locations for equipment in the data center, the layout may first be determined using these locations, and if problems arise as a result of the desired layout (i.e., lack of cooling for a rack based on equipment or supply policy requirements), then the user will be able to note any such problems once the layout is displayed and can then choose to change the layout. In one embodiment, the system may suggest one or more elements of data center equipment to be placed at one or more locations. In this case, the system may ensure that applicable data center resource supply policies, and data center resource requirements of the suggested equipment, are met. In another embodiment, the system may provide the user with one or a series of satisfactory locations for the equipment based on the data center resource requirements of the equipment and any applicable data center resource redundancy and/or runtime requirements. In another embodiment, the system may provide the user with an optimum location, followed by other satisfactory locations in a decreasing order of preference, for the new equipment based on the data center resource requirements of the equipment and any applicable data center resource redundancy and/or runtime requirements. If a particular layout is not designated by a user, then the system 200 will determine the layout in the manner discussed above with respect to stage 306 of process 300.
  • At stage 610, an updated floor model is displayed (for example, in the manner shown in FIGS. 5C and 5D), and at stage 612, a user can review the model and either provide feedback (stage 614) or indicate that the design is satisfactory. Once the floor model has been approved by a user, the retrofit design process is complete, and the equipment may be installed, relocated, or removed and parameters of the data center may be measured and updated in the manner described above in stages 318 to 322 of process 300 using for example a data center management system.
  • In one embodiment, changes to the configuration of the data center such as those designed by retrofit process 600 may be implemented by work order process 1900 as illustrated by the flowchart in FIG. 19. Initially, at stage 1900, a data center configuration change may be made using a graphical user interface such as those discussed with regard to FIGS. 5C and 5D above. Details of the change may be used to construct a set of tasks to carry out the change. These tasks may be grouped together into work orders. Work orders may be human-readable and may include textual instructions as well as pictorial representations. It should be appreciated that a single configuration change, initiated using a visual editor, may be translated into multiple work orders.
  • At stage 1902, a work order is dispatched to a user for implementation. In one embodiment, this dispatching may be automatic or manual. Automatic dispatching may occur based on a set of preconfigured rules. These rules may consider characteristics of both the work to be performed and the user to whom the work order may be dispatched. Characteristics of the user that may be considered include the user's physical location, physical area of responsibility, recent and current work load, remaining schedule availability and area of expertise. Alternatively, a work order may be automatically dispatched to one or more users in a “round robin” fashion. In another embodiment, a user, such as a data center manager, may access the dispatch UI exposed by work order module 214, discussed in relation to FIG. 3 above, to modify and/or dispatch work orders to users. In still another embodiment, a user, such as a technician, may access the dispatch UI exposed by work order module 214 to modify and/or dispatch work orders to himself.
  • At stage 1904, a user performs the work dictated by the work order and modifies the work order appropriately. Modification of the work order may include, for example, changing the details of the work performed such as recording installation of an alternative model of equipment, rearrangement of the order of equipment within the rack, reconfiguration of the flow of a data center's resources, such as power, to the equipment, etc. Modification may also include changing the status of the work order. When a user modifies a work order using a remote or mobile computing device, certain information required to complete the modification may be entered using any interface supported by the remote or mobile computing device, including barcode scanning.
  • In one embodiment, work order module 214 may monitor a data center resource provider, such as power outlet, that is targeted as part of the work order for a change in demand. When a change in demand is detected, such as an increase in the power demanded at a power outlet, the work order may be marked as complete. In another embodiment, the work order is not marked as complete until the data center resource consumption or production benchmark, such as the power consumption benchmark, of the data center equipment specified on the work order is recognized. For example, if a work order calls for the installation of a particular model of blade server, the system may not mark the work order as complete after initial activation of the blade server, but instead may wait until the power consumption history of the blade server matches a known power consumption benchmark for the model of blade server called for in the work order.
  • In the processes 300 and 600 described above, design and analysis stages are performed after all data is entered as part of an initial design or a retrofit of a facility. In another embodiment, analysis is performed real-time, and user displays are updated as the user enters data into the system.
  • In embodiments of the present invention, using the processes described above, data center operators are able to determine, in essentially real-time, whether additional equipment may be added to a data center and may also determine locations for the equipment, where data center resource requirements, such as both power and cooling requirements, of the equipment may be met. Further, reports may be generated that indicate to a user or data center manager how much capacity, redundancy and/or runtime is available for each row, for each rack, for each U space position, for each piece of data center equipment and for the facility in its entirety. Still further, as discussed above, in determining overall capacity, systems and methods are able to identify and display stranded capacity, and provide suggestions for reducing the stranded capacity.
  • In the processes and systems described above, cooling calculations for a data center and for equipment in the data center are performed as part of the design process for the data center, during operation of the data center, and during an upgrade or retrofit of the data center. In embodiments of the present invention, in determining equipment layout and performing cooling calculations, initial information on characteristics of the facility itself are identified to determine if there is sufficient cooling at the facility level. These characteristics include, for example, whether a raised floor or drop ceiling is used as an air plenum, the location and characteristics of air conditioning units (including in-row cooling units), dimensions of the room that are to house the data center, and total power draw of the data center. Based on this information, an initial determination may be made as to whether there is sufficient cooling provided by the air conditioning units for the expected power draw in the room, and if not, a recommendation may be made for additional air conditioning units. For some facilities, desired redundancy and operating margins may be included in this determination.
  • Once the determination has been made that there is sufficient cooling at the facility level, an analysis is conducted to determine if there is adequate cooling at each rack in the facility and/or at individual pieces of equipment. In at least one embodiment, the cooling capacity of a rack may be determined by increasing the power level of the rack to determine at what additional power level the airflow to the rack becomes inadequate. This can be performed individually for each of the racks in a data center. In different embodiments of the present invention, one or more of a number of different methods may be used to perform the cooling calculations. These methods include, but are not limited to, a computational fluid dynamics (CFD) analysis, a Coarse-Grid CFD analysis (designated as CGCFD), a control volume analysis (designated as CVA), and an analysis based on empirical rules and/or borrowing concepts. Further, in some embodiments, a combination of two or more of the above methods may be used to conduct portions of an overall analysis. In one embodiment, the principle of superposition is used to combine results of portions of an analysis. In particular, in many applications the airflows may be considered to be ideal, where an ideal airflow is inviscid, incompressible, irrotational without any other forces, such as buoyancy. With such an ideal airflow, a complex application can be reduced to a number of less complex applications, analysis of the less complex applications can be performed using one of the methods described herein, and superposition can be used to combine the results of each of the less complex applications to obtain analysis results for the complex application.
  • A computational fluid dynamics analysis can be used in one embodiment in association with the design and retrofit of a data center to provide detailed results of the cooling performance of a facility, including determining the availability of adequate cooling air at racks and individual pieces of equipment in the facility and determining cooling capacity for each rack. The techniques for implementing a CFD analysis of a data center are known. A CFD analysis must typically be performed by someone particularly skilled in the art, typically requires detailed information regarding the facility and the layout of equipment in the facility, and depending on the complexity of the analysis conducted, and the computing equipment used to conduct the analysis, may take hours or days to run one iteration of the analysis.
  • In another embodiment, an improved technique for conducting the cooling analysis is used. The improved technique has been developed based on computational fluid dynamics techniques. The improved technique is referred to herein as a Coarse-Grid CFD or simply CGCFD. In a typical CFD analysis, a data center to be analyzed is typically divided into non-uniform cells in the range of one to eight inches on a side. In at least one embodiment, in the CGCFD analysis, a Cartesian grid system of cells that are one foot cubes are used. The use of uniform one foot cells typically reduces the number of cells used in the calculations from a traditional CFD analysis by at least an order of magnitude. Further, uniform grid cells generally make the CFD analysis faster and more reliable relative to a comparable non-uniform cell analysis. Further, other techniques are employed in the CGCFD analysis to improve the computational efficiency of the analysis. These techniques include: the use of simple turbulence models, initializing the analysis with data obtained from the results from a prior similar solution, using two dimensional or partial two dimensional representations when possible to simplify calculations, and tailoring a CGCFD routine for a specific application. The use of two dimensional representations may be used, for example, in a raised floor or ceiling plenum, where pressure gradients in the depth direction can be neglected in the computations.
  • The tailoring of a CGCFD routine can be used in embodiments of the present invention to significantly improve computational efficiency and improve robustness (for example, so the tool can be made to work reliably in an autonomous way) of the total analysis, and multiple tailored routines may be used in combination to produce a complete analysis. For example, a first CGCFD routine may be tailored for use with different raised floor configurations to determine the output airflow at each perforated floor tile of a raised floor in a data center, and a second CGCFD routine may be tailored for use with a cluster of racks that include two rows of racks with a cold aisle between the rows. The first CGCFD routine may be run to determine the output air at perforated tiles in the cold aisle of the rack cluster, and the second CGCFD routine may use the results of the first routine to determine the airflows and temperatures at inlets and outlets of the racks. The second routine may be run a number of times to account for all of the clusters of racks located in a data center. As equipment is moved, and as different configurations are established within a cluster to optimize cooling performance, the second routine may be run to obtain new cooling data without the need to repeat the first routine, as the airflows from the perforated tiles generally would not change. In some cases, for perforated floor tiles having a large percentage of open area (for example, greater than 50%), it may be desirable to repeat the first routine as air flows may change based on room configuration.
  • Embodiments of the invention that utilize the CGCFD approach to conduct the analyses of a data center provide advantages over embodiments that utilize a traditional CFD approach. These advantages include computational efficiency and simplification of use. Iterations of cooling calculations using the CGCFD approach may be conducted in a matter of seconds or minutes versus hours or days with a full CFD analysis. Further, the CGCFD routines may be designed to operate with a limited set of input variables, allowing a less-skilled user to conduct analyses using the CGCFD approach. For example, for a CGCFD routine that is tailored to analyze only the raised floor plenum, the input variables may be limited to the height of the floor, location and type of perforated tiles, length and width of the floor, and the locations and characteristics of air conditioning units that provide cooling air to the raised floor. For a CGCFD routine that is tailored to conduct an analysis on a cluster of racks, the input data may be limited to airflow per tile (could be obtained automatically from the output of a separate CGCFD routine or using other methods), the number of racks in the cluster, the power draw of each of the racks, and room environmental details including the temperature of the surrounding environment, ceiling height, the presence of nearby walls, etc. The output data for a cluster of racks may include the input temperature at each server, or other piece of equipment in a rack. In other embodiments, the output data may simply be a measure of the amount of warm recirculated air drawn into each rack. The data may be expressed as an absolute number (e.g. in terms of cfm) or expressed as a fraction of the total air consumed by the rack. A system, such as system 200 described above, may use the output data to determine if the cooling performance of the cluster is satisfactory.
  • In another embodiment, another improved technique for conducting the cooling analysis is used. The improved technique is referred to herein as a control volume analysis or simply CVA. The control volume analysis may be used in conjunction with a CFD analysis and/or a CGCFD analysis, or may be used as a stand alone process. The CVA technique is similar in some aspects to the CGCFD technique described above, however, further simplification of the analysis process is provided. As will be discussed below, the CVA technique is a computationally efficient technique that is particularly effective for computing the three-dimensional airflow, pressure and temperature distributions in the cold aisle of a raised floor data center. However, the CVA technique is not limited in its use to this application and may be used for other applications as well. The CVA technique can provide output data in essentially real-time, allowing a user to try various locations for equipment as part of an initial design or a retrofit and obtain cooling data for the different locations in real-time.
  • The CVA technique will be described with reference to FIG. 8, which shows a subsection 700 of a data center. The subsection of the data center includes a cluster of racks that includes a first row of racks 702, and a second row of racks 704, which are located on a raised floor and separated by two rows of perforated tiles 706, 708.
  • In data centers that have clusters of racks arranged like those in FIG. 8, it is not uncommon for undesirable hot spots to develop even though the total supply of cool air to the cluster should be sufficient to meet the needs of the racks. For example, if the airflow rate through one or more perforated tiles is too great, a rack may be unable to capture all of the tile airflow and some of the cooling air may escape the cold aisle. Racks will generally draw their required air, and in this situation, if a rack can not capture cool air, it may draw hot exhaust air over the top of the rack creating a hot spot. Further, due to widely varying cooling airflow requirements, racks may compete with one another for cooling airflow. In particular, a high-power rack may borrow underutilized air from an adjacent rack, or in some cases from a rack separated by several tiles. With several racks contained in a cluster, with each having different cooling needs, the airflow patterns and temperature distribution in the cold aisle are complex functions. The CVA technique can be used to simplify the solutions to these complex functions.
  • In the CVA analysis for the rack cluster of FIG. 8, the airflow and temperature analysis is conducted on the volume of air contained in the cold aisle, between the racks, from the perforated tiles up to a height equal to the top height of the racks. The volume of air is divided into a number of control volumes equal to the number of racks in the cluster. Each control volume is defined as the volume above one of the perforated tiles extending from the perforated tile to the top of the racks. The control volume analysis includes determining for each control volume, the airflow through each of the six faces of the control volume. Once the airflows are known, temperatures and mass species concentrations can be determined for each of the control volumes. In the CVA technique, the temperature analysis can be decoupled from the airflow analysis because, as discussed above, buoyancy forces in the control volumes can be ignored. Similarly, mass species concentrations are not coupled to the airflow solution and may also be computed separately if desired in order to compute the fraction of recirculated air ingested by each rack.
  • In conducting a CVA analysis in the embodiment described herein, there are several initial assumptions made to simplify the analysis. In other embodiments, the analysis could be changed if these assumptions would not apply. The first assumption is that airflow across each face of a control volume (and therefore into the front face of a rack) is considered uniform. Resulting airflow and temperature values effectively represent an average of the airflow and temperature at each face.
  • The second assumption is that buoyancy forces within each control volume are negligible. Unless a significant hot spot develops, then there is insufficient heating of the air in a cold aisle to substantially affect airflow patterns, and even if some heating occurs, any buoyancy effects are small compared to the momentum of the airflow from typical perforated tiles.
  • The third initial assumption is that viscosity and turbulence within each control volume are negligible. In the control volumes, air is introduced through the perforated tiles and is pulled into the racks. The air is not required to change direction rapidly and there is no flow of air parallel to solid objects. Accordingly, viscosity and turbulence may be ignored and the competition of forces driving the airflow reduces to an interplay between pressure and momentum.
  • The CVA analysis may be further simplified by dividing a cluster of racks into slices of two racks separated by two perforated tiles 718A, 718B. FIG. 9 shows a cluster of six racks 710 that can be divided into three two- rack slices 712, 714, 716. The nomenclature used for identifying racks, and airflows in FIG. 9 is defined in Table 1 below, along with other variables that will be used herein in the description of the CVA technique.
  • TABLE 1
    As, At Control volume side and perforated tile area
    C1, C2 Dimensionless empirical constants in the y and x-
    momentum equations
    C Species Concentration
    CV Control volume
    N Number of 2-rack slices in cluster
    {circumflex over (n)} Outward unit normal vector
    PAi, PBi Pressure in control volume above perforated tiles Ai and Bi
    Pamb Ambient reference pressure
    ML, MR, MT z-direction momentum flux terms through left, right, and
    top surfaces of staggered CV at slice i
    TAi, TBi Temperature in control volume above perforated tiles Ai
    and Bi
    Qt Airflow rate through each perforated tile
    QAi, QBi Airflow rate through racks Ai and Bi
    QAxi, QBxi Airflow rates in the x-direction above perforated tiles Ai
    and Bi
    Qzi Airflow rates in the z-direction above perforated tiles
    between tiles Ai and Bi
    QAtopi, Airflow rates in the y-direction above perforated tiles Ai
    QBtopi and Bi at top-of-rack height
    {right arrow over (V)} Velocity Vector
    α Linear relaxation factor
    ρ Density of air
  • At the start of the CVA analysis, the rack airflows QAi, QBi and the tile airflows are known. The tile airflows can be estimated based on the mean perforated tile airflow for the entire facility or determined using a CFD analysis, a CGCFD analysis, physical measurement, or using one of a number of known programs. The rack airflows can be determined based on characteristics of equipment installed in the rack. In one embodiment, each rack airflow is determined on the basis of the power usage of the rack and using the relationship of 160 cfm per kilowatt of power as discussed above. To determine the airflow patterns, all airflows QAxi, QBxi, Qzi, QAtopi, and QBtopi, and pressures PAi, and PBi are computed based on the principle of conservation of mass and momentum. To perform this computation, a total of 7 n−2 unknowns (5 n−2 internal airflows plus 2 n pressures) may be determined using a total of 7 n−2 equations, where n is the number of 2-rack slices (or length of cluster expressed in tile or rack widths). Optionally, an energy balance or mass species balance can then be used to compute the 2 n temperatures or 2 n species concentrations based on the airflows.
  • In one embodiment, rather than solve all equations simultaneously, a semi-simultaneous approach is taken. In this embodiment, the five unknowns for each two-rack slices of a rack cluster, airflows Qzi, QAtopi, and QBtopi, and pressures PAi, and PBi, are first determined simultaneously. During these initial computations, each two-rack slice is considered in isolation, which is the equivalent of having the ends of the slices blocked, such that QAxi and QBxi are equal to zero. After an initial sweep through each two-rack slice is complete, the side-to-side airflows (QAxi, QBxi,) can be computed based on the calculated pressures within each control volume. The side-to-side airflows affect the pressures, and after computing the side to side airflows, a second computation of the airflows and pressures is conducted for each of the two-rack slices. This process is repeated until there are no significant changes in the computed variables. Once all airflows are known, all temperatures or mass species concentrations can be computed directly without the need to do multiple iterations.
  • The fundamental equations used to compute the unknowns described above, assuming steady state, incompressible and inviscid fluid flow rely on conservation of mass (m), conservation of momentum (M) conservation of energy (e) and conservation of species concentration (C), and can be written as follows:
  • A ( V -> · n ^ ) A = 0 ( m ) A ( ρ V -> · n ^ ) V -> A = - A p n ^ A ( M ) A T ( V -> · n ^ ) A = 0 ( e ) A C ( V -> · n ^ ) A = 0 ( C )
  • Applying the conservation of mass equation (m) to the two-rack slices for the conditions described above results in the following equations:

  • Q t +QA i +QAx i−1 =Qz i +QAx i +QAtopi  (1)

  • Q t +Qz i +QBx i−1 =QB i +QBx i +QBtopi  (2)
  • Where QAi is always negative based on the sign convention of FIG. 9.
  • As will now be described, staggered control volumes are used to formulate the z-momentum equations. Three staggered control volumes 730A, 730B and 730C are shown in FIG. 9A. The number of staggered control volumes is equal to the number of 2-rack slices. The staggered control volumes are the same size as the main control volumes, but are shifted so that they are located midway between opposing racks. The staggered control volumes allow pressure to be considered more easily for each face which is normal to the z-direction. If the original control volumes are used, then each control volume would have one face coplanar with a rack inlet, which is a face over which the pressure is not known and need not be known in the calculations. Applying the conservation of momentum equation (M) in the z-direction to the staggered control volume in slice i results in the following equation:

  • PA i −PB i=(ρ/(4A s 2)){(Qz i +QB i)2−(QA i +Qz i)2 }+M L +M R +M T  (3)
  • In equation (3), the first term on the right hand side of equation (3) is generally dominant, as it accounts for the effect of rack airflow rates on control volume pressures. ML, MR and MT account for losses or gains in z-momentum through the sides and the top of the control volume.
  • Using an “upwind” estimate for incoming/outgoing z-momentum and assuming that the velocity of the air in the z-direction is negligible above the racks, ML, MR and MT are determined using the equations of Table 2 below
  • TABLE 2
    IF THEN ELSE
    QAxi−1 + QBxi−1 ≧ 0 ML = −(ρ/(2 As 2)) (QAxi−1 + QBxi−1) Qzi−1 ML = −(ρ/(2 As 2)) (QAxi−1 + QBxi−1) Qzi
    QAxi + QBxi ≧ 0 MR = (ρ/(2 As 2)) (QAxi + QBxi) Qzi MR = (ρ/(2 As 2)) (QAxi + QBxi) Qzi+1
    QAtopi + QBtopi ≧ 0 MT = (ρ/(4 As 2)) (QAtopi + QBtopi) Qzi MT = 0
  • The relationship between Y-momentum and pressure may be written using equations (4) and (5) as follows:

  • PA i −P amb==(ρ/A t 2){C1[Q t+½(QAi+QAx i−1 −QAx i −Qz i)]2−½QAtopi 2}  (4)

  • PB i −P amb==(ρ/A t 2){C1[Q t+½(Qzi+QBx i−1 −QBx i −QB i)]2−½QBtopi 2}  (5)
  • In one embodiment, equations (1) through (5) are solved simultaneously for each 2-rack slice of a cluster sequentially using the process 800 shown in FIG. 10. In the first stage 802 of process 800, the user defines QT (the airflow through the perforated tiles), the number of 2-rack slices in the cluster, and the power draw of each of the racks. As discussed above, QT may be estimated as the mean perforated tile airflow rate for the entire facility or determined separately using, for example, a CFD or CGCFD analysis or other analysis or physical measurement. At stage 804, all airflow variables (except QT and the rack inlet airflows) are initialized to zero. At stage 806, equations (1) through (5) are solved simultaneously for each slice. At decision block 808 a determination is made as to whether the equations have been solved for all slices, and if not, stage 806 is repeated. Once the equations have been solved for all slices, then at stage 810, the x-direction airflow variables are updated based on the computed pressures in the control volumes, PAi and PBi as discussed below. At stage 812, a determination is made as to whether the computed pressures have changed by more than a predetermined threshold since the previous iteration and if so, stages 806 to 812 are repeated. Once there is no significant change in the computed variables, the process 800 stops at stage 814, at which point the pressures and airflows for all of the control spaces have been determined.
  • In the process 800, at stage 810, new x-direction airflow values (QAxi and QBxi) are determined based on the assumption that the pressure drop between adjacent cells is proportional to the square of the airflow rate using the equations in Table 3.
  • TABLE 3
    IF THEN ELSE
    PAi ≧ PAi+1 QAxi = As{(PAi − PAi+1)/(ρ C2)}1/2 QAxi = −As{(PAi+1 − PAi)/(ρ C2)}1/2
    PBi ≧ PBi+1 QBxi = As{(PBi − PBi+1)/(ρ C2)}1/2 QBxi = −As{(PBi+1 − PBi)/(ρ C2)}1/2
  • In one embodiment, because of non-linearities of the equations, adjustments to the x-direction airflow values at stage 810 are achieved gradually by introducing damping into the iterative process and updating the values of QAxi and QBxi using the following equations (6) and (7).

  • QAx i =αQAx i new+(1−α)QAx i old  (6)

  • QBx i =αQBx i new+(1−α)QBx i old  (7)
  • In equations (6) and (7), α is a linear relaxation factor. If α is set to zero, then no changes will occur from iteration to iteration. If α is set to 1, then there will be no damping introduced. For smaller values of α, more iterations will be required, however, the chances of obtaining a stable solution increase. The particular optimum choice of α is problem specific, however, it has been found that values of α around 0.05 work well in the process described above. Once the airflows are computed using the process above, temperatures and mass species concentrations can be calculated, if desired. It should be noted that control volumes may be used to compute temperatures or concentrations regardless of the method used to initially compute airflows.
  • The CVA technique described above can be conducted separately, one for each cluster of racks in a facility to obtain a complete cooling analysis of the facility. When a retrofit of a facility is to be done, the control volume analysis may be done for all clusters, or only for those in the vicinity of any changes to the facility.
  • Three different methods, CFD, CGCFD and CVA, have been described above for determining cooling data in embodiments of the present invention to determine placement of location of equipment in data centers. In still another embodiment, empirical rules are used either alone or in combination with one of the methods described above to determine proper placement of equipment and the adequacy of cooling air. The empirical rules that are used may take a number of different forms, and programs incorporating the empirical rules may be updated as more data is generated to support the empirical rules. In one embodiment, empirical rules are based, at least in part, on the ability of equipment racks to borrow unused capacity from surrounding neighbors. The amount that may be borrowed may be limited to an allowable fraction (or weight) of the unused capacity and the particular allowable fraction may differ depending on a number of variables such as borrower-donor separation distance, tile flow rate, and the total power draw of both the borrower and the donor.
  • In one particular embodiment, the cooling air available to a given rack is computed based on a weighted summation of the available airflows from airflow sources (i.e., supply devices, including in-row cooling units, or vents), net of airflows computed to be used by other racks, where the weights associated with the available airflows for a given rack decrease with distance between the rack and the air supply devices or vents. For example, with reference to FIG. 9, the cooling air available to each rack may initially be set equal to the cooling air supplied by the perforated tile in front of the rack, or to reflect possible losses, and provide safety margin, the cooling air available may be set equal to some amount (i.e. 90%) of the total air from the perforated tile. The cooling load for each rack is then subtracted from the available air to provide a net available cooling air figure for each perforated tile and to provide an initial indication of a lack of cooling air for any equipment rack. For each equipment rack, the available cooling air is then increased by assigning to each rack, a percentage of the net available cooling air from nearby perforated tiles. For example, the cooling air available may include 10% of the net available cooling air from a perforated tile associated with either an adjacent rack or a rack across an aisle, and 5% of the net available cooling air from a perforated tile of a diagonal rack or a rack two positions over in a row. The particular percentages or weights used may be changed based on actual results or as a result of analyses conducted. The loads of each rack may then be compared with the total available cooling air to determine remaining cooling capacity and to identify any potential problem racks.
  • In at least one embodiment, empirical rules may be used in combination with superposition to analyze data centers and provide recommended equipment layouts. Using superposition, complex problems may be broken down into simpler problems that can then be solved using empirical rules.
  • In one embodiment, empirical rules are established by initially performing a number of CFD analyses on typical rack layouts, and the results of these analyses are used to produce simple equations or look-up tables that can be used in real-time to design layouts of equipment. In such an analysis, the side-to-side airflows, such as those shown in FIG. 9 may be determined for each rack one at a time with one rack turned “on” and all other racks turned “off”. The airflows at the ends of a cluster for a number of different configurations may also be determined using CFD. The airflows may be determined for a number of different air intake values for each rack and a number of different values of air flow from the perforated tiles. The total air flows for different configurations can then be determined in real-time using superposition and the stored results. The airflows through the top (in or out) of the volume in front of each rack may then be determined based on conservation of mass. In one embodiment, when the airflow into the top of one of the volumes exceeds some percentage (i.e., 20%) of the total air flow into the rack associated with the volume, then an overheating problem may exist requiring a design around. In other embodiments, mass species concentration analyses may be used in combination with empirical rules to determine what percentage of the total air entering a control volume is recirculated air to determine when an overheating problem may exist.
  • In determining the airflows for each rack of a cluster, symmetry of the clusters can be used to reduce the number of CFD analyses that need to be performed, and the control volumes discussed above with respect to FIG. 9 may be used to establish a reference grid for the analysis. For example, with reference to the cluster of racks 710 of FIG. 9, CFD analyses need only be performed for Rack Ai−1 and Rack Ai, and the results for each other rack may be determined based on the resulting airflows and the relative position of the racks. For example, the airflows in the cluster associated with Rack Bi+1 are the same as those associated with Rack Ai−1 with the direction of some of the airflows changed for corresponding Rack A and Rack B inlet airflow and tile airflow rates.
  • In one example, which will now be described, the concepts of symmetry and superposition are used in conjunction with CFD analyses and empirical rules to provide a practical real-time solution for determining air flows in a cool aisle. Further, air flows are used to determine a recirculation index (RI) for a row of racks, which can be used to identify potential “hot spots” in a data center. As discussed above, one cooling objective in a data center is to manage the equipment rack inlet air temperatures. The rack inlet air temperatures are dominated by the airflow patterns within the cold aisle and the temperatures within and around the cold aisle. Air drawn in from outside the cold aisle is generally heated to some degree by the rack exhaust and will be hereafter referred to as “recirculated air”. While the temperature of the recirculated air is highly application dependent, air that passes directly from a perforated tile to a rack inlet will be very near the supply temperature. Thus, good cooling performance can be achieved if all of the airflow ingested by a rack comes directly from the perforated tiles.
  • A cluster of racks, which receives its required cooling air exclusively from the perforated tiles within the cluster, represents an autonomous scalable unit from which a larger facility with predictable cooling performance may be constructed. A reasonable requirement is therefore to ensure that racks are adequately cooled by air originating from the racks own cold aisle. Conversely, it is acceptable for the rack to ingest no more than a small fraction of recirculated air.
  • With the above in mind, the recirculation index (ri) is defined as the fraction of recirculated air ingested by the rack. An ri of 0% implies that all of the rack inlet air was drawn directly from the perforated tiles while an ri of 100% implies that all of the rack inlet air was drawn from outside of the cold aisle. Note that a low ri is sufficient to guarantee cool inlet temperatures; however, a high ri does not guarantee excessively high inlet temperatures.
  • The concepts of control volumes, symmetry and superposition are used in the present example to determine air flows and ultimately RI for a cold aisle. In using superposition, a sum of velocity potentials (or actual velocity components or total airflows over a consistent area) of simpler, elemental flow solutions is used to obtain a new, composite flow solutions. For example, assume we know the airflow pattern associated with only Rack A1 “on” subject to a particular tile airflow rate and we also know the airflow pattern with Rack B3 “on” subject to the same perforated tile airflow rate. The relevant horizontal airflow components can be added to obtain a solution, which approximates the airflow pattern resulting from Racks A1 and B3 both “on” simultaneously. The airflow pattern resulting from the superposition of the two separate airflow patterns is not exactly the same as the full solution—even for an ideal flow. Using superposition two solutions are added together which individually (and when added together) satisfy conservation of mass criteria. The use of superposition does not guarantee that the combined solution will be the unique solution and the difference is in the boundary conditions. As an illustration of this, consider a 2-rack example. In using superposition, the top airflow condition floats freely as a constant-pressure boundary condition in all cases. In reality, the airflow pattern constructed from superposition may not provide a perfect match to air velocity over the top surface of the cold aisle. Also, when one rack is off, an assumption is made that the face (inlet) of the rack is a symmetry boundary condition (which is consistent with an inviscid analysis). This result creates the opportunity for some flow parallel to the face of the rack, which would probably not exist when the rack is actually drawing air.
  • In the example, superposition is used to establish only the 3 n−2 internal horizontal airflows (n being equal to the length of the row in terms of racks) while end-of-row horizontal airflows are computed based on separate empirical correlations. Vertical airflow components are computed from a mass balance performed on each control volume. The horizontal airflows clearly depend on tile airflow. For example, a rack of a given airflow rate may be able to draw cooling air from a distance of many tile-widths when the perforated tile airflow rate is very low. However, this range of influence is much less as the tile flow rate is substantially increased. (As we know from the non-dimensional argument, the results would be identical if all airflows are scaled by the tile flow rate.) Therefore, the tile airflow rate is included in the analysis; the floor tiles should be “turned on” in the CFD analysis used to correlate airflow patterns. However, if the floor tiles are left “on” and the effect of each rack is considered individually, when the flows for each rack are summed, the sum would have more airflow leaving the top of the cold aisle than in reality. The answer is to correlate only the horizontal airflows and then simply compute the correct airflow into or out of the top of each control volume based on conservation of mass.
  • It is worth emphasizing that the use of the non-dimensional airflow and, in particular, superposition, simplifies the method. Without these simplifications, there would be many combinations of rack and tile airflows to evaluate and store empirically to cover a range of practical applications.
  • Based on the above discussion, the complete airflow solutions to any rack layout of interest can be constructed using superposition. Elemental building-block airflow patterns are associated with each rack and each of the four end-of-row airflows are turned on individually as illustrated in FIG. 11 for the case of a 2-rack cluster 1002. It is important to stress that FIG. 11 illustrates which airflow boundary conditions are turned on and off in each of the elemental airflow solution to yield the total solution with all airflow boundary conditions “on”. Each of arrows 1004 a to 1004 f represents one of the airflows. It is the airflow components internal to the cold aisle that are actually being combined. There are, in general, a total of 2 n+4 elemental solutions for any layout, which makes up a complete airflow solution. Obviously, fewer elemental solutions are required if some racks have zero airflow and the ends of the cold aisle are sealed (e.g. with doors).
  • The elemental airflows used with superposition may be determined in any manner including physical testing. In the present example, CFD modeling for the cold aisle is performed using the following boundary conditions:
      • Fixed velocity of air leaving the computational domain over the area of a rack face for any rack which is “on”
      • Fixed velocity entering or leaving the domain over the area of the end of the rows for any end-of-row flow “on”.
      • The top of the solution domain is “open” for air to enter or exit to the surrounding environment held at constant pressure.
      • All other surfaces are “symmetry” surfaces.
  • As stated above, there are in general 2 n+4 elemental solutions for each row length; 2 n elemental solutions associated with each rack turned on plus four end-of-row elemental solutions. Each elemental solution covers a range of dimensionless flow rates so that any arbitrary, but practical, rack or end airflow rate can be considered. So, the task is reduced to determining and storing the 3 n−2 internal horizontal control volume airflows over an appropriate range of dimensionless airflow rates.
  • Because of the geometric symmetry of a cluster of racks, only the 3 n−2 internal airflows for approximately one quarter of the 2 n+4 rack and end-of-row boundary conditions are considered and stored; n/2+1 boundary conditions if n is even and (n+1)/2+1 if n is odd. The remaining internal airflows are determined from an appropriate reinterpretation of the smaller data set by changing variable indices and signs. In addition to being efficient, this use of symmetry, forces the final output from the rack cooling performance tool to be perfectly symmetric. Each of these boundary conditions are driven individually through a range of dimensionless airflow rates while keeping track of all of the “response” airflow rates internal to the cluster. The result can be summarized in a plot of “response” airflow rates; one plot for each elemental boundary condition.
  • As an example, internal horizontal airflows associated with boundary condition Rack A1 for an n=2 cluster are shown in FIG. 12. There are 4 curves in FIG. 12 because there are 4 horizontal internal airflows associated with an n=2 cluster of racks. All of these curves can be conveniently approximated with a least-squares fit to a cubic polynomial of the generic form

  • Q*=c 1(QRA 1*)+c 2(QRA 1*)2 +c 3(QRA 1*)3  (8)
  • so that only the coefficients c1, c2, and C3 must be stored for all airflows associated with all unique boundary conditions for all n's. Storing the “response” airflow as an equation offers the additional benefit compared to a simple look-up table in that results outside the domain of FIG. 12 are automatically interpolated.
  • The process involved in compiling the curves in FIG. 12 and the constants of Equation 8 is repeated for all unique boundary conditions for all n's considered. Determining all internal airflow correlations, for example, up to n=30 requires several hundreds of CFD runs. Therefore, in at least one embodiment, the process of converting the raw CFD data into the curve-fit constants of Equation 8 is automated. In at least some examples above, the flow in the cold aisle is considered to be ideal with no viscosity or turbulence. To verify this assumption, sample CFD cases were run with turbulence and viscosity included, and little difference was detected between models that included viscosity and turbulence and those that did not. The discussion above describes a process for all internal cold-aisle airflows for any row length, perforated tile airflow, and rack airflow distribution assuming that that the end airflow is known. A process for predicting the end airflow will now be described. Unlike the airflow within the cold aisle, the end airflow is strongly coupled to the airflow in the surrounding room environment. Buoyancy forces can be significant; direct superposition of rack-induced airflows may not work well and the end airflows do not depend simply on the dimensionless rack airflow rates. The end airflow can still be determined using empirical correlations of CFD data; however, a relatively large number of CFD simulations typically should be performed in order to achieve reasonable accuracy over a useful range of actual layouts. A comprehensive model for end airflow, which takes into account different geometric and thermal environments, may be included in other embodiments. In one embodiment, described herein, a method includes predicting end airflow as a function of rack power and airflow distribution for any row length and perforated tile flow rate while assuming a fixed room environment. The example environment is large and free of other racks or objects. Air is supplied at 60° F. and is exhausted uniformly over a 14 ft-high ceiling. As discussed above, under ideal-flow conditions, we can expect air velocity at points near the racks to scale with the dimensionless rack inlet velocities. Further, as discussed above these “response” velocities vary nearly linearly with dimensionless rack flow rate (or velocity). It is, therefore, reasonable to estimate the dimensionless end airflows based on the following expression:

  • QAx 0 *=a 0 +a A1 QRA 1 *+a A2 QRA 2 *+ . . . +a An QRA n *+a B1 QRB 1 *+a B2 QRB 2 + . . . +a Bn QRB n*  (9)
  • where QAx0* is one of four dimensionless end airflows for a cluster and the coefficients aAi and aBi effectively weight the relative importance of each rack on the end airflow. The weighting coefficients associated with racks located near the end of the row will be much larger than those associated with more interior racks. Further, empirically it is found that only racks in the first four or five positions nearest the end of the row need be retained in Equation 9. For the fixed conditions considered, the constant a0 is negative, implying that the flow is “out” (driven by buoyancy) when there is zero rack airflow.
  • To determine the values of the coefficients in Equation 9 for a particular set of room environment and cluster geometry, many (on the order of 100) CFD simulations may be performed at a number of different perforated tile flow rates. A large pool of rack power values may be created from which the many CFD simulations draw rack power and airflow data from, either randomly or systematically. The rack power values may be based on the frequency distribution of actual data center racks as determined from a survey. The rack power and airflow values used in the CFD simulations may be scaled as necessary to achieve practical total supply-to-equipment airflow ratios in the range of, for example, 0.9 to 3 for each perforated tile flow rate considered. The CFD data is then used to determine a least-squares fit of the coefficients in Equation 9 for each tile flow rate considered.
  • In summary, a simple end airflow model has been described which accurately accounts for a non-uniform distribution of rack airflow and power for a fixed set of room conditions. In at least one embodiment, the model is generalized to include the effects of geometric environment, the thermal environment, and supply airflow rate. The effects of the end airflow penetrate only a few rack distances down the row; for longer row lengths predictions for the majority of the racks in the cluster will be good even if the end airflow model is not as accurate as desired.
  • The airflow into or out of the top of each control volume has been left “floating” as necessary degrees of freedom in the above example. Now, with all of the horizontal airflows computed as discussed above, the airflow at the top of each control volume is computed based on the conservation of mass. With reference to FIG. 13, using dimensional quantities, the equations for A-row and B-row control volumes are determined using equations 9(a) and 9(b).

  • QAtopi =Q T −QRA i +QAx i−1 −Qz i −QAx i  (10a)

  • QBtopi =Q T −QRB i +QBx i−1 −Qz i −QBx i  (10b)
  • Applied to all control volumes, equations 9a and 9b represent a total of 2 n equations. At this stage, there is only one unknown per equation (QAtopi and QBtopi) so they may be solved sequentially.
  • At this point, all airflows within the cold aisle are known for the example. What remains is to track the airflow into each rack so that its origin may be identified and the recirculation index (RI) can be calculated for each rack. As discussed above, RI is the fraction of recirculated air ingested by a rack. The recirculated air can enter the cold aisle at any point where there is inflow at the ends of the rows or along the top of the cold aisle. Further, the warm recirculated air need not directly enter the cold aisle via the control volume immediately adjacent to a rack of interest; it may enter anywhere, travel anywhere the airflow patterns take it, and end up at the inlet of any rack.
  • To compute RI for each rack the cool supply air is distinguished from the warm recirculated air at all points in the cold aisle. Mathematically, this is accomplished by defining the concentration of recirculated air at any point in the cold aisle using Equation 11.

  • C recirc=(mass of recirculated air)/(total mass of air)  (11)
  • It follows from Equation 11 that the supply airflow emerging from the tiles has a Crecirc=0 and that anywhere the recirculated air enters the cold aisle along the sides or top of the cold aisle and Crecirc may be set equal to 1. In practice, Crecirc may be set to a value less than 1 for the ends of the cold aisle recognizing that, on average, the top is generally much warmer than the ends of the cold aisle. Accordingly, in one embodiment, Crecirc=0.5 for any inflow at the ends of the cold aisle.
  • The recirculated air can be assumed to have the same physical properties as the cool supply air so that it has no effect, e.g. due to a density difference, on the airflow patterns in the cold aisle.
  • Now consider a small volume just covering a rack inlet. Equation 11 applied to this volume represents the average Crecirc over this volume. Dividing the numerator and denominator by a small time increment Δt and taking the limit as Δt→0, demonstrates that the average Crecirc over a rack inlet is precisely the rack recirculation index. Thus, to determine the RR's for each rack the average Crecirc over each rack inlet is determined. Referring back to FIG. 8, we can estimate the RR for each rack as the average Crecirc of the control volume immediately adjacent to the rack of interest. Crecirc over all 2 n control volumes can be computed from the conservation of mass of the recirculated air using Equation 12.
  • All CV Faces C recirc Q = 0 ( 12 )
  • where Q is the total airflow rate through each control volume face and is a known value at this stage of the calculation.
  • FIG. 13, shows control volumes 1008 and 1010 of a transverse section of a cold aisle 1006. Equation 12 is applied to the control volumes 1008 and 1010. For convenience, we label the Crecirc crossing each control volume surface with same convention used for airflows while dropping the “recirc” subscript. The result is

  • C T Q T+(CAx i−1)(QAx i−1)=(CRA i)(QRA i)+(CAx i)(QAx i)+(Cz i)(Qz i)+(CAtopi)(QAtopi)  (13a)

  • C T Q T+(CBx i−1)(QBx i−1)+(Cz i)(Qz i)=(CRB i)(QRB i)+(CBx i)(QBx i)+(CBtopi)(QBtopi)  (13b)
  • Equations 13a and 13b are not solved directly because the number of Crecirc values exceeds the number of equations. Estimating each Crecirc as the average Crecirc from the “upwind” control volume, results in a proper balance of 2 n unknown Crecirc's and 2 n equations. Based on this “upwind” approach, the appropriate Crecirc values are inserted into Equations 13a and 13b after the airflow patterns in the cold aisle have been computed thereby establishing the direction of airflow crossing each face of each control volume.
  • TABLE 4
    Crecirc Settings Based on Airflow Direction
    Upwind Value of Crecirc
    Airflow Airflow ≧ 0 Airflow < 0
    Q t 0 0
    QAxi CAi CAi+1
    QBxi CBi CBi+i
    Qzi CAi CBi
    QAtopi CAi 1
    QBtopi CBi 1

    Table 4 shows the appropriate upwind values of Crecirc to be used in Equations 13a and 13b where the CAi and CBi are the average Crecirc over the relevant “A” or “B” control volumes respectively. Not shown in the table are the settings for QAxi and QBxi at the end of the row, i.e. Qax0, QBx0, QAxn, and QBxn. In this case Crecirc may be set to 0.5 as discussed above for any “inflow”.
  • With the values of Crecirc taken from Table 4, the 2 n Equations represented by 13a and 13b may be solved simultaneously for the 2 n CAi and CBi values. These simple linear equations can be solved without iteration virtually instantaneously for any practical row length using common computing hardware. Finally, as discussed above, the computed CAi and CBi values may be directly interpreted as the recirculation index of the adjacent “A” and “B” racks respectively. In other embodiments, because of the similarity between the energy and concentration equations, bulk average temperature could be determined over each control volume instead of RI following a very similar procedure.
  • A summary of a process 1050 for determining recirculation index for a cluster of racks using the methodology described above will now be provided with reference to FIG. 14. In a first stage 1052 of the process, the row length, tile airflow, rack airflow and rack power are defined for a cold aisle to be analyzed. Next, in stage 1054, empirical data used for computing airflows is imported from a CFD analysis as described above. The end airflows are then determined at stage 1056 based on details of the cluster and details of the room environment. All horizontal airflows are then determined at stage 1058. At stage 1060, horizontal airflows induced by the 4 end airflows are computed, and at stage 1062, complete horizontal airflows are computed by adding the airflows from stages 1058 and 1060. Vertical airflows are computed at stage 1064, and then at stage 1066, the recirculation index may be determined for each rack by solving a set of conservation equations for the recirculated air as described above.
  • In one embodiment, to determine cooling capacity for a given rack based on the recirculation index, a threshold recirculation index is first established, below which a design is considered unsatisfactory. For each rack, after a satisfactory design is achieved, the power of the rack is increased until the recirculation index of the that rack (or any other rack) reaches the threshold level, and the power at which that occurs represents the maximum cooling capacity for the rack. A similar method for determining cooling capacity can be used with other analyses described herein, including the analysis using capture index values described below.
  • In other embodiments, the control volume and superposition methods described above may be modified. These modifications may include the use of more complex statistical methods (e.g., the use of neural networks) to determine end airflow conditions from large pools of CFD data. Further, the number of control volumes may be substantially increased to improve accuracy and resolution of computed variables. In particular, the latter improvement would allow airflow variations at various rack elevations (e.g., due to a variety of equipment installed in a rack) to be considered. The basic methodology could further be modified to include layouts beyond the scope discussed above including layouts involving an arbitrary number of perforated tiles of arbitrary flow rate, an arbitrary cold aisle width, arbitrary rack dimensions or other such variations from examples discussed above.
  • In processes described above, cooling analyses of a data center have focused primarily on determining airflows in the cool aisle for a cluster of racks located in a data center having a raised floor. Embodiments described above, however, are not limited for use in data centers having raised floors, and aspects of the embodiments are also applicable to data centers that do not include raised floor cooling systems. At least one embodiment described above provides a decoupling of the cold aisle from the remainder of the data center to compute airflows in the cold aisle. The effect of the room environment is then built back into the analysis using end-of-row airflows that are computed, for example, in separate CFD calculations that may be computed offline and made available through look-up tables or empirical correlations. As described below, in a similar manner to that described above, a hot aisle in a data center can be analyzed by decoupling the aisle from the remainder of the room and later building the effects of the room environment back into the analysis.
  • In additional embodiments that will now be described, processes are provided for evaluating a cluster of racks based on airflows that occur in a hot aisle for a cluster of racks. In at least one version of the additional embodiments, a raised floor data center is not used, but rather, cooling is provided using in-row cooling units as described above. In one particular process of one embodiment, a capture index (CI) is calculated and used to analyze a cluster of racks in a data center. The capture index is used in one embodiment with a row or cluster of racks having one or more in-row cooling units, and the capture index is defined as the percentage of air released by a rack into a hot aisle, which is captured by cooling units bounding the hot aisle. The CI may be considered as a complementary metric to the RI described above for use with the hot aisle. However, CI techniques may also be used with cold aisle analysis as discussed further in U.S. Provisional Application 60/897,076, filed on Jan. 24, 2007 which is hereby incorporated herein for all purposes. The CI is useful when the focus of a design is to keep the hot air within the hot aisle. As discussed above, rack inlet temperatures are typically the ultimate cooling metric, however, if all of the hot air is captured in the hot aisle, the rest of the data center (including rack inlets) can be designed and controlled to remain at “room temperature.”
  • The cold-aisle capture index is defined as the fraction of air ingested by the rack which originates from local cooling resources (e.g., perforated floor tiles or local coolers). The hot-aisle capture index is defined as the fraction of air exhausted by a rack which is captured by local extracts (e.g., local coolers or return vents). CI therefore varies between 0 and 100% with better cooling performance generally indicated by greater CI values. In a cold-aisle analysis, high CI's ensure that the bulk of the air ingested by a rack comes from local cooling resources rather than being drawn from the room environment or from air which may have already been heated by electronics equipment. In this case, rack inlet temperatures will closely track the perforated-tile airflow temperatures and, assuming these temperatures are within the desired range, acceptable cooling will be achieved. In a hot-aisle analysis, high CI's ensure that rack exhaust is captured locally and there is little heating of the surrounding room environment.
  • While good (high) CI values typically imply good cooling performance; low CI values do not necessarily imply unacceptable cooling performance. For example, consider a rack in a raised-floor environment which draws most of its airflow from the surrounding room environment rather than from the perforated tiles. The rack's cold-aisle CI will be low; however, if the surrounding room environment is sufficiently cool, the rack's inlet temperature will be acceptable anyway. However, in this case, the rack's cooling needs are met by the external room environment rather than perforated tiles within the rack's cluster. If this process is repeated many times across the data center, facility cooling will be complex and unpredictable. Thus, high CI values lead to inherently scalable cluster layouts and more predictable room environments.
  • Note that, although CI has been defined above as a rack-level quantity relative to local cooling resources, the definition of CI could be extended to any grouping of inlets and outlets, for example, a single server with reference to a single perforated tile. In this case, the metric would indicate the fraction of airflow ingested by the server which originated from the single perforated tile.
  • In one embodiment, the use of chemical concentrations with, for example a CFD analysis, can be used to quantitatively determine CI for both a hot aisle and a cold aisle. For the hot aisle, the exhaust of each rack is identified in such an analysis as a separate species having the same properties as air, so as not to change the physics of airflow. The fraction of hot air released from racki (identified as Ci) which is captured by an in-row cooler identified as coolerj may be computed using Equation 14 below.

  • f ij =C j i(Q coolerj)/Q racki)  (14)
  • where:
  • Cji is the concentration of Ci at the inlet of cooler j
  • Qcoolerj is the airflow rate (e.g. in cfm) through coolerj
  • Qracki if the airflow rate (e.g. in cfm) through racki
  • As an example, if the cooler and rack airflow are equal, and the concentration of exhaust air Ci from rack i at the cooler inlet is measured to be 0.5, then this implies that half of the exhaust air from rack i is captured by cooler j. In a hot aisle having N coolers, then the capture index (CI) is the sum of all of the fij's over all N coolers and can be expressed using Equation 15 below.
  • C I i = j = 1 N C j i Q coolerj Q racki ( 15 )
  • As will now be described, with reference to FIG. 15, which shows a cluster of racks 1080, a set of empirical rules can be used to determine the CI for each rack 1082 of the cluster. As shown in FIG. 15, the resulting CI values may be displayed on a display with the associated racks. In one example, racks having a CI less than 60% are identified in red indicating a warning, racks having a CI between 60% and 80% are indicated in yellow as a caution, and racks having a CI greater than 80% are indicated in green indicating that the CI is satisfactory.
  • In one embodiment, a large pool of CFD runs can be performed to establish and refine empirical rules. In other embodiments, neural networks and other techniques may be used to refine rules. The cluster 1080 includes two rows (row A and row B) of parallel racks that exhaust air to a common hot aisle 1084. Each Rack is labeled A1-A6 and B1-B7, identifying the row and position in the row of the rack, and for the example shown each rack has a power draw of 2 kW. The cluster also includes in-row cooling units 1086. In FIG. 15, a number of half-rack cooling units 1086 are shown, but embodiments of the invention may also be used with full-width rack cooling units or other devices. The half-rack cooling units used in the example associated with FIG. 15 have a nominal cooling capacity of 17 kW. Also shown in FIG. 15 is the CI in terms of percentage for each rack. The CI is shown as a percentage and indicates for each rack, the percentage of its exhaust air that is captured by one of the cooling units.
  • The CI is determined based on the concept that all rack-cooling unit interactions depend only on the magnitude of airflow associated with the racks and cooling units and their relative geometric positions. Each rack location can be thought of as having a certain potential to supply airflow to other rack locations. This potential varies inversely with the distance of separation. For example, rack A1 in FIG. 15 could potentially supply a large portion of its airflow to the area near rack A2. However, much less of rack A1's airflow could make its way to position A6. Further, the amount of airflow a rack can supply to other locations is in direct proportion to its own total airflow. The net airflow which can be supplied to a particular rack location A1 can be represented using Equation 16 below.
  • ( Q Ai ) net = ( Q Ai ) self + all other racks j ( Q Aj ) self A - B Δ x + c { ( Q Bi ) self + all other racks j ( Q Bj ) self A - B Δ x } ( 16 )
  • where
      • (QAi)net=The net maximum airflow that can be supplied to location Ai including contributions from all other racks.
      • (QAi)self=The actual airflow supplied by the rack at location Ai.
      • A=empirical constant.
      • B=empirical constant.
  • C=empirical “coupling” constant for accounting for effects from opposite row.
  • The net maximum airflow that can be supplied to various locations in the B row is computed using a similar expression. Finally, the same expression is used to compute the net maximum airflow, which can be captured at any rack location—with the sum over all coolers instead of racks. The CI is then estimated as the ratio of net airflow captured and net airflow supplied at any location expressed as a percentage and with values capped at 100%. The constants A, B, and C are selected to provide the best statistical fit to benchmark CFD data. Different values of the constants may be used to account for alternative configurations including different cooler types, different average rack power or peak-to-average power ratios and alternative hot-aisle spacing, row lengths, or room environments. As an example, consider a cluster of average power racks with modest rack-to-rack power variations. The cluster is 14 ft. long, contains a 3 ft. wide hot aisle, and is assumed to be in a fairly tightly packed data center environment with a 12 ft. ceiling height. In this case, reasonable predictions are made with the empirical constants taken as A=0.56, B=0.33, and C=0.65.
  • In the embodiment described above, the CI is calculated for a cluster of racks having uniform depth and width. In other embodiments, the same processes may be used for racks of non-uniform depth and width. In one embodiment, the CI calculations described above are programmed into a Microsoft Excel Spreadsheet program that allows a user to add and move cooling units to see the effect of different numbers of cooling units and their placements. In other embodiments, the process for determining CI described above may be incorporated into data center design and management systems, such as system 200 discussed above.
  • In the embodiment above, an exponential expression is used for modeling the rack and cooler interactions. In other embodiments, other expressions may be used, such as a polynomial or any other mathematical expression which contains a number of parameters which may be tuned to provide the best fit to benchmark performance data. Further, different curves and/or different coefficients may be used for the portion of the calculation associated with determining air supplied by racks than used in the portion of the calculation used for determining air captured by cooling units. In another embodiment, the rules may be further refined to address specific situations. For example, a Rack A may have no effect on another Rack B where a third rack C is located between Rack A and Rack B and has greater airflow than either Rack A or Rack B.
  • In still another embodiment, the effects of the ends of rows may be accounted for explicitly in hot aisle and cold aisle CI analysis. Separate CFD simulations may be conducted to determine the net inflow or outflow of air at each end of a row for given layouts of racks and cooling units. The results of the CFD simulations may be incorporated into the empirical methods described above to determine CI for racks in a cluster. The results of the CFD simulations can be used to provide correct airflow estimates at the ends of the row, while one of the algorithms discussed above may be used to determine CI at more interior portions of the row. Similarly, the effects of a missing rack or racks may be simulated using CFD with the results incorporated into the empirical methods.
  • The capture index method of analysis discussed above provides a rack-by-rack, or local, cooling performance metric for equipment racks in a data center. In addition to using CI as a rack-level metric in a hot aisle analysis or cold aisle analysis, in another embodiment, cluster-wide performance metrics are determined, and the overall cooling performance of the cluster may be determined based on both the CI metric and the global metric. The cluster-wide performance metrics indicate whether the cluster as a whole will have adequate cooling performance. The CI identifies which racks are not having their exhaust airflow adequately captured. If a rack having a low CI is a low power rack, this may not result in a problem. Furthermore, a rack may have a relatively high CI, yet still cause problems if it is a high power rack with a high outlet temperature. In one embodiment, for a hot aisle analysis the global metric that is used is a determination of the net power which escapes the hot aisle. The net escaped power can be determined using equation 17.
  • all racks i ( 1 - C I i ) P i ( 17 )
  • where,
    CIi=the capture index for rack i expressed as a fraction (rather than percentage), and
    Pi=the power of rack i.
  • The net escaped power determined by equation 17 may be correlated to maximum cluster inlet temperatures (e.g., a net escaped power of 25 kW may imply a maximum cluster rack inlet temperature of seventy-nine degrees F.) for clusters of particular geometric layouts (e.g., hot aisle width, row length, etc.), room environments and rack and cooler details (e.g., cooling unit flow rate and rack airflow/power (cfm/kW)). Accordingly, the net escaped power may be used to determine the highest rack inlet temperature.
  • In other embodiments, other techniques described above for calculating airflows in a cool aisle may be applied to a hot aisle or cold aisle analysis to determine CI, including the use of CFD, CGCFD and control volumes. The use of superposition may be less applicable in hot aisle analyses because the airflow patterns are not ideal. In still another embodiment, a process 1100 is provided for determining CI in a hot aisle using a CFD analysis. The CFD analysis is performed only on the hot aisle itself and end of row airflows are determined separately and may be patched into the CFD solution. Common computer hardware can be used to conduct such a CFD analysis in 10 to 20 seconds. The process 1100 is shown in flowchart form in FIG. 16. In a first stage 1102 of the process 1100 data related to the layout is loaded either manually, read from a database, or in any other manner. The data related to the layout may include row length, power draw and airflow for each rack, rack dimensions, cooling unit types, locations and flow rates, hot aisle widths and room environment details (i.e., size, temperatures).
  • At stage 1104 of the process 1100, hot-aisle end of row airflows are determined using best-fit curves, or look-up tables based on prior CFD studies. Hot aisle end of row calculations may be performed using the same techniques used to determine cold aisle end of row airflows with the input rack airflow provided as a positive value and the cooling units airflow input as a negative value. At stage 1106, a CFD analysis of just the hot aisle is performed using results of the end airflow analysis of stage 1104 and with the top of the hot aisle taken as a constant-pressure boundary. In the analysis, the exhaust air for each rack is “tagged” with a particular concentration. At stage 1108, the CI for each rack is determined based on the ratio of captured and supplied airflows computed by equation 16 and the results of stage 1106. The global cluster cooling metric may also be determined at this point using equation 17. At stage 1110, the rack CI's and the global cluster cooling metric can be used to determine if cooling for the cluster is adequate.
  • In addition to the cooling analysis methods discussed above, which provide real-time analysis of data centers, additional cooling analysis methods may be used in embodiments of the invention. These additional methods include a panel method, a potential analysis method, flow network/zonal modeling, principal component analysis or a combination of any of these and the methods discussed above. These additional cooling analysis methods are generally known to those of skill in the art.
  • Various embodiments according to the invention may be implemented on one or more computer systems as discussed above. These computer systems may include cellular phones, personal digital assistants and/or other types of mobile computing devices. Moreover, a computer system may include any sort of computing device such as a personal computer, a workstation, a mainframe, a networked client, a server, media servers and application servers. For example, system 200 may be implemented in a single computer system or in multiple computer systems. These computer systems may be, for example, general-purpose computers such as those based on Intel PENTIUM-type processor, Motorola PowerPC, Sun UltraSPARC, Hewlett-Packard PA-RISC processors, or any other type of processor.
  • For example, various aspects of the invention may be implemented as specialized software executing in a general-purpose computer system 900 such as that shown in FIG. 17. The computer system 900 may include a processor 903 connected to one or more memory devices 904, such as a disk drive, memory, or other device for storing data. Memory 904 is typically used for storing programs and data during operation of the computer system 900. The computer system 900 may also include a storage system 906 that provides additional storage capacity. Components of computer system 900 may be coupled by an interconnection mechanism 905, which may include one or more busses (e.g., between components that are integrated within a same machine) and/or a network (e.g., between components that reside on separate discrete machines). The interconnection mechanism 905 enables communications (e.g., data, instructions) to be exchanged between system components of system 900.
  • Computer system 900 also includes one or more input devices 902, for example, a keyboard, mouse, trackball, microphone, touch screen, and one or more output devices 907, for example, a printing device, display screen, speaker. In addition, computer system 900 may contain one or more interfaces (not shown) that connect computer system 900 to a communication network (in addition or as an alternative to the interconnection mechanism 905).
  • The storage system 906, shown in greater detail in FIG. 18, typically includes a computer readable and writeable nonvolatile recording medium 911 in which signals are stored that define a program to be executed by the processor or information stored on or in the medium 911 to be processed by the program to perform one or more functions associated with embodiments described herein. The medium may, for example, be a disk or flash memory. Typically, in operation, the processor causes data to be read from the nonvolatile recording medium 911 into another memory 912 that allows for faster access to the information by the processor than does the medium 911. This memory 912 is typically a volatile, random access memory such as a dynamic random access memory (DRAM) or static memory (SRAM). It may be located in storage system 906, as shown, or in memory system 904. The processor 903 generally manipulates the data within the integrated circuit memory 904, 912 and then copies the data to the medium 911 after processing is completed. A variety of mechanisms are known for managing data movement between the medium 911 and the integrated circuit memory element 904, 912, and the invention is not limited thereto. The invention is not limited to a particular memory system 904 or storage system 906.
  • The computer system may include specially-programmed, special-purpose hardware, for example, an application-specific integrated circuit (ASIC). Aspects of the invention may be implemented in software, hardware or firmware, or any combination thereof. Further, such methods, acts, systems, system elements and components thereof may be implemented as part of the computer system described above or as an independent component.
  • Although computer system 900 is shown by way of example as one type of computer system upon which various aspects of the invention may be practiced, it should be appreciated that aspects of the invention are not limited to being implemented on the computer system as shown in FIG. 17. Various aspects of the invention may be practiced on one or more computers having a different architecture or components shown in FIG. 17. Further, where functions or processes of embodiments of the invention are described herein (or in the claims) as being performed on a processor or controller, such description is intended to include systems that use more than one processor or controller to perform the functions.
  • Computer system 900 may be a general-purpose computer system that is programmable using a high-level computer programming language. Computer system 900 may be also implemented using specially programmed, special purpose hardware. In computer system 900, processor 903 is typically a commercially available processor such as the well-known Pentium class processor available from the Intel Corporation. Many other processors are available. Such a processor usually executes an operating system which may be, for example, the Windows 95, Windows 98, Windows NT, Windows 2000 (Windows ME) or Windows XP operating systems available from the Microsoft Corporation, MAC OS System X operating system available from Apple Computer, the Solaris operating system available from Sun Microsystems, or UNIX operating systems available from various sources. Many other operating systems may be used.
  • The processor and operating system together define a computer platform for which application programs in high-level programming languages are written. It should be understood that embodiments of the invention are not limited to a particular computer system platform, processor, operating system, or network. Also, it should be apparent to those skilled in the art that the present invention is not limited to a specific programming language or computer system. Further, it should be appreciated that other appropriate programming languages and other appropriate computer systems could also be used.
  • One or more portions of the computer system may be distributed across one or more computer systems coupled to a communications network. For example, as discussed above, a computer system that performs build-out functions may be located remotely from a system manager. These remote computer systems also may include remote general-purpose computer systems and/or remote computing devices. For example, various aspects of the invention may be distributed among one or more computer systems configured to provide a service (e.g., servers) to one or more client computers, or to perform an overall task as part of a distributed system. For example, various aspects of the invention may be performed on a client-server or multi-tier system that includes components distributed among one or more server systems that perform various functions according to various embodiments of the invention. These components may be executable, intermediate (e.g., IL) or interpreted (e.g., Java) code which communicate over a communication network (e.g., the Internet) using a communication protocol (e.g., TCP/IP). For example, one or more database servers may be used to store device data that is used in designing layouts, and one or more servers may be used to efficiently perform cooling calculations associated with embodiments of the present invention.
  • It should be appreciated that the invention is not limited to executing on any particular system or group of systems. Also, it should be appreciated that the invention is not limited to any particular distributed architecture, network, or communication protocol.
  • Various embodiments of the present invention may be programmed using an object-oriented programming language, such as SmallTalk, Java, C++, Ada, or C# (C-Sharp). Other object-oriented programming languages may also be used. Alternatively, functional, scripting, and/or logical programming languages may be used. Various aspects of the invention may be implemented in a non-programmed environment (e.g., documents created in HTML, XML or other format that, when viewed in a window of a browser program, render aspects of a graphical-user interface (GUI) or perform other functions). Various aspects of the invention may be implemented as programmed or non-programmed elements, or any combination thereof.
  • A variety of system elements including inputs, outputs and interfaces may exchange information with various external entities that may be information providers and/or information consumers. These external entities may include users and/or systems. Each of these system elements may both restrict exchanged information to a predefined set of values and validate any information exchanged prior to using the information or providing the information to other components.
  • In embodiments of the invention discussed above, systems and methods are described that provide indications of remaining cooling capacity for equipment enclosures. The indication of remaining cooling capacity may be a direct indication of remaining cooling in terms of, for example, kilowatts or BTU per hour, or the indication may be indirect such as providing the total capacity of cooling available to an enclosure along with an indication of how much cooling is being used, for example, in terms of percentage. Further, calculated values, including the capture index and the recirculation index may be used to determine the sufficiency of a particular design and to determine additional cooling capacity before a warning or error condition will result.
  • Embodiments of a systems and methods described above are generally described for use in relatively large data centers having numerous equipment racks, however, embodiments of the invention may also be used with smaller data centers and with facilities other than data centers. Further, as discussed above, embodiments of the present invention may be used with facilities having raised floors as well as with facilities that do not have a raised floor.
  • In embodiments of the present invention discussed above, results of analyses are described as being provided in real-time. As understood by those skilled in the art, the use of the term real-time is not meant to suggest that the results are available immediately, but rather, are available quickly giving a designer the ability to try a number of different designs over a short period of time, such as a matter of minutes.
  • Having thus described several aspects of at least one embodiment of this invention, it is to be appreciated various alterations, modifications, and improvements will readily occur to those skilled in the art. Such alterations, modifications, and improvements are intended to be part of this disclosure, and are intended to be within the spirit and scope of the invention. Accordingly, the foregoing description and drawings are by way of example only.

Claims (32)

1. A computer-implemented method for providing a representation of capacity of a data center resource, the method comprising acts of:
determining an first indication of excess capacity of a first data center resource that is unusable due to insufficient capacity of at least one other data center resource associated with the first data center resource; and
providing the first indication of excess capacity of the first data center resource to an external entity.
2. The method according to claim 1, wherein the act of determining the first indication of excess capacity includes an act of determining an amount of excess capacity of the first data center resource that is unusable due to insufficient capacity of the at least one other data center resource associated with the first data center resource.
3. The method according to claim 1, wherein the act of determining the first indication of excess capacity includes an act of determining an indication of excess capacity at a location in a rack.
4. The method according to claim 3, wherein the act of determining the first indication of excess capacity at the location in the rack includes an act of determining an indication of excess capacity at a U space position.
5. The method according to claim 1, wherein the act of determining the first indication of excess capacity includes an act of determining an indication of excess capacity of rack space that is unusable due to insufficient capacity of power.
6. The method according to claim 1, wherein the act of determining the first indication of excess capacity includes an act of determining an indication of excess capacity of power that is unusable due to insufficient capacity of cooling.
7. The method according to claim 1, wherein the act of determining the first indication of excess capacity includes an act of determining an indication of excess capacity of power distribution that is unusable due to insufficient capacity of power available for distribution.
8. The method according to claim 1, wherein the act of determining the first indication of excess capacity includes an act of determining an indication of excess capacity of physical space that is unusable due to insufficient capacity of cooling.
9. The method according to claim 1, wherein the act of determining the first indication of excess capacity includes an act of determining an indication of excess capacity of power distribution that is unusable due to insufficient capacity of network connectivity.
10. The method according to claim 1, wherein the act of determining the first indication of excess capacity includes an act of determining an indication of excess capacity of U space that is unusable due to insufficient capacity of weight support.
11. The method according to claim 1, wherein the act of providing the first indication of excess capacity includes an act of providing the indication to another system element.
12. The method according to claim 1, wherein the act of providing the first indication of excess capacity includes an act of presenting the first indication to a user of a computer system.
13. The method according to claim 12, wherein the act of presenting the first indication to the user of the computer system includes an act of presenting, to the user of the computer system, an amount of excess capacity of the first data center resource that is unusable due to insufficient capacity of the at least one other data center resource associated with the first data center resource.
14. The method according to claim 13, wherein the act of presenting, to the user of the computer system, the amount of excess capacity includes acts of:
presenting, to the user of the computer system, an identifier that identifies the first data center resource; and
presenting, to the user of the computer system, an identifier that identifies the at least one other data center resource.
15. The method according to claim 1, further comprising:
determining a configuration of data center equipment that minimizes, relative to at least one other configuration, the excess capacity of the first data center resource; and
providing the configuration of data center equipment to an external entity.
16. The method according to claim 15, wherein determining the configuration of data center equipment includes determining a location for at least one element of data center equipment.
17. The method according to claim 1, further comprising:
determining a second indication of excess capacity of a second data center resource that is unusable due to insufficient capacity of at least one other data center resource associated with the second data center resource;
determining a configuration of data center equipment that minimizes, relative to at least one other configuration, an aggregate of the first indication and the second indication; and
providing the configuration of data center equipment to an external entity.
18. A computer-implemented method for providing a representation of a data center resource, the method comprising acts of:
determining a status of capacity for the data center resource provided to either a rack or a location in a rack; and
providing the status to an external entity.
19. The method according to claim 18, wherein determining the status of capacity for the data center resource includes determining a status for the data center resource provided to a U space position.
20. The method according to claim 18, wherein the act of determining the status of capacity for the data center resource includes an act of determining a status of capacity of power.
21. The method according to claim 18, wherein the act of determining the status of capacity for the data center resource includes an act of determining a status of capacity of cooling.
22. The method according to claim 18, wherein the act of determining the status of capacity for the data center resource includes an act of determining a status of capacity of physical space.
23. The method according to claim 18, wherein the act of determining the status of capacity for the data center resource includes an act of determining a status of capacity of weight support.
24. The method according to claim 18, wherein the act of determining the status of capacity for the data center resource includes an act of determining a status of available capacity.
25. The method according to claim 18, wherein the act of determining the status of capacity for the data center resource includes an act of determining a status of utilized capacity.
26. A computer-implemented method for providing a representation of a data center resource, the method comprising the acts of:
determining a rate of change of a utilized capacity of the data center resource based at least in part on changes to a data center configuration; and
providing a predicted utilization of the data center resource based at least in part on the rate of change.
27. The method according to claim 26, wherein determining the rate of change of utilized capacity of the data center resource includes determining the rate of change using linear regression.
28. The method according to claim 26, wherein the data center resource includes power.
29. The method according to claim 26, wherein the data center resource includes cooling.
30. A system for providing a representation of a capacity of data center resource, the system comprising:
an input configured to receive a capacity of a first data center resource and a capacity of at least one other data center resource;
an output configured to provide an amount of excess capacity of the first data center resource; and
a controller coupled to the input and the output and configured to determine an amount of excess capacity of the first data center resource that is unusable due to insufficient capacity of the at least one other data center resource.
31. A system for providing a representation of a data center resource, the system comprising:
an input configured to receive capacity information for a data center resource;
an output configured to provide a status of capacity for the data center resource; and
a controller coupled to the input and the output and configured to determine the status of capacity for the data center resource at either a rack or a location in a rack.
32. A system for providing an evaluation of a data center resource, the system comprising:
an input configured to receive changes in a data center configuration;
an output configured to provide a predicted utilization of the data center resource based on a rate of change of utilized capacity of the data center resource; and
a controller configured to determine the rate of change of the utilized capacity of the data center resource based on the changes in a data center configuration.
US12/121,113 2007-05-15 2008-05-15 Methods and systems for managing facility power and cooling Abandoned US20090138313A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US12/121,113 US20090138313A1 (en) 2007-05-15 2008-05-15 Methods and systems for managing facility power and cooling
US16/019,052 US11076507B2 (en) 2007-05-15 2018-06-26 Methods and systems for managing facility power and cooling
US17/351,636 US11503744B2 (en) 2007-05-15 2021-06-18 Methods and systems for managing facility power and cooling

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US93803407P 2007-05-15 2007-05-15
US12/121,113 US20090138313A1 (en) 2007-05-15 2008-05-15 Methods and systems for managing facility power and cooling

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US16/019,052 Continuation US11076507B2 (en) 2007-05-15 2018-06-26 Methods and systems for managing facility power and cooling

Publications (1)

Publication Number Publication Date
US20090138313A1 true US20090138313A1 (en) 2009-05-28

Family

ID=39874103

Family Applications (3)

Application Number Title Priority Date Filing Date
US12/121,113 Abandoned US20090138313A1 (en) 2007-05-15 2008-05-15 Methods and systems for managing facility power and cooling
US16/019,052 Active US11076507B2 (en) 2007-05-15 2018-06-26 Methods and systems for managing facility power and cooling
US17/351,636 Active 2028-05-16 US11503744B2 (en) 2007-05-15 2021-06-18 Methods and systems for managing facility power and cooling

Family Applications After (2)

Application Number Title Priority Date Filing Date
US16/019,052 Active US11076507B2 (en) 2007-05-15 2018-06-26 Methods and systems for managing facility power and cooling
US17/351,636 Active 2028-05-16 US11503744B2 (en) 2007-05-15 2021-06-18 Methods and systems for managing facility power and cooling

Country Status (10)

Country Link
US (3) US20090138313A1 (en)
EP (1) EP2147585B1 (en)
JP (1) JP5559040B2 (en)
CN (1) CN101755495B (en)
AU (1) AU2008255030B2 (en)
BR (1) BRPI0812116B1 (en)
CA (1) CA2686564C (en)
DK (1) DK2147585T3 (en)
ES (1) ES2612328T3 (en)
WO (1) WO2008144375A2 (en)

Cited By (123)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090006222A1 (en) * 2006-01-30 2009-01-01 L'air Liquide Societe Anonyme Pour L'etude Et L'exploitation Des Procedes Georges Claude System for the Operation and Management of a Fleet of Refrigerated Autonomous Containers
US20090132097A1 (en) * 2007-11-20 2009-05-21 Ratnesh Kumar Sharma Virtual cooling infrastructure
US20090210099A1 (en) * 2008-02-15 2009-08-20 Sawczak Stephen D Systems and methods for computer equipment management
US20090243846A1 (en) * 2008-03-28 2009-10-01 Fujitsu Limited Electronic apparatus system having a plurality of rack-mounted electronic apparatuses, and method for identifying electronic apparatus in electronic apparatus system
US20090327012A1 (en) * 2008-06-30 2009-12-31 Ratnesh Kumar Sharma Cooling resource capacity allocation using lagrange multipliers
US20100106464A1 (en) * 2008-10-27 2010-04-29 Christopher Hlasny Method for designing raised floor and dropped ceiling in computing facilities
US20100115305A1 (en) * 2008-11-03 2010-05-06 Hitachi, Ltd. Methods and Apparatus to Provision Power-Saving Storage System
US20100131109A1 (en) * 2008-11-25 2010-05-27 American Power Conversion Corporation System and method for assessing and managing data center airflow and energy usage
US7752088B1 (en) * 2007-05-29 2010-07-06 Joseph Coschera On site physical inventory method
US20100179695A1 (en) * 2009-01-15 2010-07-15 Dell Products L.P. System and Method for Temperature Management of a Data Center
US20100256959A1 (en) * 2009-04-01 2010-10-07 American Power Conversion Corporation Method for computing cooling redundancy at the rack level
US20100286956A1 (en) * 2009-05-08 2010-11-11 American Power Conversion Corporation System and method for predicting cooling performance of arrangements of equipment in a data center
US20100287018A1 (en) * 2009-05-08 2010-11-11 American Power Conversion Corporation System and method for arranging equipment in a data center
US20100286955A1 (en) * 2009-05-08 2010-11-11 American Power Conversion Corporation System and method for predicting maximum cooler and rack capacities in a data center
US20100298990A1 (en) * 2009-05-21 2010-11-25 Fujitsu Limited Apparatus and method for controlling an open amount of a plurality of air transfer grilles
US20110016342A1 (en) * 2009-07-20 2011-01-20 Viridity Software, Inc. Techniques for power analysis
US20110040532A1 (en) * 2009-08-12 2011-02-17 International Business Machines Corporation Knowledge-Based Models for Data Centers
US20110040529A1 (en) * 2009-08-12 2011-02-17 International Business Machines Corporation Methods and Techniques for Creating and Visualizing Thermal Zones
US20110060571A1 (en) * 2009-09-04 2011-03-10 Fujitsu Limited Thermal-fluid-simulation analyzing apparatus
US7925746B1 (en) * 2008-10-06 2011-04-12 United Services Automobile Association (Usaa) Systems and methods for a self-contained, cooled server rack
US20110208622A1 (en) * 2010-02-25 2011-08-25 International Business Machines Corporation Data center power cost accounting system
CN102169527A (en) * 2010-02-26 2011-08-31 国际商业机器公司 Method and system for determining mounting machine frame for equipment in data center
US20110244779A1 (en) * 2010-04-02 2011-10-06 Fujitsu Limited Air conditioning system and air conditioning control method
US20110258302A1 (en) * 2010-04-20 2011-10-20 No Limits Software, LLC System And Method For Remotely Determining Identification And Physical Location Of Equipment In A Rack
US20110303406A1 (en) * 2010-06-11 2011-12-15 Fujitsu Limited Air-conditioning system and control device thereof
US20120040601A1 (en) * 2010-08-10 2012-02-16 Yamatake Corporation Air conditioning controlling device and method
US20120072195A1 (en) * 2010-09-18 2012-03-22 International Business Machines Corporation Modeling Movement Of Air Under A Floor Of A Data Center
US20120123898A1 (en) * 2010-11-17 2012-05-17 International Business Machines Corporation Generating configuration options for a computing system
US20120170205A1 (en) * 2010-12-30 2012-07-05 American Power Conversion Corporation System and method for sequential placement of cooling resources within data center layouts
US20120209573A1 (en) * 2011-02-11 2012-08-16 International Business Machines Corporation Data center design tool
US20120215359A1 (en) * 2011-02-21 2012-08-23 Amir Meir Michael Adaptive fan control based on server configuration
US20120271935A1 (en) * 2011-04-19 2012-10-25 Moon Billy G Coordinating data center compute and thermal load based on environmental data forecasts
US20120284216A1 (en) * 2009-08-12 2012-11-08 International Business Machines Corporation Knowledge-Based Models for Data Centers
US8315841B2 (en) 2005-05-02 2012-11-20 American Power Conversion Corporation Methods and systems for managing facility power and cooling
US20120303339A1 (en) * 2011-05-27 2012-11-29 International Business Machines Corporation Computational fluid dynamics modeling of a bounded domain
US8322155B2 (en) 2006-08-15 2012-12-04 American Power Conversion Corporation Method and apparatus for cooling
US8327656B2 (en) 2006-08-15 2012-12-11 American Power Conversion Corporation Method and apparatus for cooling
WO2012082985A3 (en) * 2010-12-16 2013-02-28 American Power Conversion Corporation System and methods for rack cooling analysis
US8424336B2 (en) 2006-12-18 2013-04-23 Schneider Electric It Corporation Modular ice storage for uninterruptible chilled water
US8425287B2 (en) 2007-01-23 2013-04-23 Schneider Electric It Corporation In-row air containment and cooling system and method
US20130204984A1 (en) * 2012-02-08 2013-08-08 Oracle International Corporation Management Record Specification for Management of Field Replaceable Units Installed Within Computing Cabinets
US8509959B2 (en) 2010-08-12 2013-08-13 Schneider Electric It Corporation System and method for predicting transient cooling performance for a data center
US8560677B2 (en) 2009-02-13 2013-10-15 Schneider Electric It Corporation Data center control
WO2013171757A2 (en) 2012-05-18 2013-11-21 Tata Consultancy Services Limited Method and system for determining and implementing a viable containment design of a data center
US8595515B1 (en) 2007-06-08 2013-11-26 Google Inc. Powering a data center
US8597054B2 (en) 2011-03-07 2013-12-03 Schneider Electric It Corporation CAN bus automatic line termination
US20140005982A1 (en) * 2010-06-22 2014-01-02 International Business Machines Corporation Data center physical infrastructure threshold analysis
US8628153B2 (en) 2010-05-13 2014-01-14 Pandult Corp. Aisle containment system
US8628154B2 (en) 2010-05-13 2014-01-14 Panduit Corp. Aisle containment system
US8628158B2 (en) 2010-05-13 2014-01-14 Panduit Corp. Aisle containment system
US8639482B2 (en) 2005-05-02 2014-01-28 Schneider Electric It Corporation Methods and systems for managing facility power and cooling
WO2014041554A1 (en) 2012-09-12 2014-03-20 Tata Consultancy Services Limited A method for efficient designing and operating cooling infrastructure in a data center
WO2013095623A3 (en) * 2011-12-23 2014-04-10 Schneider Electric It Corporation Dropped ceiling system cooling prediction
US8725307B2 (en) 2011-06-28 2014-05-13 Schneider Electric It Corporation System and method for measurement aided prediction of temperature and airflow values in a data center
US8744631B2 (en) 2011-01-28 2014-06-03 Hewlett-Packard Development Company, L.P. Manipulating environmental conditions in an infrastructure
GB2513141A (en) * 2013-04-17 2014-10-22 Ibm Data processing system with real-time data center air flow simulator
US8879247B2 (en) 2010-07-21 2014-11-04 International Business Machines Corporation Computer chassis cooling sidecar
US8914495B2 (en) 2011-06-07 2014-12-16 International Business Machines Corporation Automatically detecting and locating equipment within an equipment rack
US20150026621A1 (en) * 2013-07-18 2015-01-22 Netapp, Inc. System and Method for Planning an Implementation of a Computing System
US20150032283A1 (en) * 2013-07-29 2015-01-29 International Business Machines Corporation Data center cooling
US8949081B2 (en) 2010-12-21 2015-02-03 Schneider Electric It Corporation Method for computing cooling redundancy at the rack level
US8947880B2 (en) 2010-08-06 2015-02-03 Lenovo Enterprise Solutions (Singapore) Ptd. Ltd. Hot or cold aisle computer chassis
US8972217B2 (en) 2010-06-08 2015-03-03 Schneider Electric It Corporation System and method for predicting temperature values in a data center
US8996180B2 (en) 2010-09-17 2015-03-31 Schneider Electric It Corporation System and method for predicting perforated tile airflow in a data center
US9009500B1 (en) 2012-01-18 2015-04-14 Google Inc. Method of correlating power in a data center by fitting a function to a plurality of pairs of actual power draw values and estimated power draw values determined from monitored CPU utilization of a statistical sample of computers in the data center
US20150154770A1 (en) * 2008-03-11 2015-06-04 Enphase Energy, Inc. Method and apparatus for electrical power visualization
WO2015095584A1 (en) * 2013-12-19 2015-06-25 Panduit Corp. Methods and systems for identifying stranded capacity
US20150192316A1 (en) * 2012-07-06 2015-07-09 Rongliang Zhou Management of airflow provisioning to meet a cooling influence redundancy level
US20150200916A1 (en) * 2012-07-24 2015-07-16 Zuken Inc. Information Administration System
US20150250077A1 (en) * 2012-11-02 2015-09-03 Fujitsu Limited Modular data center and controlling method of modular data center
US20150256386A1 (en) * 2014-03-06 2015-09-10 Dell Products, Lp System and Method for Providing a Server Rack Management Controller
US20150253028A1 (en) * 2014-03-06 2015-09-10 Dell Products, Lp System and Method for Data Reporting in a Tile Management Controller
WO2015168565A1 (en) * 2014-05-01 2015-11-05 Amazon Technologies, Inc. Flexible data center infrastructure distribution
US20150362981A1 (en) * 2012-07-03 2015-12-17 Eaton Industries (France) Sas Grouping power sources and power loads to groups that share similar properties
US9223905B2 (en) 2011-03-25 2015-12-29 Schneider Electric It Corporation Systems and methods for predicting fluid dynamics in a data center
US20150378414A1 (en) * 2008-04-09 2015-12-31 Hitachi, Ltd. Operations management methods and devices thereof in information-processing systems
US9287710B2 (en) 2009-06-15 2016-03-15 Google Inc. Supplying grid ancillary services using controllable loads
US20160117420A1 (en) * 2008-09-29 2016-04-28 International Business Machines Corporation System and method for dynamically modeling data center partitions
US20160188764A1 (en) * 2014-12-30 2016-06-30 Schneider Electric It Corporation Raised floor plenum tool
US20160249487A1 (en) * 2013-10-04 2016-08-25 Tata Consultancy Services Limited System and method for optimizing cooling efficiency of a data center
US9448841B2 (en) * 2014-03-17 2016-09-20 Ca, Inc. Decommissioning physical host machines after relocating unproductive virtual machines therefrom
US9519517B2 (en) 2009-02-13 2016-12-13 Schneider Electtic It Corporation Data center control
US9568206B2 (en) 2006-08-15 2017-02-14 Schneider Electric It Corporation Method and apparatus for cooling
US9658662B2 (en) 2010-10-12 2017-05-23 Hewlett Packard Enterprise Development Lp Resource management for data centers
US9703665B1 (en) * 2010-02-19 2017-07-11 Acuity Holdings, Inc. Data center design process and system
US20170242446A1 (en) * 2015-04-27 2017-08-24 Hewlett Packard Enterprise Development Lp Virtualized fan speed measurement
US9753465B1 (en) 2009-07-21 2017-09-05 The Research Foundation For The State University Of New York Energy aware processing load distribution system and method
US9754055B1 (en) * 2009-05-06 2017-09-05 Amdocs Software Systems Limited System, method, and computer program product for managing an area for positioning resources, based on parameters of the resources
US9779275B2 (en) * 2014-07-03 2017-10-03 Mitsubishi Electric Corporation Control system, terminal, information setting method, and program
US9778718B2 (en) 2009-02-13 2017-10-03 Schneider Electric It Corporation Power supply and data center control
US9795063B1 (en) 2015-05-07 2017-10-17 Dhk Storage, Llc Computer server heat regulation utilizing integrated precision air flow
US20170303438A1 (en) * 2016-04-19 2017-10-19 Dell Products L.P. Systems and methods for fan performance-based scaling of thermal control parameters
US9830410B2 (en) 2011-12-22 2017-11-28 Schneider Electric It Corporation System and method for prediction of temperature values in an electronics system
US9851726B2 (en) 2013-09-04 2017-12-26 Panduit Corp. Thermal capacity management
US9939796B2 (en) 2008-09-29 2018-04-10 International Business Machines Corporation System and method to dynamically change data center partitions
US9948522B2 (en) * 2016-01-11 2018-04-17 Equinix, Inc. Associating infrastructure assets in a data center
US9952103B2 (en) 2011-12-22 2018-04-24 Schneider Electric It Corporation Analysis of effect of transient events on temperature in a data center
US10001761B2 (en) 2014-12-30 2018-06-19 Schneider Electric It Corporation Power consumption model for cooling equipment
US10082857B1 (en) * 2012-08-07 2018-09-25 Amazon Technologies, Inc. Cooling electrical systems based on power measurements
US20180359201A1 (en) 2017-06-09 2018-12-13 Equinix, Inc. Near real-time messaging service for data center infrastructure monitoring data
US20180375741A1 (en) * 2017-06-26 2018-12-27 Seagate Technology Llc Datacenter service flow optimization
US10180261B1 (en) * 2015-12-28 2019-01-15 Amazon Technologies, Inc. Model based cooling control system
US10210288B2 (en) 2012-12-27 2019-02-19 Schneider Electric It Corporation Systems and methods of visualizing airflow
US20190057181A1 (en) * 2017-08-18 2019-02-21 International Business Machines Corporation System and method for design optimization using augmented reality
WO2019051077A1 (en) * 2017-09-06 2019-03-14 Vertiv Corporation Cooling unit energy optimization via smart supply air temperature setpoint control
US10310568B2 (en) 2013-02-28 2019-06-04 Oracle International Corporation Method for interconnecting field replaceable unit to power source of communication network
US10447785B2 (en) * 2014-11-17 2019-10-15 Lg Electronics Inc. Digital device and method for controlling same
US20200106677A1 (en) * 2018-09-28 2020-04-02 Hewlett Packard Enterprise Development Lp Data center forecasting based on operation data
US10617038B2 (en) 2016-07-08 2020-04-07 Schneider Electric It Corporation Zero-equation turbulence models for large electrical and electronics enclosure applications
US10819556B1 (en) 2017-10-16 2020-10-27 Equinix, Inc. Data center agent for data center infrastructure monitoring data access and translation
US20210014998A1 (en) * 2019-07-12 2021-01-14 Hewlett Packard Enterprise Development Lp Recommending it equipment placement based on inferred hardware capabilities
US10999954B2 (en) 2019-04-23 2021-05-04 Vertiv Corporation Modular roof mounted cooling system and method for data center
US11032948B1 (en) 2019-06-17 2021-06-08 Amazon Technologies, Inc. Pre-fabricated mechanical and electrical distribution infrastructure system
US11030332B1 (en) * 2015-04-13 2021-06-08 Wells Fargo Bank, N.A. Database controlled web service type architecture
US11049052B2 (en) * 2018-05-02 2021-06-29 International Business Machines Corporation Automated managing of a data center installation
US11076509B2 (en) 2017-01-24 2021-07-27 The Research Foundation for the State University Control systems and prediction methods for it cooling performance in containment
US11076507B2 (en) 2007-05-15 2021-07-27 Schneider Electric It Corporation Methods and systems for managing facility power and cooling
US20220183191A1 (en) * 2020-12-04 2022-06-09 Schneider Electric It Corporation It-room-cooling-performance assessment
US11382232B1 (en) 2019-03-28 2022-07-05 Amazon Technologies, Inc. Self-standing modular data center infrastructure system
US11388835B1 (en) * 2018-06-27 2022-07-12 Amazon Technologies, Inc. Placement of custom servers
US11422910B2 (en) * 2016-07-21 2022-08-23 Jpmorgan Chase Bank, N.A. Method and system for implementing a data center operating system
US11599532B1 (en) * 2021-08-11 2023-03-07 Amdocs Development Limited System, method, and computer program for preventing user mistakes when making database changes
US20230083377A1 (en) * 2021-09-16 2023-03-16 Dell Products L.P. Software-defined infrastructure for identifying and remediating an airflow deficiency scenario on a rack device

Families Citing this family (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4682241B2 (en) * 2008-12-04 2011-05-11 株式会社Nttファシリティーズ Operation management server and operation management method
JP5107952B2 (en) * 2009-03-04 2012-12-26 アラクサラネットワークス株式会社 Network system, connection device, management device, management method
JP5403673B2 (en) * 2009-08-14 2014-01-29 Kddi株式会社 Node installation method and node installation support apparatus in a data center
JP5403674B2 (en) * 2009-08-14 2014-01-29 Kddi株式会社 Air conditioning management method and air conditioning management device in data center
CN101895151B (en) * 2010-06-25 2012-10-10 世纪互联智慧科技有限公司 Quadruple supply system
CN101893922B (en) * 2010-06-25 2012-09-19 世纪互联智慧科技有限公司 Quadri-generation system
CN101895150B (en) * 2010-06-25 2012-07-25 世纪互联智慧科技有限公司 Quadruple co-generation system
WO2012147131A1 (en) * 2011-04-27 2012-11-01 Hitachi, Ltd. Management computer, computer system and method for allocating it equipment in a data center
US8918794B2 (en) * 2011-08-25 2014-12-23 Empire Technology Development Llc Quality of service aware captive aggregation with true datacenter testing
IN2014DN05662A (en) 2011-12-23 2015-04-03 Schneider Electric It Corp
US20130204593A1 (en) * 2012-01-31 2013-08-08 Panduit Corp. Computational Fluid Dynamics Systems and Methods of Use Thereof
EP3902383A1 (en) 2012-09-21 2021-10-27 Schneider Electric IT Corporation Method and apparatus for characterizing server thermal transient performance
JP5433756B2 (en) * 2012-10-04 2014-03-05 アラクサラネットワークス株式会社 Display device, connection device, and display method
US10157245B2 (en) 2012-10-31 2018-12-18 Schneider Electric It Corporation System and method for fluid dynamics prediction with an enhanced potential flow model
CN105452984B (en) 2013-06-28 2019-03-08 施耐德电气It公司 Calculate the power supply service efficiency in data center
EP3140701A4 (en) * 2014-05-05 2017-05-03 Vigilent Corporation Operational performance-weighted redundancy for environmental control systems
JP6357946B2 (en) * 2014-07-25 2018-07-18 富士通株式会社 Output program, output device, and output method
US11137161B2 (en) 2017-03-30 2021-10-05 Samsung Electronics Co., Ltd. Data learning server and method for generating and using learning model thereof
CN107655140B (en) * 2017-08-17 2019-11-26 华为技术有限公司 A kind of air conditioning control method and device
US10817033B2 (en) 2017-12-14 2020-10-27 Schneider Electric It Corporation Method and system for predicting effect of a transient event on a data center
US11093563B2 (en) * 2018-02-05 2021-08-17 Microsoft Technology Licensing, Llc Sharing measured values of physical space parameters
US10970128B2 (en) * 2018-04-13 2021-04-06 Samsung Electronics Co., Ltd. Server, air conditioner and method for controlling thereof
JP2020194239A (en) * 2019-05-24 2020-12-03 株式会社日立製作所 Energy facility planning device and energy facility planning method
KR102510999B1 (en) * 2020-11-09 2023-03-16 주식회사 케이티 Server, method and computer program for recommending space layout in data center
EP4040355A1 (en) * 2021-02-08 2022-08-10 Tata Consultancy Services Limited System and method for measuring user experience of information visualizations
US11853936B2 (en) 2021-06-14 2023-12-26 Kyndryl, Inc. Minimizing the environmental impact of workloads
GB2608531A (en) * 2022-09-15 2023-01-04 Ekkosense Ltd Data centre cooling optimisation

Citations (82)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4823290A (en) * 1987-07-21 1989-04-18 Honeywell Bull Inc. Method and apparatus for monitoring the operating environment of a computer system
US4827151A (en) * 1987-02-20 1989-05-02 Kabushiki Kaisha Toshiba Uninterruptible power supply utilizing a synchronized chopper for power factor improvement
US4962734A (en) * 1990-03-14 1990-10-16 Paccar Inc. Electrically driven, circumferentially supported fan
US5367670A (en) * 1991-06-24 1994-11-22 Compaq Computer Corporation Computer system manager for monitoring events and operating parameters and generating alerts
US5381554A (en) * 1992-11-24 1995-01-10 Exide Electronics Uninterrupted power supply (UPS) system interfacing with communications network
US5404136A (en) * 1988-08-23 1995-04-04 Marsden; Derek R. Method and apparatus for monitoring the consumption of utilities in business premises
US5462225A (en) * 1994-02-04 1995-10-31 Scientific-Atlanta, Inc. Apparatus and method for controlling distribution of electrical energy to a space conditioning load
US5581478A (en) * 1995-04-13 1996-12-03 Cruse; Michael Facility environmental control system
US5657641A (en) * 1995-09-13 1997-08-19 Kooltronic, Inc. Panel mounted cooling system
US5682949A (en) * 1992-05-22 1997-11-04 Globalmic, Inc. Energy management system
US5850539A (en) * 1995-05-23 1998-12-15 Compaq Computer Corporation Automated system for facilitating creation of a rack-mountable component personal computer
US5978594A (en) * 1994-09-30 1999-11-02 Bmc Software, Inc. System for managing computer resources across a distributed computing environment by first reading discovery information about how to determine system resources presence
US5987614A (en) * 1997-06-17 1999-11-16 Vadem Distributed power management system and method for computer
US5995729A (en) * 1994-07-22 1999-11-30 Hitachi, Ltd. Method and apparatus for aiding configurating management of a computer system
US6134511A (en) * 1998-04-15 2000-10-17 Subbarao; Krishnappa Method and apparatus for improving building energy simulations
US6216956B1 (en) * 1997-10-29 2001-04-17 Tocom, Inc. Environmental condition control and energy management system and method
US6246969B1 (en) * 1998-09-08 2001-06-12 International Business Machines Corporation Method and apparatus for testing computer cooling systems
US20010005894A1 (en) * 1999-12-27 2001-06-28 Masahiro Fukui Remote power management system of information processing apparatus or the like
US20010042616A1 (en) * 2000-03-21 2001-11-22 Baer Daniel B. Method and apparatus for cooling electronic enclosures
US20020004912A1 (en) * 1990-06-01 2002-01-10 Amphus, Inc. System, architecture, and method for logical server and other network devices in a dynamically configurable multi-server network environment
US6374627B1 (en) * 2001-01-09 2002-04-23 Donald J. Schumacher Data center cooling system
US20020059804A1 (en) * 2000-02-18 2002-05-23 Toc Technology, Llc Computer room air flow
US20020062454A1 (en) * 2000-09-27 2002-05-23 Amphus, Inc. Dynamic power and workload management for multi-server system
US20020072868A1 (en) * 2000-07-13 2002-06-13 Bartone Erik J. System and method for monitoring and controlling energy usage
US6459579B1 (en) * 2001-01-03 2002-10-01 Juniper Networks, Inc. Apparatus and method for directing airflow in three dimensions to cool system components
US20030084359A1 (en) * 2001-10-31 2003-05-01 Bresniker Kirk M. System and method for intelligent control of power consumption of distributed services during periods when power consumption must be reduced
US20030084358A1 (en) * 2001-10-31 2003-05-01 Bresniker Kirk M. System and method for intelligent control of power consumption of distributed services during periods of reduced load
US20030084357A1 (en) * 2001-10-31 2003-05-01 Bresniker Kirk M. System and method for providing minimal power-consuming redundant computing hardware for distributed services
US6574104B2 (en) * 2001-10-05 2003-06-03 Hewlett-Packard Development Company L.P. Smart cooling of data centers
US20030115000A1 (en) * 2001-12-13 2003-06-19 Bodas Devadatta V. Method to estimate power and cooling requirements of systems
US20030115024A1 (en) * 2001-12-17 2003-06-19 Snevely Robert D. Designing a data center
US20030120780A1 (en) * 2001-12-21 2003-06-26 Xiaoyun Zhu Network resource assignment system and method
US20030147216A1 (en) * 2002-02-06 2003-08-07 Patel Chandrakant D. Method, apparatus, and system for cooling electronic components
US20030158718A1 (en) * 2002-02-19 2003-08-21 Nakagawa Osamu Samuel Designing layout for internet datacenter cooling
US20030193777A1 (en) * 2002-04-16 2003-10-16 Friedrich Richard J. Data center energy management system
US20030196126A1 (en) * 2002-04-11 2003-10-16 Fung Henry T. System, method, and architecture for dynamic server power management and dynamic workload management for multi-server environment
US20030204756A1 (en) * 1997-02-12 2003-10-30 Ransom Douglas S. Push communications architecture for intelligent electronic devices
US6672955B2 (en) * 2001-09-07 2004-01-06 International Business Machines Corporation Air flow management system for an internet data center
US20040020224A1 (en) * 2002-08-02 2004-02-05 Bash Cullen E. Cooling system with evaporators distributed in parallel
US6694759B1 (en) * 2003-01-27 2004-02-24 Hewlett-Packard Development Company, L.P. Pressure control of cooling fluid within a plenum using automatically adjustable vents
US6714977B1 (en) * 1999-10-27 2004-03-30 Netbotz, Inc. Method and system for monitoring computer networks and equipment
US6718277B2 (en) * 2002-04-17 2004-04-06 Hewlett-Packard Development Company, L.P. Atmospheric control within a building
US20040065104A1 (en) * 2002-10-03 2004-04-08 Bash Cullen E. Cooling of data centers
US6721672B2 (en) * 2002-01-02 2004-04-13 American Power Conversion Method and apparatus for preventing overloads of power distribution networks
US20040083012A1 (en) * 2002-10-28 2004-04-29 Miller John P. Method of modeling and sizing a heat exchanger
US20040089011A1 (en) * 2002-08-02 2004-05-13 Patel Chandrakant D. Cooling system
US20040120331A1 (en) * 2002-12-20 2004-06-24 Rhine Scott Alan Systems and methods for rapid selection of devices in a tree topology network
US20040163001A1 (en) * 2003-02-14 2004-08-19 Bodas Devadatta V. Enterprise power and thermal management
US6795928B2 (en) * 2002-03-18 2004-09-21 International Business Machines Corporation Method for managing power consumption of multiple computer servers
US20040189161A1 (en) * 2003-03-28 2004-09-30 Davis Brooks I. Zero rack unit space utilization
US6819563B1 (en) * 2003-07-02 2004-11-16 International Business Machines Corporation Method and system for cooling electronics racks using pre-cooled air
US20040240514A1 (en) * 2003-05-29 2004-12-02 Bash Cullen Edwin Air re-circulation index
US20040262409A1 (en) * 2003-06-30 2004-12-30 Crippen Martin Joseph Thermal analysis in a data processing system
US20050016189A1 (en) * 2003-07-25 2005-01-27 Honeywell International Inc. Method and apparatus for controlling humidity with an air conditioner
US20050023363A1 (en) * 2003-05-29 2005-02-03 Sharma Ratnesh K. CRAC unit control based on re-circulation index
US6859366B2 (en) * 2003-03-19 2005-02-22 American Power Conversion Data center cooling system
US6862179B2 (en) * 2002-11-26 2005-03-01 Hewlett-Packard Development Company, L.P. Partition for varying the supply of cooling fluid
US20050055590A1 (en) * 2003-09-04 2005-03-10 Farkas Keith Istvan Application management based on power consumption
US20050108582A1 (en) * 2000-09-27 2005-05-19 Fung Henry T. System, architecture, and method for logical server and other network devices in a dynamically configurable multi-server network environment
US20050177755A1 (en) * 2000-09-27 2005-08-11 Amphus, Inc. Multi-server and multi-CPU power management system and method
US20050225936A1 (en) * 2002-03-28 2005-10-13 Tony Day Cooling of a data centre
US20050228618A1 (en) * 2004-04-09 2005-10-13 Patel Chandrakant D Workload placement among data centers based on thermal efficiency
US20050251802A1 (en) * 2004-05-08 2005-11-10 Bozek James J Dynamic migration of virtual machine computer programs upon satisfaction of conditions
US6964539B2 (en) * 2002-03-18 2005-11-15 International Business Machines Corporation Method for managing power consumption of multiple computer servers
US6967283B2 (en) * 2001-03-20 2005-11-22 American Power Conversion Corporation Adjustable scalable rack power system and method
US20050267639A1 (en) * 2004-05-28 2005-12-01 Sharma Ratnesh K Data center evaluation using an air re-circulation index
US20050278075A1 (en) * 2004-05-28 2005-12-15 Neil Rasmussen Methods and apparatus for providing and distributing standby power
US20060080001A1 (en) * 2004-10-08 2006-04-13 Bash Cullen E Correlation of vent tile settings and rack temperatures
US20060112286A1 (en) * 2004-11-23 2006-05-25 Whalley Ian N Method for dynamically reprovisioning applications and other server resources in a computer center in response to power and heat dissipation requirements
US20060115586A1 (en) * 2004-11-30 2006-06-01 Baozhong Xing Process and apparatus for coating a controlled release product in a rotating drum
US20060121421A1 (en) * 2004-10-15 2006-06-08 Spitaels James S IT equipment simulation
US20060139877A1 (en) * 2004-12-29 2006-06-29 Mark Germagian Rack height cooling
US7085133B2 (en) * 2003-07-09 2006-08-01 International Business Machines Corporation Cooling using complimentary tapered plenums
US20060168975A1 (en) * 2005-01-28 2006-08-03 Hewlett-Packard Development Company, L.P. Thermal and power management apparatus
US20060214014A1 (en) * 2005-03-25 2006-09-28 Bash Cullen E Temperature control using a sensor network
US20060271700A1 (en) * 2005-05-24 2006-11-30 Fujitsu Limited Record medium with a load distribution program recorded thereon, load distribution method, and load distribution apparatus
US7155318B2 (en) * 2004-11-05 2006-12-26 Hewlett-Packard Development Company, Lp. Air conditioning unit control to reduce moisture varying operations
US20070038414A1 (en) * 2005-05-02 2007-02-15 American Power Conversion Corporation Methods and systems for managing facility power and cooling
US20070078635A1 (en) * 2005-05-02 2007-04-05 American Power Conversion Corporation Methods and systems for managing facility power and cooling
US20070150584A1 (en) * 2005-12-28 2007-06-28 Deepa Srinivasan Apparatus, system, and method for determining server utilization in hosted computing infrastructure
US20070174024A1 (en) * 2005-05-02 2007-07-26 American Power Conversion Corporation Methods and systems for managing facility power and cooling
US20070274035A1 (en) * 2003-03-19 2007-11-29 Fink James R Data center cooling

Family Cites Families (841)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US1025243A (en) 1912-02-24 1912-05-07 Lewis A Carpenter Pipe-hanger.
US1941258A (en) 1931-08-14 1933-12-26 Superheater Co Ltd Interlocking hanger
US2254914A (en) 1939-09-09 1941-09-02 Westinghouse Electric & Mfg Co Disconnecting contact
US2455022A (en) 1944-08-08 1948-11-30 Benjamin F Schmidt Submersible double-acting fluid piston deep well pump
US3102184A (en) 1960-01-05 1963-08-27 Percival H Sherron Telephone booth heating means
US3317798A (en) 1966-04-13 1967-05-02 Ibm Cooling electrical apparatus
US3559728A (en) 1968-11-29 1971-02-02 Kooltronic Fan Co Electronic equipment rack temperature control
US3643007A (en) 1969-04-02 1972-02-15 Superior Continental Corp Coaxial cable
US3681936A (en) 1970-10-26 1972-08-08 Oklahoma Mfg Co Heat exchanger
JPS5040889B1 (en) 1970-11-30 1975-12-27
US3665725A (en) 1971-01-18 1972-05-30 Thermo King Corp Capacity control for compression expansion refrigeration systems
US3742725A (en) 1971-12-06 1973-07-03 Carrier Corp Air conditioning unit
US3810138A (en) 1972-01-10 1974-05-07 Westinghouse Electric Corp Interpolative sensor output visual map display system
FR2193303B1 (en) 1972-07-24 1976-01-16 Merlin Gerin
JPS5155922A (en) 1974-11-13 1976-05-17 Hitachi Ltd HENRYUKI
US3995446A (en) 1975-07-14 1976-12-07 Eubank Marcus P Reverse air cycle air conditioner
JPS5239140A (en) 1975-09-22 1977-03-26 Mitsubishi Electric Corp Non-interrupted powe source
US4055053A (en) 1975-12-08 1977-10-25 Elfving Thore M Thermoelectric water cooler or ice freezer
US4127008A (en) 1976-11-01 1978-11-28 Lewis Tyree Jr Method and apparatus for cooling material using liquid CO2
SE418736B (en) 1976-12-23 1981-06-22 Bofors Ab WHEN IN THE MANUFACTURE OF A CEMENT USE, INCLUDING CEMENT, SAND AND THE WATER INITIATE A RECOVERY OF DISTRIBUTED AIR IN USE
US4197716A (en) 1977-09-14 1980-04-15 Halstead Industries, Inc. Refrigeration system with auxiliary heat exchanger for supplying heat during defrost cycle and for subcooling the refrigerant during a refrigeration cycle
JPS5456503A (en) 1977-10-13 1979-05-07 Unitika Ltd Method of treating photosensitive resin plate
US4352274A (en) 1978-04-10 1982-10-05 Signetics Corporation Circuit board mounting and cooling assembly
US4223535A (en) 1978-12-22 1980-09-23 Kumm Emerson L Absorption solar powered air conditioning system with storage capacity
GB2043371B (en) 1979-02-21 1983-05-25 South Eastern Elec Board Load shedding
US4285205A (en) 1979-12-20 1981-08-25 Martin Leonard I Refrigerant sub-cooling
US4338794A (en) 1980-03-17 1982-07-13 Haasis Jr Hans High efficiency ice-making system
US4275570A (en) 1980-06-16 1981-06-30 Vilter Manufacturing Corporation Oil cooling means for refrigeration screw compressor
US4568934A (en) 1981-06-09 1986-02-04 Adec, Inc. Computer controlled energy monitoring system
US4551009A (en) 1981-12-21 1985-11-05 Mita Industrial Co., Ltd. Electrostatic copying apparatus
US4419865A (en) 1981-12-31 1983-12-13 Vilter Manufacturing Company Oil cooling apparatus for refrigeration screw compressor
US4709202A (en) 1982-06-07 1987-11-24 Norand Corporation Battery powered system
JPS5938897A (en) 1982-08-27 1984-03-02 ニツタン株式会社 Abnormality monitor
US4590538A (en) 1982-11-18 1986-05-20 Cray Research, Inc. Immersion cooled high density electronic assembly
JPS59151478A (en) 1983-02-17 1984-08-29 Omron Tateisi Electronics Co Manufacture of surface light-emitting device
US4521645A (en) 1983-06-16 1985-06-04 Carroll Robert A Fire alarm system
US4747041A (en) 1983-06-27 1988-05-24 Unisys Corporation Automatic power control system which automatically activates and deactivates power to selected peripheral devices based upon system requirement
US4637020A (en) 1983-08-01 1987-01-13 Fairchild Semiconductor Corporation Method and apparatus for monitoring automated testing of electronic circuits
US4515746A (en) 1983-09-06 1985-05-07 General Electric Company Microcomposite of metal carbide and ceramic particles
US4564767A (en) 1983-11-07 1986-01-14 Tii Industries, Inc. Uninterruptible switching power supply system
US4599873A (en) 1984-01-31 1986-07-15 Hyde Robert E Apparatus for maximizing refrigeration capacity
US4718249A (en) 1984-04-16 1988-01-12 Hanson Wallace G Apparatus for heating and cooling
US4535598A (en) 1984-05-14 1985-08-20 Carrier Corporation Method and control system for verifying sensor operation in a refrigeration system
JPS615376A (en) 1984-06-19 1986-01-11 Toshiba Corp Picture processor
FR2566576B1 (en) 1984-06-20 1987-01-16 Cit Alcatel DEVICE FOR CONTROLLING THE CONNECTION OF AN ELECTRICAL CIRCUIT TO A NETWORK
JPH038038Y2 (en) 1984-11-05 1991-02-27
KR900001842B1 (en) 1984-11-15 1990-03-24 후지쓰 가부시끼가이샤 Cooling structure of electronic equipment rack
US4673826A (en) 1984-12-20 1987-06-16 The United States Of America As Represented By The Secretary Of The Air Force Autonomous uninterruptable power supply apparatus
US4718025A (en) 1985-04-15 1988-01-05 Centec Corporation Computer management control system
JPS62174397A (en) 1986-01-28 1987-07-31 Nippon Steel Corp Thin sn plated steel sheet for container having excellent corrosion resistance and weldability
US4816208A (en) 1986-02-14 1989-03-28 Westinghouse Electric Corp. Alarm management system
US4751648A (en) 1986-03-31 1988-06-14 Halliburton Company Local area network data transfer system
US4686450A (en) 1986-04-04 1987-08-11 General Signal Corporation Fluid flow sensor
US4696168A (en) 1986-10-01 1987-09-29 Roger Rasbach Refrigerant subcooler for air conditioning systems
DE3634645C1 (en) 1986-10-10 1988-03-10 Diamalt Ag Process for the production of colorless, odorless and tasteless Cassia endosperm flour
US4683529A (en) 1986-11-12 1987-07-28 Zytec Corporation Switching power supply with automatic power factor correction
JPS63127042A (en) 1986-11-17 1988-05-30 Sanki Eng Co Ltd Heat storage type air conditioning system
JPH0770853B2 (en) 1987-01-21 1995-07-31 株式会社日立製作所 Electronic device cooling system
CN1003681B (en) 1987-01-23 1989-03-22 北京工业学院 Automatic centralized remote monitoring system
US5168724A (en) 1987-02-06 1992-12-08 Reaction Thermal Systems, Inc. Ice building, chilled water system
US4851965A (en) 1987-03-09 1989-07-25 Unisys Corporation Directed air management system for cooling multiple heat sinks
US4964065A (en) 1987-03-12 1990-10-16 Decibel Products, Inc. Computer-controlled electronic system monitor
FR2614748A1 (en) 1987-04-29 1988-11-04 Omega Electronics Sa DEVICE FOR SUPPLYING A DISCHARGE LAMP
JPH0813171B2 (en) 1987-06-26 1996-02-07 株式会社ユタカ電機製作所 Stabilized power supply
US4763013A (en) 1987-09-21 1988-08-09 American Telephone And Telegraph Company, At&T Bell Laboratories Backup protection switch to prevent reverse power flow in a UPS
GB8724263D0 (en) 1987-10-15 1987-11-18 Bicc Plc Electronic enclosure cooling system
US4827733A (en) 1987-10-20 1989-05-09 Dinh Company Inc. Indirect evaporative cooling system
US4831508A (en) 1987-10-20 1989-05-16 Computer Products Inc. Power supply system having improved input power factor
US4816982A (en) 1987-11-23 1989-03-28 Viteq Corporation AC to DC power converter with integrated line current control for improving power factor
US4943902A (en) 1987-11-23 1990-07-24 Viteq Corporation AC to DC power converter and method with integrated line current control for improving power factor
FR2624684B1 (en) 1987-12-11 1990-05-04 Spie Batignolles METHOD AND DEVICE FOR COOLING ELECTRICAL PANELS
JP2575429B2 (en) 1987-12-18 1997-01-22 株式会社東芝 Indoor exhaust system for air-cooled electrical equipment
JPH01218918A (en) 1988-02-26 1989-09-01 Sanden Corp Air conditioner for vehicle
US5309146A (en) 1988-05-03 1994-05-03 Electronic Environmental Controls Inc. Room occupancy indicator means and method
US4964029A (en) 1988-05-18 1990-10-16 Viteq Corporation AC to DC power converter with input current waveform control for buck-boost regulation of output
US5047961A (en) 1988-05-31 1991-09-10 Simonsen Bent P Automatic battery monitoring system
EP0358191B1 (en) 1988-09-06 1995-04-05 Kabushiki Kaisha Toshiba PWM-controlled power supply capable of eliminating modulation-frequency signal components from ground potentials
US5173819A (en) 1988-10-05 1992-12-22 Hitachi, Ltd. Disk apparatus having an improved cooling structure
NO303200B1 (en) 1988-11-04 1998-06-08 Merlin Gerin Building engineering control unit with two-wire data and power supply line
DE3837744A1 (en) 1988-11-07 1990-05-10 Knuerr Mechanik Ag ASSEMBLY CARRIERS FOR PCBS WITH ELECTRONIC COMPONENTS
US5019717A (en) 1988-11-14 1991-05-28 Elegant Design Solutions Inc. Computer-controlled uninterruptable power supply
US5184025A (en) 1988-11-14 1993-02-02 Elegant Design Solutions, Inc. Computer-controlled uninterruptible power supply
US5086385A (en) 1989-01-31 1992-02-04 Custom Command Systems Expandable home automation system
FR2646579A1 (en) 1989-03-20 1990-11-02 Guillemot Gerard ELECTRICALLY HIGH TEMPERATURE HEATING EQUIPMENT BY REGULATED AREAS FOR THE USE OF COMPOSITE MATERIAL PRODUCTS
JPH07105949B2 (en) 1989-03-20 1995-11-13 松下電器産業株式会社 Image motion vector detection device and shake correction device
US5195706A (en) 1989-03-27 1993-03-23 Allen William M Device for holding a container upright
US5220522A (en) 1989-05-09 1993-06-15 Ansan Industries, Ltd. Peripheral data acquisition, monitor, and control device for a personal computer
US5400246A (en) 1989-05-09 1995-03-21 Ansan Industries, Ltd. Peripheral data acquisition, monitor, and adaptive control system via personal computer
US5043807A (en) 1989-05-23 1991-08-27 Zenith Electronics Corporation Three dimensional composite video motion detection
US4935861A (en) 1989-06-13 1990-06-19 Exide Electronics Uninterrupted power supply having no low frequency power magnetics
US5148043A (en) 1989-07-25 1992-09-15 Kabushiki Kaisha Toshiba Uninterruptible power supply diagnosing remaining battery capacity during normal external power source operation
JPH0385797A (en) 1989-08-30 1991-04-10 Toshiba Corp Cooling device for electric equipment
JPH0399398A (en) 1989-09-13 1991-04-24 Hitachi Ltd Remote monitoring device
US5017800A (en) 1989-09-29 1991-05-21 Wisconsin Alumni Research Foundation AC to DC to AC power conversion apparatus with few active switches and input and output control
US5057968A (en) 1989-10-16 1991-10-15 Lockheed Corporation Cooling system for electronic modules
US4980812A (en) 1989-11-09 1990-12-25 Exide Electronics Uninterrupted power supply system having improved power factor correction circuit
US5057698A (en) 1989-11-13 1991-10-15 Exide Electronics Shunt circuit for reducing audible noise at low loading conditions of a power supply employing a high frequency resonant converter
US5189394A (en) 1989-12-22 1993-02-23 Grinell Asia Pacific Pty. Limited Fire alarm display
FR2659155B1 (en) 1990-03-01 1992-05-29 Renault DEVICE FOR CONTROLLING THE ELECTRICAL SUPPLY OF A PLURALITY OF ELECTRONIC MODULES.
US5061916A (en) 1990-05-29 1991-10-29 Barber-Colman Company Event driven remote graphical reporting of building automation system parameters
US5225997A (en) 1990-06-05 1993-07-06 Sygnus Controls Inc. Automatic monitoring and remote reporting device
US5216623A (en) 1990-06-06 1993-06-01 M. T. Mcbrian, Inc. System and method for monitoring and analyzing energy characteristics
US5126585A (en) 1990-06-19 1992-06-30 Auckland Uniservices Limited Uninterruptible power supplies
US5109278A (en) 1990-07-06 1992-04-28 Commonwealth Edison Company Auto freeze frame display for intrusion monitoring system
US6105061A (en) 1990-07-26 2000-08-15 Nec Corporation Hierarchically distributed network management system using open system interconnection (OSI) protocols
JPH0485785A (en) 1990-07-27 1992-03-18 Sanyo Electric Co Ltd Video tape recorder
JP3222456B2 (en) 1990-07-30 2001-10-29 株式会社東芝 Video monitoring system, transmitting device, receiving device, and video monitoring method
CH680693A5 (en) 1990-08-07 1992-10-15 Sulzer Ag
US5440450A (en) 1990-09-14 1995-08-08 Next, Inc. Housing cooling system
US5047913A (en) 1990-09-17 1991-09-10 General Electric Company Method for controlling a power converter using an auxiliary resonant commutation circuit
US5153837A (en) 1990-10-09 1992-10-06 Sleuth Inc. Utility consumption monitoring and control system
US5097328A (en) 1990-10-16 1992-03-17 Boyette Robert B Apparatus and a method for sensing events from a remote location
US5099410A (en) 1990-11-13 1992-03-24 Wisconsin Alumni Research Foundation Single phase ac power conversion apparatus
US5349282A (en) 1990-12-11 1994-09-20 Span, Inc. Battery charging and monitoring system
US5237833A (en) 1991-01-10 1993-08-24 Mitsubishi Denki Kabushiki Kaisha Air-conditioning system
EP0497586A3 (en) 1991-01-31 1994-05-18 Sony Corp Motion detection circuit
US5168171A (en) 1991-03-04 1992-12-01 Tracewell Larry L Enclosure for circuit modules
US5589764A (en) 1991-03-05 1996-12-31 Lee; Graham S. Meter for measuring accumulated power consumption of an electrical appliance during operation of the appliance
US5150580A (en) 1991-03-08 1992-09-29 Hyde Robert E Liquid pressure amplification with superheat suppression
JP3099398B2 (en) 1991-04-05 2000-10-16 株式会社日立製作所 Constant current circuit
US5414591A (en) 1991-04-15 1995-05-09 Hitachi, Ltd. Magnetic disk storage system
US5095712A (en) 1991-05-03 1992-03-17 Carrier Corporation Economizer control with variable capacity
US5382943A (en) 1991-07-31 1995-01-17 Tanaka; Mutuo Remote monitoring unit
JP3175121B2 (en) 1991-05-14 2001-06-11 株式会社ユアサコーポレーション Uninterruptible power system
US5382893A (en) 1991-05-16 1995-01-17 Compaq Computer Corporation Maximum power regulated battery charger
US5315533A (en) 1991-05-17 1994-05-24 Best Power Technology, Inc. Back-up uninterruptible power system
US5272382A (en) 1991-06-24 1993-12-21 Compaq Computer Corporation Power supply for computer system manager
JPH0515423A (en) 1991-07-08 1993-01-26 Matsushita Electric Ind Co Ltd Electronic equipment housing furniture
GB2257598B (en) 1991-07-12 1994-11-30 Hochiki Co Surveillance monitor system using image processing
JPH0540889A (en) 1991-08-02 1993-02-19 Nec Corp Terminal equipment for image remote monitoring system
US5262758A (en) 1991-09-19 1993-11-16 Nam Young K System and method for monitoring temperature
US5281955A (en) 1991-09-20 1994-01-25 C & D Charter Power Systems, Inc. Battery charge monitoring apparatus and method
US5254928A (en) 1991-10-01 1993-10-19 Apple Computer, Inc. Power management system for battery powered computers
US5177666A (en) 1991-10-24 1993-01-05 Bland Timothy J Cooling rack for electronic devices
JPH06119581A (en) 1992-10-08 1994-04-28 Toshiba Corp Remote monitoring device
US5241217A (en) 1991-11-07 1993-08-31 Premier Power, Inc. UPS with input commutation between AC and DC sources of power
FR2684250B1 (en) 1991-11-27 1994-04-01 Merlin Gerin HIGH QUALITY ELECTRICAL ENERGY DISTRIBUTION SYSTEM.
JP2829170B2 (en) 1991-12-06 1998-11-25 株式会社東芝 Rack mount method
US5302858A (en) 1991-12-11 1994-04-12 Best Power Technology, Incorporated Method and apparatus for providing battery charging in a backup power system
FR2685526B1 (en) 1991-12-20 1994-02-04 Alcatel Nv CONNECTION NETWORK WITH MONITORING SENSORS AND DIAGNOSTIC SYSTEM, AND METHOD OF ESTABLISHING DIAGNOSTICS FOR SUCH A NETWORK.
US6400996B1 (en) 1999-02-01 2002-06-04 Steven M. Hoffberg Adaptive pattern recognition based control system and method
KR940011881B1 (en) 1991-12-23 1994-12-27 주식회사 금성사 Apparatus for detecting moving picture
JPH05181970A (en) 1991-12-27 1993-07-23 Toshiba Corp Moving image processor
US5301122A (en) 1992-02-12 1994-04-05 Measuring And Monitoring, Inc. Measuring and monitoring system
US5234185A (en) 1992-02-21 1993-08-10 General Motors Corporation Unitary pipe clamp and assembly
US5410448A (en) 1992-03-02 1995-04-25 Digital Equipment Corporation Adaptive cooling system
US5181653A (en) 1992-03-03 1993-01-26 Foster Glenn D Residential heating and air conditioning control system
DE4211759C2 (en) 1992-04-08 1997-04-03 Schroff Gmbh Fan tray
JPH0698482A (en) 1992-06-10 1994-04-08 Digital Equip Corp <Dec> Supply device of electric power
GB2267993B (en) 1992-06-15 1995-11-22 Gnb Ind Battery Co Modular battery cabinet assembly
EP0574628A1 (en) 1992-06-19 1993-12-22 Lumen Electronics Corporation A modular uninterruptible power supply
US5227262A (en) 1992-07-08 1993-07-13 Yaacov Ozer Universal camcorder battery pack
JPH06124731A (en) 1992-08-31 1994-05-06 Toshiba Corp Attachment for external battery connection, battery pack and battery discrimination and control method
US5291383A (en) 1992-09-02 1994-03-01 Exide Electronics Corporation Simplified UPS system
JPH06105376A (en) 1992-09-17 1994-04-15 Nec Corp Remote measurement equipment
FI96370C (en) 1992-10-01 1996-06-10 Fps Power Systems Oy Ab Method for checking the internal impedance of a backup power supply battery and a backup power supply
US5566339A (en) 1992-10-23 1996-10-15 Fox Network Systems, Inc. System and method for monitoring computer environment and operation
US5383339A (en) 1992-12-10 1995-01-24 Baltimore Aircoil Company, Inc. Supplemental cooling system for coupling to refrigerant-cooled apparatus
US5465011A (en) 1992-12-14 1995-11-07 Square D Company Uninterruptible power supply with improved output regulation
US5269372A (en) 1992-12-21 1993-12-14 International Business Machines Corporation Intersecting flow network for a cold plate cooling system
US5649428A (en) 1993-01-08 1997-07-22 Engelhard/Icc Hybrid air-conditioning system with improved recovery evaporator and subcool condenser coils
EP0609980B1 (en) 1993-01-11 1999-07-14 Canon Kabushiki Kaisha Motion detection method and apparatus
JP2928431B2 (en) 1993-01-13 1999-08-03 富士通株式会社 Auxiliary battery device and charge / discharge control method
US5588067A (en) 1993-02-19 1996-12-24 Peterson; Fred M. Motion detection and image acquisition apparatus and method of detecting the motion of and acquiring an image of an object
JPH06272694A (en) 1993-03-23 1994-09-27 Nippon Telegr & Teleph Corp <Ntt> Control method and device for electronic equipment cooling fan
US6069412A (en) 1993-03-29 2000-05-30 Powerware Corporation Power factor corrected UPS with improved connection of battery to neutral
JPH0744275A (en) 1993-07-30 1995-02-14 Hitachi Ltd Air cooling type information processor
US5422558A (en) 1993-05-05 1995-06-06 Astec International Ltd. Multicell battery power system
US5458991A (en) 1993-05-19 1995-10-17 Sl Waber, Inc. UPS with auto self test
US5972196A (en) 1995-06-07 1999-10-26 Lynntech, Inc. Electrochemical production of ozone and hydrogen peroxide
US5528507A (en) 1993-08-11 1996-06-18 First Pacific Networks System for utility demand monitoring and control using a distribution network
US5818725A (en) 1993-08-11 1998-10-06 First Pacific Networks System for utility demand monitoring and control
US5805458A (en) 1993-08-11 1998-09-08 First Pacific Networks System for utility demand monitoring and control
US5749237A (en) 1993-09-28 1998-05-12 Jdm, Ltd. Refrigerant system flash gas suppressor with variable speed drive
JPH0798781A (en) 1993-09-29 1995-04-11 Toshiba Corp Automatic vending machine
US5860012A (en) 1993-09-30 1999-01-12 Intel Corporation Installation of application software through a network from a source computer system on to a target computer system
US5642002A (en) 1993-10-29 1997-06-24 Alpha Technologies Apparatus and methods for generating uninterruptible AC power signals
FI98559C (en) 1993-11-09 1997-07-10 Aga Ab Method and apparatus for regulating the atmosphere in a substantially enclosed animal shelter or equivalent space
US5586250A (en) 1993-11-12 1996-12-17 Conner Peripherals, Inc. SCSI-coupled module for monitoring and controlling SCSI-coupled raid bank and bank environment
FR2713030B1 (en) 1993-11-24 1996-01-12 Merlin Gerin Uninterruptible feed through neutral through, comprising a double chopper-elevator.
FR2713305B1 (en) 1993-11-29 1996-02-09 Valeo Thermique Habitacle Quick coupling device for heat exchanger pipes.
US5504415A (en) 1993-12-03 1996-04-02 Electronic Power Technology, Inc. Method and apparatus for automatic equalization of series-connected batteries
JPH07202464A (en) 1993-12-28 1995-08-04 Toshiba Corp Electronic appliance, cooling method therefor and fan unit
US5684686A (en) 1994-01-12 1997-11-04 Deltec Electronics Corporation Boost-input backed-up uninterruptible power supply
US5491511A (en) 1994-02-04 1996-02-13 Odle; James A. Multimedia capture and audit system for a video surveillance network
JPH07269926A (en) 1994-02-14 1995-10-20 Daikin Ind Ltd Control device for air conditioner
US5845090A (en) 1994-02-14 1998-12-01 Platinium Technology, Inc. System for software distribution in a digital computer network
US5621662A (en) 1994-02-15 1997-04-15 Intellinet, Inc. Home automation system
US5395042A (en) 1994-02-17 1995-03-07 Smart Systems International Apparatus and method for automatic climate control
JPH07245955A (en) 1994-03-02 1995-09-19 Yutaka Denki Seisakusho:Kk Regulated power supply with improved power factor and uninterruptible power supply
US6274950B1 (en) 1994-03-03 2001-08-14 American Power Conversion Battery communication system
US5963457A (en) 1994-03-18 1999-10-05 Hitachi, Ltd. Electrical power distribution monitoring system and method
JP2864210B2 (en) 1994-03-31 1999-03-03 松下電器産業株式会社 Multi-circuit power management meter for high-voltage substation equipment
US5517251A (en) 1994-04-28 1996-05-14 The Regents Of The University Of California Acquisition of video images simultaneously with analog signals
US5659470A (en) 1994-05-10 1997-08-19 Atlas Copco Wagner, Inc. Computerized monitoring management system for load carrying vehicle
US5572195A (en) 1994-08-01 1996-11-05 Precision Tracking Fm, Inc. Sensory and control system for local area networks
JPH0865681A (en) 1994-08-25 1996-03-08 Sony Corp Motion vector detector and motion compensating prediction encoding system using the detector
US5497288A (en) 1994-08-31 1996-03-05 International Business Machines Corporation Apparatus for tilted serial cooling in an electronic system
IT1275336B (en) 1994-11-14 1997-08-05 Emmepi Telematica E Sicurezza DEVICE FOR REMOTE CONTROL OF A REMOTE EQUIPMENT
US5582020A (en) 1994-11-23 1996-12-10 Mainstream Engineering Corporation Chemical/mechanical system and method using two-phase/two-component compression heat pump
KR960028217A (en) 1994-12-22 1996-07-22 엘리 웨이스 Motion Detection Camera System and Method
US6072396A (en) 1994-12-30 2000-06-06 Advanced Business Sciences Apparatus and method for continuous electronic monitoring and tracking of individuals
US5579197A (en) 1995-01-24 1996-11-26 Best Power Technology, Incorporated Backup power system and method
US5553609A (en) 1995-02-09 1996-09-10 Visiting Nurse Service, Inc. Intelligent remote visual monitoring system for home health care service
US5533357A (en) 1995-02-15 1996-07-09 Carrier Corporation Air conditioning apparatus
US5602462A (en) 1995-02-21 1997-02-11 Best Power Technology, Incorporated Uninterruptible power system
US5572873A (en) 1995-03-02 1996-11-12 Emertech Incorporated Carrier method and apparatus for maintaining pharmaceutical integrity
US5732219A (en) 1995-03-17 1998-03-24 Vermeer Technologies, Inc. Computer system and computer-implemented process for remote editing of computer files
US5540548A (en) 1995-03-31 1996-07-30 Sun Microsystems, Inc. Redundant blower unit
US5844328A (en) 1995-04-06 1998-12-01 Furst; Robert Backup device for electric appliance
DE19512959A1 (en) 1995-04-10 1996-10-17 Sel Alcatel Ag Remote monitoring device
JP3198230B2 (en) 1995-04-10 2001-08-13 富士通株式会社 Fan power supply method
US5664202A (en) 1995-04-20 1997-09-02 C & C Tech Ets Intelligent power consumption monitoring and control system
JP3113793B2 (en) 1995-05-02 2000-12-04 株式会社エヌ・ティ・ティ ファシリティーズ Air conditioning system
JPH08307541A (en) 1995-05-10 1996-11-22 Hitachi Building Syst Eng & Service Co Ltd Remote monitoring device
US5742762A (en) 1995-05-19 1998-04-21 Telogy Networks, Inc. Network management gateway
US5634002A (en) 1995-05-31 1997-05-27 Sun Microsystems, Inc. Method and system for testing graphical user interface programs
US6112235A (en) 1995-06-07 2000-08-29 Spofford; Jason J. Method and apparatus for remotely managing a network hardware device having an embedded server with a client computer across a network
US6311214B1 (en) 1995-07-27 2001-10-30 Digimarc Corporation Linking of computers based on optical sensing of digital data
US5926210A (en) 1995-07-28 1999-07-20 Kalatel, Inc. Mobile, ground-based platform security system which transmits images that were taken prior to the generation of an input signal
JP3347591B2 (en) 1995-07-31 2002-11-20 キヤノン株式会社 Image processing method and apparatus
US5704219A (en) 1995-08-01 1998-01-06 Nippondenso Co., Ltd. Air conditioning apparatus
US6001065A (en) 1995-08-02 1999-12-14 Ibva Technologies, Inc. Method and apparatus for measuring and analyzing physiological signals for active or passive control of physical and virtual spaces and the contents therein
US6038879A (en) 1995-08-08 2000-03-21 Yvon Turcotte Combined air exchange and air conditioning unit
US5715160A (en) 1995-08-21 1998-02-03 Plotke; Robert Jay Motion control evaluation employing a Fourier transform
JPH0964555A (en) 1995-08-22 1997-03-07 Nitto Kogyo Kk Box for electric apparatus
CN1201514A (en) 1995-09-07 1998-12-09 大金工业株式会社 Outlet unit for underfloor air conditionor and underfloor air conditioning system using same
US5970734A (en) 1995-09-29 1999-10-26 Stillwell; Robert Method and system for creating and maintaining a frozen surface
SE506415C2 (en) 1995-10-02 1997-12-15 Ericsson Telefon Ab L M Arrangements for cooling electronic equipment
AUPN592095A0 (en) 1995-10-11 1995-11-02 Invetech Operations Pty Ltd Modular power supply
US6317775B1 (en) 1995-11-03 2001-11-13 Cisco Technology, Inc. System for distributing load over multiple servers at an internet site
US5694780A (en) 1995-12-01 1997-12-09 Alsenz; Richard H. Condensed liquid pump for compressor body cooling
US6032472A (en) 1995-12-06 2000-03-07 Carrier Corporation Motor cooling in a refrigeration system
AUPN727195A0 (en) 1995-12-21 1996-01-18 Canon Kabushiki Kaisha Motion detection method and apparatus
JP3809661B2 (en) 1995-12-28 2006-08-16 ソニー株式会社 Motion detection apparatus and motion detection method
US5732074A (en) 1996-01-16 1998-03-24 Cellport Labs, Inc. Mobile portable wireless communication system
JP3234860B2 (en) 1996-02-05 2001-12-04 株式会社日立製作所 Computer system
US6195018B1 (en) 1996-02-07 2001-02-27 Cellnet Data Systems, Inc. Metering system
AUPN828396A0 (en) 1996-02-23 1996-03-14 Queensland Rail Remote monitoring system
US5968116A (en) 1996-03-27 1999-10-19 Intel Corporation Method and apparatus for facilitating the management of networked devices
US20030206021A1 (en) 1997-07-25 2003-11-06 Laletin William H. Method and apparatus for measuring and analyzing electrical or electrochemical systems
US5794897A (en) 1996-04-22 1998-08-18 Andrew Corporation Transmission line hanger, a method of attaching the hanger and the resulting assembly
US6104868A (en) 1996-04-30 2000-08-15 Peters; Daniel G. Extendible and portable network protocol based system management architecture
JP2776369B2 (en) 1996-05-08 1998-07-16 日本電気株式会社 Electronic device cooling structure
US5735134A (en) 1996-05-30 1998-04-07 Massachusetts Institute Of Technology Set point optimization in vapor compression cycles
US5812055A (en) 1996-05-31 1998-09-22 Eskom Monitoring of a system
US6151643A (en) 1996-06-07 2000-11-21 Networks Associates, Inc. Automatic updating of diverse software products on multiple client computer systems by downloading scanning application to client computer and generating software list on client computer
US6050940A (en) 1996-06-17 2000-04-18 Cybernet Systems Corporation General-purpose medical instrumentation
US6081606A (en) 1996-06-17 2000-06-27 Sarnoff Corporation Apparatus and a method for detecting motion within an image sequence
US5892440A (en) 1997-05-14 1999-04-06 Combustion Engineering Inc. Alarm significance mapping
US7304662B1 (en) 1996-07-10 2007-12-04 Visilinx Inc. Video surveillance system and method
US6008805A (en) 1996-07-19 1999-12-28 Cisco Technology, Inc. Method and apparatus for providing multiple management interfaces to a network device
US5949974A (en) 1996-07-23 1999-09-07 Ewing; Carrell W. System for reading the status and for controlling the power supplies of appliances connected to computer networks
WO1998005060A1 (en) 1996-07-31 1998-02-05 The Board Of Trustees Of The Leland Stanford Junior University Multizone bake/chill thermal cycling module
US5872931A (en) 1996-08-13 1999-02-16 Veritas Software, Corp. Management agent automatically executes corrective scripts in accordance with occurrences of specified events regardless of conditions of management interface and management engine
US6138078A (en) 1996-08-22 2000-10-24 Csi Technology, Inc. Machine monitor with tethered sensors
US6104755A (en) 1996-09-13 2000-08-15 Texas Instruments Incorporated Motion detection using field-difference measurements
US6182157B1 (en) 1996-09-19 2001-01-30 Compaq Computer Corporation Flexible SNMP trap mechanism
US6751221B1 (en) 1996-10-04 2004-06-15 Kabushiki Kaisha Toshiba Data transmitting node and network inter-connection node suitable for home network environment
US6111036A (en) 1996-10-17 2000-08-29 Eastman Chemical Company Method for improving cooling of fluid bed polymer reactor
US5956487A (en) 1996-10-25 1999-09-21 Hewlett-Packard Company Embedding web access mechanism in an appliance for user interface functions including a web server and web browser
US5960204A (en) 1996-10-28 1999-09-28 J.D. Edwards World Source Company System and method for installing applications on a computer on an as needed basis
US7051096B1 (en) 1999-09-02 2006-05-23 Citicorp Development Center, Inc. System and method for providing global self-service financial transaction terminals with worldwide web content, centralized management, and local and remote administration
US6002982A (en) 1996-11-01 1999-12-14 Fry; William R. Sports computer with GPS receiver and performance tracking capabilities
US5731832A (en) 1996-11-05 1998-03-24 Prescient Systems Apparatus and method for detecting motion in a video signal
US5905867A (en) 1996-11-12 1999-05-18 The United States Of America As Represented By The Secretary Of The Navy Apparatus for monitoring environmental parameters at network sites
US5829130A (en) 1996-11-19 1998-11-03 Symex, Inc. Method of installing an integrated data, voice, and video communication network
US5822302A (en) 1996-11-25 1998-10-13 Mci Communications Corporation LAN early warning system
US6112237A (en) 1996-11-26 2000-08-29 Global Maintech, Inc. Electronic monitoring system and method for externally monitoring processes in a computer system
US6139177A (en) 1996-12-03 2000-10-31 Hewlett Packard Company Device access and control using embedded web access functionality
JPH10163538A (en) 1996-12-04 1998-06-19 Ngk Insulators Ltd Thermoelectric conversion device for heat exchanger
US6085243A (en) 1996-12-13 2000-07-04 3Com Corporation Distributed remote management (dRMON) for networks
US6108782A (en) 1996-12-13 2000-08-22 3Com Corporation Distributed remote monitoring (dRMON) for networks
US5974237A (en) 1996-12-18 1999-10-26 Northern Telecom Limited Communications network monitoring
US6128016A (en) 1996-12-20 2000-10-03 Nec Corporation Graphic user interface for managing a server system
US5937092A (en) 1996-12-23 1999-08-10 Esco Electronics Rejection of light intrusion false alarms in a video security system
FR2758019B1 (en) 1996-12-30 1999-01-22 Alsthom Cge Alcatel POWER CONVERTER WITH IMPROVED CONTROL OF MAIN SWITCHES
US6078253A (en) 1997-02-04 2000-06-20 Mytech Corporation Occupancy sensor and method of operating same
US6108492A (en) 1997-02-14 2000-08-22 Toshiba America Information Systems Remote monitoring system
NZ314264A (en) 1997-02-18 1999-06-29 Fisher & Paykel Ltd Substitute Refrigeration apparatus comprising at least two compartments wherein the temperature of each compartment is independently controlled and temperatures are achieved simultaneously
US5862037A (en) 1997-03-03 1999-01-19 Inclose Design, Inc. PC card for cooling a portable computer
US5764030A (en) 1997-03-14 1998-06-09 International Components Corporation Microcontrolled battery charger
GB2323433B (en) 1997-03-18 2001-04-18 Whitlenge Drink Equipment Ltd Improvements relating to cooling devices
US5978912A (en) 1997-03-20 1999-11-02 Phoenix Technologies Limited Network enhanced BIOS enabling remote management of a computer without a functioning operating system
US5923103A (en) 1997-03-31 1999-07-13 Pulizzi Engineering, Inc. Switched-output controller apparatus with repeater function and method for constructing same
US5870698A (en) 1997-05-03 1999-02-09 Atrix International, Inc. Multi-purpose machine metering/monitoring apparatus
JP3688464B2 (en) 1997-05-06 2005-08-31 株式会社東芝 Terminal device, server device, communication device, and control method
DE19719507A1 (en) 1997-05-09 1998-11-12 Alsthom Cge Alcatel Rack with subrack and ventilation device
US6324608B1 (en) 1997-05-13 2001-11-27 Micron Electronics Method for hot swapping of network components
US6189109B1 (en) 1997-05-13 2001-02-13 Micron Electronics, Inc. Method of remote access and control of environmental conditions
US5968398A (en) 1997-05-16 1999-10-19 The Lepel Corporation Apparatus and method for non-contact detection and inductive heating of heat retentive food server warming plates
JPH10322705A (en) 1997-05-21 1998-12-04 Sony Corp Motion detection and motion compensation prediction circuit
CN1282897A (en) 1997-05-27 2001-02-07 陈树锦 Intelligent phase-advancing power supply management system for computer
US5994998A (en) 1997-05-29 1999-11-30 3Com Corporation Power transfer apparatus for concurrently transmitting data and power over data wires
US6094676A (en) 1997-05-30 2000-07-25 Hilgraeve Incorporated Method and apparatus for peer-to-peer communication
JPH10332246A (en) 1997-06-03 1998-12-15 Ke Corp:Kk Cooling device
WO1999001919A1 (en) 1997-06-05 1999-01-14 Mario Pena System for drag reduction of vehicles and for battery charging
JPH10339500A (en) 1997-06-09 1998-12-22 Toshiba Corp Air conditioner
US5991885A (en) 1997-06-11 1999-11-23 Clarinet Systems, Inc. Method and apparatus for detecting the presence of a remote device and providing power thereto
CN101494646B (en) 1997-06-25 2013-10-02 三星电子株式会社 Method and apparatus for home network auto-tree builder
US6389464B1 (en) 1997-06-27 2002-05-14 Cornet Technology, Inc. Device management system for managing standards-compliant and non-compliant network elements using standard management protocols and a universal site server which is configurable from remote locations via internet browser technology
US5860280A (en) 1997-07-03 1999-01-19 Marlow Industries, Inc. Liquid cooling system with solid material formation control and method of manufacture
ATE360909T1 (en) 1997-07-03 2007-05-15 R V Holdings Corp In Trust DEVICE AND METHOD FOR MANAGING A BATTERY BACKUP POWER SUPPLY DEVICE
US6329792B1 (en) 1997-07-04 2001-12-11 Estco Energy Inc. Device and system for management of battery back up power source
JP3306651B2 (en) 1997-07-07 2002-07-24 吉田 富貴子 Remote power switching equipment
US6213194B1 (en) 1997-07-16 2001-04-10 International Business Machines Corporation Hybrid cooling system for electronics module
US5970731A (en) 1997-11-21 1999-10-26 International Business Machines Corporation Modular refrigeration system
US5963425A (en) 1997-07-16 1999-10-05 International Business Machines Corporation Combined air and refrigeration cooling for computer systems
US5954127A (en) 1997-07-16 1999-09-21 International Business Machines Corporation Cold plate for dual refrigeration system
US6034872A (en) 1997-07-16 2000-03-07 International Business Machines Corporation Cooling computer systems
US6006171A (en) 1997-07-28 1999-12-21 Vines; Caroline J. Dynamic maintenance management system
US6130620A (en) 1997-08-11 2000-10-10 Electronic Monitoring Systems, Inc. Remote monitoring system
US6101459A (en) 1997-08-15 2000-08-08 Compaq Computer Corporation System and associated method for cooling components within a computer system
US6219628B1 (en) 1997-08-18 2001-04-17 National Instruments Corporation System and method for configuring an instrument to perform measurement functions utilizing conversion of graphical programs into hardware implementations
JPH1173258A (en) 1997-08-28 1999-03-16 Toshiba Corp Low power consumption bus structure and method for controlling the same and system for synthesizing low power consumption bus structure and method therefor and portable information equipment
US6470386B1 (en) 1997-09-26 2002-10-22 Worldcom, Inc. Integrated proxy interface for web based telecommunications management tools
US6088816A (en) 1997-10-01 2000-07-11 Micron Electronics, Inc. Method of displaying system status
US6236332B1 (en) 1997-10-22 2001-05-22 Profile Systems, Llc Control and monitoring system
US6055480A (en) 1997-11-12 2000-04-25 Albert Einstein Healthcare Network Environmental monitoring system
US5914858A (en) 1997-11-13 1999-06-22 Northern Telecom Limited Baffle arrangement for an airflow balance system
EP0917034B1 (en) 1997-11-14 2002-03-06 Engel Maschinenbau Gesellschaft Mbh Method for remote monitoring and/or remote servicing of an injection moulding machine
US6175866B1 (en) 1997-11-24 2001-01-16 International Business Machines Corporation Method and system for generating unsupported network monitoring objects
US6058434A (en) 1997-11-26 2000-05-02 Acuity Imaging, Llc Apparent network interface for and between embedded and host processors
JP2001524713A (en) 1997-11-26 2001-12-04 アキュイティー イメージング エルエルシー An apparent network interface between the embedded processor and the host processor
JPH11164035A (en) 1997-11-27 1999-06-18 Meidensha Corp Remote monitoring system
TW454124B (en) 1997-11-28 2001-09-11 Accton Technology Corp Network stack with automatic switch device for the switch
US6138249A (en) 1997-12-11 2000-10-24 Emc Corporation Method and apparatus for monitoring computer systems during manufacturing, testing and in the field
US6175934B1 (en) 1997-12-15 2001-01-16 General Electric Company Method and apparatus for enhanced service quality through remote diagnostics
FR2772978B1 (en) 1997-12-18 2000-01-21 Schneider Electric Sa FIXING TERMINAL AND ELECTRICAL CONNECTION MODULE FOR PLUG-IN CIRCUIT BREAKER
US6122639A (en) 1997-12-23 2000-09-19 Cisco Technology, Inc. Network device information collection and change detection
US6173323B1 (en) 1997-12-24 2001-01-09 Lucent Technologies Inc. Adaptive polling rate algorithm for SNMP-based network monitoring
JPH11184570A (en) 1997-12-25 1999-07-09 Hitachi Ltd Device provided with power source
US6052750A (en) 1998-01-06 2000-04-18 Sony Corporation Of Japan Home audio/video network for generating default control parameters for devices coupled to the network, and replacing updated control parameters therewith
US6438978B1 (en) 1998-01-07 2002-08-27 General Electric Company Refrigeration system
GB2333864B (en) 1998-01-28 2003-05-07 Ibm Distribution of software updates via a computer network
JPH11219388A (en) 1998-02-04 1999-08-10 Fujitsu Ltd Remote management system
US5955946A (en) 1998-02-06 1999-09-21 Beheshti; Ali Alarm/facility management unit
DE19804904C1 (en) 1998-02-07 1999-11-11 Loh Kg Rittal Werk Control cabinet with facilities for cooling the interior warm air
DE19804900C2 (en) 1998-02-07 2003-04-30 Rittal Gmbh & Co Kg switch cabinet
US7047300B1 (en) 1998-02-10 2006-05-16 Sprint Communications Company L.P. Survivable and scalable data system and method for computer networks
EP0937950B1 (en) 1998-02-23 2004-10-20 Mitsubishi Denki Kabushiki Kaisha Air conditioner
DK0939387T3 (en) 1998-02-28 2004-02-02 Siemens Building Tech Ag Room monitoring device
US6922558B2 (en) 1998-03-06 2005-07-26 Don Delp Integrated building control and information system with wireless networking
GB9805556D0 (en) 1998-03-17 1998-05-13 Crimewatch Scotland Limited Alarm systems
US6088225A (en) 1998-03-17 2000-07-11 Northern Telecom Limited Cabinet with enhanced convection cooling
US6137264A (en) 1998-03-20 2000-10-24 Dallas Semiconductor Corporation Battery pack
US6218809B1 (en) 1998-03-20 2001-04-17 Dallas Semiconductor Corporation Method for monitoring operating parameters of a rechargeable power supply
US6115468A (en) 1998-03-26 2000-09-05 Cisco Technology, Inc. Power feed for Ethernet telephones via Ethernet link
US6223810B1 (en) 1998-03-31 2001-05-01 International Business Machines Extended air cooling with heat loop for dense or compact configurations of electronic components
US5953930A (en) 1998-03-31 1999-09-21 International Business Machines Corporation Evaporator for use in an extended air cooling system for electronic components
US6208345B1 (en) 1998-04-15 2001-03-27 Adc Telecommunications, Inc. Visual data integration system and method
US6347627B1 (en) 1998-04-23 2002-02-19 Pioneer Inventions, Inc. Nitrous oxide based oxygen supply system
JPH11338666A (en) 1998-05-04 1999-12-10 Hewlett Packard Co <Hp> Method for providing printable page and device distributing hard copy
US6167406A (en) 1998-05-08 2000-12-26 Allen-Bradley Company, Llc System, method and article of manufacture for building an enterprise-wide data model
CN2337456Y (en) 1998-05-08 1999-09-08 肖小松 Full-closed separating air conditioner
US6229429B1 (en) 1998-05-15 2001-05-08 Daniel J. Horon Fire protection and security monitoring system
US6298144B1 (en) 1998-05-20 2001-10-02 The United States Of America As Represented By The National Security Agency Device for and method of detecting motion in an image
US6038877A (en) 1998-05-22 2000-03-21 Bergstrom, Inc. Modular low pressure delivery vehicle air conditioning system
US6311105B1 (en) 1998-05-29 2001-10-30 Powerweb, Inc. Multi-utility energy control system
US6122603A (en) 1998-05-29 2000-09-19 Powerweb, Inc. Multi-utility energy control system with dashboard
US6054987A (en) 1998-05-29 2000-04-25 Hewlett-Packard Company Method of dynamically creating nodal views of a managed network
US6363421B2 (en) 1998-05-31 2002-03-26 Lucent Technologies, Inc. Method for computer internet remote management of a telecommunication network element
US6343320B1 (en) 1998-06-09 2002-01-29 Compaq Information Technologies Group, L.P. Automatic state consolidation for network participating devices
US6889095B1 (en) 1998-06-11 2005-05-03 Agilent Technologies, Inc. Computer network adapted for industrial environments
US6057834A (en) 1998-06-12 2000-05-02 International Business Machines Corporation Iconic subscription schedule controller for a graphic user interface
US6363422B1 (en) 1998-06-24 2002-03-26 Robert R. Hunter Multi-capability facilities monitoring and control intranet for facilities management system
US6360255B1 (en) 1998-06-25 2002-03-19 Cisco Technology, Inc. Automatically integrating an external network with a network management system
US6282546B1 (en) 1998-06-30 2001-08-28 Cisco Technology, Inc. System and method for real-time insertion of data into a multi-dimensional database for network intrusion detection and vulnerability assessment
JP2000022373A (en) 1998-07-03 2000-01-21 Toshiba Corp Electronic apparatus
US6115273A (en) 1998-07-09 2000-09-05 Illinois Tool Works Inc. Power converter with low loss switching
US5982652A (en) 1998-07-14 1999-11-09 American Power Conversion Method and apparatus for providing uninterruptible power using a power controller and a redundant power controller
JP3408424B2 (en) 1998-07-28 2003-05-19 日本電気株式会社 Electronic equipment cooling structure
JP2000047757A (en) 1998-07-30 2000-02-18 Canon Inc Information processor and its control method
KR100316647B1 (en) 1998-07-30 2002-01-15 윤종용 Power control method and apparatus therefor in computer system using wake on LAN signal
US6141762A (en) 1998-08-03 2000-10-31 Nicol; Christopher J. Power reduction in a multiprocessor digital signal processor based on processor load
EP0978780B1 (en) 1998-08-06 2003-06-04 Elonex I.P. Holdings Limited Low-power-consumption monitor standby system
US6160926A (en) 1998-08-07 2000-12-12 Hewlett-Packard Company Appliance and method for menu navigation
JP4195959B2 (en) 1998-08-07 2008-12-17 パナソニック株式会社 Uninterruptible power system
US6496862B1 (en) 1998-08-25 2002-12-17 Mitsubishi Electric Research Laboratories, Inc. Remote monitoring and control of devices connected to an IEEE 1394 bus via a gateway device
US6197444B1 (en) 1998-08-26 2001-03-06 Toshiba International Corporation Battery case
US6611866B1 (en) 1998-08-27 2003-08-26 Intel Corporation Management object for aggregated network device status
US6338094B1 (en) 1998-09-08 2002-01-08 Webtv Networks, Inc. Method, device and system for playing a video file in response to selecting a web page link
JP2000092092A (en) 1998-09-10 2000-03-31 Mitsubishi Electric Corp Power source controller and power source control system
US6202149B1 (en) 1998-09-30 2001-03-13 Ncr Corporation Automated application fail-over for coordinating applications with DBMS availability
JP2000112574A (en) 1998-10-06 2000-04-21 Hitachi Ltd Cooling fan controlling system
US6175927B1 (en) 1998-10-06 2001-01-16 International Business Machine Corporation Alert mechanism for service interruption from power loss
KR20010089326A (en) 1998-10-20 2001-09-29 지.윌리암 래그랜드 Corrugated multilayer metal foil insulation panels and methods of making
US6075698A (en) 1998-10-27 2000-06-13 Ads, The Power Resource, Inc. Removable fan for rack mounted rectifiers
JP2000134606A (en) 1998-10-29 2000-05-12 Sekyurion Nijuyon Kk Remote monitoring system, remote image reproducing method and recording medium
JP3644579B2 (en) 1998-10-29 2005-04-27 富士通株式会社 Security enhancement method and apparatus
US6191500B1 (en) 1998-11-06 2001-02-20 Kling Lindquist Partnership, Inc. System and method for providing an uninterruptible power supply to a critical load
US6157943A (en) 1998-11-12 2000-12-05 Johnson Controls Technology Company Internet access to a facility management system
US6177884B1 (en) 1998-11-12 2001-01-23 Hunt Technologies, Inc. Integrated power line metering and communication method and apparatus
JP2000151606A (en) 1998-11-16 2000-05-30 Nec Corp Network monitoring system, network monitoring method, network management device, network device to be managed and recording medium
US6078957A (en) 1998-11-20 2000-06-20 Network Alchemy, Inc. Method and apparatus for a TCP/IP load balancing and failover process in an internet protocol (IP) network clustering system
US6198628B1 (en) 1998-11-24 2001-03-06 Unisys Corporation Parallel cooling of high power devices in a serially cooled evironment
US6374296B1 (en) 1998-11-25 2002-04-16 Adc Technologies International Pte Ltd Method and system for providing cross-platform remote control and monitoring of facility access controller
US6400043B1 (en) 1998-11-30 2002-06-04 American Power Conversion Corporation Modular uninterruptable power supply
GB2344718B (en) 1998-12-07 2001-01-17 Telspec Europ Ltd Remote management system for electronic equipment
EP1009130A1 (en) 1998-12-11 2000-06-14 International Business Machines Corporation Distributed directory services for locating network resources in a very large packet switching network
US6621823B1 (en) 1998-12-21 2003-09-16 Hewlett-Packard Development Company, L.P. Network appliance combining asychronous notification with interactive network transfer protocol server
US6529936B1 (en) 1998-12-23 2003-03-04 Hewlett-Packard Company Object-oriented web server architecture suitable for various types of devices
US6720990B1 (en) 1998-12-28 2004-04-13 Walker Digital, Llc Internet surveillance system and method
US6553418B1 (en) 1999-01-02 2003-04-22 Daniel J. Collins Energy information and control system
US6437691B1 (en) 1999-01-09 2002-08-20 Heat-Timer Corporation Electronic message delivery system utilizable in the monitoring of remote equipment and method of same
US6259956B1 (en) 1999-01-14 2001-07-10 Rawl & Winstead, Inc. Method and apparatus for site management
US6631134B1 (en) 1999-01-15 2003-10-07 Cisco Technology, Inc. Method for allocating bandwidth in an optical network
US6505256B1 (en) 1999-01-15 2003-01-07 Compaq Information Technologies Group, L.P. Automatic synchronization of state colors across a web-based system
JP2000209204A (en) 1999-01-18 2000-07-28 Nec Eng Ltd Remote monitor control method and its system
SE9900303L (en) 1999-01-27 2000-07-28 Ericsson Telefon Ab L M A method enabling communication between an electronic device and a battery, an apparatus comprising an electronic device and a battery, and a battery enabling communication
US6404348B1 (en) 1999-02-11 2002-06-11 Power Quality Consultants, Inc. Modular power quality monitoring device
US6304900B1 (en) 1999-02-18 2001-10-16 International Business Machines Corporation Data processing system and method for permitting a server computer system to remotely modify operation of a client system's network hardware
US6145502A (en) 1999-03-02 2000-11-14 Heat-N-Glo Fireplace Products, Inc. Dual mode of operation fireplaces for operation in vented or unvented mode
JP2002539590A (en) 1999-03-11 2002-11-19 パワー・サーキット・イノベーションズ・インコーポレーテッド Networkable power controller
DE29904928U1 (en) 1999-03-18 1999-06-17 Fritschle Simone Housing with a fan device and fan device
US6215404B1 (en) 1999-03-24 2001-04-10 Fernando Morales Network audio-link fire alarm monitoring system and method
JP2000278267A (en) 1999-03-26 2000-10-06 Nec Corp System and method for monitoring network
JP4017282B2 (en) 1999-03-29 2007-12-05 日立造船株式会社 Operation monitoring device and remote monitoring system
AU4033100A (en) 1999-03-29 2000-10-16 Caterpillar Inc. Modular chilled fluid system and method for providing chilled fluid for cooling
MXPA01010270A (en) 1999-04-09 2002-10-23 Henry B Steen Iii Remote data access and system control.
US6449745B1 (en) 1999-04-22 2002-09-10 Synopsys, Inc. Method and apparatus for random stimulus generation
US6591279B1 (en) 1999-04-23 2003-07-08 International Business Machines Corporation System and method for computer-based notifications of real-world events using digital images
JP2000315883A (en) 1999-05-06 2000-11-14 Hitachi Ltd Radiation structure of control device
JP2000339260A (en) 1999-05-31 2000-12-08 Canon Inc Client and method for device retrieval
US6404658B1 (en) 1999-05-13 2002-06-11 American Power Conversion Method and apparatus for converting a DC voltage to an AC voltage
US6220518B1 (en) 1999-05-13 2001-04-24 Acutherm L.P. Process and apparatus for individual adjustment of the temperature set points of a plurality of VAV devices
EP1052806B1 (en) 1999-05-13 2006-03-22 Canon Kabushiki Kaisha Apparatus for searching a device on a network
US6209330B1 (en) 1999-05-17 2001-04-03 Caterpillar Inc. Modular air handling system and method for providing cooling
JP3589086B2 (en) 1999-05-17 2004-11-17 松下電器産業株式会社 Power supply
EP1101155A1 (en) 1999-05-28 2001-05-23 General Electric Company An intelligent electronic device for monitoring non-electrical characteristics
CA2376214A1 (en) 1999-06-08 2000-12-14 The Trustees Of Columbia University In The City Of New York Network telephony appliance and system for inter/intranet telephony
US6788980B1 (en) 1999-06-11 2004-09-07 Invensys Systems, Inc. Methods and apparatus for control using control devices that provide a virtual machine environment and that communicate via an IP network
US7606760B2 (en) 1999-06-18 2009-10-20 Echarge Corporation Method and apparatus for ordering goods, services and content over an internetwork using a virtual payment account
WO2000079500A1 (en) 1999-06-21 2000-12-28 Telenetics Corporation Remote meter monitoring system and method
US6553336B1 (en) 1999-06-25 2003-04-22 Telemonitor, Inc. Smart remote monitoring system and method
JP2001024638A (en) 1999-07-08 2001-01-26 Fujitsu Ltd Network controller and storage medium
US6343617B1 (en) 1999-07-09 2002-02-05 Millipore Corporation System and method of operation of a digital mass flow controller
US6164369A (en) 1999-07-13 2000-12-26 Lucent Technologies Inc. Door mounted heat exchanger for outdoor equipment enclosure
US6169669B1 (en) 1999-07-15 2001-01-02 Texas Instruments Incorporated Digital signal processor controlled uninterruptable power supply
WO2001008396A1 (en) 1999-07-23 2001-02-01 Telemetry Technologies, Inc. System and method for a virtual network operations center
US6208510B1 (en) 1999-07-23 2001-03-27 Teradyne, Inc. Integrated test cell cooling system
TW443058B (en) 1999-07-29 2001-06-23 Inventec Corp Modem with network transmission protocols
US6237353B1 (en) 1999-07-29 2001-05-29 Carrier Corporation System for removing parasitic losses in a refrigeration unit
US6184593B1 (en) 1999-07-29 2001-02-06 Abb Power T&D Company Inc. Uninterruptible power supply
US6718364B2 (en) 1999-08-10 2004-04-06 Sun Microsystems, Inc. Method and apparatus for expedited file downloads in an applet environment
US6483730B2 (en) 1999-08-13 2002-11-19 Powerware Corporation Power converters with AC and DC operating modes and methods of operation thereof
US6529230B1 (en) 1999-08-30 2003-03-04 Safe-T-Net Systems Pte Ltd Security and fire control system
US6281790B1 (en) 1999-09-01 2001-08-28 Net Talon Security Systems, Inc. Method and apparatus for remotely monitoring a site
US20070008099A1 (en) 1999-09-01 2007-01-11 Nettalon Security Systems, Inc. Method and apparatus for remotely monitoring a site
JP2001075682A (en) * 1999-09-01 2001-03-23 Canon Inc Electronic equipment, its control method, and storage medium
GB2354066B (en) 1999-09-07 2001-07-18 Lucent Technologies Inc Cabinet for units which dissipate heat
GB2354062A (en) 1999-09-13 2001-03-14 British Broadcasting Corp Cooling system for use in cooling electronic equipment
US6405216B1 (en) 1999-09-17 2002-06-11 International Business Machines Corporation Internet-based application program interface (API) documentation interface
US7149554B2 (en) 1999-09-20 2006-12-12 Motorola, Inc. Access control method utilizing a key battery
ATE339154T1 (en) 1999-09-21 2006-10-15 Honeywell Hommed Llc HOME PATIENT MONITORING SYSTEM
US6406257B1 (en) 1999-09-29 2002-06-18 Silicon Graphics, Inc. Modular air moving system and method
GB2355163A (en) 1999-10-05 2001-04-11 Inventec Corp A modem having embedded network transmission protocols
US6477667B1 (en) 1999-10-07 2002-11-05 Critical Devices, Inc. Method and system for remote device monitoring
US6954859B1 (en) 1999-10-08 2005-10-11 Axcess, Inc. Networked digital security system and methods
US6658595B1 (en) 1999-10-19 2003-12-02 Cisco Technology, Inc. Method and system for asymmetrically maintaining system operability
US6496366B1 (en) 1999-10-26 2002-12-17 Rackable Systems, Llc High density computer equipment storage system
US7330886B2 (en) 1999-10-27 2008-02-12 American Power Conversion Corporation Network appliance management
US7159022B2 (en) 2001-01-26 2007-01-02 American Power Conversion Corporation Method and system for a set of network appliances which can be connected to provide enhanced collaboration, scalability, and reliability
US7392309B2 (en) 1999-10-27 2008-06-24 American Power Conversion Corporation Network appliance management
CN1153151C (en) 1999-11-01 2004-06-09 贵州以太科技信息产业有限责任公司 Intelligent control method for different electric appliances and its universal controller
US6327143B1 (en) 1999-11-05 2001-12-04 Cray, Inc. Radial computer system and method
US6493243B1 (en) 1999-12-01 2002-12-10 Acme Electric Corporation Redundant power system and power supply therefor
GB9928455D0 (en) 1999-12-02 2000-01-26 Ascot Management Solutions Ltd Monitoring system
US8161193B1 (en) 1999-12-08 2012-04-17 Rockstar Bidco Lp System, device, and method for sending keep-alive messages in a communication network
US7206824B2 (en) 1999-12-08 2007-04-17 Sun Microsystems, Inc. Technique for configuring network deliverable pluggable components for deployment
BR9907351A (en) 1999-12-22 2001-08-07 Ericsson Telecomunicacoees S A Control method and circuit for three-level three-level elevator rectifier
CA2326278C (en) 1999-12-30 2008-06-03 At&T Corp. Remote monitoring through the brg
US6814134B1 (en) 2000-01-24 2004-11-09 Mary E. Brezinski Compact electronic cabinet cooler
JP3871569B2 (en) 2000-02-04 2007-01-24 富士通株式会社 Network control system, network device, repeater, and connection device
DE10005213A1 (en) 2000-02-05 2001-08-16 Messer Griesheim Gmbh System and method for remote monitoring of measured variables uses a measuring device to record a measurement value for a physical, chemical parameter and to visualize the measurement value on a display device.
GB0003570D0 (en) 2000-02-17 2000-04-05 Combined Power Systems Ltd Remote monitoring
US6412292B2 (en) 2000-05-09 2002-07-02 Toc Technology, Llc Computer rack heat extraction device
US6494050B2 (en) 2000-02-18 2002-12-17 Toc Technology, Llc Computer rack heat extraction device
EP1258070A2 (en) 2000-02-18 2002-11-20 Liebert Corporation Modular uninterruptible power supply
US6433444B1 (en) 2000-02-18 2002-08-13 General Electric Company Modular fault tolerant power distribution system
US6557357B2 (en) 2000-02-18 2003-05-06 Toc Technology, Llc Computer rack heat extraction device
TW550477B (en) 2000-03-01 2003-09-01 Passgate Corp Method, system and computer readable medium for Web site account and e-commerce management from a central location
US6792321B2 (en) 2000-03-02 2004-09-14 Electro Standards Laboratories Remote web-based control
JP2001257496A (en) 2000-03-13 2001-09-21 Komurakku:Kk Cabinet for electric apparatus
CA2403270C (en) 2000-03-14 2011-05-17 Joseph Robert Marchese Digital video system using networked cameras
AU2001247516A1 (en) 2000-03-16 2001-09-24 Computer Sciences Corporation System and method for long term care insurance administration
AU4216101A (en) 2000-03-17 2001-09-24 Sunarc Structures Inc Dynamic heating and cooling of a building using liquid foam
JP2001260640A (en) 2000-03-21 2001-09-26 Calsonic Kansei Corp Heating device for vehicle
US20010047387A1 (en) 2000-03-27 2001-11-29 Exoplex, Inc. Systems and methods for providing distributed cross-enterprise portals
US6310783B1 (en) 2000-03-29 2001-10-30 Powerware Corporation Modular method and apparatus for building an uninterruptible power system (UPS)
US6292379B1 (en) 2000-03-29 2001-09-18 Powerware Corporation Distributed internal fault bypass in a modular uninterruptible power supply
GB2361156B (en) 2000-04-07 2002-08-07 3Com Corp Discovering non managed devices in a network such as a LAN using HTTP
JP3860035B2 (en) 2000-04-13 2006-12-20 株式会社マキタ Adapter for charging device
US20040031282A1 (en) 2000-04-14 2004-02-19 Kopko William Leslie Desiccant air conditioner
SE517550C2 (en) 2000-04-17 2002-06-18 Micronic Laser Systems Ab Pattern generation system using a spatial light modulator
JP4403335B2 (en) 2000-04-17 2010-01-27 ソニー株式会社 Maintenance support system for video processing equipment
CA2407512A1 (en) 2000-04-25 2001-11-01 Airak, Inc. System and method for distributed monitoring using remote sensors
US6670810B2 (en) 2000-04-25 2003-12-30 Airak, Inc. System and method for distributed monitoring of surroundings using telemetry of data from remote sensors
EP1290662B1 (en) 2000-04-26 2016-07-06 Schneider Electric IT Corporation System and method for managing uninterruptible power supply systems
US6827142B2 (en) 2000-04-27 2004-12-07 Innoventor Engineering, Inc. Process and apparatus for achieving precision temperature control
US6356470B1 (en) 2000-04-29 2002-03-12 Hewlett-Packard Company Dual input, hot swappable dual redundant, enhanced N+1 AC to DC power system
US6208261B1 (en) 2000-05-13 2001-03-27 John Olstead Use of visual building alarm system to display public information to building visitors
JP2001338486A (en) 2000-05-25 2001-12-07 Hitachi Ltd Information memory device
US6761212B2 (en) 2000-05-25 2004-07-13 Liebert Corporation Spiral copper tube and aluminum fin thermosyphon heat exchanger
US7010594B2 (en) 2000-05-26 2006-03-07 Isochron, Llc System using environmental sensor and intelligent management and control transceiver for monitoring and controlling remote computing resources
KR100338604B1 (en) 2000-05-29 2002-05-27 배영주 remote network management system via serial port
US6455954B1 (en) 2000-05-31 2002-09-24 Innovations Electrical Lc Auxiliary power supply system serving as primary power source during selected times and power outages
US6344837B1 (en) 2000-06-16 2002-02-05 Andrew H. Gelsey Three-dimensional image display with picture elements formed from directionally modulated pixels
JP2002009864A (en) 2000-06-20 2002-01-11 Sony Corp Control method and communication equipment
CN1300677C (en) 2000-06-22 2007-02-14 微软公司 Distributed computing services platform
US6577901B2 (en) 2000-06-23 2003-06-10 Medtronic, Inc. Network compatible RF wireless link for medical device data management
US20020023258A1 (en) 2000-06-27 2002-02-21 Elwahab Amgad Mazen System and method for managing telecommunications devices
GB0015696D0 (en) 2000-06-28 2000-08-16 Ways Limited C Surveillance system
US6704199B2 (en) 2000-07-05 2004-03-09 Network Engines, Inc. Low profile equipment housing with angular fan
US7693976B2 (en) 2000-07-11 2010-04-06 Ciena Corporation Granular management of network resources
IT1318284B1 (en) 2000-07-31 2003-07-28 Cit Alcatel METHOD AND DEVICE FOR REMOTE CONFIGURATION AND MONITORING OF TELECOMMUNICATIONS NETWORK ELEMENTS.
US7185360B1 (en) 2000-08-01 2007-02-27 Hereuare Communications, Inc. System for distributed network authentication and access control
US6252770B1 (en) 2000-08-02 2001-06-26 Ming-Chuan Yu Electronic apparatus cooling device
US6557624B1 (en) 2000-08-09 2003-05-06 Liebert Corporation Configurable system and method for cooling a room
JP3766005B2 (en) 2000-08-09 2006-04-12 株式会社イトーキ Moving shelf equipment
EP2248536A3 (en) 2000-12-14 2010-12-08 The General Hospital Corporation doing business as Massachusetts General Hospital Topical Aminolevulinic acid-photodynamic therapy for acne vulgaris
WO2002016854A1 (en) 2000-08-23 2002-02-28 Toc Technology, Llc Computer room air flow method and apparatus
GB0021393D0 (en) 2000-08-31 2000-10-18 Imi Cornelius Uk Ltd Thermoelectric module
US6686838B1 (en) 2000-09-06 2004-02-03 Xanboo Inc. Systems and methods for the automatic registration of devices
JP3465674B2 (en) 2000-09-08 2003-11-10 株式会社ニプロン Unit insertion / removal structure for uninterruptible redundant power supply
US7292898B2 (en) 2000-09-18 2007-11-06 Balboa Instruments, Inc. Method and apparatus for remotely monitoring and controlling a pool or spa
US6317320B1 (en) 2000-09-19 2001-11-13 Marconi Communications, Inc. Cooling system for electronic components in an equipment enclosure
US20040003266A1 (en) 2000-09-22 2004-01-01 Patchlink Corporation Non-invasive automatic offsite patch fingerprinting and updating system and method
JP4013470B2 (en) 2000-09-29 2007-11-28 株式会社イトーキ Manual moving shelf device
US7043661B2 (en) 2000-10-19 2006-05-09 Tti-Team Telecom International Ltd. Topology-based reasoning apparatus for root-cause analysis of network faults
US6756998B1 (en) 2000-10-19 2004-06-29 Destiny Networks, Inc. User interface and method for home automation system
US6462944B1 (en) 2000-10-25 2002-10-08 Macase Industrial Group Ga., Inc. Computer cabinet cooling system
US6917124B2 (en) 2000-10-27 2005-07-12 Liebert Corporation Uninterruptible power supply
US6950868B1 (en) 2000-10-31 2005-09-27 Red Hat, Inc. Method of and apparatus for remote monitoring
DE10054163A1 (en) 2000-11-02 2002-06-06 Wacker Polymer Systems Gmbh Process for the production of pressed wood panels
US6330176B1 (en) 2000-11-15 2001-12-11 Powerware Corporation Multi-input power transfer and uninterruptible power supply apparatus and methods of operation thereof
US6925516B2 (en) 2001-01-19 2005-08-02 Raze Technologies, Inc. System and method for providing an improved common control bus for use in on-line insertion of line replaceable units in wireless and wireline access systems
US6829630B1 (en) 2000-11-24 2004-12-07 Xerox Corporation Mechanisms for web-object event/state-driven communication between networked devices
JP2002168551A (en) 2000-11-30 2002-06-14 Tokyo Electron Ltd Cooling device for electrode of treating device
US6368064B1 (en) 2000-12-01 2002-04-09 3Com Corporation Apparatus and method of providing redundant power and redundant fan speed control to a plurality of fans
US20020071031A1 (en) 2000-12-07 2002-06-13 Philips Electronics North America Corporation Remote monitoring via a consumer electronic appliance
WO2002047369A1 (en) 2000-12-08 2002-06-13 Ebac Limited Remote monitoring apparatus
US9058416B2 (en) 2000-12-11 2015-06-16 Peter K. Trzyna System and method for detecting and reporting online activity using real-time content-based network monitoring
US6317348B1 (en) 2000-12-20 2001-11-13 Toshiba International Corporation UPS chassis design having an internal support structure
US6795941B2 (en) 2000-12-21 2004-09-21 Honeywell International Inc. Method for diagnosing a network
EP1354446B1 (en) 2001-01-12 2006-04-12 Novar Marketing Inc. Small building automation control system
TW511731U (en) 2001-01-20 2002-11-21 Elanvital Corp Pipe fan
US6450003B1 (en) 2001-01-22 2002-09-17 Steven M. Goldstein Tool and method for straightening a door hinge
US8271626B2 (en) 2001-01-26 2012-09-18 American Power Conversion Corporation Methods for displaying physical network topology and environmental status by location, organization, or responsible party
DK174494B1 (en) 2001-01-26 2003-04-22 American Power Conversion Denm Combined AC-DC to DC converter
US6489561B2 (en) 2001-01-30 2002-12-03 American Power Conversion Battery wire lead management
US6766008B2 (en) 2001-01-31 2004-07-20 Nhc Communications System and method for switching telecommunications service in a multiple dwelling unit/multiple tenant unit environment
US7975156B2 (en) 2008-10-21 2011-07-05 Dell Products, Lp System and method for adapting a power usage of a server during a data center cooling failure
US6539725B2 (en) 2001-02-09 2003-04-01 Bsst Llc Efficiency thermoelectrics utilizing thermal isolation
US7493391B2 (en) 2001-02-12 2009-02-17 International Business Machines Corporation System for automated session resource clean-up by determining whether server resources have been held by client longer than preset thresholds
US6682100B2 (en) 2001-02-15 2004-01-27 Parker-Hannifin Corporation Connection block with clip
US6592449B2 (en) 2001-02-24 2003-07-15 International Business Machines Corporation Smart fan modules and system
US20020129355A1 (en) 2001-03-01 2002-09-12 Mike Velten Method and system for monitoring an apparatus for a computer
US7159415B2 (en) 2001-03-13 2007-01-09 Applied Design And Engineering Limited Drawer storage
US6992247B2 (en) 2002-01-02 2006-01-31 American Power Conversion Corporation Toolless mounting system and method for an adjustable scalable rack power system
US6445088B1 (en) 2001-03-20 2002-09-03 American Power Conversion Multipurpose data port
JP3870048B2 (en) 2001-03-26 2007-01-17 三星電子株式会社 Multi-room refrigerator and control method thereof
WO2002099683A1 (en) 2001-03-27 2002-12-12 Netbotz, Inc. Methods for displaying physical network topology and environmental status by location, organization, or responsible party
ITMC20010017U1 (en) 2001-03-29 2002-09-30 Compagnucci Spa EXTRACTABLE RACK FOR FURNITURE EQUIPPED WITH A SUPPORTING CARTER FOR THE PROTECTION OF THE SLIDING GUIDES.
US6535382B2 (en) 2001-04-12 2003-03-18 Johnson Controls Technology Company Cooling system for electronic equipment cabinets
US6424119B1 (en) 2001-04-19 2002-07-23 American Power Conversion Multiple energy storage device controller
ATE358803T1 (en) 2001-05-09 2007-04-15 Maersk Container Ind As COOLING UNIT AND CONTAINER WITH THIS UNIT
US6701725B2 (en) 2001-05-11 2004-03-09 Field Diagnostic Services, Inc. Estimating operating parameters of vapor compression cycle equipment
US7698389B2 (en) 2001-05-16 2010-04-13 Hewlett-Packard Development Company, L.P. Device configuration in a distributed environment
US6848989B2 (en) 2001-05-30 2005-02-01 Kongo Kabushiki Kaisha Environmental improvement device for a storage body
DE10128367A1 (en) 2001-06-12 2003-01-02 Fujitsu Siemens Computers Gmbh Computer cabinet and arrangement of a computer cabinet
US6653816B2 (en) 2001-06-24 2003-11-25 Motorola, Inc. Battery with embedded power management
US7207041B2 (en) 2001-06-28 2007-04-17 Tranzeo Wireless Technologies, Inc. Open platform architecture for shared resource access management
CN1549969A (en) 2001-07-06 2004-11-24 Method and system for correlating and determining root causes of system and enterprise events
US6718781B2 (en) 2001-07-11 2004-04-13 Thermo King Corporation Refrigeration unit apparatus and method
US6415628B1 (en) 2001-07-25 2002-07-09 Praxair Technology, Inc. System for providing direct contact refrigeration
US20030033463A1 (en) 2001-08-10 2003-02-13 Garnett Paul J. Computer system storage
US7519909B2 (en) 2001-08-10 2009-04-14 American Power Conversion Corporation Uninterruptible power supply (UPS) devices monitoring system
US6505086B1 (en) 2001-08-13 2003-01-07 William A. Dodd, Jr. XML sensor system
US20030042004A1 (en) 2001-08-29 2003-03-06 Shlomo Novotny Interchangeable cartridges for cooling electronic components
US6942018B2 (en) 2001-09-28 2005-09-13 The Board Of Trustees Of The Leland Stanford Junior University Electroosmotic microchannel cooling system
JP3829846B2 (en) 2001-10-03 2006-10-04 三菱電機株式会社 Uninterruptible power system
US6474087B1 (en) 2001-10-03 2002-11-05 Carrier Corporation Method and apparatus for the control of economizer circuit flow for optimum performance
JPWO2003034523A1 (en) 2001-10-11 2005-02-03 株式会社日立製作所 Home fuel cell system
FR2831019B1 (en) 2001-10-12 2004-07-09 Legrand Sa REFRIGERATED CABINET FOR ELECTRICAL OR ELECTRONIC APPARATUS, INCLUDING AN IMPROVED VENTILATION CIRCUIT
US6600298B2 (en) 2001-10-31 2003-07-29 Dell Products L.P. Switching DC-DC converter with the output voltage changing inversely to any change in the converter inductor current
US6983212B2 (en) 2001-11-27 2006-01-03 American Power Conversion Corporation Battery management system and method
EP1316933B1 (en) 2001-11-28 2006-08-09 Matsushita Electric Industrial Co., Ltd. Home security system
EP1490633A4 (en) 2001-11-30 2009-12-02 Choon-Kyoung Park Air conditioning apparatus
US20030130912A1 (en) * 2002-01-04 2003-07-10 Davis Tommy Lee Equipment management system
US6633835B1 (en) 2002-01-10 2003-10-14 Networks Associates Technology, Inc. Prioritized data capture, classification and filtering in a network monitoring environment
US7519000B2 (en) 2002-01-30 2009-04-14 Panduit Corp. Systems and methods for managing a network
US7024573B2 (en) 2002-02-05 2006-04-04 Hewlett-Packard Development Company, L.P. Method and apparatus for cooling heat generating components
US20030149608A1 (en) 2002-02-06 2003-08-07 Kall Jonathan J. Suite of configurable supply chain infrastructure modules for deploying collaborative e-manufacturing solutions
EP1476834A1 (en) * 2002-02-07 2004-11-17 Thinkdynamics Inc. Method and system for managing resources in a data center
US6927977B2 (en) 2002-02-12 2005-08-09 Siemens Communications, Inc. Blower exhaust for a closed frame
US20030154285A1 (en) 2002-02-13 2003-08-14 International Business Machines Corporation Method and system for assigning network addreses
CN1277178C (en) 2002-02-19 2006-09-27 株式会社东芝 Data display system, data relay device, data relay method, data system, receiving device, and data read method
US6640889B1 (en) 2002-03-04 2003-11-04 Visteon Global Technologies, Inc. Dual loop heat and air conditioning system
KR100387358B1 (en) 2002-03-22 2003-06-12 Ec Tech Co Ltd System for cooling computer component using thermoelectric element
US6625019B1 (en) 2002-04-01 2003-09-23 White Rock Networks Systems and methods for a rack-mounted communications switch component
US6668565B1 (en) 2002-04-12 2003-12-30 American Power Conversion Rack-mounted equipment cooling
JP2004007950A (en) 2002-04-15 2004-01-08 Fuji Electric Holdings Co Ltd Switching power unit
US7197553B2 (en) 2002-04-19 2007-03-27 Nortel Networks Limited Network system having a virtual-service-module
US6850426B2 (en) 2002-04-30 2005-02-01 Honeywell International Inc. Synchronous and bi-directional variable frequency power conversion systems
JP4018443B2 (en) 2002-05-13 2007-12-05 株式会社前川製作所 Thermosiphon chiller refrigerator for cold regions
US20040078708A1 (en) 2002-05-17 2004-04-22 Chuang Li Methods for facilitating the installation of computer devices
US7072960B2 (en) 2002-06-10 2006-07-04 Hewlett-Packard Development Company, L.P. Generating automated mappings of service demands to server capacities in a distributed computer system
US20030232598A1 (en) 2002-06-13 2003-12-18 Daniel Aljadeff Method and apparatus for intrusion management in a wireless network using physical location determination
US6944020B2 (en) 2002-06-20 2005-09-13 Delphi Technologies, Inc. Computer enclosure air distribution system
US7152100B2 (en) 2002-07-09 2006-12-19 Adtran, Inc. System and method for provisioning network access devices
US6819560B2 (en) 2002-07-11 2004-11-16 Storage Technology Corporation Forced air system for cooling a high density array of disk drives
US6889752B2 (en) 2002-07-11 2005-05-10 Avaya Technology Corp. Systems and methods for weatherproof cabinets with multiple compartment cooling
US6704196B1 (en) 2002-07-25 2004-03-09 Allied Systems Design, Inc. Flow-through cooling in-the-round system
KR100482004B1 (en) 2002-07-27 2005-04-13 삼성전자주식회사 Refrigerator system and upgrade methode the same
US6611428B1 (en) 2002-08-12 2003-08-26 Motorola, Inc. Cabinet for cooling electronic modules
US6695577B1 (en) 2002-08-13 2004-02-24 American Power Conversion Fan grill
US20040031280A1 (en) 2002-08-14 2004-02-19 Delaware Capital Formation, Inc. Refrigeration system
US20040075343A1 (en) 2002-09-05 2004-04-22 Paul Wareham System and method for power load management
US6662576B1 (en) 2002-09-23 2003-12-16 Vai Holdings Llc Refrigeration system with de-superheating bypass
US6807056B2 (en) 2002-09-24 2004-10-19 Hitachi, Ltd. Electronic equipment
US7007900B2 (en) 2002-10-01 2006-03-07 Andrew Corporation Cable hanger
US6825644B2 (en) 2002-11-14 2004-11-30 Fyre Storm, Inc. Switching power converter
US6775137B2 (en) 2002-11-25 2004-08-10 International Business Machines Corporation Method and apparatus for combined air and liquid cooling of stacked electronics components
US7752858B2 (en) 2002-11-25 2010-07-13 American Power Conversion Corporation Exhaust air removal system
US7500911B2 (en) 2002-11-25 2009-03-10 American Power Conversion Corporation Exhaust air removal system
WO2004049773A2 (en) 2002-11-25 2004-06-10 American Power Conversion Corporation Exhaust air removal system
US7429359B2 (en) 2002-12-19 2008-09-30 Edc Biosystems, Inc. Source and target management system for high throughput transfer of liquids
JP2004184135A (en) 2002-11-29 2004-07-02 Sanyo Electric Co Ltd Remaining-capacity calculating system of battery
US6788535B2 (en) 2002-12-12 2004-09-07 3M Innovative Properties Company Outdoor electronic equipment cabinet
US6745590B1 (en) 2003-01-13 2004-06-08 American Power Conversion Condensate removal system
US7120689B2 (en) 2003-01-23 2006-10-10 Sbc Properties, L.P. Receiving network metrics data from disparate devices and displaying in a host format
US7786616B2 (en) 2003-02-07 2010-08-31 Cummins Power Generation Inc. Generator with DC boost and split bus bidirectional DC-to-DC converter for uninterruptible power supply system or for enhanced load pickup
US6914415B2 (en) 2003-02-14 2005-07-05 Motorola, Inc. Battery adaptor to facilitate reconditioning in a smart charger
TW591870B (en) 2003-02-18 2004-06-11 Delta Electronics Inc Integrated converter with three-phase power factor correction
US6959558B2 (en) 2003-03-06 2005-11-01 American Power Conversion Corp. Systems and methods for head pressure control
US20040186905A1 (en) 2003-03-20 2004-09-23 Young Donald E. System and method for provisioning resources
US6853564B2 (en) 2003-03-25 2005-02-08 Switching Power, Inc. Switched linear inductor circuit for switched power supply
CN2612211Y (en) 2003-04-02 2004-04-14 大唐移动通信设备有限公司 Forced air cooler inside of electronic appliance cabinet
US6881142B1 (en) 2003-09-12 2005-04-19 Degree C Intelligent networked fan assisted tiles for adaptive thermal management of thermally sensitive rooms
EP1616237B1 (en) 2003-04-14 2017-10-25 Schneider Electric IT Corporation Environmental monitoring device
US7799474B2 (en) 2003-04-29 2010-09-21 Hewlett-Packard Development Company, L.P. System and method for managing electrically isolated fuel cell powered devices within an equipment rack
US7378165B2 (en) 2003-04-29 2008-05-27 Hewlett-Packard Development Company, L.P. System and method for providing electrical power to an equipment rack using a fuel cell
US6724624B1 (en) 2003-05-05 2004-04-20 Douglas A. Dodson Housing with directed-flow cooling for computer
JP4349562B2 (en) 2003-05-09 2009-10-21 独立行政法人理化学研究所 Spatial identification method and program
US7033267B2 (en) 2003-05-13 2006-04-25 American Power Conversion Corporation Rack enclosure
US7112131B2 (en) 2003-05-13 2006-09-26 American Power Conversion Corporation Rack enclosure
US7546475B2 (en) 2003-05-13 2009-06-09 Hewlett-Packard Development Company, L.P. Power-aware adaptation in a data center
US7346344B2 (en) 2003-05-30 2008-03-18 Aol Llc, A Delaware Limited Liability Company Identity-based wireless device configuration
JP4311538B2 (en) 2003-06-27 2009-08-12 株式会社日立製作所 Disk storage device cooling structure
US6975092B2 (en) 2003-07-03 2005-12-13 Dell Products L.P. Encrypted response smart battery
US7146353B2 (en) 2003-07-22 2006-12-05 Hewlett-Packard Development Company, L.P. Resource allocation for multiple applications
KR20050017350A (en) 2003-08-13 2005-02-22 삼성전자주식회사 Method for generating encryption key without an input device and apparatus therefor
US7001263B2 (en) 2003-08-14 2006-02-21 Cornell Research Foundation, Inc. Air flow monitoring and control system with reduced false alarms
US7259477B2 (en) 2003-08-15 2007-08-21 American Power Conversion Corporation Uninterruptible power supply
KR20050023758A (en) 2003-09-02 2005-03-10 엘지전자 주식회사 Condenser
US20050061013A1 (en) 2003-09-10 2005-03-24 Bond Richard C. Method and apparatus for cooling devices that in use generate unwanted heat
US7508672B2 (en) 2003-09-10 2009-03-24 Qnx Cooling Systems Inc. Cooling system
US7236363B2 (en) 2003-09-22 2007-06-26 Hewlett-Packard Development Company, L.P. Liquid cooled system module
US7012807B2 (en) 2003-09-30 2006-03-14 International Business Machines Corporation Thermal dissipation assembly and fabrication method for electronics drawer of a multiple-drawer electronics rack
US6838925B1 (en) 2003-10-07 2005-01-04 American Power Conversion Corporation Three level inverter
JP4351517B2 (en) 2003-11-14 2009-10-28 株式会社日立製作所 Data center device management method, device management server, data center device management system and program
US6832504B1 (en) 2003-11-19 2004-12-21 Be Intellectual Property, Inc. Liquid sensing system for an aircraft galley cooler using a two phase working fluid
KR100975059B1 (en) 2003-11-25 2010-08-11 삼성전자주식회사 Apparatus for controlling temperature in audio amplifier
US7106590B2 (en) 2003-12-03 2006-09-12 International Business Machines Corporation Cooling system and method employing multiple dedicated coolant conditioning units for cooling multiple electronics subsystems
US7270174B2 (en) 2003-12-16 2007-09-18 International Business Machines Corporation Method, system and program product for automatically checking coolant loops of a cooling system for a computing environment
US7000467B2 (en) 2003-12-16 2006-02-21 International Business Machines Corporation Method, system and program product for monitoring rate of volume change of coolant within a cooling system
US7096679B2 (en) 2003-12-23 2006-08-29 Tecumseh Products Company Transcritical vapor compression system and method of operating including refrigerant storage tank and non-variable expansion device
US7278273B1 (en) 2003-12-30 2007-10-09 Google Inc. Modular data center
JP2005193749A (en) 2004-01-06 2005-07-21 Sanden Corp Controller
US7612472B2 (en) 2004-01-23 2009-11-03 American Power Conversion Corporation Method and apparatus for monitoring energy storage devices
US7379305B2 (en) 2004-01-23 2008-05-27 American Power Conversion Corporation Modular UPS
US7446433B2 (en) 2004-01-23 2008-11-04 American Power Conversion Corporation Methods and apparatus for providing uninterruptible power
US7281958B2 (en) 2004-01-23 2007-10-16 American Power Conversion Corporation Power terminal block
US7432615B2 (en) 2004-01-29 2008-10-07 American Power Conversion Corporation Uninterruptable power supply system and method
US6955058B2 (en) 2004-01-30 2005-10-18 Carrier Corporation Refrigerant cycle with tandem economized and conventional compressors
DE102004008460B4 (en) 2004-02-17 2006-02-02 Rittal Gmbh & Co. Kg Arrangement for cooling control cabinets
US7050312B2 (en) 2004-03-09 2006-05-23 Eaton Power Quality Corporation Multi-mode uninterruptible power supplies and methods of operation thereof
US7475556B2 (en) 2004-03-15 2009-01-13 Parker Hannifin Corporation System and apparatus controlling a variable speed compressor system
US7062933B2 (en) 2004-03-24 2006-06-20 Intel Corporation Separate thermal and electrical throttling limits in processors
US20050213306A1 (en) 2004-03-25 2005-09-29 Lockheed Martin Corporation Environmental control method and apparatus for electronic device enclosures
US7104081B2 (en) 2004-03-30 2006-09-12 International Business Machines Corproation Condensate removal system and method for facilitating cooling of an electronics system
US8108496B2 (en) 2004-04-07 2012-01-31 American Power Conversion Corporation Method and apparatus for selecting forwarding modes
US20050237716A1 (en) 2004-04-21 2005-10-27 International Business Machines Corporation Air flow system and method for facilitating cooling of stacked electronics components
US7233492B2 (en) 2004-04-22 2007-06-19 Hewlett-Packard Development Company, L.P. Cooling systems and methods for same
US7647787B2 (en) 2004-04-22 2010-01-19 Hewlett-Packard Development Company, L.P. Upgradeable, modular data center cooling apparatus
US6967841B1 (en) 2004-05-07 2005-11-22 International Business Machines Corporation Cooling assembly for electronics drawer using passive fluid loop and air-cooled cover
US6973797B2 (en) 2004-05-10 2005-12-13 York International Corporation Capacity control for economizer refrigeration systems
US7175138B2 (en) 2004-05-10 2007-02-13 Andrew Corporation Stackable cable hanger
US7239043B2 (en) 2004-05-26 2007-07-03 Eaton Power Quality Corporation Power conversion apparatus and methods using an adaptive waveform reference
JP2006014723A (en) 2004-06-01 2006-01-19 Sumitomo Chemical Co Ltd Common marmoset-derived glyceraldehyde-3-phosphate dehydrogenase gene and use thereof
US8341965B2 (en) 2004-06-24 2013-01-01 Raytheon Company Method and system for cooling
US7536370B2 (en) 2004-06-24 2009-05-19 Sun Microsystems, Inc. Inferential diagnosing engines for grid-based computing systems
US7187549B2 (en) 2004-06-30 2007-03-06 Teradyne, Inc. Heat exchange apparatus with parallel flow
US7421575B2 (en) 2004-07-16 2008-09-02 Hewlett-Packard Development Company, L.P. Configuring a physical platform in a reconfigurable data center
US7488202B2 (en) 2004-07-28 2009-02-10 American Power Conversion Corporation Multiport cabling system and method
US7456518B2 (en) 2004-08-31 2008-11-25 American Power Conversion Corporation Method and apparatus for providing uninterruptible power
US7274112B2 (en) 2004-08-31 2007-09-25 American Power Conversion Corporation Method and apparatus for providing uninterruptible power
US7939968B2 (en) 2004-08-31 2011-05-10 American Power Conversion Corporation Method and apparatus for providing uninterruptible power
US7737580B2 (en) 2004-08-31 2010-06-15 American Power Conversion Corporation Method and apparatus for providing uninterruptible power
US7086247B2 (en) 2004-08-31 2006-08-08 International Business Machines Corporation Cooling system and method employing auxiliary thermal capacitor unit for facilitating continuous operation of an electronics rack
US7514815B2 (en) 2004-09-28 2009-04-07 American Power Conversion Corporation System and method for allocating power to loads
DE602005018669D1 (en) 2004-09-30 2010-02-11 Danfoss As MODEL FORECASTING-CONTROLLED COOLING SYSTEM
JP4169157B2 (en) * 2004-10-04 2008-10-22 インターナショナル・ビジネス・マシーンズ・コーポレーション Management server, system and program using the same
US7506768B2 (en) 2004-10-15 2009-03-24 American Power Conversion Corporation Blanking panel for equipment rack or enclosure
US20060082263A1 (en) 2004-10-15 2006-04-20 American Power Conversion Corporation Mobile data center
US7228707B2 (en) 2004-10-28 2007-06-12 Carrier Corporation Hybrid tandem compressor system with multiple evaporators and economizer circuit
JP2006132818A (en) 2004-11-04 2006-05-25 Matsushita Electric Ind Co Ltd Control method for refrigerating cycle device, and refrigerating cycle device using the same
US7293666B2 (en) 2004-11-17 2007-11-13 American Power Conversion Corporation Equipment enclosure kit and assembly method
JP4553307B2 (en) * 2004-11-19 2010-09-29 インターナショナル・ビジネス・マシーンズ・コーポレーション Information processing apparatus, control method, and program
US7165412B1 (en) 2004-11-19 2007-01-23 American Power Conversion Corporation IT equipment cooling
US7152831B2 (en) 2004-11-23 2006-12-26 The Lamson & Sessions Co. Conduit support
US6973801B1 (en) 2004-12-09 2005-12-13 International Business Machines Corporation Cooling system and method employing a closed loop coolant path and micro-scaled cooling structure within an electronics subsystem of an electronics rack
US7184269B2 (en) 2004-12-09 2007-02-27 International Business Machines Company Cooling apparatus and method for an electronics module employing an integrated heat exchange assembly
US7274566B2 (en) 2004-12-09 2007-09-25 International Business Machines Corporation Cooling apparatus for an electronics subsystem employing a coolant flow drive apparatus between coolant flow paths
US7185501B2 (en) 2004-12-16 2007-03-06 General Electric Company Cryogenic cooling system and method with backup cold storage device
KR20060077396A (en) 2004-12-30 2006-07-05 엘지전자 주식회사 Refrigerator and hybrid cooling system of refrigerator
US7426453B2 (en) 2005-01-14 2008-09-16 Hewlett-Packard Development Company, L.P. Workload placement based upon CRAC unit capacity utilizations
US7603874B2 (en) 2005-01-24 2009-10-20 American Power Conversion Corporation Split power input to chiller
US7472043B1 (en) 2005-02-18 2008-12-30 Tellabs Bedford, Inc. Mass customization configurator
JP2006250395A (en) * 2005-03-09 2006-09-21 Matsushita Electric Ind Co Ltd Cooling device and electronic apparatus comprising the same
US7522626B2 (en) 2005-03-14 2009-04-21 American Power Conversion Corporation Communications system and method
DE502005000286D1 (en) 2005-03-18 2007-02-15 Schroff Gmbh Housing for receiving electronic plug-in modules
US7385810B2 (en) 2005-04-18 2008-06-10 International Business Machines Corporation Apparatus and method for facilitating cooling of an electronics rack employing a heat exchange assembly mounted to an outlet door cover of the electronics rack
US7402921B2 (en) 2005-04-21 2008-07-22 American Power Conversion Corporation Method and apparatus for providing uninterruptible power
US7908126B2 (en) 2005-04-28 2011-03-15 Emerson Climate Technologies, Inc. Cooling system design simulator
DK1877950T3 (en) * 2005-05-02 2019-02-25 Schneider Electric It Corp Methods and systems for control of plant power and cooling
US7286351B2 (en) 2005-05-06 2007-10-23 International Business Machines Corporation Apparatus and method for facilitating cooling of an electronics rack employing a closed loop heat exchange system
US7841199B2 (en) 2005-05-17 2010-11-30 American Power Conversion Corporation Cold aisle isolation
US7315448B1 (en) 2005-06-01 2008-01-01 Hewlett-Packard Development Company, L.P. Air-cooled heat generating device airflow control system
US7254307B2 (en) 2005-06-03 2007-08-07 Telect Inc. Fiber breakout system
JP4751153B2 (en) * 2005-06-08 2011-08-17 株式会社日立製作所 Storage system
KR100679250B1 (en) 2005-07-22 2007-02-05 한국전자통신연구원 Method for automatically selecting a cluster header in a wireless sensor network and for dynamically configuring a secure wireless sensor network
US7672590B2 (en) 2005-07-28 2010-03-02 Netapp, Inc. Data center with mobile data cabinets and method of mobilizing and connecting data processing devices in a data center using consolidated data communications and power connections
US7366632B2 (en) 2005-08-02 2008-04-29 International Business Machines Corporation Method and apparatus for three-dimensional measurements
US8429630B2 (en) 2005-09-15 2013-04-23 Ca, Inc. Globally distributed utility computing cloud
US7352083B2 (en) 2005-09-16 2008-04-01 American Power Conversion Corporation Apparatus for and method of UPS operation
US7406839B2 (en) 2005-10-05 2008-08-05 American Power Conversion Corporation Sub-cooling unit for cooling system and method
US7298618B2 (en) 2005-10-25 2007-11-20 International Business Machines Corporation Cooling apparatuses and methods employing discrete cold plates compliantly coupled between a common manifold and electronics components of an assembly to be cooled
US7298617B2 (en) 2005-10-25 2007-11-20 International Business Machines Corporation Cooling apparatus and method employing discrete cold plates disposed between a module enclosure and electronics components to be cooled
US7726144B2 (en) 2005-10-25 2010-06-01 Hewlett-Packard Development Company, L.P. Thermal management using stored field replaceable unit thermal information
US7620480B2 (en) 2005-11-01 2009-11-17 Hewlett-Packard Development Company, L.P. Cooling components across a continuum
US7568360B1 (en) 2005-11-01 2009-08-04 Hewlett-Packard Development Company, L.P. Air re-circulation effect reduction system
US7310953B2 (en) 2005-11-09 2007-12-25 Emerson Climate Technologies, Inc. Refrigeration system including thermoelectric module
US7272005B2 (en) 2005-11-30 2007-09-18 International Business Machines Corporation Multi-element heat exchange assemblies and methods of fabrication for a cooling system
US20070121295A1 (en) 2005-11-30 2007-05-31 International Business Machines Corporation Hybrid liquid-air cooled module
US20070163748A1 (en) 2006-01-19 2007-07-19 American Power Conversion Corporation Cooling system and method
US8672732B2 (en) 2006-01-19 2014-03-18 Schneider Electric It Corporation Cooling system and method
US7365973B2 (en) 2006-01-19 2008-04-29 American Power Conversion Corporation Cooling system and method
US7862410B2 (en) 2006-01-20 2011-01-04 American Power Conversion Corporation Air removal unit
JP5087560B2 (en) 2006-02-10 2012-12-05 アメリカン パワー コンバージョン コーポレイション Storage rack management system and method
US8764527B2 (en) 2006-02-10 2014-07-01 Schneider Electric It Corporation Method and apparatus for providing cooling air to equipment
US7456524B2 (en) 2006-03-31 2008-11-25 American Power Conversion Corporation Apparatus for and methods of polyphase power conversion
US7558649B1 (en) 2006-05-03 2009-07-07 Hewlett-Packard Development Company, L.P. Method for predicting airflow rates
JP4800837B2 (en) 2006-05-22 2011-10-26 株式会社日立製作所 Computer system, power consumption reduction method thereof, and program thereof
CA2653808C (en) 2006-06-01 2014-10-14 Exaflop Llc Controlled warm air capture
EP2032918A4 (en) 2006-06-15 2011-09-21 Valan R Martini Energy saving system and method for cooling computer data center and telecom equipment
US7925389B2 (en) 2006-06-15 2011-04-12 International Business Machines Corporation Method of detecting recirculation of heated air within a rack enclosure
US7349213B2 (en) 2006-06-29 2008-03-25 International Business Machines Corporation Coolant control unit, and cooled electronics system and method employing the same
US7549070B2 (en) 2006-06-30 2009-06-16 Sun Microsystems, Inc. Method and apparatus for generating a dynamic power-flux map for a set of computer systems
US7844440B2 (en) 2006-07-07 2010-11-30 Edsa Micro Corporation Systems and methods for real-time dynamic simulation of uninterruptible power supply solutions and their control logic systems
US7447022B2 (en) 2006-08-09 2008-11-04 Hewlett-Packard Development Company, L.P. Rack-mount equipment bay cooling heat exchanger
US8322155B2 (en) 2006-08-15 2012-12-04 American Power Conversion Corporation Method and apparatus for cooling
US20080041077A1 (en) 2006-08-15 2008-02-21 American Power Conversion Corporation Method and apparatus for cooling
US9568206B2 (en) 2006-08-15 2017-02-14 Schneider Electric It Corporation Method and apparatus for cooling
US8327656B2 (en) 2006-08-15 2012-12-11 American Power Conversion Corporation Method and apparatus for cooling
US7397661B2 (en) 2006-08-25 2008-07-08 International Business Machines Corporation Cooled electronics system and method employing air-to-liquid heat exchange and bifurcated air flow
DE102006041788B4 (en) 2006-09-06 2012-06-14 Airbus Operations Gmbh An aircraft electronics cooling device for an aircraft with a liquid cooling system
US7705489B2 (en) 2006-09-08 2010-04-27 American Power Conversion Corporation Method and apparatus for providing uninterruptible power
US7652393B2 (en) 2006-09-14 2010-01-26 American Power Conversion Corporation Apparatus and method for employing a DC source with an uninterruptible power supply
US20080105412A1 (en) 2006-11-03 2008-05-08 American Power Conversion Corporation Continuous cooling capacity regulation using supplemental heating
US20080105753A1 (en) 2006-11-03 2008-05-08 American Power Conversion Corporation Modulating electrical reheat with contactors
US20080104985A1 (en) 2006-11-03 2008-05-08 American Power Conversion Corporation Constant temperature CRAC control algorithm
US20080142068A1 (en) 2006-12-18 2008-06-19 American Power Conversion Corporation Direct Thermoelectric chiller assembly
US7681404B2 (en) 2006-12-18 2010-03-23 American Power Conversion Corporation Modular ice storage for uninterruptible chilled water
US7433229B2 (en) 2006-12-19 2008-10-07 Phison Electronics Corp. Flash memory device with shunt
US8425287B2 (en) 2007-01-23 2013-04-23 Schneider Electric It Corporation In-row air containment and cooling system and method
AU2012201080B2 (en) * 2007-01-24 2014-07-17 Schneider Electric It Corporation System and method for evaluating equipment rack cooling performance
US7856549B2 (en) 2007-01-24 2010-12-21 Hewlett-Packard Development Company, L.P. Regulating power consumption
CN101589656B (en) 2007-01-24 2013-05-01 美国能量变换公司 System and method for evaluating equipment rack cooling performance
US7676280B1 (en) 2007-01-29 2010-03-09 Hewlett-Packard Development Company, L.P. Dynamic environmental management
US7688048B2 (en) 2007-02-21 2010-03-30 American Power Conversion Corporation 3-phase high power UPS
US8205205B2 (en) 2007-03-16 2012-06-19 Sap Ag Multi-objective allocation of computational jobs in client-server or hosting environments
US7554473B2 (en) 2007-05-02 2009-06-30 Cirrus Logic, Inc. Control system using a nonlinear delta-sigma modulator with nonlinear process modeling
EP2147585B1 (en) 2007-05-15 2016-11-02 Schneider Electric IT Corporation Method and system for managing facility power and cooling
US20090019875A1 (en) 2007-07-19 2009-01-22 American Power Conversion Corporation A/v cooling system and method
US20090030554A1 (en) 2007-07-26 2009-01-29 Bean Jr John H Cooling control device and method
US7781914B2 (en) 2007-08-10 2010-08-24 American Power Conversion Corporation Input and output power modules configured to provide selective power to an uninterruptible power supply
US8732706B2 (en) 2007-11-27 2014-05-20 Hewlett-Packard Development Company, L.P. Generating governing metrics for resource provisioning
US7832925B2 (en) 2007-12-05 2010-11-16 International Business Machines Corporation Apparatus and method for simulating heated airflow exhaust of an electronics subsystem, electronics rack or row of electronics racks
US7979250B2 (en) 2007-12-05 2011-07-12 International Business Machines Corporation Method of laying out a data center using a plurality of thermal simulators
US7707880B2 (en) 2008-02-15 2010-05-04 International Business Machines Corporation Monitoring method and system for determining rack airflow rate and rack power consumption
US8201028B2 (en) 2008-02-15 2012-06-12 The Pnc Financial Services Group, Inc. Systems and methods for computer equipment management
US7878007B2 (en) 2008-02-15 2011-02-01 International Business Machines Corporation Monitoring method and system for determining airflow rate through and heat removal rate of an air-conditioning unit
US7657347B2 (en) 2008-02-15 2010-02-02 International Business Machines Corporation Temperature-based monitoring method and system for determining first and second fluid flow rates through a heat exchanger
US7791882B2 (en) 2008-04-23 2010-09-07 International Business Machines Corporation Energy efficient apparatus and method for cooling an electronics rack
US8053926B2 (en) 2008-06-16 2011-11-08 American Power Conversion Corporation Methods and systems for managing facility power and cooling
US8306794B2 (en) 2008-06-26 2012-11-06 International Business Machines Corporation Techniques for thermal modeling of data centers to improve energy efficiency
US8849630B2 (en) 2008-06-26 2014-09-30 International Business Machines Corporation Techniques to predict three-dimensional thermal distributions in real-time
CN102077033B (en) 2008-06-30 2013-07-03 惠普发展公司,有限责任合伙企业 Method and system for optimizing fluid flow distribution in a structure
TW201106760A (en) 2008-08-27 2011-02-16 Koninkl Philips Electronics Nv Commissioning a network system
US8473265B2 (en) 2008-10-27 2013-06-25 Schneider Electric It Corporation Method for designing raised floor and dropped ceiling in computing facilities
US8209056B2 (en) 2008-11-25 2012-06-26 American Power Conversion Corporation System and method for assessing and managing data center airflow and energy usage
US9519517B2 (en) 2009-02-13 2016-12-13 Schneider Electtic It Corporation Data center control
US9778718B2 (en) 2009-02-13 2017-10-03 Schneider Electric It Corporation Power supply and data center control
JP5155922B2 (en) 2009-03-31 2013-03-06 株式会社タイトー game machine
US9904331B2 (en) 2009-04-01 2018-02-27 Schneider Electric It Corporation Method for computing cooling redundancy at the rack level
US8219362B2 (en) 2009-05-08 2012-07-10 American Power Conversion Corporation System and method for arranging equipment in a data center
US8355890B2 (en) 2009-05-08 2013-01-15 American Power Conversion Corporation System and method for predicting maximum cooler and rack capacities in a data center
US8249825B2 (en) 2009-05-08 2012-08-21 American Power Conversion Corporation System and method for predicting cooling performance of arrangements of equipment in a data center
AU2010253739B2 (en) 2009-05-29 2015-07-16 Emerson Climate Technologies Retail Solutions, Inc. System and method for monitoring and evaluating equipment operating parameter modifications
US8155922B2 (en) 2009-06-04 2012-04-10 Eaton Corporation Electrical device cooling efficiency monitoring
US8228046B2 (en) 2009-06-16 2012-07-24 American Power Conversion Corporation Apparatus and method for operating an uninterruptible power supply
US8244502B2 (en) 2009-08-12 2012-08-14 International Business Machines Corporation Knowledge-based models for data centers
US8229713B2 (en) 2009-08-12 2012-07-24 International Business Machines Corporation Methods and techniques for creating and visualizing thermal zones
CN201514313U (en) 2009-10-15 2010-06-23 江苏万工科技集团有限公司 Visualized airflow detector
US8478451B2 (en) 2009-12-14 2013-07-02 Intel Corporation Method and apparatus for dynamically allocating power in a data center
US8489745B2 (en) 2010-02-26 2013-07-16 International Business Machines Corporation Optimizing power consumption by dynamic workload adjustment
US20110215645A1 (en) 2010-03-05 2011-09-08 Active Power, Inc. Containerized continuous power system and method
US8972217B2 (en) 2010-06-08 2015-03-03 Schneider Electric It Corporation System and method for predicting temperature values in a data center
US8493221B2 (en) 2010-06-24 2013-07-23 International Business Machines Corporation Filter fouling detection using comparative temperature rise analysis
EP2593724A4 (en) 2010-07-14 2018-06-27 Harper International Corporation Airflow distribution system
US8509959B2 (en) 2010-08-12 2013-08-13 Schneider Electric It Corporation System and method for predicting transient cooling performance for a data center
US8996180B2 (en) 2010-09-17 2015-03-31 Schneider Electric It Corporation System and method for predicting perforated tile airflow in a data center
US8409807B2 (en) 2010-10-22 2013-04-02 T2 Biosystems, Inc. NMR systems and methods for the rapid detection of analytes
US8825451B2 (en) 2010-12-16 2014-09-02 Schneider Electric It Corporation System and methods for rack cooling analysis
US8688413B2 (en) 2010-12-30 2014-04-01 Christopher M. Healey System and method for sequential placement of cooling resources within data center layouts
EP2482213A1 (en) 2011-01-31 2012-08-01 Fujitsu Limited Simulation apparatus and method for simulation of energy use in a data center
US8949091B2 (en) 2011-03-09 2015-02-03 Tata Consultancy Services Limited Method and system for thermal management by quantitative determination of cooling characteristics of data center
US20120232879A1 (en) 2011-03-10 2012-09-13 International Business Machines Corporation Data center efficiency analyses and optimization
US9223905B2 (en) 2011-03-25 2015-12-29 Schneider Electric It Corporation Systems and methods for predicting fluid dynamics in a data center
US8793105B2 (en) 2011-03-31 2014-07-29 The Procter & Gamble Company Systems and methods for airborne particle contamination control
US9195243B2 (en) 2011-05-24 2015-11-24 Aten International Co., Ltd. System and method of safe and effective energy usage and conservation for data centers with rack power distribution units
US8725307B2 (en) 2011-06-28 2014-05-13 Schneider Electric It Corporation System and method for measurement aided prediction of temperature and airflow values in a data center
US9183328B2 (en) 2011-10-26 2015-11-10 Engine Simulation Partners, Llc Method and apparatus for modeling interactions of the fluid with system boundaries in fluid dynamic systems
AU2011383606A1 (en) 2011-12-22 2014-07-17 Schneider Electric It Corporation System and method for prediction of temperature values in an electronics system
CN104137105B (en) 2011-12-22 2017-07-11 施耐德电气It公司 Impact analysis on temporal event to the temperature in data center
EP2795412A4 (en) 2011-12-23 2015-08-12 Schneider Electric It Corp Dropped ceiling system cooling prediction
US9152154B2 (en) 2012-08-01 2015-10-06 International Business Machines Corporation Multi-dimensional heating and cooling system
US8913385B2 (en) 2012-10-08 2014-12-16 Hamilton Sundstrand Corporation Efficient cooling duct
US10157245B2 (en) 2012-10-31 2018-12-18 Schneider Electric It Corporation System and method for fluid dynamics prediction with an enhanced potential flow model

Patent Citations (99)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4827151A (en) * 1987-02-20 1989-05-02 Kabushiki Kaisha Toshiba Uninterruptible power supply utilizing a synchronized chopper for power factor improvement
US4823290A (en) * 1987-07-21 1989-04-18 Honeywell Bull Inc. Method and apparatus for monitoring the operating environment of a computer system
US5404136A (en) * 1988-08-23 1995-04-04 Marsden; Derek R. Method and apparatus for monitoring the consumption of utilities in business premises
US4962734A (en) * 1990-03-14 1990-10-16 Paccar Inc. Electrically driven, circumferentially supported fan
US20020004912A1 (en) * 1990-06-01 2002-01-10 Amphus, Inc. System, architecture, and method for logical server and other network devices in a dynamically configurable multi-server network environment
US20030188208A1 (en) * 1990-06-01 2003-10-02 Amphus, Inc. System, method, and architecture for dynamic server power management and dynamic workload management for multi-server environment
US20030200473A1 (en) * 1990-06-01 2003-10-23 Amphus, Inc. System and method for activity or event based dynamic energy conserving server reconfiguration
US5367670A (en) * 1991-06-24 1994-11-22 Compaq Computer Corporation Computer system manager for monitoring events and operating parameters and generating alerts
US5682949A (en) * 1992-05-22 1997-11-04 Globalmic, Inc. Energy management system
US5381554A (en) * 1992-11-24 1995-01-10 Exide Electronics Uninterrupted power supply (UPS) system interfacing with communications network
US5462225A (en) * 1994-02-04 1995-10-31 Scientific-Atlanta, Inc. Apparatus and method for controlling distribution of electrical energy to a space conditioning load
US5995729A (en) * 1994-07-22 1999-11-30 Hitachi, Ltd. Method and apparatus for aiding configurating management of a computer system
US5978594A (en) * 1994-09-30 1999-11-02 Bmc Software, Inc. System for managing computer resources across a distributed computing environment by first reading discovery information about how to determine system resources presence
US5581478A (en) * 1995-04-13 1996-12-03 Cruse; Michael Facility environmental control system
US5850539A (en) * 1995-05-23 1998-12-15 Compaq Computer Corporation Automated system for facilitating creation of a rack-mountable component personal computer
US5657641A (en) * 1995-09-13 1997-08-19 Kooltronic, Inc. Panel mounted cooling system
US20030204756A1 (en) * 1997-02-12 2003-10-30 Ransom Douglas S. Push communications architecture for intelligent electronic devices
US5987614A (en) * 1997-06-17 1999-11-16 Vadem Distributed power management system and method for computer
US6216956B1 (en) * 1997-10-29 2001-04-17 Tocom, Inc. Environmental condition control and energy management system and method
US6134511A (en) * 1998-04-15 2000-10-17 Subbarao; Krishnappa Method and apparatus for improving building energy simulations
US6246969B1 (en) * 1998-09-08 2001-06-12 International Business Machines Corporation Method and apparatus for testing computer cooling systems
US6714977B1 (en) * 1999-10-27 2004-03-30 Netbotz, Inc. Method and system for monitoring computer networks and equipment
US20010005894A1 (en) * 1999-12-27 2001-06-28 Masahiro Fukui Remote power management system of information processing apparatus or the like
US20020059804A1 (en) * 2000-02-18 2002-05-23 Toc Technology, Llc Computer room air flow
US6745579B2 (en) * 2000-02-18 2004-06-08 Toc Technology, Llc Computer room air flow method and apparatus
US20010042616A1 (en) * 2000-03-21 2001-11-22 Baer Daniel B. Method and apparatus for cooling electronic enclosures
US20020072868A1 (en) * 2000-07-13 2002-06-13 Bartone Erik J. System and method for monitoring and controlling energy usage
US20050108582A1 (en) * 2000-09-27 2005-05-19 Fung Henry T. System, architecture, and method for logical server and other network devices in a dynamically configurable multi-server network environment
US20020062454A1 (en) * 2000-09-27 2002-05-23 Amphus, Inc. Dynamic power and workload management for multi-server system
US20050177755A1 (en) * 2000-09-27 2005-08-11 Amphus, Inc. Multi-server and multi-CPU power management system and method
US6459579B1 (en) * 2001-01-03 2002-10-01 Juniper Networks, Inc. Apparatus and method for directing airflow in three dimensions to cool system components
US6374627B1 (en) * 2001-01-09 2002-04-23 Donald J. Schumacher Data center cooling system
US6967283B2 (en) * 2001-03-20 2005-11-22 American Power Conversion Corporation Adjustable scalable rack power system and method
US6672955B2 (en) * 2001-09-07 2004-01-06 International Business Machines Corporation Air flow management system for an internet data center
US6574104B2 (en) * 2001-10-05 2003-06-03 Hewlett-Packard Development Company L.P. Smart cooling of data centers
US20030084358A1 (en) * 2001-10-31 2003-05-01 Bresniker Kirk M. System and method for intelligent control of power consumption of distributed services during periods of reduced load
US20030084357A1 (en) * 2001-10-31 2003-05-01 Bresniker Kirk M. System and method for providing minimal power-consuming redundant computing hardware for distributed services
US20030084359A1 (en) * 2001-10-31 2003-05-01 Bresniker Kirk M. System and method for intelligent control of power consumption of distributed services during periods when power consumption must be reduced
US20030115000A1 (en) * 2001-12-13 2003-06-19 Bodas Devadatta V. Method to estimate power and cooling requirements of systems
US6804616B2 (en) * 2001-12-13 2004-10-12 Intel Corporation Method to estimate power and cooling requirements of systems
US20030115024A1 (en) * 2001-12-17 2003-06-19 Snevely Robert D. Designing a data center
US7020586B2 (en) * 2001-12-17 2006-03-28 Sun Microsystems, Inc. Designing a data center
US20030120780A1 (en) * 2001-12-21 2003-06-26 Xiaoyun Zhu Network resource assignment system and method
US20070150215A1 (en) * 2002-01-02 2007-06-28 American Power Conversion Corporation Method and apparatus for preventing overloads of power distribution networks
US6721672B2 (en) * 2002-01-02 2004-04-13 American Power Conversion Method and apparatus for preventing overloads of power distribution networks
US20030147216A1 (en) * 2002-02-06 2003-08-07 Patel Chandrakant D. Method, apparatus, and system for cooling electronic components
US7313503B2 (en) * 2002-02-19 2007-12-25 Hewlett-Packard Development Company, L.P. Designing layout for internet datacenter cooling
US20030158718A1 (en) * 2002-02-19 2003-08-21 Nakagawa Osamu Samuel Designing layout for internet datacenter cooling
US6964539B2 (en) * 2002-03-18 2005-11-15 International Business Machines Corporation Method for managing power consumption of multiple computer servers
US6795928B2 (en) * 2002-03-18 2004-09-21 International Business Machines Corporation Method for managing power consumption of multiple computer servers
US20050225936A1 (en) * 2002-03-28 2005-10-13 Tony Day Cooling of a data centre
US20030196126A1 (en) * 2002-04-11 2003-10-16 Fung Henry T. System, method, and architecture for dynamic server power management and dynamic workload management for multi-server environment
US20030193777A1 (en) * 2002-04-16 2003-10-16 Friedrich Richard J. Data center energy management system
US6718277B2 (en) * 2002-04-17 2004-04-06 Hewlett-Packard Development Company, L.P. Atmospheric control within a building
US6886353B2 (en) * 2002-08-02 2005-05-03 Hewlett-Packard Development Company, L.P. Cooling system
US20040089011A1 (en) * 2002-08-02 2004-05-13 Patel Chandrakant D. Cooling system
US20040020224A1 (en) * 2002-08-02 2004-02-05 Bash Cullen E. Cooling system with evaporators distributed in parallel
US20040089009A1 (en) * 2002-10-03 2004-05-13 Bash Cullen E. Cooling of data centers
US20040065097A1 (en) * 2002-10-03 2004-04-08 Bash Cullen E. Cooling of data centers
US20040075984A1 (en) * 2002-10-03 2004-04-22 Bash Cullen E. Cooling of data centers
US20040065104A1 (en) * 2002-10-03 2004-04-08 Bash Cullen E. Cooling of data centers
US20040083012A1 (en) * 2002-10-28 2004-04-29 Miller John P. Method of modeling and sizing a heat exchanger
US6862179B2 (en) * 2002-11-26 2005-03-01 Hewlett-Packard Development Company, L.P. Partition for varying the supply of cooling fluid
US20040120331A1 (en) * 2002-12-20 2004-06-24 Rhine Scott Alan Systems and methods for rapid selection of devices in a tree topology network
US6694759B1 (en) * 2003-01-27 2004-02-24 Hewlett-Packard Development Company, L.P. Pressure control of cooling fluid within a plenum using automatically adjustable vents
US20040163001A1 (en) * 2003-02-14 2004-08-19 Bodas Devadatta V. Enterprise power and thermal management
US20070076373A1 (en) * 2003-03-19 2007-04-05 James Fink Data center cooling system
US20070274035A1 (en) * 2003-03-19 2007-11-29 Fink James R Data center cooling
US6859366B2 (en) * 2003-03-19 2005-02-22 American Power Conversion Data center cooling system
US20040189161A1 (en) * 2003-03-28 2004-09-30 Davis Brooks I. Zero rack unit space utilization
US20040240514A1 (en) * 2003-05-29 2004-12-02 Bash Cullen Edwin Air re-circulation index
US20050023363A1 (en) * 2003-05-29 2005-02-03 Sharma Ratnesh K. CRAC unit control based on re-circulation index
US7051946B2 (en) * 2003-05-29 2006-05-30 Hewlett-Packard Development Company, L.P. Air re-circulation index
US20040262409A1 (en) * 2003-06-30 2004-12-30 Crippen Martin Joseph Thermal analysis in a data processing system
US6819563B1 (en) * 2003-07-02 2004-11-16 International Business Machines Corporation Method and system for cooling electronics racks using pre-cooled air
US7085133B2 (en) * 2003-07-09 2006-08-01 International Business Machines Corporation Cooling using complimentary tapered plenums
US20050016189A1 (en) * 2003-07-25 2005-01-27 Honeywell International Inc. Method and apparatus for controlling humidity with an air conditioner
US20050055590A1 (en) * 2003-09-04 2005-03-10 Farkas Keith Istvan Application management based on power consumption
US20050228618A1 (en) * 2004-04-09 2005-10-13 Patel Chandrakant D Workload placement among data centers based on thermal efficiency
US7197433B2 (en) * 2004-04-09 2007-03-27 Hewlett-Packard Development Company, L.P. Workload placement among data centers based on thermal efficiency
US20050251802A1 (en) * 2004-05-08 2005-11-10 Bozek James J Dynamic migration of virtual machine computer programs upon satisfaction of conditions
US20050267639A1 (en) * 2004-05-28 2005-12-01 Sharma Ratnesh K Data center evaluation using an air re-circulation index
US20050278075A1 (en) * 2004-05-28 2005-12-15 Neil Rasmussen Methods and apparatus for providing and distributing standby power
US7031870B2 (en) * 2004-05-28 2006-04-18 Hewlett-Packard Development Company, L.P. Data center evaluation using an air re-circulation index
US20070055409A1 (en) * 2004-05-28 2007-03-08 American Power Conversion Corporation Methods and apparatus for providing and distributing standby power
US20060080001A1 (en) * 2004-10-08 2006-04-13 Bash Cullen E Correlation of vent tile settings and rack temperatures
US7251547B2 (en) * 2004-10-08 2007-07-31 Hewlett-Packard Development Company, L.P. Correlation of vent tile settings and rack temperatures
US20060121421A1 (en) * 2004-10-15 2006-06-08 Spitaels James S IT equipment simulation
US7155318B2 (en) * 2004-11-05 2006-12-26 Hewlett-Packard Development Company, Lp. Air conditioning unit control to reduce moisture varying operations
US20060112286A1 (en) * 2004-11-23 2006-05-25 Whalley Ian N Method for dynamically reprovisioning applications and other server resources in a computer center in response to power and heat dissipation requirements
US20060115586A1 (en) * 2004-11-30 2006-06-01 Baozhong Xing Process and apparatus for coating a controlled release product in a rotating drum
US20060139877A1 (en) * 2004-12-29 2006-06-29 Mark Germagian Rack height cooling
US20060168975A1 (en) * 2005-01-28 2006-08-03 Hewlett-Packard Development Company, L.P. Thermal and power management apparatus
US20060214014A1 (en) * 2005-03-25 2006-09-28 Bash Cullen E Temperature control using a sensor network
US20070078635A1 (en) * 2005-05-02 2007-04-05 American Power Conversion Corporation Methods and systems for managing facility power and cooling
US20070038414A1 (en) * 2005-05-02 2007-02-15 American Power Conversion Corporation Methods and systems for managing facility power and cooling
US20070174024A1 (en) * 2005-05-02 2007-07-26 American Power Conversion Corporation Methods and systems for managing facility power and cooling
US20060271700A1 (en) * 2005-05-24 2006-11-30 Fujitsu Limited Record medium with a load distribution program recorded thereon, load distribution method, and load distribution apparatus
US20070150584A1 (en) * 2005-12-28 2007-06-28 Deepa Srinivasan Apparatus, system, and method for determining server utilization in hosted computing infrastructure

Cited By (222)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8639482B2 (en) 2005-05-02 2014-01-28 Schneider Electric It Corporation Methods and systems for managing facility power and cooling
US8315841B2 (en) 2005-05-02 2012-11-20 American Power Conversion Corporation Methods and systems for managing facility power and cooling
US20090006222A1 (en) * 2006-01-30 2009-01-01 L'air Liquide Societe Anonyme Pour L'etude Et L'exploitation Des Procedes Georges Claude System for the Operation and Management of a Fleet of Refrigerated Autonomous Containers
US9062908B2 (en) * 2006-01-30 2015-06-23 L'Air Liquide Société Anonyme pour l'Ètude Et l'Exploitation des Procedes Georges Claude System for the operation and management of a fleet of refrigerated autonomous containers
US8322155B2 (en) 2006-08-15 2012-12-04 American Power Conversion Corporation Method and apparatus for cooling
US9568206B2 (en) 2006-08-15 2017-02-14 Schneider Electric It Corporation Method and apparatus for cooling
US9115916B2 (en) 2006-08-15 2015-08-25 Schneider Electric It Corporation Method of operating a cooling system having one or more cooling units
US8327656B2 (en) 2006-08-15 2012-12-11 American Power Conversion Corporation Method and apparatus for cooling
US8424336B2 (en) 2006-12-18 2013-04-23 Schneider Electric It Corporation Modular ice storage for uninterruptible chilled water
US9080802B2 (en) 2006-12-18 2015-07-14 Schneider Electric It Corporation Modular ice storage for uninterruptible chilled water
US8425287B2 (en) 2007-01-23 2013-04-23 Schneider Electric It Corporation In-row air containment and cooling system and method
US11503744B2 (en) 2007-05-15 2022-11-15 Schneider Electric It Corporation Methods and systems for managing facility power and cooling
US11076507B2 (en) 2007-05-15 2021-07-27 Schneider Electric It Corporation Methods and systems for managing facility power and cooling
US7752088B1 (en) * 2007-05-29 2010-07-06 Joseph Coschera On site physical inventory method
US8601287B1 (en) 2007-06-08 2013-12-03 Exaflop Llc Computer and data center load determination
US10339227B1 (en) 2007-06-08 2019-07-02 Google Llc Data center design
US8949646B1 (en) * 2007-06-08 2015-02-03 Google Inc. Data center load monitoring for utilizing an access power amount based on a projected peak power usage and a monitored power usage
US9946815B1 (en) 2007-06-08 2018-04-17 Google Llc Computer and data center load determination
US8700929B1 (en) 2007-06-08 2014-04-15 Exaflop Llc Load control in a data center
US8595515B1 (en) 2007-06-08 2013-11-26 Google Inc. Powering a data center
US11017130B1 (en) 2007-06-08 2021-05-25 Google Llc Data center design
US8621248B1 (en) 2007-06-08 2013-12-31 Exaflop Llc Load control in a data center
US10558768B1 (en) 2007-06-08 2020-02-11 Google Llc Computer and data center load determination
US8645722B1 (en) 2007-06-08 2014-02-04 Exaflop Llc Computer and data center load determination
US20090132097A1 (en) * 2007-11-20 2009-05-21 Ratnesh Kumar Sharma Virtual cooling infrastructure
US8437881B2 (en) * 2008-02-15 2013-05-07 The Pnc Financial Services Group, Inc. Systems and methods for computer equipment management
US20090210097A1 (en) * 2008-02-15 2009-08-20 Sawczak Stephen D Systems and methods for computer equipment management
US20090210099A1 (en) * 2008-02-15 2009-08-20 Sawczak Stephen D Systems and methods for computer equipment management
US8175753B2 (en) 2008-02-15 2012-05-08 The Pnc Financial Services Group, Inc. Systems and methods for computer equipment management
US8201028B2 (en) 2008-02-15 2012-06-12 The Pnc Financial Services Group, Inc. Systems and methods for computer equipment management
US9495774B2 (en) * 2008-03-11 2016-11-15 Enphase Energy, Inc. Method and apparatus for electrical power visualization
US20150154770A1 (en) * 2008-03-11 2015-06-04 Enphase Energy, Inc. Method and apparatus for electrical power visualization
US8223015B2 (en) * 2008-03-28 2012-07-17 Fujitsu Limited Electronic apparatus system having a plurality of rack-mounted electronic apparatuses, and method for identifying electronic apparatus in electronic apparatus system
US20090243846A1 (en) * 2008-03-28 2009-10-01 Fujitsu Limited Electronic apparatus system having a plurality of rack-mounted electronic apparatuses, and method for identifying electronic apparatus in electronic apparatus system
US20150378414A1 (en) * 2008-04-09 2015-12-31 Hitachi, Ltd. Operations management methods and devices thereof in information-processing systems
US9389664B2 (en) * 2008-04-09 2016-07-12 Hitachi, Ltd. Operations management methods and devices thereof in systems
US20090327012A1 (en) * 2008-06-30 2009-12-31 Ratnesh Kumar Sharma Cooling resource capacity allocation using lagrange multipliers
US9009061B2 (en) * 2008-06-30 2015-04-14 Hewlett-Packard Development Company, L. P. Cooling resource capacity allocation based on optimization of cost function with lagrange multipliers
US9939796B2 (en) 2008-09-29 2018-04-10 International Business Machines Corporation System and method to dynamically change data center partitions
US20160117420A1 (en) * 2008-09-29 2016-04-28 International Business Machines Corporation System and method for dynamically modeling data center partitions
US10884387B2 (en) 2008-09-29 2021-01-05 International Business Machines Corporation System and method to dynamically change data center partitions
US7925746B1 (en) * 2008-10-06 2011-04-12 United Services Automobile Association (Usaa) Systems and methods for a self-contained, cooled server rack
US8626905B1 (en) * 2008-10-06 2014-01-07 United Services Automobile Association (Usaa) Monitoring a self-contained server rack system
US20100106464A1 (en) * 2008-10-27 2010-04-29 Christopher Hlasny Method for designing raised floor and dropped ceiling in computing facilities
US8473265B2 (en) 2008-10-27 2013-06-25 Schneider Electric It Corporation Method for designing raised floor and dropped ceiling in computing facilities
US20100115305A1 (en) * 2008-11-03 2010-05-06 Hitachi, Ltd. Methods and Apparatus to Provision Power-Saving Storage System
US8155766B2 (en) * 2008-11-03 2012-04-10 Hitachi, Ltd. Methods and apparatus to provision power-saving storage system
US8209056B2 (en) 2008-11-25 2012-06-26 American Power Conversion Corporation System and method for assessing and managing data center airflow and energy usage
US20100131109A1 (en) * 2008-11-25 2010-05-27 American Power Conversion Corporation System and method for assessing and managing data center airflow and energy usage
US9494985B2 (en) 2008-11-25 2016-11-15 Schneider Electric It Corporation System and method for assessing and managing data center airflow and energy usage
US8224488B2 (en) * 2009-01-15 2012-07-17 Dell Products L.P. System and method for temperature management of a data center
US20100179695A1 (en) * 2009-01-15 2010-07-15 Dell Products L.P. System and Method for Temperature Management of a Data Center
US9778718B2 (en) 2009-02-13 2017-10-03 Schneider Electric It Corporation Power supply and data center control
US9519517B2 (en) 2009-02-13 2016-12-13 Schneider Electtic It Corporation Data center control
US8560677B2 (en) 2009-02-13 2013-10-15 Schneider Electric It Corporation Data center control
US9904331B2 (en) 2009-04-01 2018-02-27 Schneider Electric It Corporation Method for computing cooling redundancy at the rack level
US20100256959A1 (en) * 2009-04-01 2010-10-07 American Power Conversion Corporation Method for computing cooling redundancy at the rack level
US9754055B1 (en) * 2009-05-06 2017-09-05 Amdocs Software Systems Limited System, method, and computer program product for managing an area for positioning resources, based on parameters of the resources
US10614194B2 (en) 2009-05-08 2020-04-07 Schneider Electric It Corporation System and method for arranging equipment in a data center
US20100286956A1 (en) * 2009-05-08 2010-11-11 American Power Conversion Corporation System and method for predicting cooling performance of arrangements of equipment in a data center
US8554515B2 (en) 2009-05-08 2013-10-08 Schneider Electric It Corporation System and method for predicting cooling performance of arrangements of equipment in a data center
US20100287018A1 (en) * 2009-05-08 2010-11-11 American Power Conversion Corporation System and method for arranging equipment in a data center
US20100286955A1 (en) * 2009-05-08 2010-11-11 American Power Conversion Corporation System and method for predicting maximum cooler and rack capacities in a data center
US8355890B2 (en) 2009-05-08 2013-01-15 American Power Conversion Corporation System and method for predicting maximum cooler and rack capacities in a data center
US9996659B2 (en) 2009-05-08 2018-06-12 Schneider Electric It Corporation System and method for arranging equipment in a data center
US8249825B2 (en) 2009-05-08 2012-08-21 American Power Conversion Corporation System and method for predicting cooling performance of arrangements of equipment in a data center
US8219362B2 (en) 2009-05-08 2012-07-10 American Power Conversion Corporation System and method for arranging equipment in a data center
US8768519B2 (en) * 2009-05-21 2014-07-01 Fujitsu Limited Apparatus and method for controlling grille aperture ratios of a plurality of air transfer grilles
US20100298990A1 (en) * 2009-05-21 2010-11-25 Fujitsu Limited Apparatus and method for controlling an open amount of a plurality of air transfer grilles
US9287710B2 (en) 2009-06-15 2016-03-15 Google Inc. Supplying grid ancillary services using controllable loads
US20110016342A1 (en) * 2009-07-20 2011-01-20 Viridity Software, Inc. Techniques for power analysis
US9116897B2 (en) 2009-07-20 2015-08-25 Schneider Electric It Corporation Techniques for power analysis
US9753465B1 (en) 2009-07-21 2017-09-05 The Research Foundation For The State University Of New York Energy aware processing load distribution system and method
US11194353B1 (en) 2009-07-21 2021-12-07 The Research Foundation for the State University Energy aware processing load distribution system and method
US11886914B1 (en) * 2009-07-21 2024-01-30 The Research Foundation For The State University Of New York Energy efficient scheduling for computing systems and method therefor
US20110040529A1 (en) * 2009-08-12 2011-02-17 International Business Machines Corporation Methods and Techniques for Creating and Visualizing Thermal Zones
US8244502B2 (en) * 2009-08-12 2012-08-14 International Business Machines Corporation Knowledge-based models for data centers
US8229713B2 (en) * 2009-08-12 2012-07-24 International Business Machines Corporation Methods and techniques for creating and visualizing thermal zones
US20110040532A1 (en) * 2009-08-12 2011-02-17 International Business Machines Corporation Knowledge-Based Models for Data Centers
US20120284216A1 (en) * 2009-08-12 2012-11-08 International Business Machines Corporation Knowledge-Based Models for Data Centers
EP2367412A1 (en) * 2009-09-04 2011-09-21 Fujitsu Limited Thermal-fluid-simulating analyzing apparatus, method and program
CN102012951A (en) * 2009-09-04 2011-04-13 富士通株式会社 Thermal-fluid-simulation analyzing apparatus
US8744818B2 (en) 2009-09-04 2014-06-03 Fujitsu Limited Thermal-fluid-simulation analyzing apparatus
US20110060571A1 (en) * 2009-09-04 2011-03-10 Fujitsu Limited Thermal-fluid-simulation analyzing apparatus
US9703665B1 (en) * 2010-02-19 2017-07-11 Acuity Holdings, Inc. Data center design process and system
US20110208622A1 (en) * 2010-02-25 2011-08-25 International Business Machines Corporation Data center power cost accounting system
US8374928B2 (en) * 2010-02-25 2013-02-12 International Business Machines Corporation Data center power cost accounting system
US20110213735A1 (en) * 2010-02-26 2011-09-01 International Business Machines Corporation Selecting An Installation Rack For A Device In A Data Center
CN102169527A (en) * 2010-02-26 2011-08-31 国际商业机器公司 Method and system for determining mounting machine frame for equipment in data center
US20110244779A1 (en) * 2010-04-02 2011-10-06 Fujitsu Limited Air conditioning system and air conditioning control method
US20110258302A1 (en) * 2010-04-20 2011-10-20 No Limits Software, LLC System And Method For Remotely Determining Identification And Physical Location Of Equipment In A Rack
US8628153B2 (en) 2010-05-13 2014-01-14 Pandult Corp. Aisle containment system
US8628154B2 (en) 2010-05-13 2014-01-14 Panduit Corp. Aisle containment system
US8628158B2 (en) 2010-05-13 2014-01-14 Panduit Corp. Aisle containment system
US8972217B2 (en) 2010-06-08 2015-03-03 Schneider Electric It Corporation System and method for predicting temperature values in a data center
US20110303406A1 (en) * 2010-06-11 2011-12-15 Fujitsu Limited Air-conditioning system and control device thereof
US20140005982A1 (en) * 2010-06-22 2014-01-02 International Business Machines Corporation Data center physical infrastructure threshold analysis
US9454449B2 (en) * 2010-06-22 2016-09-27 Globalfoundries Inc. Data center physical infrastructure threshold analysis
US8879247B2 (en) 2010-07-21 2014-11-04 International Business Machines Corporation Computer chassis cooling sidecar
US8947880B2 (en) 2010-08-06 2015-02-03 Lenovo Enterprise Solutions (Singapore) Ptd. Ltd. Hot or cold aisle computer chassis
US20120040601A1 (en) * 2010-08-10 2012-02-16 Yamatake Corporation Air conditioning controlling device and method
US8509959B2 (en) 2010-08-12 2013-08-13 Schneider Electric It Corporation System and method for predicting transient cooling performance for a data center
US9418179B2 (en) 2010-08-12 2016-08-16 Schneider Electric It Corporation System and method for predicting transient cooling performance for data center
US8996180B2 (en) 2010-09-17 2015-03-31 Schneider Electric It Corporation System and method for predicting perforated tile airflow in a data center
US8812275B2 (en) * 2010-09-18 2014-08-19 International Business Machines Corporation Modeling movement of air under a floor of a data center
US20120072195A1 (en) * 2010-09-18 2012-03-22 International Business Machines Corporation Modeling Movement Of Air Under A Floor Of A Data Center
US9658662B2 (en) 2010-10-12 2017-05-23 Hewlett Packard Enterprise Development Lp Resource management for data centers
US20120123898A1 (en) * 2010-11-17 2012-05-17 International Business Machines Corporation Generating configuration options for a computing system
WO2012082985A3 (en) * 2010-12-16 2013-02-28 American Power Conversion Corporation System and methods for rack cooling analysis
US8825451B2 (en) 2010-12-16 2014-09-02 Schneider Electric It Corporation System and methods for rack cooling analysis
US8949081B2 (en) 2010-12-21 2015-02-03 Schneider Electric It Corporation Method for computing cooling redundancy at the rack level
US20120170205A1 (en) * 2010-12-30 2012-07-05 American Power Conversion Corporation System and method for sequential placement of cooling resources within data center layouts
US8688413B2 (en) * 2010-12-30 2014-04-01 Christopher M. Healey System and method for sequential placement of cooling resources within data center layouts
US8744631B2 (en) 2011-01-28 2014-06-03 Hewlett-Packard Development Company, L.P. Manipulating environmental conditions in an infrastructure
US20120209573A1 (en) * 2011-02-11 2012-08-16 International Business Machines Corporation Data center design tool
US8630822B2 (en) * 2011-02-11 2014-01-14 International Business Machines Corporation Data center design tool
US20120215359A1 (en) * 2011-02-21 2012-08-23 Amir Meir Michael Adaptive fan control based on server configuration
US8597054B2 (en) 2011-03-07 2013-12-03 Schneider Electric It Corporation CAN bus automatic line termination
US9223905B2 (en) 2011-03-25 2015-12-29 Schneider Electric It Corporation Systems and methods for predicting fluid dynamics in a data center
US8762522B2 (en) * 2011-04-19 2014-06-24 Cisco Technology Coordinating data center compute and thermal load based on environmental data forecasts
US20120271935A1 (en) * 2011-04-19 2012-10-25 Moon Billy G Coordinating data center compute and thermal load based on environmental data forecasts
US20120303339A1 (en) * 2011-05-27 2012-11-29 International Business Machines Corporation Computational fluid dynamics modeling of a bounded domain
US8756040B2 (en) * 2011-05-27 2014-06-17 International Business Machines Corporation Computational fluid dynamics modeling of a bounded domain
US8744812B2 (en) * 2011-05-27 2014-06-03 International Business Machines Corporation Computational fluid dynamics modeling of a bounded domain
US20120303344A1 (en) * 2011-05-27 2012-11-29 International Business Machines Corporation Computational fluid dynamics modeling of a bounded domain
US8914495B2 (en) 2011-06-07 2014-12-16 International Business Machines Corporation Automatically detecting and locating equipment within an equipment rack
US8725307B2 (en) 2011-06-28 2014-05-13 Schneider Electric It Corporation System and method for measurement aided prediction of temperature and airflow values in a data center
US9952103B2 (en) 2011-12-22 2018-04-24 Schneider Electric It Corporation Analysis of effect of transient events on temperature in a data center
US9830410B2 (en) 2011-12-22 2017-11-28 Schneider Electric It Corporation System and method for prediction of temperature values in an electronics system
WO2013095623A3 (en) * 2011-12-23 2014-04-10 Schneider Electric It Corporation Dropped ceiling system cooling prediction
US9727064B2 (en) 2011-12-23 2017-08-08 Schneider Electric It Corporation Dropped ceiling system cooling prediction
US10274978B2 (en) 2011-12-23 2019-04-30 Schneider Electric It Corporation Dropped ceiling system cooling prediction
US9009500B1 (en) 2012-01-18 2015-04-14 Google Inc. Method of correlating power in a data center by fitting a function to a plurality of pairs of actual power draw values and estimated power draw values determined from monitored CPU utilization of a statistical sample of computers in the data center
US9383791B1 (en) 2012-01-18 2016-07-05 Google Inc. Accurate power allotment
US20130204984A1 (en) * 2012-02-08 2013-08-08 Oracle International Corporation Management Record Specification for Management of Field Replaceable Units Installed Within Computing Cabinets
WO2013171757A2 (en) 2012-05-18 2013-11-21 Tata Consultancy Services Limited Method and system for determining and implementing a viable containment design of a data center
EP2850566A4 (en) * 2012-05-18 2016-01-13 Tata Consultancy Services Ltd Method and system for determining and implementing a viable containment design of a data center
US10317980B2 (en) * 2012-07-03 2019-06-11 Eaton Intelligent Power Limited Grouping power sources and power loads to groups that share similar properties
US20150362981A1 (en) * 2012-07-03 2015-12-17 Eaton Industries (France) Sas Grouping power sources and power loads to groups that share similar properties
US20150192316A1 (en) * 2012-07-06 2015-07-09 Rongliang Zhou Management of airflow provisioning to meet a cooling influence redundancy level
US10228668B2 (en) * 2012-07-06 2019-03-12 Hewlett Packard Enterprise Development Lp Management of airflow provisioning to meet a cooling influence redundancy level
US20150200916A1 (en) * 2012-07-24 2015-07-16 Zuken Inc. Information Administration System
US9769125B2 (en) * 2012-07-24 2017-09-19 Zuken Inc. Information administration system
US10082857B1 (en) * 2012-08-07 2018-09-25 Amazon Technologies, Inc. Cooling electrical systems based on power measurements
WO2014041554A1 (en) 2012-09-12 2014-03-20 Tata Consultancy Services Limited A method for efficient designing and operating cooling infrastructure in a data center
AU2016259358B2 (en) * 2012-09-12 2018-03-29 Tata Consultancy Services Limited A method for efficient designing and operating cooling infrastructure in a data center
EP2895972A4 (en) * 2012-09-12 2016-05-04 Tata Consultancy Services Ltd Method for efficient designing and operating cooling infrastructure in a data center
US9839165B2 (en) * 2012-11-02 2017-12-05 Fujitsu Limited Modular data center and controlling method of modular data center
US20150250077A1 (en) * 2012-11-02 2015-09-03 Fujitsu Limited Modular data center and controlling method of modular data center
US10210288B2 (en) 2012-12-27 2019-02-19 Schneider Electric It Corporation Systems and methods of visualizing airflow
US10310568B2 (en) 2013-02-28 2019-06-04 Oracle International Corporation Method for interconnecting field replaceable unit to power source of communication network
GB2513141A (en) * 2013-04-17 2014-10-22 Ibm Data processing system with real-time data center air flow simulator
US9600604B2 (en) * 2013-07-18 2017-03-21 Netapp, Inc. System and method for planning an upgrade of a modular computing system
US20150026621A1 (en) * 2013-07-18 2015-01-22 Netapp, Inc. System and Method for Planning an Implementation of a Computing System
US20150032283A1 (en) * 2013-07-29 2015-01-29 International Business Machines Corporation Data center cooling
US9851726B2 (en) 2013-09-04 2017-12-26 Panduit Corp. Thermal capacity management
US20160249487A1 (en) * 2013-10-04 2016-08-25 Tata Consultancy Services Limited System and method for optimizing cooling efficiency of a data center
US10721845B2 (en) * 2013-10-04 2020-07-21 Tata Consultancy Services Limited System and method for optimizing cooling efficiency of a data center
WO2015095584A1 (en) * 2013-12-19 2015-06-25 Panduit Corp. Methods and systems for identifying stranded capacity
US20150256386A1 (en) * 2014-03-06 2015-09-10 Dell Products, Lp System and Method for Providing a Server Rack Management Controller
US20150253028A1 (en) * 2014-03-06 2015-09-10 Dell Products, Lp System and Method for Data Reporting in a Tile Management Controller
US9958178B2 (en) * 2014-03-06 2018-05-01 Dell Products, Lp System and method for providing a server rack management controller
US9964935B2 (en) * 2014-03-06 2018-05-08 Dell Products, Lp System and method for data reporting in a tile management controller
US9448841B2 (en) * 2014-03-17 2016-09-20 Ca, Inc. Decommissioning physical host machines after relocating unproductive virtual machines therefrom
WO2015168565A1 (en) * 2014-05-01 2015-11-05 Amazon Technologies, Inc. Flexible data center infrastructure distribution
AU2015252871B2 (en) * 2014-05-01 2017-12-14 Amazon Technologies, Inc. Flexible data center infrastructure distribution
RU2664015C2 (en) * 2014-05-01 2018-08-14 Амазон Текнолоджиз, Инк. Flexible distribution of data processing center infrastructure
US10502918B1 (en) 2014-05-01 2019-12-10 Amazon Technologies, Inc. Flexible data center infrastructure distribution
CN106464523A (en) * 2014-05-01 2017-02-22 亚马逊科技公司 Flexible data center infrastructure distribution
US11137563B2 (en) 2014-05-01 2021-10-05 Amazon Technologies, Inc. Flexible data center infrastructure distribution
US10061097B2 (en) 2014-05-01 2018-08-28 Amazon Technologies, Inc. Flexible data center infrastructure distribution
US10028405B1 (en) 2014-05-01 2018-07-17 Amazon Technologies, Inc. Flexible data center rack allocation
US9779275B2 (en) * 2014-07-03 2017-10-03 Mitsubishi Electric Corporation Control system, terminal, information setting method, and program
US10447785B2 (en) * 2014-11-17 2019-10-15 Lg Electronics Inc. Digital device and method for controlling same
US10102313B2 (en) * 2014-12-30 2018-10-16 Schneider Electric It Corporation Raised floor plenum tool
US20160188764A1 (en) * 2014-12-30 2016-06-30 Schneider Electric It Corporation Raised floor plenum tool
US10001761B2 (en) 2014-12-30 2018-06-19 Schneider Electric It Corporation Power consumption model for cooling equipment
US11030332B1 (en) * 2015-04-13 2021-06-08 Wells Fargo Bank, N.A. Database controlled web service type architecture
US10545515B2 (en) * 2015-04-27 2020-01-28 Hewlett Packard Enterprise Development Lp Virtualized fan speed measurement
US20170242446A1 (en) * 2015-04-27 2017-08-24 Hewlett Packard Enterprise Development Lp Virtualized fan speed measurement
US11291142B2 (en) 2015-05-07 2022-03-29 DHK Storage, Inc. Computer server heat regulation utilizing integrated precision air flow
US10426061B2 (en) 2015-05-07 2019-09-24 Dhk Storage, Llc Computer server heat regulation utilizing integrated precision air flow
US9832912B2 (en) 2015-05-07 2017-11-28 Dhk Storage, Llc Computer server heat regulation utilizing integrated precision air flow
US11178794B2 (en) 2015-05-07 2021-11-16 Dhk Storage, Llc Computer server heat regulation utilizing integrated precision air flow
US9795063B1 (en) 2015-05-07 2017-10-17 Dhk Storage, Llc Computer server heat regulation utilizing integrated precision air flow
US11602084B2 (en) 2015-05-07 2023-03-07 DHK Storage, Inc. Computer server heat regulation utilizing integrated precision air flow
US11432435B2 (en) 2015-05-07 2022-08-30 DHK Storage, Inc. Computer server heat regulation utilizing integrated precision air flow
US11602085B2 (en) 2015-05-07 2023-03-07 DHK Storage, Inc. Computer server heat regulation utilizing integrated precision air flow
US11291141B2 (en) 2015-05-07 2022-03-29 DHK Storage, Inc. Computer server heat regulation utilizing integrated precision air flow
US11606884B2 (en) 2015-05-07 2023-03-14 DHK Storage, Inc. Computer server heat regulation utilizing integrated precision air flow
US11602083B2 (en) 2015-05-07 2023-03-07 DHK Storage, Inc. Computer server heat regulation utilizing integrated precision air flow
US10180261B1 (en) * 2015-12-28 2019-01-15 Amazon Technologies, Inc. Model based cooling control system
US10812339B2 (en) 2016-01-11 2020-10-20 Equinix, Inc. Determining power path for data center customers
US9948522B2 (en) * 2016-01-11 2018-04-17 Equinix, Inc. Associating infrastructure assets in a data center
US11627051B2 (en) 2016-01-11 2023-04-11 Equinix, Inc. Determining asset associations for data center customers
US10574529B2 (en) 2016-01-11 2020-02-25 Equinix, Inc. Defining conditional triggers for issuing data center asset information
US10405461B2 (en) * 2016-04-19 2019-09-03 Dell Products L.P. Systems and methods for fan performance-based scaling of thermal control parameters
US20170303438A1 (en) * 2016-04-19 2017-10-19 Dell Products L.P. Systems and methods for fan performance-based scaling of thermal control parameters
US10617038B2 (en) 2016-07-08 2020-04-07 Schneider Electric It Corporation Zero-equation turbulence models for large electrical and electronics enclosure applications
US11422910B2 (en) * 2016-07-21 2022-08-23 Jpmorgan Chase Bank, N.A. Method and system for implementing a data center operating system
US11076509B2 (en) 2017-01-24 2021-07-27 The Research Foundation for the State University Control systems and prediction methods for it cooling performance in containment
US10904173B2 (en) 2017-06-09 2021-01-26 Equinix, Inc. Near real-time messaging service for data center infrastructure monitoring data
US20180359201A1 (en) 2017-06-09 2018-12-13 Equinix, Inc. Near real-time messaging service for data center infrastructure monitoring data
US20180375741A1 (en) * 2017-06-26 2018-12-27 Seagate Technology Llc Datacenter service flow optimization
US10880184B2 (en) * 2017-06-26 2020-12-29 Seagate Technology Llc Datacenter service flow optimization
US20190057180A1 (en) * 2017-08-18 2019-02-21 International Business Machines Corporation System and method for design optimization using augmented reality
US20190057181A1 (en) * 2017-08-18 2019-02-21 International Business Machines Corporation System and method for design optimization using augmented reality
US10660241B2 (en) * 2017-09-06 2020-05-19 Vertiv Corporation Cooling unit energy optimization via smart supply air temperature setpoint control
WO2019051077A1 (en) * 2017-09-06 2019-03-14 Vertiv Corporation Cooling unit energy optimization via smart supply air temperature setpoint control
US10819556B1 (en) 2017-10-16 2020-10-27 Equinix, Inc. Data center agent for data center infrastructure monitoring data access and translation
US11049052B2 (en) * 2018-05-02 2021-06-29 International Business Machines Corporation Automated managing of a data center installation
US11388835B1 (en) * 2018-06-27 2022-07-12 Amazon Technologies, Inc. Placement of custom servers
US20200106677A1 (en) * 2018-09-28 2020-04-02 Hewlett Packard Enterprise Development Lp Data center forecasting based on operation data
US11382232B1 (en) 2019-03-28 2022-07-05 Amazon Technologies, Inc. Self-standing modular data center infrastructure system
US10999954B2 (en) 2019-04-23 2021-05-04 Vertiv Corporation Modular roof mounted cooling system and method for data center
US11032948B1 (en) 2019-06-17 2021-06-08 Amazon Technologies, Inc. Pre-fabricated mechanical and electrical distribution infrastructure system
US20210014998A1 (en) * 2019-07-12 2021-01-14 Hewlett Packard Enterprise Development Lp Recommending it equipment placement based on inferred hardware capabilities
US11696418B2 (en) * 2019-07-12 2023-07-04 Hewlett Packard Enterprise Development Lp Recommending IT equipment placement based on inferred hardware capabilities
US20220183191A1 (en) * 2020-12-04 2022-06-09 Schneider Electric It Corporation It-room-cooling-performance assessment
US11599532B1 (en) * 2021-08-11 2023-03-07 Amdocs Development Limited System, method, and computer program for preventing user mistakes when making database changes
US20230083377A1 (en) * 2021-09-16 2023-03-16 Dell Products L.P. Software-defined infrastructure for identifying and remediating an airflow deficiency scenario on a rack device
US11765870B2 (en) * 2021-09-16 2023-09-19 Dell Products L.P. Software-defined infrastructure for identifying and remediating an airflow deficiency scenario on a rack device

Also Published As

Publication number Publication date
CN101755495B (en) 2013-10-16
US11076507B2 (en) 2021-07-27
JP5559040B2 (en) 2014-07-23
JP2010527491A (en) 2010-08-12
AU2008255030A1 (en) 2008-11-27
WO2008144375A3 (en) 2009-04-09
US20210329812A1 (en) 2021-10-21
EP2147585A2 (en) 2010-01-27
AU2008255030B2 (en) 2014-02-20
US11503744B2 (en) 2022-11-15
EP2147585B1 (en) 2016-11-02
CA2686564A1 (en) 2008-11-27
BRPI0812116B1 (en) 2018-12-18
DK2147585T3 (en) 2017-01-16
BRPI0812116A2 (en) 2014-11-25
CN101755495A (en) 2010-06-23
US20190008072A1 (en) 2019-01-03
ES2612328T3 (en) 2017-05-16
WO2008144375A2 (en) 2008-11-27
CA2686564C (en) 2018-04-17

Similar Documents

Publication Publication Date Title
US11503744B2 (en) Methods and systems for managing facility power and cooling
US8639482B2 (en) Methods and systems for managing facility power and cooling
US8315841B2 (en) Methods and systems for managing facility power and cooling
EP1877950B1 (en) Methods and systems for managing facility power and cooling
US8053926B2 (en) Methods and systems for managing facility power and cooling
US7596476B2 (en) Methods and systems for managing facility power and cooling
AU2008207851B2 (en) System and method for evaluating equipment rack cooling performance
EP2656165B1 (en) Method for computing cooling redundancy at the rack level
Somani Advanced thermal management strategies for energy-efficient data centers

Legal Events

Date Code Title Description
AS Assignment

Owner name: AMERICAN POWER CONVERSION CORPORATION, RHODE ISLAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:MORGAN, JANE E.;RASMUSSEN, NEIL;VANGILDER, JAMES W.;AND OTHERS;REEL/FRAME:021758/0634;SIGNING DATES FROM 20080916 TO 20081028

AS Assignment

Owner name: SCHNEIDER ELECTRIC IT CORPORATION, RHODE ISLAND

Free format text: CHANGE OF NAME;ASSIGNOR:AMERICAN POWER CONVERSION CORPORATION;REEL/FRAME:035745/0610

Effective date: 20121214

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION