I have written my own blog post http://www.greenm3.com/2010/04/google-microsoft-amazon-nokia-digital-realty-trust-dupont-fabros-vs-ashrae-standard-901-requirement-for-economizers-li.html referring to the above. I am encouraging readers to comment on this blog to allow Google to collect public opinion. I agree with the industry position Google has presented. -Dave Ohara
I strongly support Google et al and their opposition to the proposed changes to ASHRAE 901. Every data center faces unique cooling challenges that vary based on region, rack density, HVAC architecture, and more. A focus on metrics and benchmarking, as Google notes, is the most holistic means of ensuring efficiency in the data center, not a "one size fits all" methodology. - Todd Boucher, Leading Edge Design Group
Please see the following response from ASHRAEThe proposal to address data centers was made because a significant amount of energy is required to cool and ventilate computer rooms. Through this addendum, ASHRAE is proposing cost effective measures in the prescriptive path of Standard 90.1 to save energy in data centers. The addendum includes eight exceptions to requirements for the use of economizers in data centers. The addendum does not change the portion of the standard that already allows, through the Energy Cost Budget method (an alternate method of compliance), for data centers to be designed without economizers if other energy saving methodologies, including power usage effectiveness (PUE), are employed. ASHRAE is committed to excellence in the consensus standard development process and encourages anyone with comments regarding the proposed addendum regarding data centers (addendum bu) to participate in the public review process. The draft addendum and instructions for submitting comments may be found at http://www.ashrae.org/technology/page/331. Comments are due to ASHRAE by April 19, 2010. Jodi ScottASHRAE communications manager
As a matter of principle, I support a industry-wide, consensus based position as Google has presented here. Yet, it should be noted that ASHRAE does recognize and allow for various exceptions to the measuring methods prescribed in the proposed amendment to 90.1 – in so far as stipulated in the proposed language… The question would be which ought to be accepted as the “norm” and which is the “exception”. Wilhelm Wang, TransReg LLC, Westwood, NJ
PUE and DCiE are dimensionless ratios (kW/kW). They show data that may be valuable to the owner/operator, in terms of IT equipment power and total facility power, but they do not measure efficiency. For example, if there are two exact data centers side by side in one city, with the same square footage (say 100,000 sf) and the same number of servers and same HVAC equipment, if they both have the same PUE of 2.5, but for one facility it is 1000/400 kW and for the 2nd facility it is 2000/800 kW, how can they be considered to have the same "efficiency"? Or, if one data center has a PUE of 2.5 (1000/400 kW) and another one has a PUE of 2.0 (1000/500 kW), which one is more "efficient"? The only difference is that one facility uses less electricity for non-IT functions. They have the same electric bill. More useful metrics would be kWh/square foot/year (or month or day), peak Watts / square foot, overall Watts or kWh used per teraflop of server function, Btu/square foot/year (or month or day), Watts or kWh per square foot of raised floor area / year or month or day, etc.
Also, if you look at the Energy Star survey results, the EPA is supporting an "EUE" metric, which incorporates an incorrect "source energy factor" to distort what is happening. EPA does not support PUE.
ASHRAE 90.1 does have a performance path to meeting the standard. The prescriptive path that Google is commenting on is used by building owners/ design teams that do not have the resources to (or choose not to) go through an energy analysis to optimize their design. The prescriptive path attempts to determine what will work best for most projects to achieve a baseline efficiency, without requiring much analysis.Building Energy Engineer, Enermodal Engineering
I also agree that a very prescriptive approach is non constructive and only helps engineers in conformity and the ability of judgment of a process compared to a scope of works. (I’m an engineer and I certainly know the risk averse drives engineers use) Energy efficiency of a data centre is easy to understand.1- Reduce the number of servers & IT equipment you need to do a task (Initial power & cooling drop) Eg; virtual server, non rotating storage, and low loss communications ect.2- If you put power into it you have to cool it and the less energy required to cool the better. The best way it to place your data centre in cooler environment areas (less than 35degC in summer) and use outside air cooling.3- Use energy efficient support process like UPS systems, Mechanical systems, correctly or slightly over sized cabling & switchgear to reduce resistance, purchase renewable energy and quality insulation.Dave Mc
There is really no point in directly imposing a restriction on datacenter efficiency. If using too much electricity has impact on the environment, closing coal mines and raise its import tax should do the trick. This is just another attempt to redistribute power among various regulators.
'Unfortunately, the proposed ASHRAE standard is far too prescriptive. Instead of setting a required level of efficiency for the cooling system as a whole, the standard dictates which types of cooling methods must be used.'You don't have to know anything whatsoever about data centers of cooling systems to know that dictating particular methods is the wrong way to go. If you want increased efficiency, require increased efficiency. Counting on technology NOT improving has NEVER succeeded, and it never will. Google et al are right on in opposing these standards.
Great posts, all. Fascinating thread.We've been designing / refitting an 80,000 s/f, 12 - 15mW co-lo facility atop a large aquifer. 100% groundwater cooled data center, where chilled (48.5 degrees F) groundwater is pumped and circulated through CRAHs. N+X. Open loop system, with the added mechanical efficiencies of a siphon effect because of the recapture wells. For every 1mW of critical load our total cooling load is <65kW! Very efficient. Zero water loss, to boot.We look forward to proving our theoretically low PUE through operating history over time. Capital and operating costs are significantly offset.If any of you might be interested in participating in a test bed / pilot of this project in what we call Dataville, please feel free to get in touch with me directly. I'm happy to share our engineering diagrams for 100% groundwater cooling freely, open-source -style. Perhaps you might help improve / refine them.This year-round solution requires zero chiller plant, and even works around the need for free airside economizers.I would be delighted and honored if any subject matter experts and thought leaders who posted to this blog might critique or contribute thought / engineering rigors to our groundwater cooling schematic.Incidentally, we are also recipients of a $2 million CANARIE research grant to create a 100% carbon-neutral next-gen Internet / utility computing platform across Canada, and are the East Coast IX / node in this project - the GreenStar Network. The project just started.We are opening GSN to peering and benchmarking arrangements, and welcome your interest. Anton E. Self, Bastionhost Ltd.direct: 902-482-6466www.bastionhost.com
I would strongly advise Google, Amazon, MS, and similar premier companies to use the performance based compliance model offered in the same standard. The prescriptive path would be what I would expect to see applied to a small data center, under 1,000 sf tucked in the basement of an 100,000 sf office building, never to a purpose built and engineered facility. Is Google concerned that local officials will adopt only the prescriptive path and strike the performance path when implementing in local code? That is something to fear - never underestimate the ability of a local government to screw up badly...
ASHRAE's (American Society of Heating and Ventilating Engineers) forte is the use of air to heat and cool. And quite naturally they will prescribe an air based solution to cool data centers. The sad thing about the prescription is that it is far from optimum. Air cooling in any form is horribly inefficient. For example, Facebook estimates that with free cooling their energy cost for cooling is about 15% of IT load. But this IT load includes the server fan energy which accounts for about 10% of the total. Moving this to the cooling side of the equation shows that the real cost of cooling is about 38%.The other problem with air cooling is that it is difficult to manage. Baffles, screens, raised floors, actuators, humidifiers, dehumidifiers etc. are all required. Tight control loops are impossible to realize as air moves in complex ways and time constants are very long. Our tests of liquid based direct touch server cooling show that cooling overhead in a climate similar to Facebook's location (pacific northwest) would be about 5% of the real IT load. Control is also much easier as time constants are much shorter and fluids move in predictable ways through set channels.If the 90.1 changes become ratified, the only way innovative data center cooling solutions can be approved by the individual States' planning organizations is through the mechanism outlined in chapter 11. The tools available to perform the calculations are both arcane and out of date. To all intents and purposes, this kills all chance of innovation.ASHRAE is opening itself up to questions on its competence to be the arbiters of data center cooling design. Their actions to date also have the potential of hurting America's competitiveness worldwide.In the interest of full disclosure, I should point out that my company, Clustered Systems Company, Inc. designs and manufactures advanced direct touch liquid cooling systems.
The comments on this blog belong only to the person who posted them. We do, however, reserve the right to remove off-topic or inappropriate comments.