The Data Center Temperature argue

Though never directly articulated by any data center authority, the prevailing practice surrounding these basic facilities has often been “The colder, the better.” However, some leading server manufacturers and data center efficiency experts proportion the opinion that data centers can run far hotter than they do today without sacrificing uptime and with a huge savings in both cooling related costs and CO2 emissions. One server manufacturer recently announced that their rack of servers can function with inlet temperatures at 104 deg F.

Why does it feel the need to push the envelope? The cooling infrastructure is an energy hog. This system, operating 24x7x365, consumes a lot of electricity to create the optimal computing ecosystem, which may hover anywhere between 55 to 65 deg F. (The current “recommended” range from ASHRAE is 18-27 C or 64.4 deg F by 80.6 deg F)

To unprotected to efficiencies, a number of influential end users are running their data centers warmer and are advising their contemporaries to follow suit. But the time of action isn’t as simple as raising the thermostat in your home. Here are some of the meaningful arguments and considerations:

Position: Raising server inlet temperature will realize meaningful energy savings.

Arguments for:

o Sun Microsystems, both a noticeable hardware manufacturer and data center operator, estimates a 4% savings in energy costs for every one (1) degree increase in server inlet temperature. (Miller, 2007)

o A higher temperature setting can average more hours of “free-cooling” possible by air-side or water side economizers. This information is especially powerful to an area like San Jose, California, where outside air (dry-bulb) temperatures are at or below 70 deg F for 82% of the year. Depending on geography, the annual savings from economization could go beyond six figures.

Arguments Against:

o The cooling infrastructure has certain design setpoints. How do we know that raising server inlet temperature won’t consequence in false economy, causing additional, unnecessary consumption in other elements like the server fans, pumps, or compressors?

o Free-cooling, while great for new data centers, is an expensive proposition for existing ones. The complete cooling infrastructure would require re-engineering and may be cost prohibitive and unnecessarily complicate.

o Costs from thermal-related equipment failures or downtime will offset the savings realized from a higher temperature setpoint.

Position: Raising server inlet temperature complicates reliability, recovery, and equipment warranties.

Arguments for:

o Inlet air and exhaust air frequently mix in a data center. Temperatures are kept low to offset this mixing and to keep the server inlet temperature within ASHRAE’s recommended range. Raising the temperature could strengthen already-existing hotspots.

o Cool temperatures provide an envelope of cool air in the room, an asset in the case of a cooling system failure. The staff may have more time to diagnose and repair the problem and, if necessary, shut down equipment gracefully.

o In the case of the 104 degree F server, what’s the chance every piece of equipment-from storage to networking-would perform reliability? Would all warranties keep valid at 104 deg F?

Arguments Against:

o Raising the data center temperature is part of an efficiency program. The temperature increase must follow best practices in airflow management: using blanking panels, sealing cable cutouts, eliminating cable obstructions under the raised floor, and implementing some form of air containment. These measures can effectively reduce the mixing of hot and cold air and allow for a safe, functional temperature increase.

o The 104 degree F server is an extreme case that encourages thoughtful discussion and basic inquiry among data center operators. After their study, perhaps a facility that once operated at 62 deg now operates at 70 deg F. These changes can considerably enhance energy efficiency, while not compromising availability or equipment warranties.

Position: Servers are not as fragile and sensitive as one may think. Studies performed in 2008 underscore the resiliency of modern hardware.

Arguments For:

o Microsoft ran servers in a tent in the damp Pacific Northwest from November 2007 by June 2008. They experienced no failures.

o Using an air side economizer, Intel subjected 450 high density servers to the elements-temperatures as high as 92 deg and relative humidity ranges from 4 to 90%. The server failure rate during this experiment was only marginally higher than Intel’s enterprise facility.

o Data centers can function with a temperature in the 80s and nevertheless be ASHRAE compliant. The upper limit of their recommended temperature range increased to 80.6 deg F (up from 77 deg F).

Arguments Against:

o High temperatures, over time, affect server performance. Server fan speed, for example, will increase in response to higher temperatures. This use and tear can shorten the device’s life.

o Studies from data center behemoths like Microsoft and Intel may not be applicable to all businesses:

o Their enormous data center footprint is more immune to an occasional server failure that may consequence from excessive heat.

o They can leverage their buying strength to receive gold-plated warranties that permit higher temperature settings.

o They are most likely uncommon their hardware at a more rapid speed than other businesses. If that server is completely spent after 3 years, no big deal. A smaller business may need that server to last longer than 3 years.

Position: Higher Inlet Temperatures may consequence in uncomfortable working conditions for data center staff and visitors.

Arguments for:

o Consider the 104 degree F rack. The hot aisle could be anywhere from 130 deg to 150 deg F. already the higher end of ASHRAE’s operating range (80.6 deg F) would consequence in hot aisle temperatures around 105-110 deg F. Staff servicing these racks would persevere very uncomfortable working conditions.

o Responding to higher temperatures, the server fan speed will increase to dissipate more air. The increased fan speed would increase the noise level in the data center. The noise may approach or go beyond OSHA sound limits, requiring occupants to use ear protection.

Arguments Against

o It goes without saying that as the server inlet temperature increases, so does the hot aisle temperature. Businesses must carefully balance worker comfort and energy efficiency efforts in the data center.

o Not all data center environments have high user quantity. Some high performance/supercomputing applications function in a lights-out ecosystem and contain a homogeneous collection of hardware. These applications are well suited for higher temperature setpoints.

o The definition of data center is more fluid than ever. The traditional brick and mortar facility can add instantaneous compute strength by a data center container without a costly construction project. The container, segregated from the rest of the building, can function at higher temperatures and unprotected to greater efficiencies (Some close-coupled cooling products function similarly).

Conclusions

The movement to raise data center temperatures is gaining but it will confront opposition until the concerns are addressed. Reliability and availability are at the top of any IT specialized’s performance plan. For this reason, most to date have decided to error on the side of caution: to keep it cool at all costs. however, higher temperatures and reliability are not mutually exclusive. There are ways to safeguard your data center investments and become more energy efficient.

Temperature is inseparable from airflow management; data center professionals must understand how the air gets around, into, and by their server racks. Computational fluid dynamics (CFDs) can help by analyzing and charting projected airflow on the data center floor, but as cooling equipment doesn’t always perform to spec and the data you go into could miss some meaningful obstructions, onsite monitoring and adjustments are basic requirements to insure that your CFD data and calculations are accurate.

Data centers with excess cooling are chief environments to raise the temperature setpoint. Those with hotspots or insufficient cooling can start with low-cost remedies like blanking panels and grommets. Close-coupled cooling and containment strategies are especially applicable, as server exhaust air, so often the cause of thermal challenges, is secluded and extremely from entering the cold aisle.

With airflow addressed, users can focus on finding their “sweet identify”-the ideal temperature setting which aligns with business requirements and improves energy efficiency. Finding it requires proactive measurement and examination. But the rewards-smaller energy bills, improved carbon footprints and a message of corporate responsibility-are well worth the effort.

Leave a Reply