Cooling method for supercomputers to save millions of gallons of water

1st September 2016
Posted By : Enaie Azambuja
Cooling method for supercomputers to save millions of gallons of water


In different parts of the country, people discuss gray-water recycling and rainwater capture to minimise the millions of gallons of groundwater required to cool large data centres. But the simple answer in many climates, said Sandia National Laboratories researcher David J. Martinez, is to use liquid refrigerant.

Based on that principle, Martinez—engineering project lead for Sandia's infrastructure computing services—is helping design and monitor a cooling system expected to save 4 million to 5 million gallons annually in New Mexico if installed next year at Sandia's computing center, and hundreds of millions of gallons nationally if the method is widely adopted.

It's now being tested at the National Renewable Energy Laboratory in Colorado, which expects to save a million gallons annually. The system, built by Johnson Controls and called the Thermosyphon Cooler Hybrid System, cools like a refrigerator without the expense and energy needs of a compressor.

Currently, many data centers use water to remove waste heat from servers. The warmed water is piped to cooling towers, where a separate stream of water is turned to mist and evaporates into the atmosphere. Like sweat evaporating from the body, the process removes heat from the piped water, which returns to chill the installation.

But large-scale replenishment of the evaporated water is needed to continue the process. Thus, an increasing amount of water will be needed worldwide to evaporate heat from the growing number of data centers, which themselves are increasing in size as more users put information into the cloud.

"My job is to eventually put cooling towers out of business," Martinez said. "Ten years ago, I gave a talk on the then-new approach of using water to directly cool supercomputers. There were 30 people at the start of my lecture and only 10 at the end. "'Dave,' they said, 'no way water can cool a supercomputer. You need air.'

"So now most data centers use water to cool themselves, but I'm always looking at the future and I see refrigerant cooling coming in for half the data centers in the U.S., north and west of Texas, where the climate will make it work."

The prototype method uses a liquid refrigerant instead of water to carry away heat. The system works like this: Water heated by the computing center is pumped within a closed system into proximity with another system containing refrigerant.

The refrigerant absorbs heat from the water so that the water, now cooled, can circulate to cool again. Meanwhile the heated refrigerant vaporises and rises in its closed system to exchange heat with the atmosphere. As heat is removed from the refrigerant, it condenses and sinks to absorb more heat, and the cycle repeats.

"There's no water loss like there is in a cooling tower that relies on evaporation," Martinez said. "We also don't have to add chemicals such as biocides, another expense. This system does not utilise a compressor, which would incur more costs. The system utilises phase-changing refrigerant and only requires outside air that's cool enough to absorb the heat."

In New Mexico, that would occur in spring, fall and winter, saving millions of gallons of water. In summer, the state's ambient temperature is high enough that a cooling tower or some method of evaporation could be used.

But more efficient computer architectures can raise the acceptable temperature for servers to operate and make the occasional use of cooling towers even less frequent.

"If you don't have to cool a data center to 45ºF but instead only to 65 to 80ºF, then a warmer outside air temperature—just a little cooler than the necessary temperature in the data center—could do the job," Martinez said.

For indirect air cooling in a facility, better design brings the correct amount of cooling to the right location, allowing operating temperatures to be raised and allowing the refrigerant cycle to be used more during the year.

"At Sandia, we used to have to run at 45ºF. Now we're at 65 to 78ºF. We arranged for air to flow more smoothly instead of ignoring whorls as it cycled in open spaces. We did that by working with supercomputer architects and manufacturers of cooling units so they designed more efficient air-flow arrangements.

Also, we installed fans sensitive to room temperature, so they slow down as the room cools from decreased computer usage and go faster as computer demand increases. This results in a more efficient and economical way to circulate air in a data center." Big jobs that don't have to be completed immediately can be scheduled at night when temperatures are cooler.

You must be logged in to comment

Write a comment

No comments

Sign up to view our publications

Sign up

Sign up to view our downloads

Sign up

The Digital Healthcare Show 2019
26th June 2019
United Kingdom EXCEL, London
unbound london 2019
17th July 2019
United Kingdom Old Truman Brewery, London
DSEI 2019
10th September 2019
United Kingdom EXCEL, London
EMO Hannover 2019
16th September 2019
Germany Hannover
European Microwave Week 2019
29th September 2019
France Porte De Versailles Paris