In 2005, problems in the data centre at Pacific Northwest National Laboratory came to a head.
Unscheduled outages were occurring almost monthly, bringing down the data centre for hours at a time. Groups were buying an increasing number of rack-mounted servers - which had recently become cheaper at the time - to boost the computing resources, says Ralph Wescott, data centre services manager for the government laboratory, which is managed by the U.S. Department of Energy. In July, 2005, the server room had reached its capacity limit.
"Groups would go buy a server and throw it over the wall to me, saying, 'Hey, install this,'" Wescott says. "But I didn't have any space, power or cooling (capacity) left. If I installed (one more), the whole room would go dark."
Wescott and PNNL embarked on a broad project to revamp their data centre without breaking the budget. Every quarter for three years, the data centre group spent a weekend shutting down the server room and replacing a row of old servers and tangled network cables under the floor with more efficient, yet more powerful servers connected by fewer cables running in the ceiling. The new configuration allowed for more efficient cooling under the floor.
The result? PNNL moved from 500 applications on 500 servers to 800 applications running on 150 servers.
During a tight economy, tackling such information-technology projects require a tight grip on the purse strings, says Joseph Pucciarelli, the program director of technology, financial and executive strategies for analyst firm IDC, a sister company to CIO.com.
"The situation is a very common one," he says. "Companies are making just-in-time investments. They have a problem, and they are looking at the problem in a constrained way."
Here are some lessons PNNL learned in bringing their data centre back from the brink.
1. Plan, don't react The first problem Wescott needed to solve was the data centre group's habit of reacting to each small problem as it arose, rather than seeing the systematic issues and creating a plan to create a sustainable service. In addition to the 500 servers, the data centre had some 33,000 cables connecting those servers to power, networking and security systems.
"We decided what the data centre should look like and what its capacity should be," he says.
The group concluded that the current trajectory would result in 3,000 applications, each running on its own server, in 10 years. Now, the data centre has 81 percent of applications virtualized - and average of 17 per server - and Wescott plans to reach the 90 percent mark.
Companies should focus on three areas to increase capacity, says IDC's Pucciarelli. Reducing the number of physical servers and running applications on virtual systems helps reduce power requirements, as does more efficient cooling systems and improvements in electrical distribution.
"That's typically the one-two-three that you go to when updating the data centre," he says.
Pucciarelli has encountered many companies that have replaced up to 50 servers with just two or three larger capacity systems and used virtualization to run their applications.
2. Measure to manage Data centre managers need ways to monitor the state of the data centre, but all too frequently they don't have the right tools, PNNL's Wescott says. Prior to the changes, Pacific Northwest National Labs had no way to measure the efficiency of its data centre. Power problems were discovered when the room went dark, or though a more seat-of-your-pants method.
"If there was too much amperage through our power supplies, the way I found out was to put my hand on the circuit breaker and if it was warm, then I knew we had a problem," he says. "That's proof that you need tools."
Now, PNNL has sensors in place on every fourth cabinet at the low, medium and high points to create a 3-D heat map of the server room. The data allowed Wescott to change the way he cools the data centre, increasing overall temperatures and applying cooling where he needed it.
"I think that is going to save me a lot of money, and wear and tear, on my air conditioners," he says, adding that current estimates are that the data centre will be 40 percent more efficient with cooling.