A list of 10 common surprises for data center/IT managers was released at AFCOM Data Center World Spring. The list includes information on a surprising cause of data center downtime, what data center managers might not know about that next server refresh, and the growing trend sneaking up on virtually every IT manager.
Ten common surprises:
- Those high-density predictions finally are coming true: After rapid growth early in the century, projections of double-digit rack densities have been slow to come to fruition. Average densities hovered between 6.0 and 7.4 kW per rack from 2006 to 2009, but the most recent Data Center Users’ Group (DCUG) survey predicted average rack densities will reach 12.0 kW within three years. That puts a premium on adequate UPS capacity and power distribution as well as cooling to handle the corresponding heat output.
- Data center managers will replace servers three times before they replace UPS or cooling systems: Server refreshes happen approximately every three years. Cooling and UPS systems are expected to last much longer—sometimes decades. That means the infrastructure organizations invest in today must be able to support—or, more accurately, scale to support—servers that may be two, three or even four generations removed from today’s models. What does that mean for today’s data center manager? It makes it imperative that today’s infrastructure technologies have the ability to scale to support future needs. Modular solutions can scale to meet both short- and long-term requirements.
- Downtime is expensive: Everyone understands downtime is bad, but the actual costs associated with an unplanned outage are stunning. According to a Ponemon Institute study, an outage can cost an organization an average of about $5,000 per minute. That’s $300,000 in just an hour. The same study indicates the most common causes of downtime are UPS battery failure and exceeding UPS capacity. Avoid those problems by investing in the right UPS—adequately sized to support the load—and proactively monitoring and maintaining batteries.
- Water and the data center do not mix – but we keep trying: The first part of this probably isn’t a surprise. Sensitive IT equipment does not respond well to water. However, the aforementioned Ponemon study indicates 35% of all unplanned outages are a result of some type of water incursion. These aren’t just leaky valves; in fact, many water-related outages are the result of a spilled coffee or diet soda. The fix: Check those valves, but more importantly, check the drinks at the door.
- New servers use more power than old servers: Sever consolidation and virtualization can shrink server inventory by huge numbers, but that doesn’t exactly equate to huge energy savings. New virtualized servers, especially powerful blade servers, can consume four or five times as much energy as those from the preceding generation (although they usually do it more efficiently). The relatively modest savings at the end of that consolidation project may come as a surprise. There is no fix for this, but prepare for it by making sure the infrastructure is adequate to support the power and cooling needs of these new servers.
- Monitoring is a mess: IT managers have more visibility into their data centers than ever before, but accessing and making sense of the data that comes with that visibility can be a daunting task. According to an Emerson Network Power survey of data center professionals, data center managers use, on average, at least four different software platforms to manage their physical infrastructure. More than 40% of those surveyed say they produce three or more reports for their supervisors every month, and 34% say it takes three hours or more to prepare those reports. The solution? Move toward a single monitoring and management platform. Today’s DCIM solutions can consolidate that information and proactively manage the infrastructure to improve energy and operational efficiency and even availability.
- The IT guy is in charge of the building’s HVAC system: The gap between IT and Facilities is shrinking, and the lion’s share of the responsibility for both pieces is falling on the IT professionals. Traditionally, IT and data center managers have had to work through Facilities when they need more power or cooling to support increasing IT needs. That process is being streamlined, thanks in large part to those aforementioned DCIM solutions that increase visibility and control over all aspects of a building’s infrastructure. Forward-thinking data center managers are developing a DCIM strategy to help them understand this expansion of their roles and responsibilities.
- That patchwork data center needs to be a quilt: In the past, data center managers freely mixed and matched components from various vendors because those systems worked together only tangentially. That is changing. The advent of increasingly intelligent, dynamic infrastructure technologies and monitoring and management systems has increased the amount of actionable data across the data center, delivering real-time modeling capabilities that enable significant operational efficiencies. IT and infrastructure systems still can work independently, but to truly leverage the full extent of their capabilities, integration is imperative.
- Data center on demand is a reality: The days of lengthy design, order and deployment delays are over. Today there are modular, integrated, rapidly deployable data center solutions for any space. Integrated, virtually plug-and-play solutions that include rack, server and power and cooling can be installed easily in a closet or conference room. On the larger end, containerized data centers can be used to quickly establish a network or to add capacity to an existing data center. The solution to most problems is a phone call away.
- IT loads vary – a lot: Many industries see extreme peaks and valleys in their network usage. Financial institutions, for example, may see heavy use during traditional business hours and virtually nothing overnight. Holiday shopping and tax seasons also can create unusual spikes in IT activity. Businesses depending on their IT systems during these times need to have the capacity to handle those peaks, but often can operate inefficiently during the valleys. A scalable infrastructure with intelligent controls can adjust to those highs and lows to ensure efficient operation. BD+C
Related Stories
| Aug 19, 2011
Enhanced acoustical design
Ambient noise levels in some facility types are trending up and becoming a barrier to clear communication between building occupants.
| Jul 22, 2011
The Right Platform for IPD
Workstations for successful integrated project delivery, a white paper by Dell and BD+C.
| Jul 22, 2011
High-performance windows and doors
Learning objectives After reading this article, you should be able to: Understand issues of thermal performance and energy efficiency in relation to window and door systems; describe optimal detailing of the window-wall interface and how it contributes to building performance, sustainability, and occupant well-being; understand how durability contributes to sustainable windows/doors; and list sustainable O&M requirements for window and door systems.
| Jul 21, 2011
Falling Architecture Billings Index reflects decrease in design demands
This months Architecture Billings Index (ABI), provided by the American Institute of Architects, is almost a full point lower than last month’s reported score. June’s reading of 47.2 was short of the required 50 to achieve billings increases, making July’s reading of 46.3 an unwelcome sign of market tidings.
| Jun 29, 2011
New leadership role for architects in net-zero design
BD+C Editorial Director Rob Cassidy talks with RNL Design's Tom Hootman, AIA, about the changing role of architects in net-zero designs.