Fire Detection and Suppression Technology

Sometimes the unimaginable happens. A fire can threaten to destroy a data center. To protect the valuable equipment and information housed in the facility, it is critical to install a fire suppression system adequate to the size, type and operational responsibilities of the complex. By definition, a fire suppression system is a combination of fire detection and extinguishing devices designed to circumvent catastrophic business loss as a result of a fire. This loss includes not only the cost of equipment replacement, but also the cost of recovering lost data or business-specific applications.

Detection Systems: The First Line of Defense

A critical component of any suppression system is smoke detectors. Depending on the application, they can be of the photoelectric or ionization type. Detectors perform several vital functions:

  • Warn facility occupants of possible fire.
  • Shut down all electrical service to the equipment so as not to “fuel the fire.”
  • Activate the suppression medium.

If it is properly designed, the detection system can also be used to limit business loss due to power-off interfaces by detecting a system failure rather than an actual smoke condition.

A highly effective detection system is one we call an “intelligent” system. It uses a software-based early warning system to provide an accurate means of detection and verification at the ceiling plane and underfloor plenum.

Water and Clean Agent Gas: Common Suppression Media

Suppression medium is activated if a true emergency is detected. The two most commonly used media to put out a fire are water and clean agent gas such as FM200, Inergen, and NAFS-III.

Determining which type of suppression medium to use depends in large part on the requirements of local code enforcement authorities, building and/or landlord stipulations, and input from insurance underwriters. It also depends on user preference, which is influenced by such factors as cost, business risk relative to data recovery, existing systems, and so forth.

Water sprinkler systems

Water sprinkler systems are found in most buildings regardless of the presence of a data center. As a general rule, where sprinkler systems exist, it is less expensive to convert to a pre-action sprinkler system than to install a clean agent system. Pre-action sprinklers are the water-based choice for data centers and refer to systems that control the flow of water to pipes in the ceiling plane. Smoke and heat activate a valve that advances the water to the ceiling plane. That way, inadvertent damage to equipment from leakage or accidental discharge is prevented. (By comparison, with an ordinary sprinkler system, water is contained in pipes in the ceiling plane at all times.)

Water is highly effective at putting out fires and is well suited for areas like printer rooms that contain combustible materials like paper and toner. The downside of water-based systems is the messy and lengthy clean up and recovery time after a water discharge.

Clean agents

There are primarily three clean agents presently vying for acceptance in the marketplace, FM200, NAF S-III, and Inergen. These agents were developed in response to the phase-out of Halon and the development of NFPA 2001, which was adopted in the Fall of 1994.

Consideration of these agents as alternatives to CO2 in under floor applications is viable. The costs of these systems has dropped in recent years due to more competition in the market place with competing vendors offering these various gas options.

  1. FM-200 (Heptafluoropropane – HFC-227EA) is a colorless, liquefied compressed gas. It is stored as a liquid and dispensed into the hazard as a colorless,FM-200 tanks electrically non-conductive vapor. It leaves no residue. It has acceptable toxicity for use in occupied spaces when used as specified in the United States Environmental Protection Agency (EPA) proposed Significant New Alternatives Policy (SNAP) program rules. FM-200 extinguishes a fire by a combination of chemical and physical mechanisms.

    FM-200 is an effective fire-extinguishing agent that can be used on many types of fires. It is effective for use on Class A Surface-Burning Fires, Class B Flammable Liquid, and Class C Electrical Fires.

    On a weight of agent basis, FM-200 is a very effective gaseous extinguishing agent. The minimum design concentration for total flood applications in accordance with NFPA 2001 shall be 7.0%.

  2. NAF S-III is a clean, non-conductive media used for the protection of a variety of potential fire hazards, including electrical and electronic equipment. NAF S-III is a clean gaseous agent at atmospheric pressure and does not leave a residue. It is colorless and non-corrosive.

    NAF S-III acts as a fire-extinguishing agent by breaking the free radical chain reaction that occurs in the flame during combustion and pyrolysis. Like Halon 1301, NAF S-III has a better efficiency with flaming liquids than with deep-seated Class A fires.

    NAF S-III fire extinguishing systems have the capability to rapidly suppress surface-burning fires within enclosures. The extinguishing agent is a specially developed chemical that is a gas at atmospheric pressure and is effective in an enclosed risk area. NAF S-III extinguishes most normal fires at the design concentration by volume of 8.60% at 20° C.

    NAF S-III is stored in high-pressure containers and super-pressurized by dry nitrogen to provide additional energy to ensure rapid discharge. At the normal operating pressure of 360 psi (24.8 bar) or 600 psi (42 bar), NAF is in liquid form in the container.

    Once the system is activated, the container valves are opened and the nitrogen propels the liquid under pressure through the pipe work to the nozzles, where it vaporizes. The high rate of the discharge through the nozzles ensures a homogeneous mixture with the air. Sufficient quantities of NAF S-III should be discharged to meet the concentration required and the pressure at each nozzle must be located to achieve uniform mixing.

  3. Inergen is composed of naturally occurring gases already found in Earth’s atmosphere (nitrogen, argon, and CO2). Inergen suppresses fire by displacing the oxygen in the environment. Inergen, however, is not toxic to the occupants because of the way it interacts with the human body. The level of CO2 in Inergen stimulates the rate of respiration and increases the body’s use of oxygen. This compensates for the lower oxygen levels that are present when Inergen is discharged.

    Inergen is stored as a dry, compressed gas and is released through piping systems similar to those utilized in other gaseous suppression systems.

  4. FE-25 fire suppression agent is environmentally acceptable replacement for Halon 1301. FE-25 is an odorless, colorless, liquefied compressed gas. It is stored as a liquid and dispensed into the hazard as a colorless, electrically non-conductive vapor that is clear and does not obscure vision. It leaves no residue and has acceptable toxicity for use in occupied spaces at design concentrations. FE-25 extinguishes a fire by a combination of chemical and physical mechanisms. FE-25 does not displace oxygen and therefore is safe for use in occupied spaces without fear of oxygen deprivation.

    FE-25 has zero ozone depleting potential, a low global warming potential, and a short atmospheric lifetime.

    FE-25 closely matches Halon 1301 in terms of physical properties such as flow characteristics and vapor pressure. The pressure traces, vaporization, and spray patterns for FE-25 nearly duplicate that of Halon 1301. The minimum design concentration for FE-25 systems is 8.0% meaning that about 25% more of FE-25 agent will be required. Fe-25 requires about 1.3 times the storage area of Halon.

    When retrofitting existing Halon 1301 system, the nozzles and cylinder assembly will need to be upgraded, however, the piping system likely will not need to be changed, which is cost-effective retrofit that minimizes business interruption.

  5. FE-13 is a clean, high-pressure agent that leaves no residue when discharged. FE-13 efficiently suppresses fire by the process of physiochemical thermal transfer. The presence of FE-13 absorbs heat from the fire as a sponge absorbs liquid. FE-13 is safe for use in occupied spaces up to a 24% concentration. Design concentration for total flood application is 16%.
  6. Novec 1230 is the newest clean-agent gas available on the market. It is marketed as a long-term sustainable alternative to FM-200 and Halon. Novec 1230 has a 0.0 ozone depletion potential (equivalent to FM-200), but has an atmospheric lifetime of only five days, compared to FM-200’s half life of over 20 years. Novec 1230 has a zero global warming potential. Novec 1230 is designed to a concentration level of 4-6%, which will require less gas than other clean agent. Novec 1230 extinguishes the fire by heat absorption, and is heavier than air, so the gas will sink in the room. Novec 1230 is also safe for electronic equipment, so the data center may not have to be shut down in the event of a gas discharge.

    Novec 1230 will require the same amount of tanks as FM-200, and is stored as a liquid under pressure. Under normal atmospheric conditions, it will exist as a gas. The system is approximately 5-7% more expensive than FM-200.

Table – Relative Cost Comparison of Extinguishing Methods
Scenario Characteristics:

  • Occupied Room
  • Housing electrical equipment
  • 10,000 cu-ft room volume
  • Room fully enclosed and building is fully sprinklered
Design Basis:

(1) Total flooding.

(2) Does not include the cost of fire alarm and detection system. Probable cost < $4,000.

(3) Assume a fully sprinklered building and

(4) Includes the cost of the extinguishing agent.

Extinguishing Agent Design Concentration, Density Agent Quantity Installation Cost (4) Recharge Cost Design Basis
FM-200 7.44 % by volume 364 lbs 20% more than Inergen Almost twice the cost of Inergen (1) + (2)
FE-25 96% by volume 335 lbs Parallel to FM-200 less gas 20%-25% less than FM-200 (1) + (2)
Inergen 37.5 % by volume 4780-cu-ft (1) + (2)
NAF S-III 8.60 % by volume ___ ___ ___ (1) + (2)
Pre-Action Sprinklers 0.1 gpm/s.f water N/A 1/4 the cost of Halon or Inergen N/A (2) + (3)

Note: NAF S-III does not appear to have the market presence to be a viable alternative.

Mobile Computing and Driving Cloud Applications

How Cloud Computing is Driving Mobile Data Growth and a Next Generation of Data Center Strategies.

At the 2010 Techonomy Conference in Lake Tahoe, Eric Schmidt quoted a figure that most data professionals can relate to:

Every two days, there is as much ‘data’ produced globally as was produced from the dawn of civilization up to 2003 —  approximately five exabytes every 48 hours.

Certainly, the rapid growth in user-generated content is a major driver of the expansion. But, user-generated content is really a reflection of a bigger shift in the IT landscape that is being driven by changes in the way business interacts with customers in both B2B and B2C markets, and an attendant revolution in the tools that consumers and businesses use to interact with applications and data.

According to Gartner, data growth remains on the top of list of IT’s biggest challenges for 2011, with 47% of respondents to a summer 2010 survey of over 1,000 large enterprises citing data growth in their top three challenges. (i)  (N.B.  37% of listed system performance and scalability and 36% cited network congestion and connectivity issues.) Gartner’s research uncovered data growth rate between 40% and 60% y/y. For enterprise IT professionals this includes rapid expansion in unstructured data, such as e-mail and regulatory and compliance documents. But that is just part of the story.

In November 2009, technology writer Robert Cringely wrote:

“We’re in the middle of a huge platform shift in computing and most of us don’t even know it.  The transition is from desktop to mobile and is as real as earlier transitions from mainframes to minicomputers to personal computers to networked computers with graphical interfaces.”  (ii) 

According to ABI Research, quoted in IEEE Spectrum, monthly data transmission by mobile computing devices, including cellphones, tablets and portable computers, will increase 1400% by 2014, with the number of people subscribing to cloud-based applications increasing from 71 million to just under 1 billion in the same timeframe. (iii)  Cloud applications, in this case, are defined as applications that are delivered with data storage or processing power not primarily resident on the mobile device. The leading applications in this shift include consumer-friendly utility software (such as maps), but also business productivity tools (especially for sales, data sharing, and collaboration), rapidly growing social networking applications, and now-ubiquitous search functionality.

From consumers to business professionals, we are rapidly becoming addicted to real-time access of all forms of data. Rich-media (audio, video) garners the headlines, but today, we expect to log on to find order status, place new orders, make stock trades or bank deposits, explore real-time inventory, or review engineering designs — all through a growing array of mobile devices.

What does all this mean to data center professionals?

Mobile applications carry new demands on data types, access control and traffic management, especially in cloud architectures. In the cloud, network resources are consumed by virtual machines, and network management and monitoring must intelligently correlate virtual machine traffic with physical network components and resources. Rapidly increasing demand on transactional resources dictates completely new QoS approaches to bandwidth allocation and traffic shaping. And data resources must be dynamically provisioned and carefully mirrored to ensure durable performance under widely varying loads.

BRUNS-PAK Cloud Computing Programs

BRUNS-PAK engineers are acutely aware this dramatic evolution in the IT market.  Our cloud computing team helps clients:

  • Organize and document objectives
  • Understand data center implications for both private cloud deployment or public cloud integration
  • Plan for long-term optimized data center performance though load modeling for high-utilization, virtualized deployment architectures
  • Develop a data center deployment model that is adaptive to rapid growth in both network access and data volume and can respond to the need for capacity expansion on a rapid basis through either increased site utilization or modular expansion

For more information on our Cloud Computing capabilities and offerings, click here to access our Contact Us page and call or request information.


(i)  “Data Growth Remains IT’s Biggest Challenge, Gartner Says.”  Computerworld Online.  http://www.computerworld.com/s/article/9194283/Data_growth_remains_IT_s_biggest_challenge_Gartner_says. November 2, 2010.

(ii)  Cringely, Robert X, “Pictures In Our Heads.” I, Cringely. 11/17/2009.

(iii) “Cloud Computing Drives Mobile Data Growth”.  IEEE Spectrum Online. http://spectrum.ieee.org/telecom/wireless/cloud-computing-drives-mobile-data-growth. October 2009

The Importance of Being Agile

IT/Line of Business IT SpendingIn a Nov 2012 survey by IDG Research Services1, CIOs got yet another wakeup call regarding the price for being non-responsive to increasingly demanding line-of-business owners in the enterprise. In the study, the majority of respondents indicated that line-of-business teams control over 20% of the IT budget. Further, 20% of respondents indicated that these teams now dictate the majority of IT spending. For CIO’s used to having rigid control of their IT infrastructure, this is a discomforting reminder of the changes being wrought on IT by rapidly growing demand for agility that supports greater alignment with business operations.

The accessibility of software as a service solutions, growing acceptance of mobile applications running on personally or company owned devices, and shortening time to market for new services-oriented applications are combining to make it easier than ever for non-IT professionals to identify lightweight solutions to business problems that can be deployed without IT assistance and charged to company credit cards. The trend may have started in the shadows, but it is a mainstream approach, with applications like Salesforce.com, Workday and Evernote now mainstays in enterprise operations. The upside? Speed, reduced capital costs and flexible deployment options. The downside? Greater enterprise risk, compliance concerns, and barriers to sharing data across application environments.

How can IT adapt? In today’s market, CIOs must eschew old strategies favoring tight management of infrastructure and services, in favor of a new services-driven orchestration model in which the IT department focuses on the delivery of strategic value from IT assets, independent of where those assets reside and who owns them. This model makes concepts like just-in-time IT a focus, and forces the integration of new strategies for enterprise IT delivery, including:

  • cloud computing
  • containers/PODS/ultramicro data centers
  • advanced networking techniques for minimal latency
  • SDN (software defined networking)
  • OPEX/CAPEX rebalancing

Making sense of these trends and technologies can be confusing. BRUNS-PAK Consulting Services is an integral part of BRUNS-PAK’s comprehensive data center services offerings. Our consulting services team is expert at helping customers to plan and implement complex strategies for alternative infrastructures and dynamic IT deployment. By helping IT management understand and optimize the following critical infrastructure considerations, we can make it easier to align IT strategy with business needs, and reduce the rise of shadow IT initiatives:

  • Value of current facilities renovation/expansion (CAPEX vs. OPEX)
  • New data center build options (CAPEX)
  • Alternative financing options/leaseback (OPEX)
  • Co-location design and optimization
  • Cloud integration
  • Containers/Pods
  • Network/WiFi design and management
  • Migration/relocation options
  • Hybrid computing environment design and deployment

To learn more about how BRUNS-PAK Consulting Services can help you address emerging challenges in your data center strategy, contact Paul Evanko, Vice President at 732-248-4455, or via e-mail at pdevanko@bruns-pak.com.

Request More Information

[1] IDG Research Services, “Clouds, business issues and time management dominate the CIO’s world in 2013”, Nov. 2012 http://www.enterprisecioforum.com/en/whitepaper/it-and-cios-2013-will-look-awful-lot-201

Why CFD for Energy Efficiency?

While winter temperatures make it a little easier to distract yourself from the costs of data center cooling, the realities are that for many companies, data center cooling remains a topic of high importance. At BRUNS-PAK, we have long championed design options that can make significant difference in your data center HVAC costs, including:

  • Airside Economization: the use of “free” outside air in your cooling plan
  • Heat Wheel Integration: integration of heat wheel exchange systems for optimizing energy efficiency
  • Higher Data Center Ambient Temperature: following the guidelines in ASHRAE 9.9 means real savings
  • Hot Aisle/Cold Aisle Configuration: reducing hot/cold mixing can produce measurable improvements in cooling efficiency

However, one item that companies do not take regular advantage of is CFD Modeling. Computational Flow Dynamics is often used in data center design projects, but its use in understanding airflow and cooling efficiency in existing data centers can yield measurable improvements in the optimized configuration of your data center assets, along with recommendations for HVAC improvements.

As leaders in the use of CFD modeling, BRUNS-PAK can provide expert consultation on ways to leverage this technique to support both short-term energy efficiency optimization modifications, and long-term strategic options for improving your data center sustainability profile.

For more information, contact Paul Evanko, Vice President at 732-248-4455, or via e-mail at pdevanko@bruns-pak.com.

Request More Information

Alternative Financing Strategies for Data Center Expansion

The rising reliance on real-time, data-informed decision making in the enterprise, is placing new demands on the CIO to increase capacity and quality of service to knowledge workers throughout the enterprise. The CIO challenge in many organizations, however, is how to deliver that increased capability, capacity and quality of service while dealing with rising pressure to cut costs or forego major capital expenditures.

Traditionally, this has meant a strategic decision between:

  • Renovation of existing data center facilities (dominantly OPEX)
  • Expansion of existing data center facilities (balance of OPEX and CAPEX)
  • Building new data center facilities (CAPEX program)

BRUNS-PAK Data Center Design/Build Leaseback Programs offer a secure way to finance new data center capacity through allocation of operating dollars instead of capital dollars. Backed by one of the nation’s leading financial services institutions, BRUNS-PAK leaseback options are integrated with the BRUNS-PAK design/build methodology which offers both the data center owner and the financing organization, a clear, well-documented, fixed price plan for data center construction projects. Financing options are available for both large-scale and moderate-scale programs.

If you are looking to evaluate OPEX options for an upcoming data center project, contact Paul Evanko, Vice President at 732-248-4455, or via e-mail at pdevanko@bruns-pak.com.

Request More Information

Data Growth In Always On World Driving Data Center Demand

At the Techonomy Conference, Google’s then-CEO Eric Schmidt noted that every two days, people create as much information as civilization did from the dawn of time until 2003 — an estimated five exabytes every two day. The explosion of corporate, academic, and governmental data, combined with explosive growth in user-generated content (Twitter, Facebook, YouTube, etc.), is changing the applications landscape, as companies and customers begin to understand the potent mix that online data, high-performance computing, and high-bandwidth networking create. We are already an Always On society, and the future only becomes even more On.

As semiconductor miniaturization enters the deep submicron era, the density of electrical/mechanical loads continues to increase in server farms. Virtualization has helped to increase processor utilization leading to even greater demands on cooling services. And the unique demands of always on computing create new challenges in mirroring, distributed content distribution networks, and cost-efficient peak load servicing.

At BRUNS-PAK, we recognize that the evolution of the data center, and with it our design-build methodologies, is a never-ending process. As the information technology industry continues to evolve and strive ultrareliable performance complemented by increasingly efficient operation and scalable deployment, only BRUNS-PAK has committed the resources to ensure that our technical team will remain ready to design and deliver “state-of-the- art” facilities even as the “art” changes.

High Impact Measures to Boost Data Center Efficiency (Part 1)

With Data Center energy consumption at an all time high, maintaining the lowest possible total cost of ownership has become increasingly difficult. We’ve detailed some high impact measures to help improve efficiency, and reduce power and cooling requirements to create a greener, more cost effective Data Center.

The first step in energy-efficiency planning is measuring current energy usage. The power system is a critical element in the facilities infrastructure, and knowing where that energy is used and by which specific equipment is essential when creating, expanding, or optimizing a Data Center.

In order to understand how energy efficiency measures affect energy consumption in the Data Center, a baseline needs to be established for the current energy used. There are currently two primary metrics being used by a number of organizations such as the Green Grid to promote the notion of measuring Data Center energy efficiency. The first is Power Usage Effectiveness (PUE) which is defined as the total facility power consumed divided by the IT equipment power consumption. The second metric is PUE’s reciprocal known as Data Center Infrastructure Efficiency (DCiE) which is defined as the IT equipment power consumed divided by the total facility power consumption.

Total facility power is defined as power measured from the utility meter or switch gear solely dedicated to the operation of the Data Center infrastructure in the building if the building is a shared facility with other functions. This includes power consumed by electrical equipment such as switchgear, UPSs (uninterruptible power system) and batteries, PDUs (power distribution units), and stand-by generators. Mechanical equipment dedicated to the HVAC needs of the Data Center such as CRACs (computer room air conditioning units), chillers, DX (direct expansion) air handler units, drycoolers, pumps, and cooling towers. IT equipment power includes the loads associated with IT processes including server, storage, network, tape and other processing equipment fed through Data Center infrastructure support equipment such as PDUs, RPPs (remote power panels), or other distribution means fed from a UPS.

To collect the information noted above, an effective building management system (BMS) should be employed to help collect, categorize, and trend the data gathered. Most systems offered by BMS providers such as Johnson Controls, Andover, Automated Logic, Honeywell, Siemens, and others can allow monitoring of energy consumption for both the IT equipment and the associated infrastructure equipment serving the Data Center. Metering and other DCPs (data collection points) should be provided at all switchgear relating to power and mechanical needs of the Data Center. Also metering should be provided at the output side of the UPS modules or better yet the PDUs. This will provide the energy consumption rates of both the facility power and IT equipment power.

The types of electrical monitoring which can be employed to measure this type of information can be broken down into three basic forms:

  • Amperage-only monitoring
  • Estimated Wattage monitoring
  • True RMS Wattage monitoring

Amperage-only and Estimated Wattage monitoring means can be flawed in the information they provide due to the inaccuracies of measuring the sine wave and its form. Should a sine wave be produced inaccurately, as many double conversion UPS systems do, averaging means of formulating power consumption can prove to be flawed. True RMS Wattage monitoring provides a much more accurate means of understanding the idiosyncrasies of power consumption relating to data processing power sources. BMS systems which employ measures such as wave form capture sampling using real time updating provide a very high degree of accuracy. It should be pointed out that this type of monitoring can be expensive at implementation based on the number of locations it is determined to be used. Should the decision be made to measure power at the distribution level of PDUs and CRAC units to determine power consumption, the cost at this level can be higher than if monitoring was to be placed at the distribution panel boards feeding these types of devices. As long as all the IT equipment and associated infrastructure equipment is being fed from a singular (or dual) location, this monitoring may be far less expensive while still providing nearly the same information for the distributed systems required out in the Data Center.

Traditional Data Centers which are not currently enacting any type of energy efficiency measures are operating with an average PUE of over 3.  A Data Center which is actively pursuing energy efficient measures can achieve much lower ratings, and in return realize substantial energy savings.

High Impact Measures to Boost Data Center Efficiency (Part 2)

While typical energy audits focus on the mechanical and electrical infrastructure, in data centers the facility framework is only one factor in the cost equation. Often times improving in other areas can be even more rewarding. For example, consideration of the actual kilowatts consumed by servers and other IT hardware is crucial when examining energy efficiency in a data center.

Data processing equipment accounts for most of the energy consumption in a data center, and because of this, facility executives really need to start by thinking ‘inside the box’. Best practices in the type, usage, and configuration of deployment all can significantly reduce the overall energy needs for this equipment.

Pull the Plug on Idle Servers

It’s a simple concept really, if it’s not doing anything, unplug it.  However, in many data centers, up to 15% of the servers should be decommissioned and yet are left running for no other reason than lack of drive to clean up outdated equipment.  Some estimates indicate that the cost of each idle server can exceed $1,000 annually when considering total data center energy usage.  That’s a lot of wasted capital!  Addressing the issue can have an immediate impact on the bottom line.

The solution is to establish a rigorous program to decommission obsolete hardware.

Maintaining an asset management database is a necessity to help enterprises ensure that they are consuming resources efficiently. This database should contain accurate, up-to-date information on server location and configuration, enabling IT staff to easily identify variables of power, cooling, and available rack space when planning future server and storage deployments and identifying potential systems to retire.

Upgrade to Energy Efficient Servers

Another simple measure to reducing energy consumption is to buy more energy-efficient servers. The bulk of IT departments ignore energy efficiency ratings when selecting new hardware, focusing on performance and up-front costs rather than total cost of ownership. However, if just one server uses 50 watts less than another it will equate to a savings of more than $250 over a three year period, and an even more profound savings of $1,500+ on facility infrastructure expenditures can be realized.

Data processing equipment all rely on power supplies to take incoming power to the device and distribute it accordingly throughout its internal components as required. These power supplies are typically specified by the manufacturer to provide for the worse case conditions of the device under a maximized configuration. In the past, these power supplies typically were rated far beyond the components capabilities to provide a “safety factor” in the device. As more pressure is being brought to the forefront on energy efficiency in computing, manufacturers have been striving to match their power supplies more closely to the components capabilities, or power parity.

One of the more power consuming components in most IT processing equipment are the fans required to provide air for proper cooling internally within the equipment. These fans run continuously as long as the device is running. Both equipment and chip manufacturers have been making strides to better pair fan use with actual equipment needs. As chip development continues, heat tolerance is being increased.  Also, fans are being created which can be run in stages depending on the processing load of the equipment. This means that fans can run at lower speeds when processing is at a lower state, thus consuming less power.

Processing equipment developed within the last 3-5 years (depending on the manufacturer) is likely to be relatively energy efficient, anything older than that certainly should be evaluated.

Consolidate and Virtualize

Another “low-hanging fruit” in many data centers is server consolidation and virtualization. Typical utilization rates for non-virtualized servers is measured between 5 and 10 percent of their total physical capacity, wasting hardware, space, and electricity. By moving to virtualized servers, data centers will be fully supported with less hardware, resulting in lower equipment costs, lower electrical consumption (thanks to reduced server power and cooling), and less physical space required to house the server farm.

It is important to remember that not all applications and servers are good candidates for virtualization which adds complexity to the endeavor.

Along with consolidated server applications, associated storage for these systems is also becoming more consolidated as well. Storage Area Networks (SAN) and Network Attached Storage (NAS) solutions are becoming the norm in data center typologies. Virtualized tape systems are also replacing larger tape storage devices of the past. As these systems become more regularized, they have also been increasing in density. This allows more storage in the same footprint with only marginal increases in energy consumption. The advent of solid state storage devices (SSD’s) will likely only create higher densities with lower overall power consumption in the future. Although these devices are not yet in production on central storage equipment, it will only be a matter of time before they are utilized.

The Bottom Line

A comprehensive efficiency strategy that targets IT processing equipment in addition to other tactics can substantially reduce energy consumption and net large savings. A facility-wide energy audit from an experienced partner will help to identify the areas where the most immediate impact can be achieved.

High Impact Measures to Boost Data Center Efficiency (Part 3)

Energy efficiency in electrical systems can be achieved through some measures to limit losses through devices among these components. Power parity (the amount of power put into a device equaling the amount of power provided to the device) provides for the most efficient use of power. Transformers and equipment which utilize transformers (such as UPS systems and PDU’s) tend to have some losses in efficiency due to the friction losses in the windings of these transformers. As equipment vendors apply more stringent manufacturing techniques to their products, improvements can be made to efficiencies of this type of equipment. UPS vendors now provide UPS systems which operate at a .95 or higher power factor. This means that there is only a 5% loss of power into the device compared to power supplied by the device. It should be noted that these power factors are generally based on a load limit on the device no lower than around 30% of the rated maximum for the device, although some of the newer UPS systems can maintain their power factor down to as low as 20% of the rated maximum. As equipment is replaced due to changes in a system, end of life, or equipment failure, higher efficiency equipment should be specified and provided to improve on energy efficiency for these systems.

Measurement and Recording Data

We mentioned in part 1 of this series that in order to understand the consumption of power related to the data center, metering of these systems needs to be provided. Further, trending of this information is invaluable to understanding a baseline of energy use as well as the outcome of changes implemented to improve efficiency. The Power Usage Effectiveness (PUE) of the systems is an indicator of how efficient the data center operates. It is very important to understand where your data center ranks for PUE in order to know what measures should be taken to improve efficiency. This means that recording power usage at the main switchgear supporting both the electrical and mechanical equipment supplying the data center, and at the distribution side of the UPS systems distribution (preferably at the 120/208 volt level at the PDU’s) is ideal to achieve the simplest means of calculating the PUE.

Lighting

Lighting systems have been moving towards more energy efficient components in recent years.  These systems have moved away from the use of incandescent and T12 luminaires to compact fluorescent and LED fixtures. ENERGY STAR has reported savings of 42% by switching from T12 fluorescent luminaires with magnetic ballasts to high efficiency T8 luminaires with electronic ballasts. It should be noted that oftentimes these higher efficiency luminaires actually produce higher lighting levels in addition to using less power. The more recent introduction of LED lighting luminaires, which can be retrofit into current fluorescent fixtures, is driving these efficiencies even higher.

Lighting Controls

Another energy savings measure which can be implemented in the data center is lighting controls. The notion of “lights out” data center operations refers to personnel not being normally stationed in the data center space. As operational controls of data processing applications become more network driven, and remotely accessed, less time is required in the data center to perform these activities. As a result of this reduced time spent in the data center, lighting becomes less necessary to operate under non-manned periods. Lighting controls utilizing occupancy sensors as a means of controlling lighting offers a reasonable solution to taking control of shutting off the lights out of the personnel entering and using the space. However, occupancy sensors do not allow for continued presence in the space when personnel are out of sensory contact with a motion or occupancy sensor due to working within or at the lower portions of equipment racks. In order to better accommodate these specialized circumstances in the data center, a combination of occupancy/motion sensors in conjunction with card access systems allows for a highly effective and efficient lighting controls strategy.

The Bottom Line

Once the proper metering components are in place and baselines are established, it’s relatively simple to determine which electrical infrastructure equipment will benefit from an upgrade and what the payback for the investment will be. Also, paying attention to lighting controls can improve energy efficiency in the data center.  No matter what the situation is in your data center, a facility-wide energy audit from an experienced partner will help to identify the areas where the most immediate impact can be achieved.

High Impact Measures to Boost Data Center Efficiency (Part 4)

Mechanical cooling, depending on the efficiencies of the systems being used, can consume as high as 50% of the total power used in a data center. Good engineering practice, equipment efficiencies, and solid operational understanding can all benefit in a lower cost of ownership and operations.

Mechanical Economization or “Free Cooling”

The advent of “green” data center practices has ushered in a heightened interest in reducing mechanical systems energy use. As part of these efforts and in conjunction with data center design “best practices”, a means of mechanical economization or “free cooling” has become a design standard rather than a luxury.

Mechanical economization utilizes the ambient temperature of the local climate to provide an alternative means of heat rejection from standard mechanical systems. Two means of creating this ambient usage are through waterside or airside systems:

Waterside economization utilizes a liquid medium which is run through an outdoor series of coils to be cooled to a lower temperature.  If the ambient cooling meets the necessary set point required for the supply water temperature, the chiller barrel never needs to run, thus greatly reducing the power required for the chiller. During periods where the ambient temperature are not at levels to provide 100% economization, partial “free cooling” can still be provided reducing the overall power needs of the chiller, while still providing some mechanical cooling to reduce the return water to a proper supply temperature.

Airside economization utilizes an air exchange through either a cross stream configuration (which mixes return air with outside air passing through a filter media to help create the supply air stream) or a heat wheel (also known as an enthalpy wheel which nearly eliminates outside air mixing, but typically requires a much larger footprint). This system essentially eliminates the water system medium requirements. These systems can outperform waterside economization in colder climates.

Mechanical Systems Controls and Monitoring

With mechanical systems improving their efficiencies through the equipment improvements and added systems designs noted above, controls and monitoring of these systems becomes more critical in order to maintain these efficient operations. CRAC unit manufacturers have added better controls for these units that allow for a more systematic approach for data center HVAC concerns. Units now communicate with one another throughout the data center and share individual operating conditions to assure a more singular response to the general room conditions.

Monitoring of these systems and trending data also benefit operations and maintenance personnel associated with the data center to better understand the effects of things like economization, maintenance, and other conditions which may affect the data center mechanical systems.

Motors and Drives

Because of reliability requirements in the data center, oftentimes mechanical systems are running at 50% or less their rated capacities during normal operation. This allows for failover scenarios to provide design loads even when a component in the system is not in operation. To further hamper efficient operation, most data center loads are typically below their maximum design capacities in order to plan for growth in the space.

In order to help alleviate the power consumption on equipment running at lower loads and help this equipment maintain better efficiency (as well as better life expectancy), the use of Variable Frequency Drives (VFDs) has provided a simple solution to allow better performance at lower loading while also reducing power consumption. A VFD is an electrical controlling device for motors varying the frequency to consume less power at
lower speeds when loads are not at their rated capacities. A motor can consume as low as 25% of the power required at 60% of the speed compared to 100% loading and speed. Additional benefits include reduced wear at start up and reduced over all motor wear by running it at lower rates than at its single speed maximum.

VFDs can be used in chillers, pumps, and cooling towers of a central system. VFDs can also be used on the air handler systems in the data center as well. VFDs can further provide more recordable information about power consumption for mechanical equipment. The recent improvements in the design technology of the new Variable Frequency Drives (VFDs) over the past few years have been substantial.  The operation of HVAC equipment, especially the pumps and CRAC Unit Fans, at reduced speed can produce cost saving of almost 20 percent.

The New Normal in Data Center Infrastructure Strategy

IT/Line of Business IT SpendingCloud computing is a top-of-mind initiative for organizations in all industries. The promise of scalable, on-demand infrastructure, consumption-based pricing that reduces capex demands, and faster time-to-market for new solutions constitutes an intoxicating potion for requirements-challenged, cash-strapped IT executives.

However, for many IT executives, the migration to the cloud is not a simple decision for one big reason security. When you own and manage your own infrastructure or employ traditional colo or managed hosting services, there are established policies, practices and risk mitigation strategies that are widely accepted. In the murky waters of the cloud, entirely new risks emerge, including:

  • Less transparency on infrastructure security practices, especially in below-the-hypervisor assets
  • New multi-tenancy considerations that are not as well documented or understood
  • Greater delegation of governance, risk and compliance demands to the cloud services provider

Despite these considerations, the financial lure of the cloud is inescapable. Public cloud services providers (CSPs) like Amazon and Microsoft have created massive economies of scale and are increasingly focused on segmented private cloud services that set a new normal in terms of cost-effectiveness, scalability and the ability to deliver truly agile IT infrastructure.

This has forced many IT departments to begin to look at workload segmentation in a new light. Beyond the questions of transactional vs. archival or batch vs. real-time workloads, organizations now need to look at applications that are “cloud adaptable”, both in terms of performance/technical readiness and in terms of governance, risk and compliance. New, business-driven applications like social CRM, human capital management, collaborative procurement and predictive analytics are all strong candidates for migration to on-demand cloud architecture.

This leads to another ‘new normal’ in IT infrastructure hybrid architectures. Hybrid IT infrastructure bridges public and private clouds, managed services providers and on-premise data centers. This composite fabric needs to be secured and managed for optimized performance, compliance and risk, opening up entirely new challenges and ushering in whole new classes of automation and management toolkits, such as internal cloud services brokers. It also forces greater emphasis on internal plans for virtualization or on-premise cloud deployments that can be integrated seamlessly in these complex architectures.

Making sense of this trend and its associated technologies can be confusing. BRUNS-PAK Consulting Services is a growing part of BRUNS-PAK’s comprehensive data center services offerings. Our consulting services team is expert at helping customers to plan and implement complex strategies for alternative infrastructures and dynamic IT deployment. By helping IT management understand and optimize the following critical infrastructure considerations, we can make it easier to align IT strategy with business needs, and reduce the rise of shadow IT initiatives:

  • Value of current facilities renovation/expansion (CAPEX vs. OPEX)
  • New data center build options (CAPEX)
  • Alternative financing options/leaseback (OPEX)
  • Co-location design and optimization
  • Cloud integration
  • Containers/Pods
  • Network/WiFi design and management
  • Migration/relocation options
  • Hybrid computing environment design and deployment

GE and EMC Pivotal: Three Things Every CIO Can Learn From Them.

Recently, General Electric announced a $105 Million investment in EMC Pivotal. The investment reflects the companies growing commitment to smart systems/devices under their industrial Internet initiative. From locomotives to turbines to household appliances, GE sees a world where the ‘internet of things’ delivers measurable value to users of these increasingly intelligent systems.

They are not alone in their strategy. Apple ex-pats Tony Fadell and Matt Rogers took their knowledge of design engineering and online connectivity to create Nest, which sells smart building thermostats. Nest is more than a programmable thermostat, however. This web-connected device learns from a homeowner’s behavioral patterns and creates a temperature-setting schedule from them. It is also a data-use giant…compiling data on its users to drive smarter energy utilization. More important, it shows how entrepreneurs are beginning to embrace technology to do to other common devices what Apple has done to our portable music devices (iPod) and phones (iPhone)—namely make them stylish, fun and easy to use.

So, at GE, drawing on the trend, the company is rethinking how turbines can talk to their owners to drive smarter operation…or more reliable operation. How locomotives can talk to controllers to ensure timely services and ensure maintenance schedules are maintained. And for IT teams at GE this means tons of diverse data streams, structure, unstructured and semi-structured that need storage and interpretation. If this is your business, as GE increasingly deems it is, then the investment in EMC Pivotal makes sense.

But what can we all learn from GE? Here are three important takeaways from the GE investment for CIOs in all business, academic and government segments:

Data Volume Will Grow.

In conversation with IT executives, we still see a tendency to talk about data in traditional terms. That is, we think of applications in our traditional departments (HR, sales, finance, manufacturing, etc) as being our data sources. However, overlooking the explosion in data volumes likely to come from marketing, social media and from customer devices like the Nest thermostats could leave IT teams scrambling for resource when the tsunami from these sources hit.

CIOs Must Drive Business Value…Not Just IT.

GE is slowly and methodically betting its business on data and they are not alone. The key takeaway is the rapid shift from CIO as owner of IT services to broker of services supporting business value. This shift requires CIOs to rethink their facilities and infrastructure strategy in order to ensure, nimble, scalable, secure on-demand, affordable resources for the business.

Data Center Facilities Are Not What They Used To Be.

The Microsoft Azure cloud facility in Quincy, WA includes three distinct architectural approaches to data center design, from traditional raised floor integrated facility to a novel, open air modular form factor that redefines what it means to be a data center. This one facility single-handedly demonstrates the complex decisions facing IT executives looking to plot data center facility strategy for the next decade. Building out data center resources to support consumer-grade data processing (i.e. Google or Amazon class price/performance), you need to consider groundbreaking concepts.

The BRUNS-PAK Data Center Methodology

Over the past 44 years, BRUNS-PAK has quietly assembled one of the most diverse, skilled teams of professionals focused on the strategies and implementation tactics required to craft durable data center strategies in this new era. From strategic planning to design/build support, construction and commissioning, BRUNS-PAK is helping clients craft solutions that balance the myriad decisions underpinning effective data center strategy, including:

  • Renovation vs. expansion options (CAPEX v. OPEX)
  • Build and own
  • Build and leaseback
  • Migration/relocation options
  • Co-Location
  • Cloud integration / Private cloud build out
  • Container/Pod deployment
  • Network optimization
  • Business impact analysis
  • Hybrid computing architecture

With over 6,000 customers in all industry, government and academic sectors, BRUNS-PAK has a proven process for designing, constructing, commissioning and managing data center facilities, including LEED-certified, high efficiency facilities in use by some of the world’s leading companies and institutions.

Guided by the Green Grid

As business demands increase, so too does the number of data center facilities and the amount of IT equipment they house. With escalating demand for data center operations and rising energy costs, it is essential for data center owners and operators to monitor, assess and continually improve performance using energy efficiency and environmental impact metrics. “Overall, global data center traffic is estimated to grow threefold from 2012 to 2017 and although data centers are becoming more efficient, their total energy use is projected to grow,” said Deva Bodas, principal engineer and lead architect for Server Power Management at Intel Corporation and board member for The Green Grid. Government and industry regulators are now adding increased pressure for energy-efficient computing in order to reduce the carbon footprint while data center managers fear they may reach a point of resource limitations.

The ever-present issue is that with such a diverse range of efficiency assessment approaches, many organizations are unclear of what exactly their efficiency assessments should entail. The Green Grid Association, a global consortium, provides a forum where IT, facilities and other C-level executives come together to discuss different options for implementing standardized data center measurement systems. Through data collection and analysis, assessment of emerging technologies and exploration of top data center operation practices, industry-leading metrics are collaboratively devised by end users, policy makers, technology providers, facility architects and utility companies. Many data center efficiency metrics established by the Green Grid task force are now industry-standard, including Power Usage Effectiveness (PUE™), Data Center Infrastructure Efficiency (DCiE™), Carbon Usage Effectiveness (CUE™), Water Usage Effectiveness (WUE™) and Data Center Productivity (DCP). These globally adapted metrics are employed by BRUNS-PAK as a dependable way to measure specific data center results against comparable organizations, improve existing data center efficiencies and make intelligent decisions in new data center deployments.

BRUNS-PAK fully maximizes the metrics, technical resources and educational tools that the Green Greed provides to accurately assess various key elements of data center efficiency. Standardized life cycle assessments such as the Green Grid’s Data Center Maturity Model (DCMM) and Data Center Life Cycle Analysis are essential resources used by BRUNS-PAK in conversations with data center owners to give the knowledge required in deciding whether to rebuild or renovate, predict expected returns and identify areas of IT operations that require improvement. In addition to the standard PUE metric, BRUNS-PAK also leverages Green Grid’s DCeP (Data Center Energy Productivity), a new equation that quantifies useful work that a data center produces based on the amount of energy it consumes and allows each organization to define “useful work” as it relates to their unique business. Additionally, the EDE (Electronic Disposal Efficiency) metric is implemented by BRUNS-PAK to help data center operators evaluate how their outdated electronic equipment is managed and disposed. It the combination of these recognized metrics that guide the design of all of BRUNS-PAK facilities.

Following the five core tenants of the Green Grid Design Guide, a new architectural approach to how data centers are built and modernized that focuses on energy efficiency, BRUNS-PAK takes a holistic approach to data center design that leverages such efficiency metrics from start to finish.

The Green Grid Design Guide is described as “a guide for the standardization and evolution of key capabilities” and is based on five core tenants that BRUNS-PAK factors into every data center design:

Fully Scalable:

ƒAll systems/subsystems scale energy consumption and performance to use the minimal energy required to accomplish workload.

Fully Instrumented:

ƒAll systems/subsystems within the datacenter are instrumented and provide real time operating power and performance data through standardized management interfaces.

Fully Announced:

ƒAll systems/subsystems are discoverable and report minimum and maximum energy used, performance level capabilities and location.

Enhanced Management Infrastructure:

ƒCompute, network, storage, power, cooling and facilities utilize standardized management/interoperability interfaces and language.

Policy Driven:

Operations are automated at all levels via policies set through management infrastructure.

Standardized Metrics/Measurements:

ƒEnergy efficiency is monitored at all levels within the datacenter from individual subsystems to complete datacenter and is reported using standardized metrics during operation.

The 2014 Data Center Transformation

Welcome to BRUNS-PAK’s Mission Critical Data Center Blog. We hope to provide you with interesting and timely information to help you navigate the ever changing Data Center in your facility. With that said, lets shed some light on one of our clients most basic and pressing questions:

“WHAT IS CAUSING THE 2014 DATA CENTER TRANSFORMATION?”

  1. Worldwide economic downturn from 2008 – 2014 to conserve capex spending – do more for less.
  2. The availability of the emerging co-location / cloud / container / disaster recovery strategies.
  3. The maturation of the total cost of ownership models (15 elements) revealing a “hybrid” solution (in many cases) that optimize return on investment while balancing risk (warning: risk models evolving 2014+).
  4. Improve the “quarterly bottom line” at all costs.
  5. In the private sectors, focus on the short term stock appreciation, is in many cases, priority one.