Is It Time to Adjust the Temperature of Your Data Center’s Chilled Water Temperature?

Ways to Cool a Data CenterData center cooling is a topic that could be discussed endlessly.  What works best for one data center may not work well for another depending on a variety of factors including data center location, size of data center and type of building.  Cooling with water is an eco-friendly and exceptionally effective means of cooling and what many are finding is that chilled water may be even more effective.  It remains the goal of most data centers to effectively cool while also being efficient and eco-friendly.  When using a chilled water system, a water chiller is used to produce chilled water which is then pumped into the CRAH (computer room air handler) and then then circulates around chilled coils and cools the air in the computer room by removing the heat from a room.  It circulates out and then gets chilled again and sent back through the system, making it a very efficient means of cooling a data center.

In the event of an outage, air cooled chillers can actually return to operation more quickly, making redundancy easier to achieve as well.  Additionally, chilled water cooling is easily scalable and adaptable to the ever-changing needs of a data center.  In an effort to improve efficiency, many data centers are more closely examining just how cool the chilled water cooling system needs to be.  If it can be adjusted by even a degree or two, a significant improvement in energy efficiency can be made.   Schneider Electric further examines the advantage of opting to adjust chilled water cooling temperatures in data centers, “In a nutshell, that means many data centers don’t need to be as cool as they used to. Most data centers will find temperatures of 24°-25°C (75°-77°F) will suffer no difference in reliability vs. cooler temperatures… If temperatures inside the data center are higher than in the past, that means the temperature of the chilled water used to cool it – known as the set point for the chillers – can also be higher. As it turns out, that has a profound effect on cooling system efficiency. Raising the chilled water set point from the usual values of 7° to 10°C used in comfort cooling chilled water plants up to 18° to 20°C or higher can result in an operational expense savings of about 40%. That’s because less energy is required to cool the water year-round. In summer, higher evaporating temperatures mean compressors don’t have to work as hard, resulting in improved efficiency. In cooler months, users benefit from many more hours of economizer or “free cooling” operation. A higher set point also results in a capital expense savings of some 30% because chillers don’t have to be as large as at traditional temperatures.”  With a re-examination of what temperature your data center needs to maintain to maximize uptime, data centers may be able to adjust their chilled water cooling temperature to save a significant amount of expense and dramatically improve data center energy efficiency.

Posted in computer room construction, Computer Room Design, data center cooling | Tagged , | Comments Off

Active Security Monitoring

AdobeStock_93793795

Security Risks

In the wake of many high profile data breaches, from government institutions to retailers, there is an evolving environment in the data management world. An environment that requires more active security policy to be established in order to reduce the amount of time that sensitive data is unknowingly exposed to malicious sources. Having strictly preventive security policy although at times effective opens the door to a flood of destructive malware as relaxed policy in regards to monitoring of data movement can allow compromised systems to be unpatched for indeterminate periods of time, unnecessarily exposing data systems.

Active Monitoring Systems

Although preventative maintenance is an essential part of security, actively monitoring data systems can result in quicker detection of penetration by malicious software, these breaches may go unnoticed for long periods of time if only preventative security measures are taken in the data center. It’s a given that systems should should be monitored on a daily basis, but dealing with a large flood of data and knowing how to prioritize it is near impossible for a large data center. Especially in the face of remote access by authorized staff from various locations, of whom may unknowingly bring security risks into the operational environment. With such big data coming in so quickly from a variety of secure and insecure networks the only answer to monitoring such a large scale system of data transfer and accompanying network activity is software based analytics. Big data can be sorted and actively monitored in a meaningful manner through the analytics derived from computational algorithms, algorithms of which can sort malicious activity based on potential risk to reduce false positives or non factor threats that will be blocked by preventative security systems, giving security personnel a more focused view of malicious activity in the network. Any and all detection can be stored and logged for future reference to increase efficiency of automated detection systems. Software based analysis and monitoring of network activity can help identify issues as they stream in, with sorting of priority and potential risk security personnel are able to catch threats immedietly. This reduces liability as security breaches are detected on the fly, reducing exposure of sensitive data and the time of which malicious software has access to said data systems.

Posted in Cloud Computing | Comments Off

Managing Thermal Systems During Winter

Networking communication technology concept, network and internet telecommunication equipment in server room, data center interior with computers and Earth globe in blue lightWinter is soon approaching and with it comes the concern of not just managing the drop in temperature, but also managing the low humidity that comes with it. Ensuring temperature and humidity control systems are set to the industry recommended ranges and are receiving routine maintenance in the coming winter months will help prevent any unexpected impairment from these environmental conditions. Below are some considerations to be made in order to prevent downtime or loss of of data in the winter months.

Temperature Control Systems and Maintenance

Problems arising from improper preparation for these yearly temperature lows can be minor or in some cases catastrophic. Being one step ahead of the incoming chill is an essential preparation to ensure data center up-time. Maintaining thermal management systems is critically important as low temperatures and the accompanying low humidity produces static electricity, which can potentially damaging sensitive electronics systems, creating a situation where indiscriminate data loss is a real possibility. However, maintaining environmental conditions within industry standard guidelines is not not as simple as running systems until the desired effects are achieved. With the large fluctuations in temperature and humidity that come during the winter months air conditioning units and humidity control systems will be under duress attempting to keep the desired stable conditions within the data center. If these systems don’t receive proper servicing they can cause leakage of coolant or water into the data center through failure of internal components, creating a multitude of issues as not only will flooding or leakage have to be dealt with, but also unregulated thermal conditions during their repair. Receiving routine maintenance of these systems to check for faulty hardware is essential as always, but a pre-winter diagnostic for thermal management systems during this time of year can help prevent a potential disaster and ensure operations continue unhindered.

Automated Thermal Management Systems

Keeping the temperature and humidity in a narrow range is not an easy accomplished task. Luckily there are automated control systems that can act as hive-mind for a network of thermal management systems, ensuring the ideal temperature and humidity zones are reached. These systems bring a big picture to temperature and humidity control by ensuring individual systems in the building don’t counteract each other’s purpose and they can give proper warning of inefficiency, such as in the case of one unit futility counteracting another by humidifying while another is dehumidifying. This expansive view into the cooling systems within a data center is a great diagnostics system and helps reduce the workload and potential for human error, in turn reducing costs through better overall efficiency of the system.

Posted in Facility Maintenance | Comments Off

Human Error and Data Center Maintenance

datacenter45

The most underrated force that leads to downtime or inefficiency in the data center is personnel. Even when systems are functioning optimally human error can lead to unexpected consequences due to carelessness or forgetfulness. As systems become more automated and self reliant the human factor is stronger than ever in the reasons why downtime is experienced. There are some considerations that should be made in how to manage human error in the data center and reduce downtime associated with it.

Documentation

Proper documentation in a task mannered step by step checklist is a great way to reduce risks associated with routine tasks. Even the most experienced IT worker may fall out of step if procedures become too visceral, leaving room for mistakes that result in downtime. This is why it’s critical that guidelines are made for all tasks, in order to ensure that there is reference for anyone who needs it, especially in emergency situations or to rectify a mistake. All equipment, should be labeled properly and diagrams drawn up to ensure procedures can be followed without unneeded time to find the referenced items or areas said in documentation. At a bare minimum critical items such as the emergency off and switching devices should be labeled.

Training and Consistent Policy

Training personnel to follow a standard set of practices within the business is essential for those with access to the facility or data systems. All personnel should be familiar with essential equipment in order to avoid an unexpected shutdown as even with proper documentation carelessness or lack of understanding of importance of systems can lead to mistakes. Security should be tight with a sign in policy that requires observation of non-essential personnel not just to protect the equipment, but to ensure nothing is inadvertently damaged. As with any electronics liquids and foods are a huge risk on a daily basis and should be kept away from any rooms with critical equipment, proper signage should be in place and this policy should be enforced thoroughly.

 

Posted in data center maintenance | Comments Off

Managing Data Center Power Costs

Data center with network servers in futuristic room.Providing energy to servers is a substantial part of a data center’s costs. In many cases this is due to servers running consistently at peak performance in preparation for peak capacity. This creates a lot of unnecessary expenditure as these systems are not always needed to run at such high performance levels. This creates inflated minimum power requirements for maintaining critical systems, forcing unnecessary expansion of power management systems and further increased costs.

Dynamic or Scheduled Performance

Protecting functionality while ensuring peak performance is a huge challenge, but one that should not be ignored as there are potentially substantial returns in capital from adjustable performance in these systems. Servers waste a tremendous amount of power by running at peak performance under times of low demand, especially if the suite of applications are technically demanding and require powerful systems. Deactivating servers or resizing clusters on a schedule of known usage or under dynamically controlled systems, of which can detect potential shifts in usage and need for more functionality, can help dramatically reduce power consumption. Reducing idle power consumption is a significant way to cut costs and even the largest business can benefit through dynamic management of performance. These methods reduce servers drain on power while ensuring there is no downside on the user end, even if there is unusually large loads of traffic.

Load balancing with Multiple Data Centers

As a business grows and it’s pool of users expand it may become beneficial to have data centers strategically located, running applications only in areas of which are located in times of off peak hour. Off peak hours provide significantly reduced prices in power due to less demand outside of the typical business hours. Concerns with latency might make this an issue for some businesses, but for the majority of computational tasks running applications with a few hundred milliseconds of latency is not a concern. In certain cases redirecting traffic with needs for low latency to a first tier of high cost data centers and redirecting those with no latency concerns to cheap power areas would be ideal. Integrating functionality across multiple data centers allows capacity and latency to be shifted with user demand determining performance, saving capital in the process. Such varying location can also offer more stability as systems aren’t isolated in their power supplier and servers draw power independently increasing the reliability of these platforms.

Posted in Data Center Infrastructure Management | Comments Off

Optimal Locations for a Data Center

With the increase in remotely performed operations, and the need for less work crew on site due to automated procedures, there are a wide range of locations that can be selected for construction of your data center. Below are some considerations to be made in regards to choosing a site for future construction.

Cheap Primary Power SourceAdobeStock_87909563

The most crucial part of any data center is it’s primary power source. Being connected to a large power grid and the accompanying infrastructure can help reduce costs of kilowatt per hour usage. The downside is that searching for a cheap power source may restrict a multitude of other factors that should be considered when choosing a site for construction, but under certain conditions the potential benefits from cheap power may make up for other potential detriments. The extra working capital from a cheap power source would be of primary concern for data centers which have an unexpected rate of growth and may require substantially more power in the future. 

Low Population Areas

Typically areas out of the urban environments allow for easier construction and expansion due to more relaxed zoning laws. Less population density also allows for easier security monitoring and better isolation of the data center, providing more acute risk management, as less human factor is introduced. These areas may also have lower living expenses and in turn potentially provide cheaper labor. 

Damage To Operations From Climate

It’s ideal to find climates that are consistent when selecting a site for your data center. Being on the lower side of the temperature range without being in the extremes is also a great way to allow for passive cooling, which affords the opportunity to cut costs further. Avoiding areas prone to devastating acts of nature is a crucial concern as these events can devastate and destroy not just potentially your data center, but also the surrounding infrastructure. Therefore Excessively rainy climates prone to flooding should be avoided,  and considerations of other environmental hazards unique to some regions should be made. In most cases arid cool environments in low population areas with access to proper roadways are the preferred choice for a data center.

 

Posted in Data Center Construction | Comments Off

Industrial Machines Are Coming To The Cloud

AdobeStock_93793795

According to Gartner, Inc 20.8 billion objects will have connectivity by 2020. The Internet of Things has been highly spoken of as the force behind this predicted growth in devices and data, but silently behind it the Industrial Internet of Things has churned along waiting to revolutionize the world of industry through the cloud.

The Industrial Internet of Things

As various sectors of business attempt to automate or enhance productivity through embedded electronics the integration of physical devices into cloud based computation and storage environments is becoming more common everyday. The Industrial Internet of Things brings the brains of embedded electronics and their cloud based systems to the largest machines in a way not seen before. An article on the GE blog titled “Industrial Internet Success: Union Pacific Railroad” references a particularly interesting utilization of the industrial internet by Union Pacific, of whom installed systems on their trains to provide real time sensory information, which was in turn coupled with weather data, to process and evaluate train line safety by giving warnings of potential derailments weeks before they happen.

Big Data Reduces Costs

The case examined above doesn’t necessarily show the entire scope of the Industrial Internet of Things, but it’s an indication that quantifiable data from hardware systems in an industrial setting is potentially revolutionary. Analysis of big data provided from industrial systems has huge implications for capital investment. This is due to the fact that such systems will lead to a  reduction in working costs, through identification of inefficiency and elimination of potential risks, through objective forecasting of environmental hazards. Through collaboration of data from industrial machines and cloud based storage and computational environments there is great potential for optimizations of business capabilities through the Industrial Internet of Things, as it’s been shown industrial data can cut liabilities dramatically.

These cloud based systems could prove a pivotal moment for large scale industry as issues in the operating environment that have gone unnoticed for extended periods of time, at great cost, can be identified by such big data analytics. It’s only a matter of time before the Industrial Internet of Things becomes a dominant force in the cloud computing world as businesses operating industrial machines begin seeking efficiency provided by analytics of industrial machines operating in cloud based environments.

 

Posted in Cloud Computing | Comments Off

Protecting Data Center Hardware and Avoiding Unexpected Downtime

Data centerEnsuring hardware is protected from unexpected physical consequences is an essential and often overlooked part of maintaining a data center. The strategies below will reduce the risk of downtime and potential damage to hardware.

Power Management

The Uninterruptible Power Supply (UPS) provides power when utility services fail, keeping essential equipment functional at all times. In order to guarantee uptime and adequate power needs in a prolonged outage backup UPS systems should also be installed on essential systems. Any remote monitoring equipment should also be fitted with a UPS to allow for continued monitoring of data center operations. A Power Distribution Unit (PDU) should also be installed to the main power source to protect against critical loads. Power distribution to all systems should run through a UPS, connected to a PDU. Together the UPS and PDU ensure equipment is protected and active throughout any power fluctuations.

Cooling

Cooling equipment needs are dependent on the particular environment at hand, although equipment may handle a large range of temperatures such fluctuation should be avoided. To ensure components are adequately cooled or heated keep temperatures in a range of high 60’s to low 70’s. Finding and eliminating hotspots in racks with tray fans is a great way to cut down on unnecessary use of air conditioning, but with proper organization and management of equipment this should not be needed. Airflow considerations should be made around all equipment, but particular devices such as the UPS are more prone to potential degradation if exposed to excess heat. If the room is too compact to allow for cool airflow to the UPS it should be kept outside the area.

Organization

Organization and setup is an essential part of protecting critical equipment and avoiding long term problems. If pressed for time and proper organization during setup isn’t possible, then arrangements should be made for a fully configured system. Poorly arranged systems can lead to over cooling situations in which spot fans are unable to maintain ideal temperatures and over usage of air conditioning is needed. Considerations should also be made in regards to the potential for water damage from the environment. Avoiding any room with water pipes or any areas that could potentially flood, such as basements, is a must. If in a rainy climate having an umbrella installed over server racks can protect from any unexpected leaks from the roof.

In short, well prepared power management and cooling systems are the best insurance against any unexpected downtime.

 

Posted in data center maintenance | Comments Off

Why Hyper-Converged Infrastructure Is the Future of IT

Technology, Network Server, Data.

As the Internet of Things (IoT) continues to develop, the wisdom of converged infrastructure is making more and more sense. The basic idea behind it, developed during the early days of device integration, was the idea that bringing together device infrastructure within a single, stable ecosystem would provide the streamlining necessary to handle the data from multiple inputs across the whole system. As an approach, it worked well enough to provide much of the backbone for today’s networked home and business experience.

As businesses move toward more and more reliance on the Internet of Things to manage their daily affairs, though, the influx of data from those devices is becoming overwhelming. That is leading to a new concept, hyper-converged infrastructure, which promises to provide solutions for handling this ever-growing influx of information.

Why It Is Necessary

According to the research firm IDC, the amount of data produced but IoT endpoints, devices capable of communicating information back to a central server via the network, is doubling every 2 years. This means that by 2020 there will be 10 times as much data being processed worldwide as there was in 2013. The demands for data processing at that point will be such that each year, meaning there will be nearly as many points of information collected and communicated as there are stars in the known universe, annually. Nor is that expansion likely to slow, with the number of IoT endpoint devices being projected to grow from 10.3 billion to 29.5 billion in that same time frame. This kind of data glut necessitates a new way of sorting information if data processing is going to keep pace with the demand for it.

Features of Hyper-Converged Infrastructure

Where converged infrastructure brought together the server, storage, and networking components into a streamlined package to be more cost-effective and efficient, hyper-converged infrastructure goes a step further, combining converged infrastructure with a hypervisor and management software to create a system that manages its resources as effectively as possible in every moment. The savings, in terms of time committed, are substantial:

100 hours shorter set-up time on average

Up to 10 hours per month in management time reduced

When this is delivered in a self-contained enclosure to complete the hyper-converged infrastructure, it also eliminates the need for dedicated server rooms. The result is a lean, efficient server setup that is designed to handle a large influx of information, and that can be easily replicated as your data needs grow. That is what makes it the future of information technology in the workplace.

Posted in Data Center Design | Comments Off

Well-Maintained Cooling Equipment Helps Avoid Data Center Downtime

Trouble in data center

According to the Gartner Group, provider of technology related information and statistics, $5,600 a minute is the average cost of data center downtime, which amounts to a whopping $336,000 per hour. Obviously, this is devastating in terms of lost revenue and squandered productivity. In order to avoid downtime, data center customers say that regular service for the center’s cooling system is as important as keeping the UPS (uninterruptible power supply) humming.

Struggling to Keep Up

The cooling equipment has many mechanical parts, which will eventually break down and require replacement. A considerable amount of energy is also used, since the data center must meet the standards required by ASHRAE (the American Society of Heating, Refrigerating and Air-Conditioning Engineers) to operate at a temperature of up to 81 degrees Fahrenheit. Consequently, in addition to being subject to equipment failure, an aging cooling system may not be able to keep up with the data center’s current energy requirements.

Regular Maintenance is Key

Many data centers contract with service providers in order to keep essential equipment in top shape. A cooling equipment professional, for example, can examine your system and identify trouble spots. Parts that will need eventual replacement include loose wires, dirty filters and humidifier bottles. Those that are approaching the end of their useful life can be changed out before they have a chance to break down.

Making the Most of Your Cooling System

Regular cooling system maintenance will benefit data center performance, which depends on this component’s efficiency. Because cooling equipment is so energy intensive, there have been some significant improvements in system design and efficiency in the past several years so that much less energy is used. Efforts as small as cleaning away leaves and other debris from condenser coils can be helpful, but replacing certain parts will also increase energy efficiency. When this is handled through a regular maintenance agreement, energy savings is often the result.

Choosing a Cooling Service Contractor

Because data center cooling systems are designed with many complex components, the individuals who service the equipment should be factory-trained, certified technicians. These professionals can properly diagnose any problems and resolve them before they have the opportunity to turn into outright disasters. Certified technicians will not only see that your cooling system operates at maximum efficiency, they will ensure that the product warranties remain effective. They can also recommend any changes or improvements that will enable your data center to perform at its best so as to avoid the threat of debilitating downtime.

 

Posted in data center maintenance | Comments Off