I received a press release from a vendor the other day which was riddled with the words “always-on” and “zero downtime” to describe what its offerings bring to IT environments. These terms are being used more frequently by service providers and IT vendors, but is zero downtime even possible?
Circuit board with clock picture from Shutterstock
The concept of improving uptime has become a popular topic as disaster recovery is declining in popularity. Disaster recovery is a source of consternation for IT managers, according to Gartner research director for technology and services, Michael Warrilow.
“It’s a very capital intensive and frustrating activity,” he told Lifehacker Australia. “You’re spending a lot of money on the off-chance that something goes wrong with your IT and even when it does go wrong, its might not even work properly. So the conversation is shifting from preparing for disaster to moving towards high levels of availability.”
This explains the surge in the use of “zero downtime” in the IT space. You’d think that zero downtime would mean that an IT service or infrastructure will never suffer an outage, effectively providing 100 per cent uptime. Not the case, according to IDC senior market analyst, Prabhitha Sheethal Dcruz.
“Zero downtime frequently translates to 99.999 per cent uptime, which equates to 5.26 minutes of downtime per year,” she told us. “While short outages may be acceptable for non-criticali workloads, the same is not true for business critical and mission critical workloads where the downtime stakes can be very high – consider a stock exchange where a single lost transaction may incur a significant financial cost or a medical system downtime that can cost lives.”
IDC has done global research into the financial implication of IT system downtime and found:
- For Fortune 1000 companies, the average total cost of unplanned application downtime per year is US$1.25 billion — US$2.5 billion.
- The average hourly cost of an infrastructure failure is US$100,000 per hour.
- The average hourly cost of a critical application failure is US$500,000 — US$1 million.
Downtime can be costly, but if you’re an organisation that uses third-party cloud services it’s difficult to expect 100 per cent uptime from your providers.
“I guess it’s technically feasible, but financially impossible,” Warrilow said. The term “zero downtime” in the service provider space is essentially marketing speak, he said, and the best you’re going to get from them is quote on a service level agreement (SLA) offering a certain level of uptime.
“They usually offer 99.5 per cent or higher uptime and it’s a matter of horses for courses,” he said. “Cloud is most often used for general purpose workloads rather than mission-critical traffic, not for important things like key financial systems. You can also add additional layers of resilience to get downtime close to zero per cent.”
For organisations seeking a that zero downtime nirvana, Warrilow and Dcruz have a few tips:
- Figure out if you actually need zero downtime
A lot of times you don’t need it. Considering the higher the uptime you seek, the higher the cost incurred on IT so going with a lower uptime offering can save a lot of money.
- Make service providers and IT vendors stress test their zero downtime claims
Get them to come in and demonstrate their offerings in your environment and consider negotiating penalties into your contract if they don’t live up to their claims.
- Plan well
Invariably, things will break. Impact analysis should forms part of an organisation’s greater business continuity and disaster recovery (BCDR) plan.
Does your company strive for 100 per cent uptime with its IT systems? Let us know in the comments.