Hard Lessons of Public Cloud: Designing for When AWS Goes Down
March 07, 2017

Eric Wright
Turbonomic

Share this

So, your site is down because AWS S3 went away. Too soon? It's never too soon to talk about why the responsibility for designing resilient infrastructure belongs in your camp. It's like when Smokey the Bear used to say that "only you can prevent forest fires." The difference is that it's Jeff Bezos saying it this time.

We have some real insight into what design for cloud resiliency really means thanks to a chat that I had recently.

Cloud Goes Down, so Design for It

There is no special text in the terms and conditions. These are hard facts. AWS designs its infrastructure to be as resilient as possible, but clearly tells you that you should design with the intention of surviving partial service outages. It isn't that AWS plans on being down a lot, but they have been hit by specific DDoS attacks, and also have had to reboot EC2 hosts in order to patch for security vulnerabilities.

At the time I was writing this, AWS S3 was fighting its way back to life in the US-east-1 Region. This means that there were multiple Availability Zones in the throes of recovery, and that potentially hundreds of thousands of web sites, and applications were experiencing issues retrieving objects from the widely used object storage platform.

So, how do we do this better? Let's ask someone who does design and see how the developers think about things. With that, I wanted to share a great discussion that I had with former Disney lead architect and current Principal Software Architect at Turbonomic, Steve Haines.

Q&A: Understanding the Developer's Reaction to the AWS Outage

EW: What does it mean to think about designing across regions inside the public cloud?

SH: Designing an application to run across multiple AWS regions is not a trivial task. While you can deploy stateless services or micro-services to multiple regions and then configure Route53 (Amazon's DNS Service) to point to Elastic Load Balancers (ELBs) in each region, that doesn't completely solve the problem.

First, it's crucial to consider the cost of redundancy. How many regions and how many availability zones (AZ) in each region do we want to deploy to? From historical outages, you're probably safe with two regions, but you do not want to keep a full copy of your application deployed in another region just for disaster recovery: you want to use it and distribute workloads across those regions!

For some use cases this will be easy, but for others you will need to design your application so that it is close to the resources it needs to access. If you design your application with failure in mind and to run in multiple regions then you can manage the cost because both regions will be running your workloads.

EW: That seems to be a bit of the cost of doing business for design and resiliency, but what is the impact below the presentation layers? It feels like that is the sort of "low hanging fruit" as we know it, but there is much more to the application architecture than that, right?

SH: Exactly! That leads to the next challenge: resources, such as databases and files. While AWS provides users multi-A to Z database replication free of charge for databases running behind RDS, users are still paying for storage, IOPS, etc. However, this model changes if a user wants to replicate across regions. For example, Oracle provides a product called GoldenGate for performing cross-region replication, which is a great tool but can significantly impact your IT budget.

Alternatively, you can consider one of Amazon's native offerings, Aurora, which supports cross- region replication out-of-the-box, but that needs to be a design decision you make when you're building or refactoring your application. And, if you store files in S3, be sure that you enable cross- region replication, it will cost you more, but it will ensure that files stored in one region will be available in the event of a regional outage.

EW: Sounds like we have already got some challenges in front of us with just porting our designs to cloud platforms, but when you're already leaning into the cloud as a first-class destination for your apps we have to already think about big outages. We do disaster recovery testing on-premises because that's something we can control. How do we do that type of testing out in the public cloud?

SH: Good question. It's important to remember that while designing an application to run in a cross-region capacity is one thing, having the confidence that it will work when you lose a region is another beast altogether!

This is where I'll defer to Netflix's practice of designing for failure and regularly testing failure scenarios. They have a "Simian Army" (https://github.com/Netflix/SimianArmy) that simulates various failure scenarios in production and ensures that everything continues to work. One of the members of the Simian Army is the Chaos Gorilla that regularly kills a region and ensures that Netflix continues to function, which is one of the reasons they were able to sustain the previous full region outage.

If you're serious about running across regions then you need to regularly validate that it works!

But maybe we should think bigger than cross-region – what if we could design across clouds for the ultimate protection?

EW: Thanks for the background and advice, Steve. Good food for thought for all of us in the IT industry. I'm sure there are a lot of people having this discussion in the coming weeks after the recent outage.

Eric Wright is Principal Solutions Engineer at Turbonomic.

Share this

The Latest

March 27, 2017

Monitoring a business means monitoring an entire business – not just IT or application performance. If businesses truly care about differentiating themselves from the competition, they must approach monitoring holistically. Separate, siloed monitoring systems are quickly becoming a thing of the past ...

March 24, 2017

A growing IT delivery gap is slowing down the majority of the businesses surveyed and directly putting revenue at risk, according to MuleSoft's 2017 Connectivity Benchmark Report on digital transformation initiatives and the business impact of APIs ...

March 23, 2017

Why containers are growing in popularity is no surprise — they’re extremely easy to spin up or down, but come with an unforeseen issue. Without the right foresight, DevOps and IT teams may lose a lot of visibility into these containers resulting in operational blind spots and even more haystacks to find the presumptive performance issue needle ...

March 22, 2017

Much emphasis is placed on servers and storage when discussing Application Performance, mainly because the application lives on a server and uses storage. However, the network has considerable importance, certainly in the case of WANs where there are ways of speeding up the transmission of data of a network ...

March 21, 2017

The majority of IT executives believe investment in IT Service Management (ITSM) is important to gain the agility needed to compete in an era of global, cross-industry disruption and digital transformation, according to Delivering Value to Today’s Digital Enterprise: The State of IT Service Management 2017, a report by BMC, conducted in association with Forbes ...

March 17, 2017

Let’s say your company has examined all the potential pros and cons, and moved your critical business applications to the cloud. The advertised benefits of the cloud seem like they’ll work out great. And in many ways, life is easier for you now. But as often happens when things seem too good to be true, reality has a way of kicking in to reveal just exactly how many things can go wrong with your cloud setup – things that can directly impact your business ...

March 16, 2017

IT leadership is more driven to be innovative than ever, but also more in need of justifying costs and showing value than ever. Combining the two is no mean feat, especially when individual technologies are put forward as the single tantalizing answer ...

March 15, 2017

The move to Citrix 7.X is in full swing. This has improved the centralizing of Management and reduction of costs, but End User Experience is becoming top of the business objectives list. However, delivering that is not something to be considered after the upgrade ...

March 14, 2017

As organizations understand the findings of the Cyber Monday Web Performance Index and look to improve their site performance for the next Cyber Monday shopping day, I wanted to offer a few recommendations to help any organization improve in 2017 ...

March 13, 2017

Online retailers stand to make a lot of money on Cyber Monday as long as their infrastructure can keep up with customers. If your company's site goes offline or substantially slows down, you're going to lose sales. And even top ecommerce sites experience performance or stability issues at peak loads, like Cyber Monday, according to Apica's Cyber Monday Web Performance Index ...