Post on 27-Jun-2015
description
© 2014 Amazon.com, Inc. and its affiliates. All rights reserved. May not be copied, modified, or distributed in whole or in part without the express consent of Amazon.com, Inc.
More Nines for Your Dimes : Improving Availability and Lowering Costs
using Auto Scaling
Jinesh Varia, Technology Evangelist (@jinman)
Derek Pai, Principal Product Manager (@derekpai)Date: 3/26/2014
Topics We’ll Cover Today• Auto Scaling introduction
• Console demo
• Maintaining application response times and fleet utilization
• Handling cyclical demand, unexpected “weather events”
• Auto Scaling for 99.9% Uptime
• Single-instance groups
• Cost control and asymmetric scaling responses
• Cloudformation, custom scripts, and multiple inputs
• Using performance testing to choose scaling strategies
• Dealing with bouncy or steep curves
AWS
The Weather Channel
Nokia
Adobe
Dreambox
Ways You Can Use Auto Scaling
Launch EC2 instances
and groups from
reusable templates
Scale up and down as
needed automatically
Auto-replace
Instances and
maintain EC2 capacity
Common Scenarios
• Schedule a one-time scale out and flip to production
• Follow daily, weekly, or monthly cycles
• Provision capacity dynamically by scaling on CPU, memory,
request rate, queue depth, users, etc.
• Auto-tag instances with cost center, project, version, stage
• Auto-replace instances that fail ELB or EC2 checks
• Auto-balance instances across multiple zones.
Prepare for a Big Launch
Fit Capacity to Demand
Be Ready for Spikes
Simplify Cost Allocation
Maintain Stable Capacity
Go Multi-AZ
Demo
Learn the new terms:
Launch Configuration
Auto Scaling Group
Scaling Policy
Amazon CloudWatch Alarm
Amazon SNS Notification
Instance Lifecycle
What’s New in Auto Scaling
Better integration
• EC2 console support
• Scheduled scaling policies in
CloudFormation templates
• ELB connection draining
• Auto-assign public IPs in VPC
• Spot + Auto Scaling
More APIs
• Create groups based on running
instances
• Create launch configurations based
on running instances
• Attach running instances to a group
• Describe account limits for groups
and launch configs
Why Auto Scaling?
Scale Up Control CostsImprove Availability
Why Auto Scaling?
Scale Up Control CostsImprove Availability
The Weather Company
• Top 30 web property in the U.S.• 2nd most viewed television
channel in the U.S.• 85% of U.S. airlines depend on
our forecasts• Major retailers base marketing
spend and store displays based on our forecasts
• 163 million unique visitors across TV and web
Wunderground Radar and
Maps
100 million hits a day
One Billion data points per day
Migrated real-time radar mapping system wunderground.com to
AWS Cloud
30,000Personal
Weather
StationsSource: Wunderground, Inc. 2013
Why Auto Scaling?
Why Auto Scaling?
Why Auto Scaling?
Why Auto Scaling?
Why Auto Scaling?
Hurricane Sandy
Before Migration – Traditional IT Model doesn’t scale well
Server Count(110 Servers)
Avg. CPU Load HTTP Response Latency(~6000 ms)
HTTP Response Latency(5-15ms)
Server Count(from 110 to 170 Instances)
Avg. CPU Load
After Migration - Wunderground Radar App
Radar on AWS Auto Scaling Architecture
Radar on AWS
CPU Utilization
Radar on AWS
Host Count
Radar on AWS
Radar on AWS
Radar on AWS
Scale up to ensure consistent
performance during high-demand
Why Auto Scaling?
Scale Up Control CostsImprove Availability
Auto Scaling
for 99.9%
Uptime
Here.com Local Search Application
• Local Search app
• First customer facing
application on AWS
• Obvious need for
Uptime
Here.com Local Search Architecture
US-East-1
US-West-2
EU-West-1
US-East-1a
Zookeeper1
Zookeeper2
Zookeeper3
Frontend
Group
Backend
Groups
US-East-1b
Zookeeper1
Zookeeper2
Zookeeper3
Frontend Group
Backend Groups
AP-Southeast-1
Here.com Local Search Architecture
US-East-1
US-West-2
EU-West-1
US-East-1a
Zookeeper1
Zookeeper2
Zookeeper3
Frontend
Group
Backend
Groups
US-East-1b
Zookeeper1
Zookeeper2
Zookeeper3
Frontend Group
Backend Groups
AP-Southeast-1
Single-Instance Auto Scaling
Groups (Zookeeper)
1. Auto-healing: Instances auto-register in
DNS via Route53
2. Dynamic: Auto Scaling Group Names
are used for cluster-node lookups
(cluster1-zookeeper1)
3. Used Standard Tools such as DNS
instead of Queries or Elastic IPs
Auto Scaling when upgrading
without any downtime
Map Data on S3
US-East-1a
Zookeeper1
cluster1
old old
Map Data on S3
US-East-1a
Zookeeper1
cluster1
old old
New Data
V2
Map Data on S3
US-East-1a
Zookeeper1
cluster1
old old
New Data
V2
Map Data on S3
US-East-1a
Zookeeper1
cluster1
old old
New Data
V2
Map Data on S3
US-East-1a
Zookeeper1
cluster1
old old
New Data
V2
Map Data on S3
US-East-1a
Zookeeper1
cluster1
old old
New Data
V2
New
v2New
V2
Map Data on S3
US-East-1a
Zookeeper1
cluster1
old old
New Data
V2
New
v2New
V2
“Auto scaling”
Web Server Fleet(Amazon EC2)
Database Fleet(RDS or DB on EC2)
Load Balancing(ELB)
v1.1 v1.1
v1.1 v1.1
v1.2
v1.2
v1.2
v1.2
Auto scalingMax instances
Min instances
Scaling Trigger
Custom Metrics
Upper Threshold
Lower Threshold
Increment by
Common scenario: Blue Green Deployments
Using Auto Scaling
Here.com Local Search Success
• Increased Uptime 99.9%
• All detected health problems have been successfully replaced by Auto Scaling with zero intervention.
• Zookeeper setup has performed flawlessly
“We’ve been
paranoid so it still
pages us; It’s
beginning to feel
silly.”
Why Auto Scaling?
Scale Up Control CostsImprove Availability
Adobe Creative
Cloud Runs on
AWS
Adobe Shared
Cloud Architecture
on AWS
Auto Scaling the Web Layer
Based on Number of HTTP requestsAverage CPU loadNetwork in/out
Auto Scaling the Web Layer
Auto Scaling the Worker Layer
Based on SQS queue length
Based on Number of HTTP requestsAverage CPU loadNetwork in/out
Scale up fast, scale down slow
Cost Control
• Scheduled scaling: we analyzed our traffic and
picked numbers. – scale up in the morning, scale down in the evening
• Policies for slow scale down
• Stage environments: downscale everything to
“min-size” daily (or more)
CloudFormation + Auto Scaling"ScaleUpPolicy" : {
"Type" : "AWS::Auto Scaling::ScalingPolicy", "Properties" : {
"AdjustmentType" : "ChangeInCapacity", "Auto ScalingGroupName" : { "Ref" : "WorkerAuto ScalingGroup" }, "Cooldown" : {"Ref": "cooldown"}, "ScalingAdjustment" : { "Ref" : "adjustup" }
} },"WorkerAlarmScaleUp": {
"Type": "AWS::CloudWatch::Alarm", "Properties": {
"EvaluationPeriods":{"Ref" : "evalperiod"}, "Statistic": "Sum", "Threshold": {"Ref" : "upthreshold"}, "AlarmDescription": "Scale up if the work load of transcode queue is high", "Period": {"Ref" : "period"}, "AlarmActions": [ { "Ref": "ScaleUpPolicy" }, { "Ref" : "scalingSNStopic" } ], "Namespace": "AWS/SQS", "Dimensions": [ { "Name": "QueueName", "Value": {"Ref" : "queuename" }}], "ComparisonOperator": "GreaterThanThreshold", "MetricName": "ApproximateNumberOfMessagesVisible"
How – Custom Metrics. . .
Sat Oct 6 05:51:03 UTC 2012
Number of AZs: 4
Number of Web Servers: 16
Number of Healthy Web Servers: 16
ELB Request Count: 9523.0
Request Count Per Healthy Web Server: 595.1875
Network In Per Healthy Web Server: 51 MB
Network Out Per Healthy Web Server: 1 MB
CPU Per Healthy Web Server: 25.23875
Publishing Custom Metrics: InstanceRequestCount, HealthyWebServers, InstanceNetworkIn, InstanceNetworkOut, InstanceCPUUtilization to namespace WebServer in us-east-1
. . .
How – multi-input scaling
Scale up
Scale down
+2 instances if more than 50 visible messages for >5 min
+50% instances if more than 1000 msg for >2 min
+ fixed 100 instances if more than 10000 msg for >1 min
-10 instance if 0 msg for more than 10 min
-25% if 0 msg for more than 30 min
Adobe’s Advice
• Use CloudFormation!
• Know your system, thresholds
• Watch your scaling history
• Scaling up is easy, scaling down not so much
• Mantra: scale up fast; scale down slow
Brandon Adams,
Dreambox Learning
Scaling strategies we use
Scaling with
CloudWatch alarmsScheduled scaling (onetime, recurring)
A little background on our application
• Ruby on Rails
• Unicorn
• We teach kids math!
A workload well suited for auto scaling
Scaling with CloudWatch alarms
Performance test to get a baseline• Discover the ideal number of
worker processes per server– Too few and resources go
unused
– Too many and performance suffers under load
• Obtain the maximum load sustainable per server– Our performance tests
measures number of concurrent users
• Find the chokepoint– For us, this was CPU utilization
Performance testing
Identify the breaking point
Breaking point was at about 400 users per server
Our first method to find scale points• Provision a static amount
of servers that we know can handle peak load
• Adjust scale up and scale down alarms based on observed highs and lows
• This worked, but was super inefficient, both in time and money spent
Let’s do some math – identify variables
Independent
• Concurrent users
Dependent
• CPU utilization
• Memory utilization
• Disk I/O
• Network I/O
Let’s do some math – find the slope• Adding about 1600 users per hour
• Which is about 27 per minute
• We know that we can handle a max of
about 400 users per server at 80% CPU
usage
• Which is about 0.2% CPU usage per user
Let’s do some math – when to scale?
• We know (from other testing) that it takes
us about 5 minutes for a new node to
come online
• We’re adding 27 users per minute
• Which means we need to start spinning
up new nodes when we’re about 135
users ( 27 x 5 ) per node short of max
• Which is at about 53% utilization:
(80% - (0.2% * 135))
Simple Math - Scaling point equations𝑠𝑐𝑎𝑙𝑖𝑛𝑔 𝑝𝑜𝑖𝑛𝑡 = 𝑚𝑎𝑥𝑖𝑚𝑢𝑚 𝑢𝑠𝑒𝑟𝑠 𝑝𝑒𝑟 𝑛𝑜𝑑𝑒 − (𝑢𝑠𝑒𝑟𝑠 𝑎𝑑𝑑𝑒𝑑 𝑝𝑒𝑟 𝑚𝑖𝑛𝑢𝑡𝑒 ∗ 𝑠𝑝𝑖𝑛 𝑢𝑝 𝑡𝑖𝑚𝑒)
𝑠𝑐𝑎𝑙𝑖𝑛𝑔 𝑝𝑜𝑖𝑛𝑡 = 400 − (27 ∗ 5)
𝑠𝑐𝑎𝑙𝑖𝑛𝑔 𝑝𝑜𝑖𝑛𝑡 = 265 users per node
𝑠𝑐𝑎𝑙𝑖𝑛𝑔 𝑝𝑜𝑖𝑛𝑡 = 𝑢𝑠𝑒𝑟𝑠 𝑝𝑒𝑟 𝑛𝑜𝑑𝑒 ∗ 𝑐𝑝𝑢 𝑝𝑒𝑟 𝑢𝑠𝑒𝑟𝑠𝑐𝑎𝑙𝑖𝑛𝑔 𝑝𝑜𝑖𝑛𝑡 = 265 ∗ .2𝑠𝑐𝑎𝑙𝑖𝑛𝑔 𝑝𝑜𝑖𝑛𝑡 = 53% cpu per node
𝑐𝑝𝑢
𝑛𝑜𝑑𝑒=𝑢𝑠𝑒𝑟𝑠
𝑛𝑜𝑑𝑒∗𝑐𝑝𝑢
𝑢𝑠𝑒𝑟
How much to scale up by?
• The lowest we can scale up by is 1 node per AZ,
otherwise we would be unbalanced
• For us, this is an extra 800 users of capacity in
five minutes, plenty enough to keep up with our
rate of adding 1600 users per hour
• Adding 800 users of capacity every five minutes,
we could support 9600 additional users per hour
Evaluate your predictions• In the real world, we’ve inched up from
scaling at 53%
• Our perf test is a little harsher than the
real world
• Numbers derived from the perf test are
only as accurate as the simulation of
traffic you in your perf test.
Scheduled scaling
Acceleration in load is not constantRequest count for a 24 hour period
We can’t use one size fits all• Scale too aggressively
– Overprovisioning: increases
cost
– Bounciness: we add more
than we need and have to
partially scale back shortly
after scaling up, which
increases cost
• Scale too timidly
– Poor performance
– Outages due to lack of
capacity
Bounciness and steepness
• Add scheduled scaling points to eliminate bounciness
• Scheduled scale for the steepest points of your demand curve
• Let dynamic scaling take care of the less steep parts of the curve
Scaling curve before…
minmin
min
min
min
…and after
min
min
min
min
min
Putting it all together
The opportunity cost of NOT scaling
• Our usage curve
from 3/20
• Low of about 5
concurrent users
• High of about
10,000 concurrent
users
The opportunity cost of NOT scaling
• No autoscaling
• 672 instance hours
• $302.40 at on-
demand prices
The opportunity cost of NOT scaling
• Autoscaling four
times per day
• 360 instance hours
• $162 at on-
demand prices
• 46% savings vs no
autoscaling
The opportunity cost of NOT scaling
• Autoscaling as
needed, twelve
times per day
• 272 instance hours
• $122.40 at on-
demand prices
• 24% savings vs
scaling 4 times per
day
• 60% savings vs no
autoscaling
The opportunity cost of NOT scaling
$302/day
$162/day
$122/day
Demand curve hugs the usage curve…
…and a (mostly) flat response curve
Auto Scaling saves us a lot of money;
With a little bit of math, flexibility of AWS
allows us to further save by aligning our
demand curve with usage curve.
Why Auto Scaling?
Scale Up Control CostsImprove Availability
Key Takeaways
• Maintaining application response times and fleet utilization
• Scaling up and handling unexpected “weather events”
• Auto Scaling for 99.9% Uptime
• Single-instance groups
• Cost control and asymmetric scaling responses
• Cloudformation, custom scripts, and multiple inputs
• Using performance testing to choose scaling strategies
• Dealing with bouncy or steep curves
The Weather Channel
Nokia
Adobe
Dreambox
Thank You!
Jinesh Variajvaria@amazon.com
@jinman
Derek Paiderekp@amazon.com
@derekpai