# Design a system that scales to millions of users on AWS

## Step 1: Outline use cases and constraints

> Gather requirements and scope the problem. Ask questions to clarify use cases and constraints. Discuss assumptions.

### Use cases

Solving this problem takes an iterative approach of: 1) Benchmark/**Load Test**; 2) Profile for **bottlenecks**; 3) address bottlenecks while **evaluating alternatives** and **trade-offs**; 4) repeat, which is good pattern for **evolving** basic designs to scalable designs.

Unless you have a background in AWS or are applying for a position that requires AWS knowledge, AWS-specific details are not a requirement. However, much of the principles discussed in this exercise can apply more generally outside of the AWS ecosystem.

We'll scope the problem to handle only the following use cases

* Phase 1: User **makes a read or write request**
  * Service does processing, stores user data, then returns the results
* Phase 2: Service needs to evolve from serving a small amount of users to **millions of users**
  * Discuss general scaling patterns as we evolve an architecture to handle a large number of users and requests
* Phase 3: Service has **high availability**

### State assumptions

* Traffic is not evenly distributed
* Need for relational data
* Scale from 1 user to tens of millions of users
  * Denote increase of users as:
    * Users+
    * Users++
    * Users+++
    * ...
  * 10 million users
  * 1 billion writes per month
  * 100 billion reads per month
  * 100:1 read to write ratio
  * 1 KB content per write

### Calculate usage

> Clarify with your interviewer if you should run back-of-the-envelope usage calculations.

#### base: 1 billion writes per month

#### Storage

* new content per month：

$$1 kb \* 1 billion = 1 TB$$

* store 3 years in total

$$12 \* 3 \* 1000 GB = 36 TB$$

* Assume most writes are from new content instead of updates to existing ones

#### Traffic

* write tps:&#x20;

$$1 billion / (30 \* 24 \* 3600) \~= 400 TPS$$

* read tps:

$$400 \* 100 = 40000 TPS$$

## Step 2: Create a high level design

> 永远先把最重要的 component 放上来，然后一点一点去 scale

![](/files/-M92PfybnvfIKe2i5QU8)

## Step 3: Design core components

> Dive into details for each core component.

### User makes a read or write request

### Goal

> 从最简单的开始: single box -> vertical scaling

* With only 1-2 users, you only need a basic setup
  * **Single box** for simplicity
  * **Vertical scaling** when needed
  * Monitor to determine **bottlenecks**

### Start with a single box

* **Web server** on EC2
  * Storage for user data
  * MySQL Database (RDBMS)

### Vertical Scaling

> 这是最初的 setup，如果搭一个 1, 2 个人能用的系统

* Simply choose a **bigger box**
* Keep an eye on metrics to determine how to scale up
  * Use basic monitoring to determine bottlenecks: **CPU, memory, IO, network, etc**
  * **CloudWatch, top, nagios, statsd, graphite, etc**
* Scaling vertically can get very expensive
* No redundancy/failover
* The alternative to Vertical Scaling is Horizontal scaling

### Start with SQL, consider NoSQL

The constraints assume there is a need for relational data. We can start off using a MySQL Database on the single box.

### Assign a public static IP

> 让 customer 可以来访问这个终端

* Elastic IPs provide a public endpoint whose IP doesn't change on reboot
* Helps with **failover**, just point the domain to a new IP

### Use a DNS

Add a DNS such as Route 53 to **map the domain** to the instance's public IP.

### Secure the web server

Open up only necessary ports:

* Allow the web server to respond to incoming requests from:
  * 80 for HTTP
  * 443 for HTTPS
  * 22 for SSH to only whitelisted IPs
* Prevent the web server from initiating outbound connections

> 至此，第一阶段就做完了。涉及了 web server，database，vertical scaling，DNS，security

## Step 4 - Scale up: Users+

> Identify and address bottlenecks, given the constraints.

![](/files/-M97o6hVErFeYav5Gj-2)

> 所有的 scale up 都是因为有了 bottleneck，需要解决它

### Assumptions

Our user count is starting to pick up and the load is increasing on our single box. Our **Benchmarks/Load Tests** and **Profiling** are pointing to the **MySQL Database** taking up more and more memory and CPU resources, while the user content is filling up disk space.

We've been able to address these issues with Vertical Scaling so far. Unfortunately, this has become quite expensive and it doesn't allow for independent scaling of the MySQL Database and Web Server.

> 第一个 bottleneck：MySQL DB。解决方法：把 DB 从 WebServer 里分离出来，把 object store 也分离出来

### Goals

* **Lighten load** on the single box and allow for independent scaling
  * Store static content separately in an Object Store
  * Move the MySQL Database to a separate box
* Disadvantages
  * These changes would **increase complexity** and would require changes to the Web Server to point to the Object Store and the MySQL Database
  * **Additional security measures** must be taken to secure the new components
  * AWS **costs** could also increase, but should be weighed with the costs of managing similar systems on your own

### Store static content separately

* Consider using a managed Object Store like S3 to store static content
  * Highly scalable and reliable
  * Server side encryption
* Move static content to S3
  * User files
  * JS
  * CSS
  * Images
  * Videos

### Move the MySQL database to a separate box

* Consider using a service like **RDS** to manage the MySQL Database
  * Simple to administer, scale
  * Multiple availability zones
  * Encryption at rest

### Secure the system

* Encrypt data in transit and at rest
* Use a **Virtual Private Cloud (VPC)**
  * Create a public subnet for the single Web Server so it can send and receive traffic from the internet
  * Create a private subnet for everything else, preventing outside access
  * Only open ports from whitelisted IPs for each component
* These same patterns should be implemented for new components in the remainder of the exercise

## Step 5 - Scale up: Users++

![](/files/-M9GIV6EfWnj_kU-y8Lw)

### Assumptions

Our Benchmarks/Load Tests and Profiling show that our **single Web Server bottlenecks** during peak hours, resulting in slow responses and in some cases, downtime. As the service matures, we'd also like to move towards higher **availability** and **redundancy**.

> 上一个阶段，分离出来了 DB。这个阶段，我们要 scale up web server，提高 availablility 和 reliability（redundancy）

### Goals

* The following goals attempt to address the scaling issues with the Web Server
  * Based on the Benchmarks/Load Tests and Profiling, you might only need to implement one or two of these techniques
* Use **Horizontal Scaling** to handle increasing loads and to address single points of failure
  * Add a **Load Balancer** such as Amazon's ELB or HAProxy
    * ELB is highly available
    * If you are configuring your own Load Balancer, setting up multiple servers in **active-active** or **active-passive** in multiple availability zones will improve availability
    * **Terminate SSL** on the Load Balancer to reduce computational load on backend servers and to simplify certificate administration
  * Use **multiple Web Servers** spread out over multiple availability zones
  * Use **multiple MySQL** instances in [Master-Slave Failover](https://github.com/donnemartin/system-design-primer#master-slave-replication) mode across multiple availability zones to **improve redundancy**

    > Herizontal scaling: 加 hosts -> load balancer; master-slave DB
* Separate out the Web Servers from the [Application Servers](https://github.com/donnemartin/system-design-primer#application-layer)
  * Scale and configure **both layers independently**
  * Web Servers can run as a [Reverse Proxy](https://github.com/donnemartin/system-design-primer#reverse-proxy-web-server)
  * For example, you can add Application Servers handling **Read APIs** while others handle **Write APIs**

    > 把 application layer 分离出来；把 read API 和 write API 分离开；
* Move static (and some dynamic) content to a Content Delivery Network (**CDN**) such as CloudFront to reduce load and latency

## Step 6 - Scale up: Users+++

![](/files/-M9GRMTG62LNI9kr4z1O)

### Assumptions

Our Benchmarks/Load Tests and Profiling show that we are **read-heavy** (**100:1** with writes) and our database is suffering from poor performance from the high read requests.

> 我们是一个 read-heavy 的 system。在这个阶段，我们需要加 cache 来提高 performance 了。

### Goals

* The following goals attempt to address the scaling issues with the **MySQL Database**
  * Based on the Benchmarks/Load Tests and Profiling, you might only need to implement one or two of these techniques
* Move the following data to a [Memory Cache](https://github.com/donnemartin/system-design-primer#cache) such as Elasticache to reduce load and latency
  * Frequently accessed content from MySQL
    * First, try to configure the **MySQL Database cache** to see if that is sufficient to relieve the bottleneck before implementing a **Memory Cache**
  * **Session data** from the Web Servers
    * The Web Servers become **stateless**, allowing for Autoscaling
  * Reading **1 MB** sequentially from memory takes about 250 microseconds, while reading from **SSD takes 4x** and from **disk takes 80x longer**.
* Add [MySQL Read Replicas](https://github.com/donnemartin/system-design-primer#master-slave-replication) to reduce load on the write master
* Add more **Web Servers** and **Application Servers** to improve responsiveness

### Add MySQL read replicas

* In addition to adding and scaling a Memory Cache, MySQL Read Replicas can also help relieve load on the MySQL Write Master
* Add logic to Web Server to **separate out writes and reads**
* Add **Load Balancers** in front of MySQL Read Replicas (not pictured to reduce clutter)
* Most services are **read-heavy** vs write-heavy

## Step 7 - Scale up: Users++++

![](/files/-M9GRMTHSVHCx1JMpZf3)

### Assumptions

Our Benchmarks/Load Tests and Profiling show that our traffic spikes during regular business hours in the U.S. and drop significantly when users leave the office. We think we can cut costs by automatically spinning up and down servers based on actual load. We're a small shop so we'd like to automate as much of the DevOps as possible for **Autoscaling** and for the general operations.

> 这个阶段，我们 works on autoscaling

### Goals

* Add **Autoscaling** to provision capacity as needed
  * Keep up with traffic spikes
  * Reduce costs by powering down unused instances
* Automate DevOps
  * Chef, Puppet, Ansible, etc
* Continue monitoring metrics to address bottlenecks
  * **Host level** - Review a single EC2 instance
  * **Aggregate level** - Review load balancer stats
  * **Log analysis** - CloudWatch, CloudTrail, Loggly, Splunk, Sumo
  * **External site performance** - Pingdom or New Relic
  * **Handle notifications and incidents** - PagerDuty
  * **Error Reporting** - Sentry

### Add autoscaling

* Consider a managed service such as **AWS Autoscaling**
  * Create **one group for each Web Server** and one for each Application Server type, place each group in **multiple availability zones**
  * Set a min and max number of instances
  * Trigger to scale up and down through CloudWatch
    * Simple time of day metric for predictable loads or
    * Metrics over a time period:
      * CPU load
      * Latency
      * Network traffic
      * Custom metric
  * Disadvantages
    * Autoscaling can introduce **complexity**
    * It could **take some time** before a system appropriately scales up to meet increased demand, or to scale down when demand drops

## Step 8 - Scale up: Users+++++

![](/files/-M9GldNKrehB68aPpiSc)

**Note**: Autoscaling groups not shown to reduce clutter

### Assumptions

As the service continues to grow towards the figures outlined in the constraints, we iteratively run Benchmarks/Load Tests and Profiling to uncover and address new bottlenecks.

> 在这个阶段，我们要进行 SQL 的 scaling

### Goals

We'll continue to address scaling issues due to the problem's constraints:

* If our **MySQL** Database starts to **grow too large**, we might consider only storing a limited time period of data in the database, while storing the rest in a **data warehouse such as Redshift**
  * A data warehouse such as Redshift can comfortably handle the constraint of 1 TB of new content per month
* With **40,000** average read requests per second, read traffic for popular content can be addressed by scaling the **Memory Cache**, which is also useful for handling the unevenly distributed traffic and traffic spikes
  * The **SQL Read Replicas** might have trouble handling the **cache misses**, we'll probably need to employ **additional SQL scaling patterns**
* **400** average writes per second (with presumably significantly higher peaks) might be **tough for a single SQL Write Master-Slave**, also pointing to a need for additional scaling techniques

### SQL scaling patterns include:

* Federation
* Sharding
* Denormalization
* SQL Tuning

### Consider NoSQL

* To further address the high read and write requests, we should also consider moving appropriate data to a **NoSQL Database** such as **DynamoDB**.

### Asynchronous

We can further separate out our Application Servers to allow for independent scaling. Batch processes or computations that do not need to be done in real-time can be done **Asynchronously with Queues and Workers**:

* For example, in a photo service, the photo upload and the thumbnail creation can be separated:
  * Client uploads photo&#x20;
  * Application Server **puts a job in a Queue** such as **SQS**
  * The Worker Service on EC2 or Lambda pulls work off the Queue then:
    * Creates a thumbnail
    * Updates a **Database**
    * Stores the thumbnail in the **Object Store**

## Additional talking points

### SQL scaling patterns

* Read replicas
* Federation
* Sharding
* Denormalization
* SQL Tuning

### NoSQL

* Key-value store
* Document store
* Wide column store
* Graph database
* SQL vs NoSQL

### Caching

* Where to cache
  * Client caching
  * CDN caching
  * Web server caching
  * Database caching
  * Application caching
* What to cache
  * Caching at the database query level
  * Caching at the object level
* When to update the cache
  * Cache-aside
  * Write-through
  * Write-behind (write-back)
  * Refresh ahead

### Asynchronism and microservices

* Message queues
* Task queues
* Back pressure
* Microservices

### Communications

* Discuss tradeoffs:
  * External communication with clients - [HTTP APIs following REST](https://github.com/donnemartin/system-design-primer#representational-state-transfer-rest)
  * Internal communications - [RPC](https://github.com/donnemartin/system-design-primer#remote-procedure-call-rpc)
* [Service discovery](https://github.com/donnemartin/system-design-primer#service-discovery)

### [Security](https://github.com/donnemartin/system-design-primer#security)

### Latency numbers

* See [Latency numbers](https://github.com/donnemartin/system-design-primer#latency-numbers-every-programmer-should-know) every programmer should know.


---

# Agent Instructions: Querying This Documentation

If you need additional information that is not directly available in this page, you can query the documentation dynamically by asking a question.

Perform an HTTP GET request on the current page URL with the `ask` query parameter:

```
GET https://liuyang89116.gitbook.io/system-design/chapter-2/scaling_aws.md?ask=<question>
```

The question should be specific, self-contained, and written in natural language.
The response will contain a direct answer to the question and relevant excerpts and sources from the documentation.

Use this mechanism when the answer is not explicitly present in the current page, you need clarification or additional context, or you want to retrieve related documentation sections.
