Design Pastebin.com (or Bit.ly)
Design Pastebin.com (or Bit.ly)
Step 1: Outline use cases and constraints
Gather requirements and scope the problem. Ask questions to clarify use cases and constraints. Discuss assumptions.
Use cases
User enters a block of text and gets a randomly generated link
Expiration
Default setting does not expire
Can optionally set a timed expiration
User enters a paste's url and views the contents
User is anonymous
Service tracks analytics of pages
Monthly visit stats
Service deletes expired pastes
Service has high availability
Out of scope
User registers for an account
User verifies email
User logs into a registered account
User edits the document
User can set visibility
User can set the shortlink
Constraints and assumptions
State assumptions
Traffic is not evenly distributed
Following a short link should be fast
Pastes are text only
Page view analytics do not need to be realtime
10 million users
10 million paste writes per month
100 million paste reads per month
10:1 read to write ratio
Calculate usage
Size per paste
1 KB content per paste (1 KB = 1000 bytes)
shortlink
- 7 bytesexpiration_length_in_minutes
- 4 bytescreated_at
- 5 bytespaste_path
- 255 bytestotal = ~1.27 KB
12.7 GB of new paste content per month
1.27 KB per paste * 10 million pastes per month
~450 GB of new paste content in 3 years
360 million
shortlink
s in 3 yearsAssume most are new pastes instead of updates to existing ones
4 paste writes per second on average
40 read requests per second on average
Handy conversion guide:
2.5 million seconds per month
1 request per second = 2.5 million requests per month
40 requests per second = 100 million requests per month
400 requests per second = 1 billion requests per month
Step 2: Create a high level design
Step 3: Design core components
Use case: User enters a block of text and gets a randomly generated link
We could use a relational database as a large hash table, mapping the generated url to a file server and path containing the paste file.
Instead of managing a file server, we could use a managed Object Store such as Amazon S3 or a NoSQL document store.
An alternative to a relational database acting as a large hash table, we could use a NoSQL key-value store. We should discuss the tradeoffs between choosing SQL or NoSQL. The following discussion uses the relational database approach.
The Client sends a create paste request to the Web Server, running as a reverse proxy
The Web Server forwards the request to the Write API server
The Write API server does the following:
Generates a unique url
Checks if the url is unique by looking at the SQL Database for a duplicate
If the url is not unique, it generates another url
If we supported a custom url, we could use the user-supplied (also check for a duplicate)
Saves to the SQL Database
pastes
tableSaves the paste data to the Object Store
Returns the
url
The pastes
table could have the following structure:
Setting the primary key to be based on the shortlink
column creates an index that the database uses to enforce uniqueness. We'll create an additional index on created_at
to speed up lookups (log-time instead of scanning the entire table) and to keep the data in memory. Reading 1 MB sequentially from memory takes about 250 microseconds, while reading from SSD takes 4x and from disk takes 80x longer.
To generate the unique url, we could:
Take the MD5 hash of the user's ip_address + timestamp
MD5 is a widely used hashing function that produces a 128-bit hash value
MD5 is uniformly distributed
Alternatively, we could also take the MD5 hash of randomly-generated data
Base 62 encode the MD5 hash
Base 62 encodes to
[a-zA-Z0-9]
which works well for urls, eliminating the need for escaping special charactersThere is only one hash result for the original input and Base 62 is deterministic (no randomness involved)
Base 64 is another popular encoding but provides issues for urls because of the additional
+
and/
charactersThe following Base 62 pseudocode runs in
O(k)
time wherek
is the number of digits = 7:
Take the first 7 characters of the output, which results in 62^7 possible values and should be sufficient to handle our constraint of 360 million shortlinks in 3 years:
We'll use a public REST API:
Response:
For internal communications, we could use Remote Procedure Calls.
Use case: User enters a paste's url and views the contents
The Client sends a get paste request to the Web Server
The Web Server forwards the request to the Read API server
The Read API server does the following:
Checks the SQL Database for the generated url
If the url is in the SQL Database, fetch the paste contents from the Object Store
Else, return an error message for the user
REST API:
Response:
Use case: Service tracks analytics of pages
Since realtime analytics are not a requirement, we could simply MapReduce the Web Server logs to generate hit counts.
Clarify with your interviewer how much code you are expected to write.
Use case: Service deletes expired pastes
To delete expired pastes, we could just scan the SQL Database for all entries whose expiration timestamp are older than the current timestamp. All expired entries would then be deleted (or marked as expired) from the table.
Step 4: Scale the design
Important: Do not simply jump right into the final design from the initial design!
State you would do this iteratively: 1. Benchmark/Load Test; 1. Profile for bottlenecks; 1. address bottlenecks while evaluating alternatives and trade-offs; 1. repeat.
See Design a system that scales to millions of users on AWS as a sample on how to iteratively scale the initial design.
It's important to discuss what bottlenecks you might encounter with the initial design and how you might address each of them. For example, what issues are addressed by adding a Load Balancer with multiple Web Servers? CDN? Master-Slave Replicas? What are the alternatives and Trade-Offs for each?
We'll introduce some components to complete the design and to address scalability issues. Internal load balancers are not shown to reduce clutter.
To avoid repeating discussions, refer to the following system design topics for main talking points, tradeoffs, and alternatives:
The Analytics Database could use a data warehousing solution such as Amazon Redshift or Google BigQuery.
An Object Store such as Amazon S3 can comfortably handle the constraint of 12.7 GB of new content per month.
To address the 40 average read requests per second (higher at peak), traffic for popular content should be handled by the Memory Cache instead of the database. The Memory Cache is also useful for handling the unevenly distributed traffic and traffic spikes. The SQL Read Replicas should be able to handle the cache misses, as long as the replicas are not bogged down with replicating writes.
4 average paste writes per second (with higher at peak) should be do-able for a single SQL Write Master-Slave. Otherwise, we'll need to employ additional SQL scaling patterns:
We should also consider moving some data to a NoSQL Database.
Additional talking points
Additional topics to dive into, depending on the problem scope and time remaining.
NoSQL
Caching
When to update the cache
Asynchronism and microservices
Communications
Discuss tradeoffs:
External communication with clients - HTTP APIs following REST
Internal communications - RPC
Security
Refer to the security section.
Latency numbers
See Latency numbers every programmer should know.
Ongoing
Continue benchmarking and monitoring your system to address bottlenecks as they come up
Scaling is an iterative process
Last updated