Back to top

Design Dating Platform - Tinder

Disclaimer: This is our take on designing a location-based social search application that allows users to use a swiping motion to like or dislike other users similar to Tinder. We have created this design based on our research going through the Tinder engineering blog. These articles are quite informative and detailed. We have provided the list of the tech talks and blogs in our reference section.

Quick Video Glance

Problem Statement

Design a location-based social search application similar to Tinder, which is often used as a dating service. It allows users to use a swiping motion to like (swipe right) or dislike (swipe left) other users and allows users to chat if both parties like each other(a “match”).

Gathering Requirements

In Scope

The application should be able to support the following requirements.

  • User should be able to create their Tinder profile by adding their bio and uploading photos.

  • User should be able to view recommendations of other users in geographically nearby regions.

  • Users should be able to like(swipe right) or dislike(swipe left) other recommended users.

  • Users should get notifications when matched with other users.

  • Users should be able to move to a different location and still get recommendations of nearby users.

Out of Scope

  • Sending and receiving messages from other users. We have covered it in our article on designing WhatsApp.

High Level Design


Fig 0: Tinder Architecture

There will be a fleet of micro-services behind the Gateway, which will be serving the user requests. The Profile Creator Service will be invoked when the user profile gets created. This service will store the user information in a database and add the user to the corresponding geo-sharded index so that the user shows up in recommendations of nearby users. This index gets queried by the Recommendation Service when it receives the request to generate recommendations for other users. Once the user starts swiping through those recommendations, the Swipes Service receives those swipes and places them in a data-streams(e.g., AWS Kinesis/ SQS). There is a fleet of workers which read data from those streams for generating matches. The workers do this by querying the LikesCache to determine if it’s a match, in which case the match notification is sent to both the users using technologies such as WebSockets.

Component Design

User Profile Creation

Fig 1: Sequence Diagram for User Profile Creation

The sequence diagram above shows the sequence of operations which gets executed when a user creates a profile on Tinder. Within the synchronous process, the user media (e.g., photos) is uploaded on a file server, and the user information, including the user’s location, is persisted in a key-value store like Amazon DynamoDB. Additionally, this user is added to a queue for adding the user to a geo-sharded index.

The asynchronous process reads the user information from the queue and passes this information to the GeoShardingIndexer. The indexer uses geo libraries like Google’s S2 library to map the user’s location to a geo-shard and add the user to the index associated with that shard. This helps the user to show up in the recommendations of other nearby users. For instance, in the image below, we have shown how a user from North America gets mapped to the corresponding index so that the user gets shown in recommendations of nearby users.

Fig 2: User from North America will be mapped to the corresponding shard(Image: Tinder Engineering Blog)

UserProfileInfo – Sample Data Model

We have shown below a json blob for storing the user profile information. We can use a key-value store such as Amazon DynamoDB or Riak for maintaining this data.

  "userId": "AWDGT567RTH",
  "name": "Julie",
  "age": 25,
  "gender": "F",
  "location": {
    "latitude": 123123,
    "longitude": 123123
  "media": {
    "images": [
  "recommendationPreferences": {
    "ageRange": {
      "min": 21,
      "max": 31
    "radius": 50

Fetch User Recommendations

Fig 3: Sequence Diagram for Fetching User Recommendations

In the previous section, we saw how users get added to the geo-sharded index. Let’s see how the user gets shown in the recommendation of other users. When a user request comes to the Recommendation Engine, it forwards the request to the GeoShardedIndexer. The indexer determines the geo-shards to be queried based on user location and radius using geo-libraries like Google’s S2. After that, the indexer queries all the geo-sharded indexes(more details in the next section) mapped to the shards returned by Google S2 to fetch the list of all the users in those indexes and returns that list to the RecommendationEngine. The engine applies to filter on the list based on user preferences and returns the final list of recommendations to the user.

Geo-Sharded Index

A naïve approach to maintaining this index would be to have an Elasticsearch cluster with one index and the default number of shards. However, this approach won’t hold up to the scaling expectations which an application like Tinder requires. We should leverage the fact that Tinder’s recommendations are location-based. For instance, when we are serving a user from India, we don’t need to include the users in the USA. This fact can be used by keeping an optimal index size for better performance. We can optimize the index size by sharding the existing user records based on their geo-locations so that the active user count remains balanced across shards. We can represent the balance of a geo-sharding configuration with N shards by the standard deviation of active user counts across shards, as mentioned below.

Balance(Shard1, Shard2,…, ShardN) = standard-deviation(Active User Count of Shard1, Shard2,…, ShardN)

The geo-sharding configuration with the minimal standard deviation would be best balanced. We can use geo-libraries like Google’s S2 library, which is based on a hierarchical decomposition of the sphere into “cells” using Quad-Trees. We have shown below a visualization of the generated geo-sharded map for our use-case. We can infer from the graph below that geo-shards are physically closer and larger for areas having a lower number of active users. For instance, in the image below, shards are more massive on water bodies like seas and oceans as they only have users from some islands. However, shards are smaller on land. In the image below, we can see that North America has three shards. However, entire England and Greenland along-with a large portion of Atlantic Ocean share a single shard due to the lesser density of active users.

Fig 4: Geo-sharded map generated b S2, each shard is an S2 cell__(__Image:_ Tinder Engineering Blog)

The S2 library provides two primary functions: i) given a location point(lat, long), return the S2 cell that contains it ii) given a circle(lat, long, radius), return the S2 cells that cover the circle. Each S2 cell can be represented by a geo-shard, which will be mapped to an index in our system. When a profile gets created, the user gets added to the search index for that corresponding S2 cell. To fetch recommendations for a user, we query the indexes of the nearby S2 cells depending on the circle radius, as shown in the image below.

Fig 5: Fetching recommend__ations for a user from nearby shards(Image: Tinder Engineering Blog)

Swipes and Matches

Fig 6: Sequence of operations for user swipes and matching

In the image above, we have shown the sequence of operations that gets executed when a user swipes left/right. The swipes ingester processes the swipes and puts the left swipes into a stream which persist those swipes to a low-cost data storage (e.g., Amazon S3). These left swipes can be used for data analysis for some use cases.

On the other hand, the right swipes are put in a separate stream and are ultimately read by the matcher worker thread. The matcher worker threads read the likes message from the stream and checks if the corresponding entry exists in the LikesCache. For instance, in the image above, Alice likes Bob, and the match worker checks if an entry exists for Bob liking Alice in the cache. If both Alice and Bob like each other, then it’s called a match, and a match notification is sent to both the users using server push mechanism like Websockets. If Bob hasn’t liked Alice yet, an entry is made in the LikesCache for Alice liking Bob.

Matches Data Model

We can use a key-value store (e.g., Amazon DynamoDB) to persist the information about matches (users liking each other). The hash key used for this data store can be a composite key of the unique identifiers of the users who liked each other. The value in the data-store will contain metadata information related to the match.

Key Value
userId1_userId2 {
(e.g. AWDGT567RTH_ ARTHT567WDG) “matchTimestamp”: “T2”,
“likes”: [
“likerId”: “AWDGT567RTH”,
“userLikedId”: “ARTHT567WDG”,
“timestamp”: “T1”
“likerId”: “ARTHT567WDG”,
“userLikedId”: “AWDGT567RTH”,
“timestamp”: “T2”

User Switching Locations

Fig 7: Sequence of operations for user switching location

When a user switches locations, we want to ensure that we provide recommendations to the user from the new site and vice-versa. The user location gets updated to the new location so that the updated location is used for fetching recommendations for the user. Additionally, we also update the index mapped to the user’s new location with the user’s information, so the user shows up in recommendations at the location. This process gets executed asynchronously.

The elastic search cluster (explained below) containing the geo-sharded indexes reads the message from a queue to update the user’s index. The elastic search coordinating nodes moves the user’s information from the index mapped to the user’s old location to the index mapped to the user’s new location. This will ensure that the user shows up in the recommendations of other users in the new location.

Elastic Search Cluster

Fig 8:_ Geo-sharded cluster architecture_

The cluster will comprise of multiple master nodes, each having two auto-scaling groups(ASG), one containing only coordinating nodes(this is where all the requests are sent), and another listing all the data nodes. Each data node will contain a certain number of indexes(a combination of primaries and replicas) of randomly distributed shards. For each user query, the responsibility of the coordinating node is to query the data-nodes of the target shards for handling the user query. We increase the reliability and robustness of the elastic search cluster by sharding the user data using their geographical locations and creating replicas of those shards.


One of the most critical aspects of an application like Tinder is the recommendations (of potential matches) it provides to a user. In one of the sections above, we saw how we generate recommendations for a user by querying indexes corresponding to the nearby geo-shards of a user. We can optimize the system by applying machine learning to rank the recommendations. The machine learning model will maximize the user’s potential to right swipe recommended potential matches. We have listed below some of the features which may impact the user’s decision to swipe left or right.

  • User demographics data: Age, Gender, Race, Location, Profession and so forth
  • User’s Tinder History data: Swipe Left, Swipe Rights, Historical geo-locations, Daily usage time
  • Extracted Information from User’s Bio: Likes, Dislikes, Preferences
  • Extracted Information from User’s Pictures: Facial features, Hair color, Body type

We can frame a regression problem by using these features to find the probability that the user will swipe right a recommendation. We can then leverage algorithms such as Logistic Regression to compute those probabilities, which will be used for ranking the recommendations.

In addition to this, we can also optimize the mechanism to send match notifications to the user by prefetching the information that a user has been swiped right by a recommended user. Pre-fetching this information, we can notify the user about a match (if and when it occurs) right-away, hence preventing the network call. For instance, if Alice is shown in Bob’s recommendation and Alice has already swiped Bob right then, Bob gets an instantaneous match notification(without any network hop) in case Bob swipes Alice right too.