This is a brief overview of the stats that we captured by running Backend on different AWS machine configurations that we hope gives a rough idea for users in making the decisions to provision infrastructure to host Rudder. All numbers capture below are by using metrics described in Monitoring and Metrics.

Load Test Results

All tests are done using db.m4.xlarge Amazon RDS instance for hosting the postgres database

Gateway

MachineLoad

Response Time (ms)

gateway.response_time

Throughput

gateway.write_key_requests

Dangling Tables
m4.2xLarge
8Core 32GB
2.5K/s--2.5K/sNo
m4.2xLarge
8Core 32GB
5K/s--3K/sYes
m4.2xLarge
8Core 32GB
3K/s--2.7K/sNo

m5.xLarge

4Core 16GB

2.5K/s31.9K/sNo

m5.large

2Core 8GB

2.5K/s4.21.7K/sNo
Backend migrates and drops tables that have a threshold of jobs processed. Gateway tables are backed up to object storage (S3, MinIO etc.) if configured by user. Dangling tables indicate tables are ready for drop at a rate greater than the rate at which tables are backed up to object storage. Concurrent uploads to object storage is in the roadmap for upcoming versions of Backend.

Transformer

Machine

Gateway Throughput

gateway.write_key_requests

Throughput

processor.transformer_received

m4.2xLarge
8Core 32GB
2.7k/s2.7K/s

m5.xLarge

4Core 16GB

1.9K/s1.9K/s

m5.large

2Core 8GB

1.7K/s1.6K/s
Transformer is a NodeJS/Koa server launced as cluster of node processes, processses count equal to the number of cores of the machine. Choosing an instance with lower number of core the number of cores in instance processor might reduce the throughput of transformer.

Batch Router - S3 Destination

Machine

Gateway Throughput

gateway.write_key_requests

Throughput

batch_router.dest_successful_events

m4.2xLarge
8Core 32GB
2.7K/s2.8K/s

m5.xLarge

4Core 16GB

1.9K/s1.8K/s

m5.large

2Core 8GB

1.7K/s1.6K/s

Below is an image captured in CloudWatch Metrics showing the captured stats

screenshot 2019 10 25 at 12 22 22 pm Gateway Requests and Batch Router Throughputs

Database Requirements

Rudder recommends using a database with at least 1TB allocated storage as there could be downtime to increase storage realtime depending on your database service provider.

Estimating Storage

If you want to dig deeper and figure out the right storage size, go through the following example.
Following variables should be considered to come with a right storage size for your use case.

VariableDescriptionProduction sample data
numSourcesTotal number of sources2
numEventsPerSecNumber of events per sec for a given source2500
avgGwEventSizeEvent size that is captured at the gateway by Rudder2.1 KB
gwEventOverheadSize of extra metadata that Rudder stores at Gateway to process the event300 B
numDestsNumber of enabled destinations for a given source3
avgRtEventSizeThe payload size that needs to be sent by the router to the destination after applying transformations1.2 KB
rtEventOverheadSize of extra metadata that Rudder stores to process the event300 B
gatewayStorage=numEventsPerSec(avgGwEventSize+gwEventOverhead)gatewayStorage = numEventsPerSec * (avgGwEventSize + gwEventOverhead)
routerStorage=numEventsPerSecnumDests(avgRtEventSize+rtEventOverhead)routerStorage = numEventsPerSec * numDests * (avgRtEventSize + rtEventOverhead)
totalStoragePerHour=3600firstSourcelastSource(gatewayStorage+routerStorage)totalStoragePerHour = 3600 * \sum_{firstSource}^{lastSource} (gatewayStorage + routerStorage)

In the above production example, after substituting the values, totalStoragePerHour adds up to 120 GB

Sample your peak load in production to estimate the storage requirements and substitute your values to get an estimate of the storage needed per 1 hour of data.

Event data and tables are ephemeral. In a happy path, we would have only a few minutes of event data being stored.

We recommend at least 10 hours worth of event storage computed above to gracefully handle destinations going down for a few hours.

If you want to prepare for a destination going for down for days, accommodate them into your storage capacity.

Estimating Connections

Rudder batches requests efficiently to write data. Under heavy load, backend can be configured (batchTimeoutInMS and maxBatchSize ) to batch more requests to limit concurrent connections to the database. If write latencies to the database are not in permissible thresholds, a new data set needs to be added i.e., backend server and database server.

Rudder reads the data back from the database at a constant rate. A sudden spike in user traffic will not result in more read DB requests.

RAM Requirements

Rudder does not cache aggressively and hence does not need huge amount of memory. Load tests were performed on 4 GB and 8 GB memory instances.

Rudder caches active user events by default to form configurable user sessions server side. The length of any user session can be configured with sessionThresholdInS and sessionThresholdEvents. Once a user's session is formed, that user events are cleared from the cache. If you don't need sessions, this can be disabled by setting processSessions to false.

VariableDescriptionSample data
numActiveUsersNumber of active users during a session (2 min) in your application during peak hours10000
avgGwEventSizeEvent size that is captured at the gateway by Rudder2.1 KB
userEventsInThresholdNumber of user events in the given threshold i.e., 40 user events in 2 min40
memoryNeeded=numActiveUsersuserEventsInThresholdavgGwEventSizememoryNeeded = numActiveUsers * userEventsInThreshold * avgGwEventSize

Memory required in the above example would be 840 MB.

The memory estimate does not include the default RAM required for running the OS and the required processes.

Contact us

For more information on the topics covered on this page, email us or start a conversation in our Slack community.

Contents