Split lets companies fulfill the promise of continuous delivery: with Split’s safe launch and experimentation, teams can take any feature into production with fine-grain control. In order to precisely release features (and gather data on their engagement), Split’s SDKs utilize different streams of data. And many of our customers run Split with SDKs installed in multiple languages and throughout their application stack. By default, Split’s SDKs keep Split segment and Split targeting data synchronized as your users navigate across large distributed systems, treatments and conditions. Some languages, however, do not have a native capability to keep a shared local cache of this data to properly serve treatments.
Today, we’re giving these languages the power to do so by introducing a new feature: the Split Synchronizer. Let’s take a look at how it works, and the performance boost it can give your app in serving treatments with Split.
How the Split Synchronizer Works
The Split Synchronizer coordinates the sending and receiving of data to a remote data store that all of your processes can share when updating definitions and rules for the evaluation of treatments. Its an easy-to-install job that sits on your servers and is designed to run entirely in the background so that it has zero impact on the performance of your application servers,. Split supports Redis as a remote data store out-of-the box, so it’s no surprise the Split Synchronizer utilizes Redis as the cache for your SDKs when evaluating treatments. It also posts impression data and metrics generated by the SDKs back to Split’s servers, for exposure in the web console or sending to the data integration of your choice.
The Split Synchronizer works with most of the languages Split supports:
Those not currently listed already have the ability to keep a local data cache for Split’s SDKs, but if you’re already a Split customer and curious about using the Synchronizer with a language not listed above, drop us a note.
Why take the Synchronizer approach? Whether you’re working with one Split SDK or many, your team will see the benefits of:
- Data Consistency: Split-synchronizer and associated SDKs implement a producer-consumer pattern, using a Redis cache as a remote data store to guarantee data consistency. This allows you to take ownership of the data store, giving your team the power to improve the performance when feature flag definitions are requested from the SDK.
- Security: Since the Redis instance is located on either your local network or behind a virtual private network (VPN), any data contained within your features flags will remain in your app and secure.
- Fault Tolerance: If for some reason a connection issue is experienced on your network, the last consistent state of your feature flags will be stored in the Redis cache, and the Split Synchronizer will synchronize it once that the connection issue is restored.
- Error Handling: In addition to logging errors in the file system or even in the stdout, Split Synchronizer has the ability to send errors to a customized Slack channel or Slack user.
This service performs four different actions:
- Fetch Splits: retrieve the split definitions.
- Fetch Segments: retrieve your segments lists and memberships to targeting groups.
- Post Impressions: send the impressions generated by the SDK to Split servers.
- Post Metrics: send different metrics (like latencies) from the SDK to Split servers.
Performance Improvements and Benchmark
The Split Synchronizer is written in Go and introduces notable performance improvements over previous language-specific synchronizers bundled with our SDKs. Now, tracked data (Split’s impressions log) will be posted at Split servers in a near real-time window.
The following numbers show an SDK’s impression-handling capability over time, after installing the Split Synchronizer:
|~3M /minute||~180M /hour||~4.32B /day|
These numbers belong to the test scenario below:
|Spliti-sync instance:||m4.large (2 CPU – 8 GB Memory)|
|Redis instance:||cache.m3.large (2 vCPU – 6 GB Memory)|
|Number of features flags:||200|
|Impressions to send per post:||500k|
|Impressions generated per feature flag:||100k|
|Insertion rate in Redis:||1M each ~5s|
|Duration of the benchmark:||48 hours|
Analyzing Performance Improvements
Let’s look at performance by analyzing some infrastructure numbers for the duration of this benchmark .
The rate of successfully posted impressions remained stable throughout the duration of the test:
Memory consumption remained stable during the benchmark:
The CPU remained constant and within boundaries (~25% to 55%):
Throughput remained constant at ~20MB/s:
Stay up to date
Don’t miss out! Subscribe to our digest to get the latest about feature flags, continuous delivery, experimentation, and more.
With feature flags, you can control the percentage allocation of users you want to be exposed to a specific feature. This process provides risk mitigation and confirms both usability and scalability. Canary releases, or controlled rollouts, serve as an added layer of protection in case something goes wrong. What is…
Software testing is usually a point of friction for software development teams. Feature flags create an added layer of complexity that you need to account for. So how can we make the process of testing in a feature-flagged codebase easier? Isolate Tests to Reduce the Scope of the System Under…
Our mission is to help our customers develop impactful products that ultimately create a world-class experience for their customers. A big part of this means ensuring speed and reliability from development to production, and from your application to your end users. To that end, we’re excited to share some recent…