Skip to content

Instantly share code, notes, and snippets.

@JonCole
Last active March 13, 2025 14:30
Show Gist options
  • Save JonCole/925630df72be1351b21440625ff2671f to your computer and use it in GitHub Desktop.
Save JonCole/925630df72be1351b21440625ff2671f to your computer and use it in GitHub Desktop.
Redis Best Practices

Some of the Redis best practices content has moved

This content from this markdown file has moved a new, happier home where it can serve more people. Please check it out : https://docs.microsoft.com/azure/azure-cache-for-redis/cache-best-practices.

NOTE: Client specific guidance listed below is still valid and should still be considered. I will update this document once all content has been moved.

StackExchange.Redis

General Guidance

  1. Set AbortConnect to false, then let the ConnectionMultiplexer reconnect automatically. See here for details

  2. Reuse the ConnectionMultiplexer - do not create a new one for each request. The Lazy<ConnectionMultiplexer> pattern shown here is strongly recommended.

  3. Configure your ThreadPool settings to avoid timeouts.

  4. Be aware of the performance costs associated with different operations you are running. For instance, the "KEYS" command is an O(n) operation and should be avoided. The redis.io site has details around the time complexity for each operation that it supports.

  5. Consider turning on "Server GC". "Workstation GC" is the default and can impact the latencies when garbage collection is in happening.

  6. Most customers find that a single ConnectionMultiplexer is sufficient for their needs. However, if you have high throughput requirements, you may consider slowly increasing the number of connections to see if you get an improvement in throughput. Avoid setting it to an arbitrarily large number of connections as it may not give the desired benefit.

  7. Configure supported TLS settings. If you are targeting .NET 4.7 or later, you should not have to do anything because StackExchange.Redis will automatically use the OS level settings when deciding which TLS versions to support (which is a good thing in most cases). If you are targeting an older version of .NET, then you should configure the client to use the highest TLS version that your client system supports (typically TLS 1.2). Once you move to a newer version of the .NET framework, then you should probably remove this configuration and let the OS settings take precedence. This can configured through the sslProtocols connection string entry (requires NuGet package version 1.2.3 or later), or through the ConfigurationOptions class as show here:

    var options = ConfigurationOptions.Parse(connectionString);

    options.SslProtocols = System.Security.Authentication.SslProtocols.Tls12;

    ConnectionMultiplexer.Connect(options);

Reconnecting with Lazy<T> pattern

We have seen a few rare cases where StackExchange.Redis fails to reconnect after a connection blip (for example, due to patching). Restarting the client or creating a new ConnectionMultiplexer will fix the issue. Here is some sample code that still uses the recommended Lazy<ConnectionMultiplexer>pattern while allowing apps to force a reconnection periodically. Make sure to update code calling into the ConnectionMultiplexer so that they handle any ObjectDisposedException errors that occur as a result of disposing the old one.

using System;
using System.Threading;
using StackExchange.Redis;
// Source Code Usage License: https://gist.github.com/JonCole/34ca1d2698da7a1aa65ff781c37ecdea
static class Redis
{
static long lastReconnectTicks = DateTimeOffset.MinValue.UtcTicks;
static DateTimeOffset firstError = DateTimeOffset.MinValue;
static DateTimeOffset previousError = DateTimeOffset.MinValue;
static object reconnectLock = new object();
// In general, let StackExchange.Redis handle most reconnects,
// so limit the frequency of how often this will actually reconnect.
public static TimeSpan ReconnectMinFrequency = TimeSpan.FromSeconds(60);
// if errors continue for longer than the below threshold, then the
// multiplexer seems to not be reconnecting, so re-create the multiplexer
public static TimeSpan ReconnectErrorThreshold = TimeSpan.FromSeconds(30);
static string connectionString = "TODO: CALL InitializeConnectionString() method with connection string";
static Lazy<ConnectionMultiplexer> multiplexer = CreateMultiplexer();
public static ConnectionMultiplexer Connection { get { return multiplexer.Value; } }
public static void InitializeConnectionString(string cnxString)
{
if (string.IsNullOrWhiteSpace(cnxString))
throw new ArgumentNullException(nameof(cnxString));
connectionString = cnxString;
}
/// <summary>
/// Force a new ConnectionMultiplexer to be created.
/// NOTES:
/// 1. Users of the ConnectionMultiplexer MUST handle ObjectDisposedExceptions, which can now happen as a result of calling ForceReconnect()
/// 2. Don't call ForceReconnect for Timeouts, just for RedisConnectionExceptions or SocketExceptions
/// 3. Call this method every time you see a connection exception, the code will wait to reconnect:
/// a. for at least the "ReconnectErrorThreshold" time of repeated errors before actually reconnecting
/// b. not reconnect more frequently than configured in "ReconnectMinFrequency"
/// </summary>
public static void ForceReconnect()
{
var utcNow = DateTimeOffset.UtcNow;
var previousTicks = Interlocked.Read(ref lastReconnectTicks);
var previousReconnect = new DateTimeOffset(previousTicks, TimeSpan.Zero);
var elapsedSinceLastReconnect = utcNow - previousReconnect;
// If mulitple threads call ForceReconnect at the same time, we only want to honor one of them.
if (elapsedSinceLastReconnect > ReconnectMinFrequency)
{
lock (reconnectLock)
{
utcNow = DateTimeOffset.UtcNow;
elapsedSinceLastReconnect = utcNow - previousReconnect;
if (firstError == DateTimeOffset.MinValue)
{
// We haven't seen an error since last reconnect, so set initial values.
firstError = utcNow;
previousError = utcNow;
return;
}
if (elapsedSinceLastReconnect < ReconnectMinFrequency)
return; // Some other thread made it through the check and the lock, so nothing to do.
var elapsedSinceFirstError = utcNow - firstError;
var elapsedSinceMostRecentError = utcNow - previousError;
var shouldReconnect =
elapsedSinceFirstError >= ReconnectErrorThreshold // make sure we gave the multiplexer enough time to reconnect on its own if it can
&& elapsedSinceMostRecentError <= ReconnectErrorThreshold; //make sure we aren't working on stale data (e.g. if there was a gap in errors, don't reconnect yet).
// Update the previousError timestamp to be now (e.g. this reconnect request)
previousError = utcNow;
if (shouldReconnect)
{
firstError = DateTimeOffset.MinValue;
previousError = DateTimeOffset.MinValue;
var oldMultiplexer = multiplexer;
CloseMultiplexer(oldMultiplexer);
multiplexer = CreateMultiplexer();
Interlocked.Exchange(ref lastReconnectTicks, utcNow.UtcTicks);
}
}
}
}
private static Lazy<ConnectionMultiplexer> CreateMultiplexer()
{
return new Lazy<ConnectionMultiplexer>(() => ConnectionMultiplexer.Connect(connectionString));
}
private static void CloseMultiplexer(Lazy<ConnectionMultiplexer> oldMultiplexer)
{
if (oldMultiplexer != null)
{
try
{
oldMultiplexer.Value.Close();
}
catch (Exception)
{
// Example error condition: if accessing old.Value causes a connection attempt and that fails.
}
}
}
}
// CONTENT MOVED:
// https://github.com/Azure/AzureCacheForRedis/blob/main/Redis-SampleCode-Java-JedisPool.java
@amarwadi
Copy link

Hello, is there an example on how to consume this ForceReconnect functionality? I also have a issue and I've evaluated every single item and I know that I must be doing something wrong in setting up my connection: https://gist.github.com/anonymous/c7c741b8a97e2c77e4bc3eeb9d4781ec

Do you have any recommendations?

@BoushZhou
Copy link

Good!

@BoushZhou
Copy link

Cloud you push python code in Azure Redis.

@VictorioBerra
Copy link

VictorioBerra commented Apr 2, 2019

I would really appreciate a C# Serverless Function example. Nowhere can I find good practice information on that. Especially if I should be using the Lazy pattern as described here. Ideally, the StackExchange.Redis client would be mentioned here in addition to the HttpClient and DocumentClient

@balukantu
Copy link

balukantu commented Apr 5, 2019

I am planning to use Redis cache in my application and the data size may vary from 10KB - 500 MB. If I use the large key/value entries from Redis, Reding and Saving data might get performance issue, it means I have to transfer more data over the network from the server to the client and I might get below consequences. (Got below from @stackoverflow)

  1. It takes more time to transfer the data, so client may need to have a higher timeout value configured to allow for this additional transfer time.
  2. Requests made to the server can get stuck behind the big transfer and cause other requests to timeout.
  3. Network buffers used to transfer this data can impact available memory on the client or server, which can aggravate the available memory issues described around fragmentation.
  4. If these large key/value items are accessed frequently, this magnifies the impacts if we repeatedly transferring this data over and over again.
    So I would like to do partitioning in Redis cache.

Example: I have an endpoint(GetEmployees(Some computation happens to get all employees)) which returns around 500 records with size 400 MB ( Example). I get the same data until I add any new employee so I want to have cache mechanism so that I can store these records and get it from the cache instead of doing calculations again and again.

If I store all records in the single shard of Redis, will face above mentioned issues.

Key/Value : GetEmplyeesTodayDate/ {List Of Records}

So, I would like to split the data and across shards and get all records when reading from Redis.

What is the best way to store all 500 records in Redis with same/multiple keys (s) and retrieve them?

@sagarsitap596
Copy link

I believe if you want to store all 500 records under one key and you may also want to access one record at a time go for HSET
EG. hset employees(key) employee1(field) data(value).

Sharding works on keys. So if you are trying to store 500 reocrds under single key. All the records will reside on single shard.
If you store each record seperately as a key value then all 500 records may span across available shards.

@dehaaneric
Copy link

Hi Jon, while reading all new Best Practices on the Microsoft documention pages, I can't seem to find any ForceReconnect recommendations like this gist. Are you aware of any implementations of this force-reconnect feature inside the StackExchange.Redis v2+?

@JonCole
Copy link
Author

JonCole commented Sep 12, 2019

The official documentation still points to this gist for the ForceReconnect example code. It is above in this same page, under the https://gist.github.com/JonCole/925630df72be1351b21440625ff2671f#file-redis-lazyreconnect-cs part.

@TheSoftwareArchitect
Copy link

"> I would really appreciate a C# Serverless Function example. Nowhere can I find good practice information on that. Especially if I should be using the Lazy pattern as described here. Ideally, the StackExchange.Redis client would be mentioned here in addition to the HttpClient and DocumentClient"

Yes. You should. Here's some code I'm using for a thread-safe singleton that leverages the Lazy (T) model.
"
public class MyRedisConnectorHelper
{

    private static readonly Lazy<ConnectionMultiplexer> LazyConnection = new Lazy<ConnectionMultiplexer>(() =>
    {
        string cacheConnection = ConfigurationManager.AppSettings["CacheConnection"].ToString();
        return ConnectionMultiplexer.Connect(cacheConnection);
    });

    public static ConnectionMultiplexer Connection
    {
        get
        {
            return LazyConnection.Value;
        }
    }

}"

You'll still need to add a config file as described in the Azure documentation. I'm actually using mine for a local Redis server I'm running on Ubuntu, but it works great!

I don't have a serverless Redis example right now. If I work on one later or tomorrow, I'll post it too. Good luck!

@javaadpatel
Copy link

This guide is really amazing, I am just wondering what the best way of handling the ObjectDisposedException is?

@JonCole
Copy link
Author

JonCole commented Oct 31, 2019

This guide is really amazing, I am just wondering what the best way of handling the ObjectDisposedException is?

If I were to build this, I would create a wrapper around the IConnectionMultiplexer/IDatabase interfaces that handles the ObjectDisposedExceptions that are thrown, such that as few callers have to be aware of this as possible. I expect that when you see these errors, retrying immediately on the new connection should be fine.

@javaadpatel
Copy link

That makes sense, the call that I have that is failing is this.database.ScriptEvaluate... which is IDatabase. Do know of any examples I could look at to see how to wrap this interface?

@maca88
Copy link

maca88 commented Feb 27, 2020

I recently made wrappers around StackExchange.Redis interfaces which are able to do a force reconnect based on the logic posted here and can be found at: https://github.com/maca88/StackExchange.Redis.Resilience
Do note that these wrappers were not yet tested in a real application, so please use them with caution.

@danlarson
Copy link

ForceReconnect needs something to reset the firstErrror on successful use of the multiplexer, something like firstError = DateTimeOffset.MinValue; so that a connection blip doesn't trigger destroying and recreating the multiplexer. The code as is doesn't let the multiplexer just handle the reconnection itself very well, and needs a bit of a test harness to ensure it works as designed.

@JonCole
Copy link
Author

JonCole commented May 21, 2020

ForceReconnect needs something to reset the firstErrror on successful use of the multiplexer, something like firstError = DateTimeOffset.MinValue; so that a connection blip doesn't trigger destroying and recreating the multiplexer. The code as is doesn't let the multiplexer just handle the reconnection itself very well, and needs a bit of a test harness to ensure it works as designed.

@danlarson Yeah, looking at the code again, I can see how we have a problem here when we hit this code but we don't end up doing the actual forced reconnect. Seems like this would mean that this would sometimes (every other time?) force a reconnect when it doesn't need to. Another problem that this code doesn't handle very well is the case where the multiplexer is connected to a clustered redis instance. If the connection to just one shard is broken, this logic forces a reconnect to all shards to fix the one broken connection. Not ideal. Ultimately, the underlying issue needs to be fixed inside StackExchange.Redis. Clustering makes this a harder problem to solve - we can't say that one successful request means that the connection issue is fixed because the success call might be from connection to a different shard. Maybe we reset the firstError if elapsed time is greater than some threshold? Thoughts?

@pooran
Copy link

pooran commented Jun 14, 2020

image
same code, no change.. what might be wrong?

@hustlestar
Copy link

Is there any way to access or log Redis Connection Pool while using it through Spring Redis Data template?

@upen4a3
Copy link

upen4a3 commented Oct 20, 2020

How can we force reconnect on spring-boot reactive application which is using Lettuce connection on azure.We see it is taking 15 min to reconnect in azure ,where as local docker image takes less than 5 min to reconnect.

@balazzii
Copy link

balazzii commented Dec 8, 2020

https://gist.github.com/JonCole/925630df72be1351b21440625ff2671f#file-redis-lazyreconnect-cs-L38

I am planning on calling ForceReconnect if RedisTimeoutException occurs continuously for more than 5 minutes, but it says that "Don't call ForceReconnect for Timeouts".
I would like to know the reason for this.
Is there some problem on forcefully reconnecting on RedisTimeoutException?

@coultonluke
Copy link

coultonluke commented Jan 8, 2021

Most customers find that a single ConnectionMultiplexer is sufficient for their needs. However, if you have high throughput requirements, you may consider slowly increasing the number of connections to see if you get an improvement in throughput. Avoid setting it to an arbitrarily large number of connections as it may not give the desired benefit.

How does one increase the number of connections? The provided static class allows for only a single ConnectionMultiplexer and ForceReconnection of only a single ConnectionMultiplexer...?

Thanks for your help

@JonCole
Copy link
Author

JonCole commented Jan 8, 2021

How does one increase the number of connections?

You would have to create a pool of multiplexers to draw from, but I don't have example code for that currently. ForceReconnect is designed for a single connection multiplexer. If you do create a pool, make sure that each instance in the pool has its own set of variables for the ForceReconnect logic.

@AdvanDizdarevic
Copy link

"> I would really appreciate a C# Serverless Function example. Nowhere can I find good practice information on that. Especially if I should be using the Lazy pattern as described here. Ideally, the StackExchange.Redis client would be mentioned here in addition to the HttpClient and DocumentClient"

Yes. You should. Here's some code I'm using for a thread-safe singleton that leverages the Lazy (T) model.
"
public class MyRedisConnectorHelper
{

    private static readonly Lazy<ConnectionMultiplexer> LazyConnection = new Lazy<ConnectionMultiplexer>(() =>
    {
        string cacheConnection = ConfigurationManager.AppSettings["CacheConnection"].ToString();
        return ConnectionMultiplexer.Connect(cacheConnection);
    });

    public static ConnectionMultiplexer Connection
    {
        get
        {
            return LazyConnection.Value;
        }
    }

}"

You'll still need to add a config file as described in the Azure documentation. I'm actually using mine for a local Redis server I'm running on Ubuntu, but it works great!

I don't have a serverless Redis example right now. If I work on one later or tomorrow, I'll post it too. Good luck!

image
same code, no change.. what might be wrong?

image
It's probably telling you that the value for cacheConnection is null. Just pass the string to the Multiplexer as in this example

@yapaxi
Copy link

yapaxi commented May 29, 2021

I am investigating locking\deadlocking\thread-pool-starvation issues related to this code. I am not a redis expert so some help is appreciated.

Concerns:

  • Lazy with sync IO inside: Connect has Task.Wait() and .Result
  • Reconnect closes multiplexer under lock while Close has .Result inside

#Lazy

If async await code calls Lazy.Value and initialization happens then all thread-pool threads are synchronousely blocked on waiting for initialization. If for any reason intialization stucks or is just a bit slow a service which goes through high-load will be simply paralyzed and thread pool will be forced to grow and add more more and threads which all will get stuck on Lazy<T>.Value. Recommendation to increase ThreadPool.MinThreads is like to try to quench the fire with gasoline.

#Reconnect

I looked into Close implementation and found few cases of .Result inside. Given that close happens under lock the same issue as with Lazy value happens.

The code inside StackExchange.Redis is a mixture of await, .Wait, .Result which is a very dangerous approach. Is the code in this article still recommended to be used in high-loaded production services or maybe there are better alternatives?

@coultonluke
Copy link

I am investigating locking\deadlocking\thread-pool-starvation issues related to this code. I am not redis expert so some help is appreciated.

A C# web application that uses this code intermittently gets killed off and complains of thread pool starvation, so it's very interesting that you are suggesting that this might be a problem caused by the implementation here. I'd be very interested to know the outcome of this.

@yapaxi
Copy link

yapaxi commented May 29, 2021

Our case was like this:

  • service is restarted
  • after start service receives heavy async-io load spawning many tasks
  • the load hangs in Monitor (lock) and Lazy.Value
  • thread-pool quickly grows but it does not help - application is still stuck, even tasks unrelated to Redis are stuck because they cannot be completed

Absence of sync-context in modern ASP.NET Core and improved thread pool scaling in .NET does not mean that mixing await and .Result.Wait is now a good idea. It is not and it will never be. From my point of view the code in this article should be deprecated or at least rewritten using SemaphoreSlim instead of Lazy and Lock. Also only async-IO may be used (ConnectAsync, CloseAsync).

@JonCole
Copy link
Author

JonCole commented Jun 1, 2021

If I understand the concern correctly, the main problem is that many threads are blocked on the locking at the same time, and the high level solution would be to make the creation of the multiplexer be asynchronous. Is that correct? At some point, it seems like you have to have a lock if you want to ensure only one connection object is created. I suppose you could create multiple then discard the extras but that has its own set of problems it can create (for instance, flooding the server with connection requests, many of which will just get closed).

@yapaxi
Copy link

yapaxi commented Jun 1, 2021

The problem

Yes, there is a problem with Lazy and Lock.

Here is a simplified scenario for starvation: https://gist.github.com/yapaxi/cfc529bee9bd5361482051d45a62a571.

In a nutshell it has:

  • initialization task (get-google) which is pseudo-synchronousely waited in Lazy
  • loop that adds fake "load" that synchronousely waits for Lazy
  • in the end "load" waits for lazy and lazy waits for load to release thread-pool threads
  • eventhough thread pool grows new thread-pool threads are not used to finish get-google until some kind of timeout happens (in the test case)

On sunny days get-google finishes in 30 seconds, sometimes in 1 minute, sometimes never. In the example case most likely DNS resolve locks or deadlocks. In real case it is almost impossible to tell why Multiplexer.Connect is slow or near-deadlocks - too many things happen.
In real world 30+ seconds is sometimes enough for cluster to consider a node dead.

A solution

The best way I see is to use awaitable SemaphoreSlim instead of lazy and lock. That will ensure only single Multiplexer is created.

I do not know anything about Redis (I am only a part of outage investigation team) but I can help with multithreading.

Here is one approach: https://gist.github.com/yapaxi/9526ca734c8cc30173276c616324d33b

Be careful:

  • I changed timeouts and thresholds for testing purposes
  • I do not know how to handle Redis connection timeout properly so I just added timeout tasks
  • I kept logic in reconnected but reduced nesting a bit
  • It is hard to guess if you could use the same lock in reconnect because this is an incomplete solution so I decided to make Reconnect non-waiting to completely avoid deadlocks and transfer connection responsibility to GetConnection() method: reconnect only nullifies the state;

Hope this helps. Questions are welcome.

@mayankiitg
Copy link

In the Redis-LazyReconnect.cs, at line number 57, inside the lock, we should re-read the value of previousReconnect.
elapsedSinceLastReconnect = utcNow - previousReconnect;

As this thread might have waited to acquire a lock and some other thread might have got it first and reconnected successfully, thus updating previousReconnect. And we are reading stale value in all other threads for this variable.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment