This content from this markdown file has moved a new, happier home where it can serve more people. Please check it out : https://docs.microsoft.com/azure/azure-cache-for-redis/cache-best-practices.
-
Star
(424)
You must be signed in to star a gist -
Fork
(85)
You must be signed in to fork a gist
-
-
Save JonCole/925630df72be1351b21440625ff2671f to your computer and use it in GitHub Desktop.
-
Set AbortConnect to false, then let the ConnectionMultiplexer reconnect automatically. See here for details
-
Reuse the ConnectionMultiplexer - do not create a new one for each request. The
Lazy<ConnectionMultiplexer>
pattern shown here is strongly recommended. -
Configure your ThreadPool settings to avoid timeouts.
-
Be aware of the performance costs associated with different operations you are running. For instance, the "KEYS" command is an O(n) operation and should be avoided. The redis.io site has details around the time complexity for each operation that it supports.
-
Consider turning on "Server GC". "Workstation GC" is the default and can impact the latencies when garbage collection is in happening.
-
Most customers find that a single ConnectionMultiplexer is sufficient for their needs. However, if you have high throughput requirements, you may consider slowly increasing the number of connections to see if you get an improvement in throughput. Avoid setting it to an arbitrarily large number of connections as it may not give the desired benefit.
-
Configure supported TLS settings. If you are targeting .NET 4.7 or later, you should not have to do anything because StackExchange.Redis will automatically use the OS level settings when deciding which TLS versions to support (which is a good thing in most cases). If you are targeting an older version of .NET, then you should configure the client to use the highest TLS version that your client system supports (typically TLS 1.2). Once you move to a newer version of the .NET framework, then you should probably remove this configuration and let the OS settings take precedence. This can configured through the sslProtocols connection string entry (requires NuGet package version 1.2.3 or later), or through the ConfigurationOptions class as show here:
var options = ConfigurationOptions.Parse(connectionString);
options.SslProtocols = System.Security.Authentication.SslProtocols.Tls12;
ConnectionMultiplexer.Connect(options);
We have seen a few rare cases where StackExchange.Redis fails to reconnect after a connection blip (for example, due to patching). Restarting the client or creating a new ConnectionMultiplexer will fix the issue. Here is some sample code that still uses the recommended Lazy<ConnectionMultiplexer>
pattern while allowing apps to force a reconnection periodically. Make sure to update code calling into the ConnectionMultiplexer so that they handle any ObjectDisposedException
errors that occur as a result of disposing the old one.
using System; | |
using System.Threading; | |
using StackExchange.Redis; | |
// Source Code Usage License: https://gist.github.com/JonCole/34ca1d2698da7a1aa65ff781c37ecdea | |
static class Redis | |
{ | |
static long lastReconnectTicks = DateTimeOffset.MinValue.UtcTicks; | |
static DateTimeOffset firstError = DateTimeOffset.MinValue; | |
static DateTimeOffset previousError = DateTimeOffset.MinValue; | |
static object reconnectLock = new object(); | |
// In general, let StackExchange.Redis handle most reconnects, | |
// so limit the frequency of how often this will actually reconnect. | |
public static TimeSpan ReconnectMinFrequency = TimeSpan.FromSeconds(60); | |
// if errors continue for longer than the below threshold, then the | |
// multiplexer seems to not be reconnecting, so re-create the multiplexer | |
public static TimeSpan ReconnectErrorThreshold = TimeSpan.FromSeconds(30); | |
static string connectionString = "TODO: CALL InitializeConnectionString() method with connection string"; | |
static Lazy<ConnectionMultiplexer> multiplexer = CreateMultiplexer(); | |
public static ConnectionMultiplexer Connection { get { return multiplexer.Value; } } | |
public static void InitializeConnectionString(string cnxString) | |
{ | |
if (string.IsNullOrWhiteSpace(cnxString)) | |
throw new ArgumentNullException(nameof(cnxString)); | |
connectionString = cnxString; | |
} | |
/// <summary> | |
/// Force a new ConnectionMultiplexer to be created. | |
/// NOTES: | |
/// 1. Users of the ConnectionMultiplexer MUST handle ObjectDisposedExceptions, which can now happen as a result of calling ForceReconnect() | |
/// 2. Don't call ForceReconnect for Timeouts, just for RedisConnectionExceptions or SocketExceptions | |
/// 3. Call this method every time you see a connection exception, the code will wait to reconnect: | |
/// a. for at least the "ReconnectErrorThreshold" time of repeated errors before actually reconnecting | |
/// b. not reconnect more frequently than configured in "ReconnectMinFrequency" | |
/// </summary> | |
public static void ForceReconnect() | |
{ | |
var utcNow = DateTimeOffset.UtcNow; | |
var previousTicks = Interlocked.Read(ref lastReconnectTicks); | |
var previousReconnect = new DateTimeOffset(previousTicks, TimeSpan.Zero); | |
var elapsedSinceLastReconnect = utcNow - previousReconnect; | |
// If mulitple threads call ForceReconnect at the same time, we only want to honor one of them. | |
if (elapsedSinceLastReconnect > ReconnectMinFrequency) | |
{ | |
lock (reconnectLock) | |
{ | |
utcNow = DateTimeOffset.UtcNow; | |
elapsedSinceLastReconnect = utcNow - previousReconnect; | |
if (firstError == DateTimeOffset.MinValue) | |
{ | |
// We haven't seen an error since last reconnect, so set initial values. | |
firstError = utcNow; | |
previousError = utcNow; | |
return; | |
} | |
if (elapsedSinceLastReconnect < ReconnectMinFrequency) | |
return; // Some other thread made it through the check and the lock, so nothing to do. | |
var elapsedSinceFirstError = utcNow - firstError; | |
var elapsedSinceMostRecentError = utcNow - previousError; | |
var shouldReconnect = | |
elapsedSinceFirstError >= ReconnectErrorThreshold // make sure we gave the multiplexer enough time to reconnect on its own if it can | |
&& elapsedSinceMostRecentError <= ReconnectErrorThreshold; //make sure we aren't working on stale data (e.g. if there was a gap in errors, don't reconnect yet). | |
// Update the previousError timestamp to be now (e.g. this reconnect request) | |
previousError = utcNow; | |
if (shouldReconnect) | |
{ | |
firstError = DateTimeOffset.MinValue; | |
previousError = DateTimeOffset.MinValue; | |
var oldMultiplexer = multiplexer; | |
CloseMultiplexer(oldMultiplexer); | |
multiplexer = CreateMultiplexer(); | |
Interlocked.Exchange(ref lastReconnectTicks, utcNow.UtcTicks); | |
} | |
} | |
} | |
} | |
private static Lazy<ConnectionMultiplexer> CreateMultiplexer() | |
{ | |
return new Lazy<ConnectionMultiplexer>(() => ConnectionMultiplexer.Connect(connectionString)); | |
} | |
private static void CloseMultiplexer(Lazy<ConnectionMultiplexer> oldMultiplexer) | |
{ | |
if (oldMultiplexer != null) | |
{ | |
try | |
{ | |
oldMultiplexer.Value.Close(); | |
} | |
catch (Exception) | |
{ | |
// Example error condition: if accessing old.Value causes a connection attempt and that fails. | |
} | |
} | |
} | |
} |
// CONTENT MOVED: | |
// https://github.com/Azure/AzureCacheForRedis/blob/main/Redis-SampleCode-Java-JedisPool.java |
Good!
Cloud you push python code in Azure Redis.
I am planning to use Redis cache in my application and the data size may vary from 10KB - 500 MB. If I use the large key/value entries from Redis, Reding and Saving data might get performance issue, it means I have to transfer more data over the network from the server to the client and I might get below consequences. (Got below from @stackoverflow)
- It takes more time to transfer the data, so client may need to have a higher timeout value configured to allow for this additional transfer time.
- Requests made to the server can get stuck behind the big transfer and cause other requests to timeout.
- Network buffers used to transfer this data can impact available memory on the client or server, which can aggravate the available memory issues described around fragmentation.
- If these large key/value items are accessed frequently, this magnifies the impacts if we repeatedly transferring this data over and over again.
So I would like to do partitioning in Redis cache.
Example: I have an endpoint(GetEmployees(Some computation happens to get all employees)) which returns around 500 records with size 400 MB ( Example). I get the same data until I add any new employee so I want to have cache mechanism so that I can store these records and get it from the cache instead of doing calculations again and again.
If I store all records in the single shard of Redis, will face above mentioned issues.
Key/Value : GetEmplyeesTodayDate/ {List Of Records}
So, I would like to split the data and across shards and get all records when reading from Redis.
What is the best way to store all 500 records in Redis with same/multiple keys (s) and retrieve them?
I believe if you want to store all 500 records under one key and you may also want to access one record at a time go for HSET
EG. hset employees(key) employee1(field) data(value).
Sharding works on keys. So if you are trying to store 500 reocrds under single key. All the records will reside on single shard.
If you store each record seperately as a key value then all 500 records may span across available shards.
Hi Jon, while reading all new Best Practices on the Microsoft documention pages, I can't seem to find any ForceReconnect recommendations like this gist. Are you aware of any implementations of this force-reconnect feature inside the StackExchange.Redis v2+?
The official documentation still points to this gist for the ForceReconnect example code. It is above in this same page, under the https://gist.github.com/JonCole/925630df72be1351b21440625ff2671f#file-redis-lazyreconnect-cs part.
"> I would really appreciate a C# Serverless Function example. Nowhere can I find good practice information on that. Especially if I should be using the Lazy pattern as described here. Ideally, the StackExchange.Redis client would be mentioned here in addition to the HttpClient and DocumentClient"
Yes. You should. Here's some code I'm using for a thread-safe singleton that leverages the Lazy (T) model.
"
public class MyRedisConnectorHelper
{
private static readonly Lazy<ConnectionMultiplexer> LazyConnection = new Lazy<ConnectionMultiplexer>(() =>
{
string cacheConnection = ConfigurationManager.AppSettings["CacheConnection"].ToString();
return ConnectionMultiplexer.Connect(cacheConnection);
});
public static ConnectionMultiplexer Connection
{
get
{
return LazyConnection.Value;
}
}
}"
You'll still need to add a config file as described in the Azure documentation. I'm actually using mine for a local Redis server I'm running on Ubuntu, but it works great!
I don't have a serverless Redis example right now. If I work on one later or tomorrow, I'll post it too. Good luck!
This guide is really amazing, I am just wondering what the best way of handling the ObjectDisposedException is?
This guide is really amazing, I am just wondering what the best way of handling the ObjectDisposedException is?
If I were to build this, I would create a wrapper around the IConnectionMultiplexer/IDatabase interfaces that handles the ObjectDisposedExceptions that are thrown, such that as few callers have to be aware of this as possible. I expect that when you see these errors, retrying immediately on the new connection should be fine.
That makes sense, the call that I have that is failing is this.database.ScriptEvaluate...
which is IDatabase. Do know of any examples I could look at to see how to wrap this interface?
I recently made wrappers around StackExchange.Redis
interfaces which are able to do a force reconnect based on the logic posted here and can be found at: https://github.com/maca88/StackExchange.Redis.Resilience
Do note that these wrappers were not yet tested in a real application, so please use them with caution.
ForceReconnect needs something to reset the firstErrror on successful use of the multiplexer, something like firstError = DateTimeOffset.MinValue;
so that a connection blip doesn't trigger destroying and recreating the multiplexer. The code as is doesn't let the multiplexer just handle the reconnection itself very well, and needs a bit of a test harness to ensure it works as designed.
ForceReconnect needs something to reset the firstErrror on successful use of the multiplexer, something like
firstError = DateTimeOffset.MinValue;
so that a connection blip doesn't trigger destroying and recreating the multiplexer. The code as is doesn't let the multiplexer just handle the reconnection itself very well, and needs a bit of a test harness to ensure it works as designed.
@danlarson Yeah, looking at the code again, I can see how we have a problem here when we hit this code but we don't end up doing the actual forced reconnect. Seems like this would mean that this would sometimes (every other time?) force a reconnect when it doesn't need to. Another problem that this code doesn't handle very well is the case where the multiplexer is connected to a clustered redis instance. If the connection to just one shard is broken, this logic forces a reconnect to all shards to fix the one broken connection. Not ideal. Ultimately, the underlying issue needs to be fixed inside StackExchange.Redis. Clustering makes this a harder problem to solve - we can't say that one successful request means that the connection issue is fixed because the success call might be from connection to a different shard. Maybe we reset the firstError if elapsed time is greater than some threshold? Thoughts?
Is there any way to access or log Redis Connection Pool while using it through Spring Redis Data template?
How can we force reconnect on spring-boot reactive application which is using Lettuce connection on azure.We see it is taking 15 min to reconnect in azure ,where as local docker image takes less than 5 min to reconnect.
https://gist.github.com/JonCole/925630df72be1351b21440625ff2671f#file-redis-lazyreconnect-cs-L38
I am planning on calling ForceReconnect if RedisTimeoutException occurs continuously for more than 5 minutes, but it says that "Don't call ForceReconnect for Timeouts".
I would like to know the reason for this.
Is there some problem on forcefully reconnecting on RedisTimeoutException?
Most customers find that a single ConnectionMultiplexer is sufficient for their needs. However, if you have high throughput requirements, you may consider slowly increasing the number of connections to see if you get an improvement in throughput. Avoid setting it to an arbitrarily large number of connections as it may not give the desired benefit.
How does one increase the number of connections? The provided static class allows for only a single ConnectionMultiplexer and ForceReconnection of only a single ConnectionMultiplexer...?
Thanks for your help
How does one increase the number of connections?
You would have to create a pool of multiplexers to draw from, but I don't have example code for that currently. ForceReconnect is designed for a single connection multiplexer. If you do create a pool, make sure that each instance in the pool has its own set of variables for the ForceReconnect logic.
"> I would really appreciate a C# Serverless Function example. Nowhere can I find good practice information on that. Especially if I should be using the Lazy pattern as described here. Ideally, the StackExchange.Redis client would be mentioned here in addition to the HttpClient and DocumentClient"
Yes. You should. Here's some code I'm using for a thread-safe singleton that leverages the Lazy (T) model.
"
public class MyRedisConnectorHelper
{private static readonly Lazy<ConnectionMultiplexer> LazyConnection = new Lazy<ConnectionMultiplexer>(() => { string cacheConnection = ConfigurationManager.AppSettings["CacheConnection"].ToString(); return ConnectionMultiplexer.Connect(cacheConnection); }); public static ConnectionMultiplexer Connection { get { return LazyConnection.Value; } } }"
You'll still need to add a config file as described in the Azure documentation. I'm actually using mine for a local Redis server I'm running on Ubuntu, but it works great!
I don't have a serverless Redis example right now. If I work on one later or tomorrow, I'll post it too. Good luck!
It's probably telling you that the value for cacheConnection is null. Just pass the string to the Multiplexer as in this example
I am investigating locking\deadlocking\thread-pool-starvation issues related to this code. I am not a redis expert so some help is appreciated.
Concerns:
- Lazy with sync IO inside:
Connect
hasTask.Wait()
and.Result
- Reconnect closes multiplexer under
lock
whileClose
has .Result inside
#Lazy
If async await code calls Lazy.Value and initialization happens then all thread-pool threads are synchronousely blocked on waiting for initialization. If for any reason intialization stucks or is just a bit slow a service which goes through high-load will be simply paralyzed and thread pool will be forced to grow and add more more and threads which all will get stuck on Lazy<T>.Value
. Recommendation to increase ThreadPool.MinThreads is like to try to quench the fire with gasoline.
#Reconnect
I looked into Close implementation and found few cases of .Result inside. Given that close happens under lock the same issue as with Lazy value happens.
The code inside StackExchange.Redis is a mixture of await, .Wait, .Result which is a very dangerous approach. Is the code in this article still recommended to be used in high-loaded production services or maybe there are better alternatives?
I am investigating locking\deadlocking\thread-pool-starvation issues related to this code. I am not redis expert so some help is appreciated.
A C# web application that uses this code intermittently gets killed off and complains of thread pool starvation, so it's very interesting that you are suggesting that this might be a problem caused by the implementation here. I'd be very interested to know the outcome of this.
Our case was like this:
- service is restarted
- after start service receives heavy async-io load spawning many tasks
- the load hangs in Monitor (lock) and Lazy.Value
- thread-pool quickly grows but it does not help - application is still stuck, even tasks unrelated to Redis are stuck because they cannot be completed
Absence of sync-context in modern ASP.NET Core and improved thread pool scaling in .NET does not mean that mixing await and .Result.Wait is now a good idea. It is not and it will never be. From my point of view the code in this article should be deprecated or at least rewritten using SemaphoreSlim instead of Lazy and Lock. Also only async-IO may be used (ConnectAsync, CloseAsync).
If I understand the concern correctly, the main problem is that many threads are blocked on the locking at the same time, and the high level solution would be to make the creation of the multiplexer be asynchronous. Is that correct? At some point, it seems like you have to have a lock if you want to ensure only one connection object is created. I suppose you could create multiple then discard the extras but that has its own set of problems it can create (for instance, flooding the server with connection requests, many of which will just get closed).
The problem
Yes, there is a problem with Lazy and Lock.
Here is a simplified scenario for starvation: https://gist.github.com/yapaxi/cfc529bee9bd5361482051d45a62a571.
In a nutshell it has:
- initialization task (get-google) which is pseudo-synchronousely waited in Lazy
- loop that adds fake "load" that synchronousely waits for Lazy
- in the end "load" waits for lazy and lazy waits for load to release thread-pool threads
- eventhough thread pool grows new thread-pool threads are not used to finish get-google until some kind of timeout happens (in the test case)
On sunny days get-google finishes in 30 seconds, sometimes in 1 minute, sometimes never. In the example case most likely DNS resolve locks or deadlocks. In real case it is almost impossible to tell why Multiplexer.Connect is slow or near-deadlocks - too many things happen.
In real world 30+ seconds is sometimes enough for cluster to consider a node dead.
A solution
The best way I see is to use awaitable SemaphoreSlim instead of lazy and lock. That will ensure only single Multiplexer is created.
I do not know anything about Redis (I am only a part of outage investigation team) but I can help with multithreading.
Here is one approach: https://gist.github.com/yapaxi/9526ca734c8cc30173276c616324d33b
Be careful:
- I changed timeouts and thresholds for testing purposes
- I do not know how to handle Redis connection timeout properly so I just added timeout tasks
- I kept logic in reconnected but reduced nesting a bit
- It is hard to guess if you could use the same lock in reconnect because this is an incomplete solution so I decided to make Reconnect non-waiting to completely avoid deadlocks and transfer connection responsibility to GetConnection() method: reconnect only nullifies the state;
Hope this helps. Questions are welcome.
In the Redis-LazyReconnect.cs, at line number 57, inside the lock, we should re-read the value of previousReconnect.
elapsedSinceLastReconnect = utcNow - previousReconnect;
As this thread might have waited to acquire a lock and some other thread might have got it first and reconnected successfully, thus updating previousReconnect. And we are reading stale value in all other threads for this variable.
Hello, is there an example on how to consume this ForceReconnect functionality? I also have a issue and I've evaluated every single item and I know that I must be doing something wrong in setting up my connection: https://gist.github.com/anonymous/c7c741b8a97e2c77e4bc3eeb9d4781ec
Do you have any recommendations?