f5c93fc993
This fixes #8518 by adding a conditional check on `SyncResult` in a function when `prev_stream_token == current_stream_token`, as a sanity check. In `CachedResponse.set.<remove>()`, the result is immediately popped from the cache if the conditional function returns "false". This prevents the caching of a timed-out `SyncResult` (that has `next_key` as the stream key that produced that `SyncResult`). The cache is prevented from returning a `SyncResult` that makes the client request the same stream key over and over again, effectively making it stuck in a loop of requesting and getting a response immediately for as long as the cache keeps those values. Signed-off-by: Jonathan de Jong <jonathan@automatia.nl> |
||
---|---|---|
.. | ||
__init__.py | ||
cached_call.py | ||
deferred_cache.py | ||
descriptors.py | ||
dictionary_cache.py | ||
expiringcache.py | ||
lrucache.py | ||
response_cache.py | ||
stream_change_cache.py | ||
treecache.py | ||
ttlcache.py |