Fix 'Unhandled error in Deferred' (#12089)

* Fix 'Unhandled error in Deferred'

Fixes a CRITICAL "Unhandled error in Deferred" log message which happened when
a function wrapped with `@cachedList` failed

* Minor optimisation to cachedListDescriptor

we can avoid re-using `missing`, which saves looking up entries in
`deferreds_map`, and means we don't need to copy it.

* Improve type annotation on CachedListDescriptor
pull/12111/head
Richard van der Hoff 2022-03-01 09:51:38 +00:00 committed by GitHub
parent 9d11fee8f2
commit 5458eb8551
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
3 changed files with 37 additions and 36 deletions

1
changelog.d/12089.bugfix Normal file
View File

@ -0,0 +1 @@
Fix occasional 'Unhandled error in Deferred' error message.

View File

@ -18,6 +18,7 @@ import inspect
import logging import logging
from typing import ( from typing import (
Any, Any,
Awaitable,
Callable, Callable,
Dict, Dict,
Generic, Generic,
@ -346,15 +347,15 @@ class DeferredCacheListDescriptor(_CacheDescriptorBase):
"""Wraps an existing cache to support bulk fetching of keys. """Wraps an existing cache to support bulk fetching of keys.
Given an iterable of keys it looks in the cache to find any hits, then passes Given an iterable of keys it looks in the cache to find any hits, then passes
the tuple of missing keys to the wrapped function. the set of missing keys to the wrapped function.
Once wrapped, the function returns a Deferred which resolves to the list Once wrapped, the function returns a Deferred which resolves to a Dict mapping from
of results. input key to output value.
""" """
def __init__( def __init__(
self, self,
orig: Callable[..., Any], orig: Callable[..., Awaitable[Dict]],
cached_method_name: str, cached_method_name: str,
list_name: str, list_name: str,
num_args: Optional[int] = None, num_args: Optional[int] = None,
@ -385,13 +386,13 @@ class DeferredCacheListDescriptor(_CacheDescriptorBase):
def __get__( def __get__(
self, obj: Optional[Any], objtype: Optional[Type] = None self, obj: Optional[Any], objtype: Optional[Type] = None
) -> Callable[..., Any]: ) -> Callable[..., "defer.Deferred[Dict[Hashable, Any]]"]:
cached_method = getattr(obj, self.cached_method_name) cached_method = getattr(obj, self.cached_method_name)
cache: DeferredCache[CacheKey, Any] = cached_method.cache cache: DeferredCache[CacheKey, Any] = cached_method.cache
num_args = cached_method.num_args num_args = cached_method.num_args
@functools.wraps(self.orig) @functools.wraps(self.orig)
def wrapped(*args: Any, **kwargs: Any) -> Any: def wrapped(*args: Any, **kwargs: Any) -> "defer.Deferred[Dict]":
# If we're passed a cache_context then we'll want to call its # If we're passed a cache_context then we'll want to call its
# invalidate() whenever we are invalidated # invalidate() whenever we are invalidated
invalidate_callback = kwargs.pop("on_invalidate", None) invalidate_callback = kwargs.pop("on_invalidate", None)
@ -444,39 +445,38 @@ class DeferredCacheListDescriptor(_CacheDescriptorBase):
deferred: "defer.Deferred[Any]" = defer.Deferred() deferred: "defer.Deferred[Any]" = defer.Deferred()
deferreds_map[arg] = deferred deferreds_map[arg] = deferred
key = arg_to_cache_key(arg) key = arg_to_cache_key(arg)
cache.set(key, deferred, callback=invalidate_callback) cached_defers.append(
cache.set(key, deferred, callback=invalidate_callback)
)
def complete_all(res: Dict[Hashable, Any]) -> None: def complete_all(res: Dict[Hashable, Any]) -> None:
# the wrapped function has completed. It returns a # the wrapped function has completed. It returns a dict.
# a dict. We can now resolve the observable deferreds in # We can now update our own result map, and then resolve the
# the cache and update our own result map. # observable deferreds in the cache.
for e in missing: for e, d1 in deferreds_map.items():
val = res.get(e, None) val = res.get(e, None)
deferreds_map[e].callback(val) # make sure we update the results map before running the
# deferreds, because as soon as we run the last deferred, the
# gatherResults() below will complete and return the result
# dict to our caller.
results[e] = val results[e] = val
d1.callback(val)
def errback(f: Failure) -> Failure: def errback_all(f: Failure) -> None:
# the wrapped function has failed. Invalidate any cache # the wrapped function has failed. Propagate the failure into
# entries we're supposed to be populating, and fail # the cache, which will invalidate the entry, and cause the
# their deferreds. # relevant cached_deferreds to fail, which will propagate the
for e in missing: # failure to our caller.
key = arg_to_cache_key(e) for d1 in deferreds_map.values():
cache.invalidate(key) d1.errback(f)
deferreds_map[e].errback(f)
# return the failure, to propagate to our caller.
return f
args_to_call = dict(arg_dict) args_to_call = dict(arg_dict)
# copy the missing set before sending it to the callee, to guard against args_to_call[self.list_name] = missing
# modification.
args_to_call[self.list_name] = tuple(missing)
cached_defers.append( # dispatch the call, and attach the two handlers
defer.maybeDeferred( defer.maybeDeferred(
preserve_fn(self.orig), **args_to_call preserve_fn(self.orig), **args_to_call
).addCallbacks(complete_all, errback) ).addCallbacks(complete_all, errback_all)
)
if cached_defers: if cached_defers:
d = defer.gatherResults(cached_defers, consumeErrors=True).addCallbacks( d = defer.gatherResults(cached_defers, consumeErrors=True).addCallbacks(

View File

@ -673,14 +673,14 @@ class CachedListDescriptorTestCase(unittest.TestCase):
self.assertEqual(current_context(), SENTINEL_CONTEXT) self.assertEqual(current_context(), SENTINEL_CONTEXT)
r = yield d1 r = yield d1
self.assertEqual(current_context(), c1) self.assertEqual(current_context(), c1)
obj.mock.assert_called_once_with((10, 20), 2) obj.mock.assert_called_once_with({10, 20}, 2)
self.assertEqual(r, {10: "fish", 20: "chips"}) self.assertEqual(r, {10: "fish", 20: "chips"})
obj.mock.reset_mock() obj.mock.reset_mock()
# a call with different params should call the mock again # a call with different params should call the mock again
obj.mock.return_value = {30: "peas"} obj.mock.return_value = {30: "peas"}
r = yield obj.list_fn([20, 30], 2) r = yield obj.list_fn([20, 30], 2)
obj.mock.assert_called_once_with((30,), 2) obj.mock.assert_called_once_with({30}, 2)
self.assertEqual(r, {20: "chips", 30: "peas"}) self.assertEqual(r, {20: "chips", 30: "peas"})
obj.mock.reset_mock() obj.mock.reset_mock()
@ -701,7 +701,7 @@ class CachedListDescriptorTestCase(unittest.TestCase):
obj.mock.return_value = {40: "gravy"} obj.mock.return_value = {40: "gravy"}
iterable = (x for x in [10, 40, 40]) iterable = (x for x in [10, 40, 40])
r = yield obj.list_fn(iterable, 2) r = yield obj.list_fn(iterable, 2)
obj.mock.assert_called_once_with((40,), 2) obj.mock.assert_called_once_with({40}, 2)
self.assertEqual(r, {10: "fish", 40: "gravy"}) self.assertEqual(r, {10: "fish", 40: "gravy"})
def test_concurrent_lookups(self): def test_concurrent_lookups(self):
@ -729,7 +729,7 @@ class CachedListDescriptorTestCase(unittest.TestCase):
d3 = obj.list_fn([10]) d3 = obj.list_fn([10])
# the mock should have been called exactly once # the mock should have been called exactly once
obj.mock.assert_called_once_with((10,)) obj.mock.assert_called_once_with({10})
obj.mock.reset_mock() obj.mock.reset_mock()
# ... and none of the calls should yet be complete # ... and none of the calls should yet be complete
@ -771,7 +771,7 @@ class CachedListDescriptorTestCase(unittest.TestCase):
# cache miss # cache miss
obj.mock.return_value = {10: "fish", 20: "chips"} obj.mock.return_value = {10: "fish", 20: "chips"}
r1 = yield obj.list_fn([10, 20], 2, on_invalidate=invalidate0) r1 = yield obj.list_fn([10, 20], 2, on_invalidate=invalidate0)
obj.mock.assert_called_once_with((10, 20), 2) obj.mock.assert_called_once_with({10, 20}, 2)
self.assertEqual(r1, {10: "fish", 20: "chips"}) self.assertEqual(r1, {10: "fish", 20: "chips"})
obj.mock.reset_mock() obj.mock.reset_mock()