2017-03-22 14:54:20 +01:00
|
|
|
# Copyright 2016 OpenMarket Ltd
|
2018-04-05 17:24:04 +02:00
|
|
|
# Copyright 2018 New Vector Ltd
|
2017-03-22 14:54:20 +01:00
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
2017-03-30 14:22:24 +02:00
|
|
|
import logging
|
2020-10-16 22:32:52 +02:00
|
|
|
from typing import Set
|
2021-04-09 19:44:38 +02:00
|
|
|
from unittest import mock
|
2018-07-09 08:09:20 +02:00
|
|
|
|
|
|
|
from twisted.internet import defer, reactor
|
2021-11-04 15:45:34 +01:00
|
|
|
from twisted.internet.defer import Deferred
|
2018-07-09 08:09:20 +02:00
|
|
|
|
2017-03-30 14:22:24 +02:00
|
|
|
from synapse.api.errors import SynapseError
|
2019-07-03 16:07:04 +02:00
|
|
|
from synapse.logging.context import (
|
2020-03-24 15:45:33 +01:00
|
|
|
SENTINEL_CONTEXT,
|
2019-07-03 16:07:04 +02:00
|
|
|
LoggingContext,
|
|
|
|
PreserveLoggingContext,
|
2020-03-24 15:45:33 +01:00
|
|
|
current_context,
|
2019-07-03 16:07:04 +02:00
|
|
|
make_deferred_yieldable,
|
|
|
|
)
|
2017-03-22 14:54:20 +01:00
|
|
|
from synapse.util.caches import descriptors
|
2020-10-30 12:43:17 +01:00
|
|
|
from synapse.util.caches.descriptors import cached, lru_cache
|
2018-07-09 08:09:20 +02:00
|
|
|
|
2017-03-22 14:54:20 +01:00
|
|
|
from tests import unittest
|
2020-10-30 12:43:17 +01:00
|
|
|
from tests.test_utils import get_awaitable_result
|
2017-03-22 14:54:20 +01:00
|
|
|
|
2017-03-30 14:22:24 +02:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
2017-03-22 14:54:20 +01:00
|
|
|
|
2020-10-30 12:43:17 +01:00
|
|
|
class LruCacheDecoratorTestCase(unittest.TestCase):
|
|
|
|
def test_base(self):
|
|
|
|
class Cls:
|
|
|
|
def __init__(self):
|
|
|
|
self.mock = mock.Mock()
|
|
|
|
|
|
|
|
@lru_cache()
|
|
|
|
def fn(self, arg1, arg2):
|
|
|
|
return self.mock(arg1, arg2)
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
obj.mock.return_value = "fish"
|
|
|
|
r = obj.fn(1, 2)
|
|
|
|
self.assertEqual(r, "fish")
|
|
|
|
obj.mock.assert_called_once_with(1, 2)
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# a call with different params should call the mock again
|
|
|
|
obj.mock.return_value = "chips"
|
|
|
|
r = obj.fn(1, 3)
|
|
|
|
self.assertEqual(r, "chips")
|
|
|
|
obj.mock.assert_called_once_with(1, 3)
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# the two values should now be cached
|
|
|
|
r = obj.fn(1, 2)
|
|
|
|
self.assertEqual(r, "fish")
|
|
|
|
r = obj.fn(1, 3)
|
|
|
|
self.assertEqual(r, "chips")
|
|
|
|
obj.mock.assert_not_called()
|
|
|
|
|
|
|
|
|
2018-07-04 10:35:40 +02:00
|
|
|
def run_on_reactor():
|
|
|
|
d = defer.Deferred()
|
|
|
|
reactor.callLater(0, d.callback, 0)
|
2019-07-03 16:07:04 +02:00
|
|
|
return make_deferred_yieldable(d)
|
2018-07-04 10:35:40 +02:00
|
|
|
|
|
|
|
|
2017-03-22 14:54:20 +01:00
|
|
|
class DescriptorTestCase(unittest.TestCase):
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def test_cache(self):
|
2020-09-04 12:54:56 +02:00
|
|
|
class Cls:
|
2017-03-22 14:54:20 +01:00
|
|
|
def __init__(self):
|
|
|
|
self.mock = mock.Mock()
|
|
|
|
|
|
|
|
@descriptors.cached()
|
|
|
|
def fn(self, arg1, arg2):
|
|
|
|
return self.mock(arg1, arg2)
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
|
2019-06-20 11:32:02 +02:00
|
|
|
obj.mock.return_value = "fish"
|
2017-03-22 14:54:20 +01:00
|
|
|
r = yield obj.fn(1, 2)
|
2019-06-20 11:32:02 +02:00
|
|
|
self.assertEqual(r, "fish")
|
2017-03-22 14:54:20 +01:00
|
|
|
obj.mock.assert_called_once_with(1, 2)
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# a call with different params should call the mock again
|
2019-06-20 11:32:02 +02:00
|
|
|
obj.mock.return_value = "chips"
|
2017-03-22 14:54:20 +01:00
|
|
|
r = yield obj.fn(1, 3)
|
2019-06-20 11:32:02 +02:00
|
|
|
self.assertEqual(r, "chips")
|
2017-03-22 14:54:20 +01:00
|
|
|
obj.mock.assert_called_once_with(1, 3)
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# the two values should now be cached
|
|
|
|
r = yield obj.fn(1, 2)
|
2019-06-20 11:32:02 +02:00
|
|
|
self.assertEqual(r, "fish")
|
2017-03-22 14:54:20 +01:00
|
|
|
r = yield obj.fn(1, 3)
|
2019-06-20 11:32:02 +02:00
|
|
|
self.assertEqual(r, "chips")
|
2017-03-22 14:54:20 +01:00
|
|
|
obj.mock.assert_not_called()
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def test_cache_num_args(self):
|
|
|
|
"""Only the first num_args arguments should matter to the cache"""
|
|
|
|
|
2020-09-04 12:54:56 +02:00
|
|
|
class Cls:
|
2017-03-22 14:54:20 +01:00
|
|
|
def __init__(self):
|
|
|
|
self.mock = mock.Mock()
|
|
|
|
|
|
|
|
@descriptors.cached(num_args=1)
|
|
|
|
def fn(self, arg1, arg2):
|
|
|
|
return self.mock(arg1, arg2)
|
|
|
|
|
|
|
|
obj = Cls()
|
2019-06-20 11:32:02 +02:00
|
|
|
obj.mock.return_value = "fish"
|
2017-03-22 14:54:20 +01:00
|
|
|
r = yield obj.fn(1, 2)
|
2019-06-20 11:32:02 +02:00
|
|
|
self.assertEqual(r, "fish")
|
2017-03-22 14:54:20 +01:00
|
|
|
obj.mock.assert_called_once_with(1, 2)
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# a call with different params should call the mock again
|
2019-06-20 11:32:02 +02:00
|
|
|
obj.mock.return_value = "chips"
|
2017-03-22 14:54:20 +01:00
|
|
|
r = yield obj.fn(2, 3)
|
2019-06-20 11:32:02 +02:00
|
|
|
self.assertEqual(r, "chips")
|
2017-03-22 14:54:20 +01:00
|
|
|
obj.mock.assert_called_once_with(2, 3)
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# the two values should now be cached; we should be able to vary
|
|
|
|
# the second argument and still get the cached result.
|
|
|
|
r = yield obj.fn(1, 4)
|
2019-06-20 11:32:02 +02:00
|
|
|
self.assertEqual(r, "fish")
|
2017-03-22 14:54:20 +01:00
|
|
|
r = yield obj.fn(2, 5)
|
2019-06-20 11:32:02 +02:00
|
|
|
self.assertEqual(r, "chips")
|
2017-03-22 14:54:20 +01:00
|
|
|
obj.mock.assert_not_called()
|
2017-03-30 14:22:24 +02:00
|
|
|
|
2019-07-25 16:59:45 +02:00
|
|
|
def test_cache_with_sync_exception(self):
|
2021-02-16 23:32:34 +01:00
|
|
|
"""If the wrapped function throws synchronously, things should continue to work"""
|
2019-07-25 16:59:45 +02:00
|
|
|
|
2020-09-04 12:54:56 +02:00
|
|
|
class Cls:
|
2019-07-25 16:59:45 +02:00
|
|
|
@cached()
|
|
|
|
def fn(self, arg1):
|
|
|
|
raise SynapseError(100, "mai spoon iz too big!!1")
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
|
|
|
|
# this should fail immediately
|
|
|
|
d = obj.fn(1)
|
|
|
|
self.failureResultOf(d, SynapseError)
|
|
|
|
|
|
|
|
# ... leaving the cache empty
|
|
|
|
self.assertEqual(len(obj.fn.cache.cache), 0)
|
|
|
|
|
|
|
|
# and a second call should result in a second exception
|
|
|
|
d = obj.fn(1)
|
|
|
|
self.failureResultOf(d, SynapseError)
|
|
|
|
|
2020-10-16 22:32:52 +02:00
|
|
|
def test_cache_with_async_exception(self):
|
2021-02-16 23:32:34 +01:00
|
|
|
"""The wrapped function returns a failure"""
|
2020-10-16 22:32:52 +02:00
|
|
|
|
|
|
|
class Cls:
|
|
|
|
result = None
|
|
|
|
call_count = 0
|
|
|
|
|
|
|
|
@cached()
|
|
|
|
def fn(self, arg1):
|
|
|
|
self.call_count += 1
|
|
|
|
return self.result
|
|
|
|
|
|
|
|
obj = Cls()
|
2021-07-13 12:52:58 +02:00
|
|
|
callbacks: Set[str] = set()
|
2020-10-16 22:32:52 +02:00
|
|
|
|
|
|
|
# set off an asynchronous request
|
|
|
|
obj.result = origin_d = defer.Deferred()
|
|
|
|
|
|
|
|
d1 = obj.fn(1, on_invalidate=lambda: callbacks.add("d1"))
|
|
|
|
self.assertFalse(d1.called)
|
|
|
|
|
|
|
|
# a second request should also return a deferred, but should not call the
|
|
|
|
# function itself.
|
|
|
|
d2 = obj.fn(1, on_invalidate=lambda: callbacks.add("d2"))
|
|
|
|
self.assertFalse(d2.called)
|
|
|
|
self.assertEqual(obj.call_count, 1)
|
|
|
|
|
|
|
|
# no callbacks yet
|
|
|
|
self.assertEqual(callbacks, set())
|
|
|
|
|
|
|
|
# the original request fails
|
|
|
|
e = Exception("bzz")
|
|
|
|
origin_d.errback(e)
|
|
|
|
|
|
|
|
# ... which should cause the lookups to fail similarly
|
|
|
|
self.assertIs(self.failureResultOf(d1, Exception).value, e)
|
|
|
|
self.assertIs(self.failureResultOf(d2, Exception).value, e)
|
|
|
|
|
|
|
|
# ... and the callbacks to have been, uh, called.
|
|
|
|
self.assertEqual(callbacks, {"d1", "d2"})
|
|
|
|
|
|
|
|
# ... leaving the cache empty
|
|
|
|
self.assertEqual(len(obj.fn.cache.cache), 0)
|
|
|
|
|
|
|
|
# and a second call should work as normal
|
|
|
|
obj.result = defer.succeed(100)
|
|
|
|
d3 = obj.fn(1)
|
|
|
|
self.assertEqual(self.successResultOf(d3), 100)
|
|
|
|
self.assertEqual(obj.call_count, 2)
|
|
|
|
|
2017-03-30 14:22:24 +02:00
|
|
|
def test_cache_logcontexts(self):
|
|
|
|
"""Check that logcontexts are set and restored correctly when
|
|
|
|
using the cache."""
|
|
|
|
|
|
|
|
complete_lookup = defer.Deferred()
|
|
|
|
|
2020-09-04 12:54:56 +02:00
|
|
|
class Cls:
|
2017-03-30 14:22:24 +02:00
|
|
|
@descriptors.cached()
|
|
|
|
def fn(self, arg1):
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def inner_fn():
|
2019-07-03 16:07:04 +02:00
|
|
|
with PreserveLoggingContext():
|
2017-03-30 14:22:24 +02:00
|
|
|
yield complete_lookup
|
2019-07-23 15:00:55 +02:00
|
|
|
return 1
|
2017-03-30 14:22:24 +02:00
|
|
|
|
|
|
|
return inner_fn()
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def do_lookup():
|
2021-04-20 15:19:00 +02:00
|
|
|
with LoggingContext("c1") as c1:
|
2017-03-30 14:22:24 +02:00
|
|
|
r = yield obj.fn(1)
|
2020-03-24 15:45:33 +01:00
|
|
|
self.assertEqual(current_context(), c1)
|
2019-07-23 15:00:55 +02:00
|
|
|
return r
|
2017-03-30 14:22:24 +02:00
|
|
|
|
|
|
|
def check_result(r):
|
|
|
|
self.assertEqual(r, 1)
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
|
|
|
|
# set off a deferred which will do a cache lookup
|
|
|
|
d1 = do_lookup()
|
2020-03-24 15:45:33 +01:00
|
|
|
self.assertEqual(current_context(), SENTINEL_CONTEXT)
|
2017-03-30 14:22:24 +02:00
|
|
|
d1.addCallback(check_result)
|
|
|
|
|
|
|
|
# and another
|
|
|
|
d2 = do_lookup()
|
2020-03-24 15:45:33 +01:00
|
|
|
self.assertEqual(current_context(), SENTINEL_CONTEXT)
|
2017-03-30 14:22:24 +02:00
|
|
|
d2.addCallback(check_result)
|
|
|
|
|
|
|
|
# let the lookup complete
|
|
|
|
complete_lookup.callback(None)
|
|
|
|
|
|
|
|
return defer.gatherResults([d1, d2])
|
|
|
|
|
|
|
|
def test_cache_logcontexts_with_exception(self):
|
|
|
|
"""Check that the cache sets and restores logcontexts correctly when
|
|
|
|
the lookup function throws an exception"""
|
|
|
|
|
2020-09-04 12:54:56 +02:00
|
|
|
class Cls:
|
2017-03-30 14:22:24 +02:00
|
|
|
@descriptors.cached()
|
|
|
|
def fn(self, arg1):
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def inner_fn():
|
2018-07-04 10:35:40 +02:00
|
|
|
# we want this to behave like an asynchronous function
|
|
|
|
yield run_on_reactor()
|
2017-03-30 14:22:24 +02:00
|
|
|
raise SynapseError(400, "blah")
|
|
|
|
|
|
|
|
return inner_fn()
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def do_lookup():
|
2021-04-20 15:19:00 +02:00
|
|
|
with LoggingContext("c1") as c1:
|
2017-03-30 14:22:24 +02:00
|
|
|
try:
|
2018-07-04 10:35:40 +02:00
|
|
|
d = obj.fn(1)
|
|
|
|
self.assertEqual(
|
2021-02-16 23:32:34 +01:00
|
|
|
current_context(),
|
|
|
|
SENTINEL_CONTEXT,
|
2018-07-04 10:35:40 +02:00
|
|
|
)
|
|
|
|
yield d
|
2017-03-30 14:22:24 +02:00
|
|
|
self.fail("No exception thrown")
|
|
|
|
except SynapseError:
|
|
|
|
pass
|
|
|
|
|
2020-03-24 15:45:33 +01:00
|
|
|
self.assertEqual(current_context(), c1)
|
2017-03-30 14:22:24 +02:00
|
|
|
|
2019-07-25 16:59:45 +02:00
|
|
|
# the cache should now be empty
|
|
|
|
self.assertEqual(len(obj.fn.cache.cache), 0)
|
|
|
|
|
2017-03-30 14:22:24 +02:00
|
|
|
obj = Cls()
|
|
|
|
|
|
|
|
# set off a deferred which will do a cache lookup
|
|
|
|
d1 = do_lookup()
|
2020-03-24 15:45:33 +01:00
|
|
|
self.assertEqual(current_context(), SENTINEL_CONTEXT)
|
2017-03-30 14:22:24 +02:00
|
|
|
|
|
|
|
return d1
|
2017-03-28 12:19:15 +02:00
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def test_cache_default_args(self):
|
2020-09-04 12:54:56 +02:00
|
|
|
class Cls:
|
2017-03-28 12:19:15 +02:00
|
|
|
def __init__(self):
|
|
|
|
self.mock = mock.Mock()
|
|
|
|
|
|
|
|
@descriptors.cached()
|
|
|
|
def fn(self, arg1, arg2=2, arg3=3):
|
|
|
|
return self.mock(arg1, arg2, arg3)
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
|
2019-06-20 11:32:02 +02:00
|
|
|
obj.mock.return_value = "fish"
|
2017-03-28 12:19:15 +02:00
|
|
|
r = yield obj.fn(1, 2, 3)
|
2019-06-20 11:32:02 +02:00
|
|
|
self.assertEqual(r, "fish")
|
2017-03-28 12:19:15 +02:00
|
|
|
obj.mock.assert_called_once_with(1, 2, 3)
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# a call with same params shouldn't call the mock again
|
|
|
|
r = yield obj.fn(1, 2)
|
2019-06-20 11:32:02 +02:00
|
|
|
self.assertEqual(r, "fish")
|
2017-03-28 12:19:15 +02:00
|
|
|
obj.mock.assert_not_called()
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# a call with different params should call the mock again
|
2019-06-20 11:32:02 +02:00
|
|
|
obj.mock.return_value = "chips"
|
2017-03-28 12:19:15 +02:00
|
|
|
r = yield obj.fn(2, 3)
|
2019-06-20 11:32:02 +02:00
|
|
|
self.assertEqual(r, "chips")
|
2017-03-28 12:19:15 +02:00
|
|
|
obj.mock.assert_called_once_with(2, 3, 3)
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# the two values should now be cached
|
|
|
|
r = yield obj.fn(1, 2)
|
2019-06-20 11:32:02 +02:00
|
|
|
self.assertEqual(r, "fish")
|
2017-03-28 12:19:15 +02:00
|
|
|
r = yield obj.fn(2, 3)
|
2019-06-20 11:32:02 +02:00
|
|
|
self.assertEqual(r, "chips")
|
2017-03-28 12:19:15 +02:00
|
|
|
obj.mock.assert_not_called()
|
2018-06-10 23:38:50 +02:00
|
|
|
|
2019-07-25 16:59:45 +02:00
|
|
|
def test_cache_iterable(self):
|
2020-09-04 12:54:56 +02:00
|
|
|
class Cls:
|
2019-07-25 16:59:45 +02:00
|
|
|
def __init__(self):
|
|
|
|
self.mock = mock.Mock()
|
|
|
|
|
|
|
|
@descriptors.cached(iterable=True)
|
|
|
|
def fn(self, arg1, arg2):
|
|
|
|
return self.mock(arg1, arg2)
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
|
|
|
|
obj.mock.return_value = ["spam", "eggs"]
|
|
|
|
r = obj.fn(1, 2)
|
2019-10-30 12:35:46 +01:00
|
|
|
self.assertEqual(r.result, ["spam", "eggs"])
|
2019-07-25 16:59:45 +02:00
|
|
|
obj.mock.assert_called_once_with(1, 2)
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# a call with different params should call the mock again
|
|
|
|
obj.mock.return_value = ["chips"]
|
|
|
|
r = obj.fn(1, 3)
|
2019-10-30 12:35:46 +01:00
|
|
|
self.assertEqual(r.result, ["chips"])
|
2019-07-25 16:59:45 +02:00
|
|
|
obj.mock.assert_called_once_with(1, 3)
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# the two values should now be cached
|
|
|
|
self.assertEqual(len(obj.fn.cache.cache), 3)
|
|
|
|
|
|
|
|
r = obj.fn(1, 2)
|
2019-10-28 14:33:04 +01:00
|
|
|
self.assertEqual(r.result, ["spam", "eggs"])
|
2019-07-25 16:59:45 +02:00
|
|
|
r = obj.fn(1, 3)
|
2019-10-28 14:33:04 +01:00
|
|
|
self.assertEqual(r.result, ["chips"])
|
2019-07-25 16:59:45 +02:00
|
|
|
obj.mock.assert_not_called()
|
|
|
|
|
|
|
|
def test_cache_iterable_with_sync_exception(self):
|
2021-02-16 23:32:34 +01:00
|
|
|
"""If the wrapped function throws synchronously, things should continue to work"""
|
2019-07-25 16:59:45 +02:00
|
|
|
|
2020-09-04 12:54:56 +02:00
|
|
|
class Cls:
|
2019-07-25 16:59:45 +02:00
|
|
|
@descriptors.cached(iterable=True)
|
|
|
|
def fn(self, arg1):
|
|
|
|
raise SynapseError(100, "mai spoon iz too big!!1")
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
|
|
|
|
# this should fail immediately
|
|
|
|
d = obj.fn(1)
|
|
|
|
self.failureResultOf(d, SynapseError)
|
|
|
|
|
|
|
|
# ... leaving the cache empty
|
|
|
|
self.assertEqual(len(obj.fn.cache.cache), 0)
|
|
|
|
|
|
|
|
# and a second call should result in a second exception
|
|
|
|
d = obj.fn(1)
|
|
|
|
self.failureResultOf(d, SynapseError)
|
|
|
|
|
2020-10-30 12:43:17 +01:00
|
|
|
def test_invalidate_cascade(self):
|
|
|
|
"""Invalidations should cascade up through cache contexts"""
|
|
|
|
|
|
|
|
class Cls:
|
|
|
|
@cached(cache_context=True)
|
|
|
|
async def func1(self, key, cache_context):
|
|
|
|
return await self.func2(key, on_invalidate=cache_context.invalidate)
|
|
|
|
|
|
|
|
@cached(cache_context=True)
|
|
|
|
async def func2(self, key, cache_context):
|
|
|
|
return self.func3(key, on_invalidate=cache_context.invalidate)
|
|
|
|
|
|
|
|
@lru_cache(cache_context=True)
|
|
|
|
def func3(self, key, cache_context):
|
|
|
|
self.invalidate = cache_context.invalidate
|
|
|
|
return 42
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
|
|
|
|
top_invalidate = mock.Mock()
|
|
|
|
r = get_awaitable_result(obj.func1("k1", on_invalidate=top_invalidate))
|
|
|
|
self.assertEqual(r, 42)
|
|
|
|
obj.invalidate()
|
|
|
|
top_invalidate.assert_called_once()
|
|
|
|
|
2018-06-10 23:38:50 +02:00
|
|
|
|
2020-10-16 23:31:16 +02:00
|
|
|
class CacheDecoratorTestCase(unittest.HomeserverTestCase):
|
|
|
|
"""More tests for @cached
|
|
|
|
|
|
|
|
The following is a set of tests that got lost in a different file for a while.
|
|
|
|
|
|
|
|
There are probably duplicates of the tests in DescriptorTestCase. Ideally the
|
|
|
|
duplicates would be removed and the two sets of classes combined.
|
|
|
|
"""
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def test_passthrough(self):
|
|
|
|
class A:
|
|
|
|
@cached()
|
|
|
|
def func(self, key):
|
|
|
|
return key
|
|
|
|
|
|
|
|
a = A()
|
|
|
|
|
|
|
|
self.assertEquals((yield a.func("foo")), "foo")
|
|
|
|
self.assertEquals((yield a.func("bar")), "bar")
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def test_hit(self):
|
|
|
|
callcount = [0]
|
|
|
|
|
|
|
|
class A:
|
|
|
|
@cached()
|
|
|
|
def func(self, key):
|
|
|
|
callcount[0] += 1
|
|
|
|
return key
|
|
|
|
|
|
|
|
a = A()
|
|
|
|
yield a.func("foo")
|
|
|
|
|
|
|
|
self.assertEquals(callcount[0], 1)
|
|
|
|
|
|
|
|
self.assertEquals((yield a.func("foo")), "foo")
|
|
|
|
self.assertEquals(callcount[0], 1)
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def test_invalidate(self):
|
|
|
|
callcount = [0]
|
|
|
|
|
|
|
|
class A:
|
|
|
|
@cached()
|
|
|
|
def func(self, key):
|
|
|
|
callcount[0] += 1
|
|
|
|
return key
|
|
|
|
|
|
|
|
a = A()
|
|
|
|
yield a.func("foo")
|
|
|
|
|
|
|
|
self.assertEquals(callcount[0], 1)
|
|
|
|
|
|
|
|
a.func.invalidate(("foo",))
|
|
|
|
|
|
|
|
yield a.func("foo")
|
|
|
|
|
|
|
|
self.assertEquals(callcount[0], 2)
|
|
|
|
|
|
|
|
def test_invalidate_missing(self):
|
|
|
|
class A:
|
|
|
|
@cached()
|
|
|
|
def func(self, key):
|
|
|
|
return key
|
|
|
|
|
|
|
|
A().func.invalidate(("what",))
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def test_max_entries(self):
|
|
|
|
callcount = [0]
|
|
|
|
|
|
|
|
class A:
|
|
|
|
@cached(max_entries=10)
|
|
|
|
def func(self, key):
|
|
|
|
callcount[0] += 1
|
|
|
|
return key
|
|
|
|
|
|
|
|
a = A()
|
|
|
|
|
|
|
|
for k in range(0, 12):
|
|
|
|
yield a.func(k)
|
|
|
|
|
|
|
|
self.assertEquals(callcount[0], 12)
|
|
|
|
|
|
|
|
# There must have been at least 2 evictions, meaning if we calculate
|
|
|
|
# all 12 values again, we must get called at least 2 more times
|
|
|
|
for k in range(0, 12):
|
|
|
|
yield a.func(k)
|
|
|
|
|
|
|
|
self.assertTrue(
|
|
|
|
callcount[0] >= 14, msg="Expected callcount >= 14, got %d" % (callcount[0])
|
|
|
|
)
|
|
|
|
|
|
|
|
def test_prefill(self):
|
|
|
|
callcount = [0]
|
|
|
|
|
|
|
|
d = defer.succeed(123)
|
|
|
|
|
|
|
|
class A:
|
|
|
|
@cached()
|
|
|
|
def func(self, key):
|
|
|
|
callcount[0] += 1
|
|
|
|
return d
|
|
|
|
|
|
|
|
a = A()
|
|
|
|
|
2020-10-16 13:34:55 +02:00
|
|
|
a.func.prefill(("foo",), 456)
|
2020-10-16 23:31:16 +02:00
|
|
|
|
2020-10-16 13:34:55 +02:00
|
|
|
self.assertEquals(a.func("foo").result, 456)
|
2020-10-16 23:31:16 +02:00
|
|
|
self.assertEquals(callcount[0], 0)
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def test_invalidate_context(self):
|
|
|
|
callcount = [0]
|
|
|
|
callcount2 = [0]
|
|
|
|
|
|
|
|
class A:
|
|
|
|
@cached()
|
|
|
|
def func(self, key):
|
|
|
|
callcount[0] += 1
|
|
|
|
return key
|
|
|
|
|
|
|
|
@cached(cache_context=True)
|
|
|
|
def func2(self, key, cache_context):
|
|
|
|
callcount2[0] += 1
|
|
|
|
return self.func(key, on_invalidate=cache_context.invalidate)
|
|
|
|
|
|
|
|
a = A()
|
|
|
|
yield a.func2("foo")
|
|
|
|
|
|
|
|
self.assertEquals(callcount[0], 1)
|
|
|
|
self.assertEquals(callcount2[0], 1)
|
|
|
|
|
|
|
|
a.func.invalidate(("foo",))
|
|
|
|
yield a.func("foo")
|
|
|
|
|
|
|
|
self.assertEquals(callcount[0], 2)
|
|
|
|
self.assertEquals(callcount2[0], 1)
|
|
|
|
|
|
|
|
yield a.func2("foo")
|
|
|
|
|
|
|
|
self.assertEquals(callcount[0], 2)
|
|
|
|
self.assertEquals(callcount2[0], 2)
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def test_eviction_context(self):
|
|
|
|
callcount = [0]
|
|
|
|
callcount2 = [0]
|
|
|
|
|
|
|
|
class A:
|
|
|
|
@cached(max_entries=2)
|
|
|
|
def func(self, key):
|
|
|
|
callcount[0] += 1
|
|
|
|
return key
|
|
|
|
|
|
|
|
@cached(cache_context=True)
|
|
|
|
def func2(self, key, cache_context):
|
|
|
|
callcount2[0] += 1
|
|
|
|
return self.func(key, on_invalidate=cache_context.invalidate)
|
|
|
|
|
|
|
|
a = A()
|
|
|
|
yield a.func2("foo")
|
|
|
|
yield a.func2("foo2")
|
|
|
|
|
|
|
|
self.assertEquals(callcount[0], 2)
|
|
|
|
self.assertEquals(callcount2[0], 2)
|
|
|
|
|
|
|
|
yield a.func2("foo")
|
|
|
|
self.assertEquals(callcount[0], 2)
|
|
|
|
self.assertEquals(callcount2[0], 2)
|
|
|
|
|
|
|
|
yield a.func("foo3")
|
|
|
|
|
|
|
|
self.assertEquals(callcount[0], 3)
|
|
|
|
self.assertEquals(callcount2[0], 2)
|
|
|
|
|
|
|
|
yield a.func2("foo")
|
|
|
|
|
|
|
|
self.assertEquals(callcount[0], 4)
|
|
|
|
self.assertEquals(callcount2[0], 3)
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def test_double_get(self):
|
|
|
|
callcount = [0]
|
|
|
|
callcount2 = [0]
|
|
|
|
|
|
|
|
class A:
|
|
|
|
@cached()
|
|
|
|
def func(self, key):
|
|
|
|
callcount[0] += 1
|
|
|
|
return key
|
|
|
|
|
|
|
|
@cached(cache_context=True)
|
|
|
|
def func2(self, key, cache_context):
|
|
|
|
callcount2[0] += 1
|
|
|
|
return self.func(key, on_invalidate=cache_context.invalidate)
|
|
|
|
|
|
|
|
a = A()
|
|
|
|
a.func2.cache.cache = mock.Mock(wraps=a.func2.cache.cache)
|
|
|
|
|
|
|
|
yield a.func2("foo")
|
|
|
|
|
|
|
|
self.assertEquals(callcount[0], 1)
|
|
|
|
self.assertEquals(callcount2[0], 1)
|
|
|
|
|
|
|
|
a.func2.invalidate(("foo",))
|
2021-05-27 11:33:56 +02:00
|
|
|
self.assertEquals(a.func2.cache.cache.del_multi.call_count, 1)
|
2020-10-16 23:31:16 +02:00
|
|
|
|
|
|
|
yield a.func2("foo")
|
|
|
|
a.func2.invalidate(("foo",))
|
2021-05-27 11:33:56 +02:00
|
|
|
self.assertEquals(a.func2.cache.cache.del_multi.call_count, 2)
|
2020-10-16 23:31:16 +02:00
|
|
|
|
|
|
|
self.assertEquals(callcount[0], 1)
|
|
|
|
self.assertEquals(callcount2[0], 2)
|
|
|
|
|
|
|
|
a.func.invalidate(("foo",))
|
2021-05-27 11:33:56 +02:00
|
|
|
self.assertEquals(a.func2.cache.cache.del_multi.call_count, 3)
|
2020-10-16 23:31:16 +02:00
|
|
|
yield a.func("foo")
|
|
|
|
|
|
|
|
self.assertEquals(callcount[0], 2)
|
|
|
|
self.assertEquals(callcount2[0], 2)
|
|
|
|
|
|
|
|
yield a.func2("foo")
|
|
|
|
|
|
|
|
self.assertEquals(callcount[0], 2)
|
|
|
|
self.assertEquals(callcount2[0], 3)
|
|
|
|
|
|
|
|
|
2018-06-10 23:38:50 +02:00
|
|
|
class CachedListDescriptorTestCase(unittest.TestCase):
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def test_cache(self):
|
2020-09-04 12:54:56 +02:00
|
|
|
class Cls:
|
2018-06-10 23:38:50 +02:00
|
|
|
def __init__(self):
|
|
|
|
self.mock = mock.Mock()
|
|
|
|
|
|
|
|
@descriptors.cached()
|
|
|
|
def fn(self, arg1, arg2):
|
|
|
|
pass
|
|
|
|
|
2020-08-19 13:09:07 +02:00
|
|
|
@descriptors.cachedList("fn", "args1")
|
|
|
|
async def list_fn(self, args1, arg2):
|
2021-04-08 14:01:14 +02:00
|
|
|
assert current_context().name == "c1"
|
2018-06-10 23:38:50 +02:00
|
|
|
# we want this to behave like an asynchronous function
|
2020-08-19 13:09:07 +02:00
|
|
|
await run_on_reactor()
|
2021-04-08 14:01:14 +02:00
|
|
|
assert current_context().name == "c1"
|
2019-07-23 15:00:55 +02:00
|
|
|
return self.mock(args1, arg2)
|
2018-06-10 23:38:50 +02:00
|
|
|
|
2021-04-08 14:01:14 +02:00
|
|
|
with LoggingContext("c1") as c1:
|
2018-06-10 23:38:50 +02:00
|
|
|
obj = Cls()
|
2019-06-20 11:32:02 +02:00
|
|
|
obj.mock.return_value = {10: "fish", 20: "chips"}
|
2021-05-14 12:12:36 +02:00
|
|
|
|
|
|
|
# start the lookup off
|
2018-06-10 23:38:50 +02:00
|
|
|
d1 = obj.list_fn([10, 20], 2)
|
2020-03-24 15:45:33 +01:00
|
|
|
self.assertEqual(current_context(), SENTINEL_CONTEXT)
|
2018-06-10 23:38:50 +02:00
|
|
|
r = yield d1
|
2020-03-24 15:45:33 +01:00
|
|
|
self.assertEqual(current_context(), c1)
|
2021-05-14 12:12:36 +02:00
|
|
|
obj.mock.assert_called_once_with((10, 20), 2)
|
2019-06-20 11:32:02 +02:00
|
|
|
self.assertEqual(r, {10: "fish", 20: "chips"})
|
2018-06-10 23:38:50 +02:00
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# a call with different params should call the mock again
|
2019-06-20 11:32:02 +02:00
|
|
|
obj.mock.return_value = {30: "peas"}
|
2018-06-10 23:38:50 +02:00
|
|
|
r = yield obj.list_fn([20, 30], 2)
|
2021-05-14 12:12:36 +02:00
|
|
|
obj.mock.assert_called_once_with((30,), 2)
|
2019-06-20 11:32:02 +02:00
|
|
|
self.assertEqual(r, {20: "chips", 30: "peas"})
|
2018-06-10 23:38:50 +02:00
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# all the values should now be cached
|
|
|
|
r = yield obj.fn(10, 2)
|
2019-06-20 11:32:02 +02:00
|
|
|
self.assertEqual(r, "fish")
|
2018-06-10 23:38:50 +02:00
|
|
|
r = yield obj.fn(20, 2)
|
2019-06-20 11:32:02 +02:00
|
|
|
self.assertEqual(r, "chips")
|
2018-06-10 23:38:50 +02:00
|
|
|
r = yield obj.fn(30, 2)
|
2019-06-20 11:32:02 +02:00
|
|
|
self.assertEqual(r, "peas")
|
2018-06-10 23:38:50 +02:00
|
|
|
r = yield obj.list_fn([10, 20, 30], 2)
|
|
|
|
obj.mock.assert_not_called()
|
2019-06-20 11:32:02 +02:00
|
|
|
self.assertEqual(r, {10: "fish", 20: "chips", 30: "peas"})
|
2018-07-27 17:17:17 +02:00
|
|
|
|
2021-05-14 12:12:36 +02:00
|
|
|
# we should also be able to use a (single-use) iterable, and should
|
|
|
|
# deduplicate the keys
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
obj.mock.return_value = {40: "gravy"}
|
|
|
|
iterable = (x for x in [10, 40, 40])
|
|
|
|
r = yield obj.list_fn(iterable, 2)
|
|
|
|
obj.mock.assert_called_once_with((40,), 2)
|
|
|
|
self.assertEqual(r, {10: "fish", 40: "gravy"})
|
|
|
|
|
2021-11-04 15:45:34 +01:00
|
|
|
def test_concurrent_lookups(self):
|
|
|
|
"""All concurrent lookups should get the same result"""
|
|
|
|
|
|
|
|
class Cls:
|
|
|
|
def __init__(self):
|
|
|
|
self.mock = mock.Mock()
|
|
|
|
|
|
|
|
@descriptors.cached()
|
|
|
|
def fn(self, arg1):
|
|
|
|
pass
|
|
|
|
|
|
|
|
@descriptors.cachedList("fn", "args1")
|
|
|
|
def list_fn(self, args1) -> "Deferred[dict]":
|
|
|
|
return self.mock(args1)
|
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
deferred_result = Deferred()
|
|
|
|
obj.mock.return_value = deferred_result
|
|
|
|
|
|
|
|
# start off several concurrent lookups of the same key
|
|
|
|
d1 = obj.list_fn([10])
|
|
|
|
d2 = obj.list_fn([10])
|
|
|
|
d3 = obj.list_fn([10])
|
|
|
|
|
|
|
|
# the mock should have been called exactly once
|
|
|
|
obj.mock.assert_called_once_with((10,))
|
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# ... and none of the calls should yet be complete
|
|
|
|
self.assertFalse(d1.called)
|
|
|
|
self.assertFalse(d2.called)
|
|
|
|
self.assertFalse(d3.called)
|
|
|
|
|
|
|
|
# complete the lookup. @cachedList functions need to complete with a map
|
|
|
|
# of input->result
|
|
|
|
deferred_result.callback({10: "peas"})
|
|
|
|
|
|
|
|
# ... which should give the right result to all the callers
|
|
|
|
self.assertEqual(self.successResultOf(d1), {10: "peas"})
|
|
|
|
self.assertEqual(self.successResultOf(d2), {10: "peas"})
|
|
|
|
self.assertEqual(self.successResultOf(d3), {10: "peas"})
|
|
|
|
|
2018-07-27 17:17:17 +02:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def test_invalidate(self):
|
|
|
|
"""Make sure that invalidation callbacks are called."""
|
2018-08-10 15:54:09 +02:00
|
|
|
|
2020-09-04 12:54:56 +02:00
|
|
|
class Cls:
|
2018-07-27 17:17:17 +02:00
|
|
|
def __init__(self):
|
|
|
|
self.mock = mock.Mock()
|
|
|
|
|
|
|
|
@descriptors.cached()
|
|
|
|
def fn(self, arg1, arg2):
|
|
|
|
pass
|
|
|
|
|
2020-08-19 13:09:07 +02:00
|
|
|
@descriptors.cachedList("fn", "args1")
|
|
|
|
async def list_fn(self, args1, arg2):
|
2018-07-27 17:17:17 +02:00
|
|
|
# we want this to behave like an asynchronous function
|
2020-08-19 13:09:07 +02:00
|
|
|
await run_on_reactor()
|
2019-07-23 15:00:55 +02:00
|
|
|
return self.mock(args1, arg2)
|
2018-07-27 17:17:17 +02:00
|
|
|
|
|
|
|
obj = Cls()
|
|
|
|
invalidate0 = mock.Mock()
|
|
|
|
invalidate1 = mock.Mock()
|
|
|
|
|
|
|
|
# cache miss
|
2019-06-20 11:32:02 +02:00
|
|
|
obj.mock.return_value = {10: "fish", 20: "chips"}
|
2018-07-27 17:17:17 +02:00
|
|
|
r1 = yield obj.list_fn([10, 20], 2, on_invalidate=invalidate0)
|
2021-05-14 12:12:36 +02:00
|
|
|
obj.mock.assert_called_once_with((10, 20), 2)
|
2019-06-20 11:32:02 +02:00
|
|
|
self.assertEqual(r1, {10: "fish", 20: "chips"})
|
2018-07-27 17:17:17 +02:00
|
|
|
obj.mock.reset_mock()
|
|
|
|
|
|
|
|
# cache hit
|
|
|
|
r2 = yield obj.list_fn([10, 20], 2, on_invalidate=invalidate1)
|
|
|
|
obj.mock.assert_not_called()
|
2019-06-20 11:32:02 +02:00
|
|
|
self.assertEqual(r2, {10: "fish", 20: "chips"})
|
2018-07-27 17:17:17 +02:00
|
|
|
|
|
|
|
invalidate0.assert_not_called()
|
|
|
|
invalidate1.assert_not_called()
|
|
|
|
|
|
|
|
# now if we invalidate the keys, both invalidations should get called
|
|
|
|
obj.fn.invalidate((10, 2))
|
|
|
|
invalidate0.assert_called_once()
|
|
|
|
invalidate1.assert_called_once()
|