1206 lines
		
	
	
		
			43 KiB
		
	
	
	
		
			Python
		
	
	
			
		
		
	
	
			1206 lines
		
	
	
		
			43 KiB
		
	
	
	
		
			Python
		
	
	
| # Copyright 2018 New Vector Ltd
 | ||
| #
 | ||
| # Licensed under the Apache License, Version 2.0 (the 'License');
 | ||
| # you may not use this file except in compliance with the License.
 | ||
| # You may obtain a copy of the License at
 | ||
| #
 | ||
| #     http://www.apache.org/licenses/LICENSE-2.0
 | ||
| #
 | ||
| # Unless required by applicable law or agreed to in writing, software
 | ||
| # distributed under the License is distributed on an 'AS IS' BASIS,
 | ||
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | ||
| # See the License for the specific language governing permissions and
 | ||
| # limitations under the License.
 | ||
| 
 | ||
| import datetime
 | ||
| from typing import Dict, List, Tuple, Union
 | ||
| 
 | ||
| import attr
 | ||
| from parameterized import parameterized
 | ||
| 
 | ||
| from twisted.test.proto_helpers import MemoryReactor
 | ||
| 
 | ||
| from synapse.api.constants import EventTypes
 | ||
| from synapse.api.room_versions import (
 | ||
|     KNOWN_ROOM_VERSIONS,
 | ||
|     EventFormatVersions,
 | ||
|     RoomVersion,
 | ||
| )
 | ||
| from synapse.events import _EventInternalMetadata
 | ||
| from synapse.rest import admin
 | ||
| from synapse.rest.client import login, room
 | ||
| from synapse.server import HomeServer
 | ||
| from synapse.storage.database import LoggingTransaction
 | ||
| from synapse.types import JsonDict
 | ||
| from synapse.util import Clock, json_encoder
 | ||
| 
 | ||
| import tests.unittest
 | ||
| import tests.utils
 | ||
| 
 | ||
| 
 | ||
| @attr.s(auto_attribs=True, frozen=True, slots=True)
 | ||
| class _BackfillSetupInfo:
 | ||
|     room_id: str
 | ||
|     depth_map: Dict[str, int]
 | ||
| 
 | ||
| 
 | ||
| class EventFederationWorkerStoreTestCase(tests.unittest.HomeserverTestCase):
 | ||
|     servlets = [
 | ||
|         admin.register_servlets,
 | ||
|         room.register_servlets,
 | ||
|         login.register_servlets,
 | ||
|     ]
 | ||
| 
 | ||
|     def prepare(self, reactor: MemoryReactor, clock: Clock, hs: HomeServer) -> None:
 | ||
|         self.store = hs.get_datastores().main
 | ||
| 
 | ||
|     def test_get_prev_events_for_room(self):
 | ||
|         room_id = "@ROOM:local"
 | ||
| 
 | ||
|         # add a bunch of events and hashes to act as forward extremities
 | ||
|         def insert_event(txn, i):
 | ||
|             event_id = "$event_%i:local" % i
 | ||
| 
 | ||
|             txn.execute(
 | ||
|                 (
 | ||
|                     "INSERT INTO events ("
 | ||
|                     "   room_id, event_id, type, depth, topological_ordering,"
 | ||
|                     "   content, processed, outlier, stream_ordering) "
 | ||
|                     "VALUES (?, ?, 'm.test', ?, ?, 'test', ?, ?, ?)"
 | ||
|                 ),
 | ||
|                 (room_id, event_id, i, i, True, False, i),
 | ||
|             )
 | ||
| 
 | ||
|             txn.execute(
 | ||
|                 (
 | ||
|                     "INSERT INTO event_forward_extremities (room_id, event_id) "
 | ||
|                     "VALUES (?, ?)"
 | ||
|                 ),
 | ||
|                 (room_id, event_id),
 | ||
|             )
 | ||
| 
 | ||
|         for i in range(0, 20):
 | ||
|             self.get_success(
 | ||
|                 self.store.db_pool.runInteraction("insert", insert_event, i)
 | ||
|             )
 | ||
| 
 | ||
|         # this should get the last ten
 | ||
|         r = self.get_success(self.store.get_prev_events_for_room(room_id))
 | ||
|         self.assertEqual(10, len(r))
 | ||
|         for i in range(0, 10):
 | ||
|             self.assertEqual("$event_%i:local" % (19 - i), r[i])
 | ||
| 
 | ||
|     def test_get_rooms_with_many_extremities(self):
 | ||
|         room1 = "#room1"
 | ||
|         room2 = "#room2"
 | ||
|         room3 = "#room3"
 | ||
| 
 | ||
|         def insert_event(txn, i, room_id):
 | ||
|             event_id = "$event_%i:local" % i
 | ||
|             txn.execute(
 | ||
|                 (
 | ||
|                     "INSERT INTO event_forward_extremities (room_id, event_id) "
 | ||
|                     "VALUES (?, ?)"
 | ||
|                 ),
 | ||
|                 (room_id, event_id),
 | ||
|             )
 | ||
| 
 | ||
|         for i in range(0, 20):
 | ||
|             self.get_success(
 | ||
|                 self.store.db_pool.runInteraction("insert", insert_event, i, room1)
 | ||
|             )
 | ||
|             self.get_success(
 | ||
|                 self.store.db_pool.runInteraction("insert", insert_event, i, room2)
 | ||
|             )
 | ||
|             self.get_success(
 | ||
|                 self.store.db_pool.runInteraction("insert", insert_event, i, room3)
 | ||
|             )
 | ||
| 
 | ||
|         # Test simple case
 | ||
|         r = self.get_success(self.store.get_rooms_with_many_extremities(5, 5, []))
 | ||
|         self.assertEqual(len(r), 3)
 | ||
| 
 | ||
|         # Does filter work?
 | ||
| 
 | ||
|         r = self.get_success(self.store.get_rooms_with_many_extremities(5, 5, [room1]))
 | ||
|         self.assertTrue(room2 in r)
 | ||
|         self.assertTrue(room3 in r)
 | ||
|         self.assertEqual(len(r), 2)
 | ||
| 
 | ||
|         r = self.get_success(
 | ||
|             self.store.get_rooms_with_many_extremities(5, 5, [room1, room2])
 | ||
|         )
 | ||
|         self.assertEqual(r, [room3])
 | ||
| 
 | ||
|         # Does filter and limit work?
 | ||
| 
 | ||
|         r = self.get_success(self.store.get_rooms_with_many_extremities(5, 1, [room1]))
 | ||
|         self.assertTrue(r == [room2] or r == [room3])
 | ||
| 
 | ||
|     def _setup_auth_chain(self, use_chain_cover_index: bool) -> str:
 | ||
|         room_id = "@ROOM:local"
 | ||
| 
 | ||
|         # The silly auth graph we use to test the auth difference algorithm,
 | ||
|         # where the top are the most recent events.
 | ||
|         #
 | ||
|         #   A   B
 | ||
|         #    \ /
 | ||
|         #  D  E
 | ||
|         #  \  |
 | ||
|         #   ` F   C
 | ||
|         #     |  /|
 | ||
|         #     G ´ |
 | ||
|         #     | \ |
 | ||
|         #     H   I
 | ||
|         #     |   |
 | ||
|         #     K   J
 | ||
| 
 | ||
|         auth_graph = {
 | ||
|             "a": ["e"],
 | ||
|             "b": ["e"],
 | ||
|             "c": ["g", "i"],
 | ||
|             "d": ["f"],
 | ||
|             "e": ["f"],
 | ||
|             "f": ["g"],
 | ||
|             "g": ["h", "i"],
 | ||
|             "h": ["k"],
 | ||
|             "i": ["j"],
 | ||
|             "k": [],
 | ||
|             "j": [],
 | ||
|         }
 | ||
| 
 | ||
|         depth_map = {
 | ||
|             "a": 7,
 | ||
|             "b": 7,
 | ||
|             "c": 4,
 | ||
|             "d": 6,
 | ||
|             "e": 6,
 | ||
|             "f": 5,
 | ||
|             "g": 3,
 | ||
|             "h": 2,
 | ||
|             "i": 2,
 | ||
|             "k": 1,
 | ||
|             "j": 1,
 | ||
|         }
 | ||
| 
 | ||
|         # Mark the room as maybe having a cover index.
 | ||
| 
 | ||
|         def store_room(txn):
 | ||
|             self.store.db_pool.simple_insert_txn(
 | ||
|                 txn,
 | ||
|                 "rooms",
 | ||
|                 {
 | ||
|                     "room_id": room_id,
 | ||
|                     "creator": "room_creator_user_id",
 | ||
|                     "is_public": True,
 | ||
|                     "room_version": "6",
 | ||
|                     "has_auth_chain_index": use_chain_cover_index,
 | ||
|                 },
 | ||
|             )
 | ||
| 
 | ||
|         self.get_success(self.store.db_pool.runInteraction("store_room", store_room))
 | ||
| 
 | ||
|         # We rudely fiddle with the appropriate tables directly, as that's much
 | ||
|         # easier than constructing events properly.
 | ||
| 
 | ||
|         def insert_event(txn):
 | ||
|             stream_ordering = 0
 | ||
| 
 | ||
|             for event_id in auth_graph:
 | ||
|                 stream_ordering += 1
 | ||
|                 depth = depth_map[event_id]
 | ||
| 
 | ||
|                 self.store.db_pool.simple_insert_txn(
 | ||
|                     txn,
 | ||
|                     table="events",
 | ||
|                     values={
 | ||
|                         "event_id": event_id,
 | ||
|                         "room_id": room_id,
 | ||
|                         "depth": depth,
 | ||
|                         "topological_ordering": depth,
 | ||
|                         "type": "m.test",
 | ||
|                         "processed": True,
 | ||
|                         "outlier": False,
 | ||
|                         "stream_ordering": stream_ordering,
 | ||
|                     },
 | ||
|                 )
 | ||
| 
 | ||
|             self.hs.datastores.persist_events._persist_event_auth_chain_txn(
 | ||
|                 txn,
 | ||
|                 [
 | ||
|                     FakeEvent(event_id, room_id, auth_graph[event_id])
 | ||
|                     for event_id in auth_graph
 | ||
|                 ],
 | ||
|             )
 | ||
| 
 | ||
|         self.get_success(
 | ||
|             self.store.db_pool.runInteraction(
 | ||
|                 "insert",
 | ||
|                 insert_event,
 | ||
|             )
 | ||
|         )
 | ||
| 
 | ||
|         return room_id
 | ||
| 
 | ||
|     @parameterized.expand([(True,), (False,)])
 | ||
|     def test_auth_chain_ids(self, use_chain_cover_index: bool):
 | ||
|         room_id = self._setup_auth_chain(use_chain_cover_index)
 | ||
| 
 | ||
|         # a and b have the same auth chain.
 | ||
|         auth_chain_ids = self.get_success(self.store.get_auth_chain_ids(room_id, ["a"]))
 | ||
|         self.assertCountEqual(auth_chain_ids, ["e", "f", "g", "h", "i", "j", "k"])
 | ||
|         auth_chain_ids = self.get_success(self.store.get_auth_chain_ids(room_id, ["b"]))
 | ||
|         self.assertCountEqual(auth_chain_ids, ["e", "f", "g", "h", "i", "j", "k"])
 | ||
|         auth_chain_ids = self.get_success(
 | ||
|             self.store.get_auth_chain_ids(room_id, ["a", "b"])
 | ||
|         )
 | ||
|         self.assertCountEqual(auth_chain_ids, ["e", "f", "g", "h", "i", "j", "k"])
 | ||
| 
 | ||
|         auth_chain_ids = self.get_success(self.store.get_auth_chain_ids(room_id, ["c"]))
 | ||
|         self.assertCountEqual(auth_chain_ids, ["g", "h", "i", "j", "k"])
 | ||
| 
 | ||
|         # d and e have the same auth chain.
 | ||
|         auth_chain_ids = self.get_success(self.store.get_auth_chain_ids(room_id, ["d"]))
 | ||
|         self.assertCountEqual(auth_chain_ids, ["f", "g", "h", "i", "j", "k"])
 | ||
|         auth_chain_ids = self.get_success(self.store.get_auth_chain_ids(room_id, ["e"]))
 | ||
|         self.assertCountEqual(auth_chain_ids, ["f", "g", "h", "i", "j", "k"])
 | ||
| 
 | ||
|         auth_chain_ids = self.get_success(self.store.get_auth_chain_ids(room_id, ["f"]))
 | ||
|         self.assertCountEqual(auth_chain_ids, ["g", "h", "i", "j", "k"])
 | ||
| 
 | ||
|         auth_chain_ids = self.get_success(self.store.get_auth_chain_ids(room_id, ["g"]))
 | ||
|         self.assertCountEqual(auth_chain_ids, ["h", "i", "j", "k"])
 | ||
| 
 | ||
|         auth_chain_ids = self.get_success(self.store.get_auth_chain_ids(room_id, ["h"]))
 | ||
|         self.assertEqual(auth_chain_ids, {"k"})
 | ||
| 
 | ||
|         auth_chain_ids = self.get_success(self.store.get_auth_chain_ids(room_id, ["i"]))
 | ||
|         self.assertEqual(auth_chain_ids, {"j"})
 | ||
| 
 | ||
|         # j and k have no parents.
 | ||
|         auth_chain_ids = self.get_success(self.store.get_auth_chain_ids(room_id, ["j"]))
 | ||
|         self.assertEqual(auth_chain_ids, set())
 | ||
|         auth_chain_ids = self.get_success(self.store.get_auth_chain_ids(room_id, ["k"]))
 | ||
|         self.assertEqual(auth_chain_ids, set())
 | ||
| 
 | ||
|         # More complex input sequences.
 | ||
|         auth_chain_ids = self.get_success(
 | ||
|             self.store.get_auth_chain_ids(room_id, ["b", "c", "d"])
 | ||
|         )
 | ||
|         self.assertCountEqual(auth_chain_ids, ["e", "f", "g", "h", "i", "j", "k"])
 | ||
| 
 | ||
|         auth_chain_ids = self.get_success(
 | ||
|             self.store.get_auth_chain_ids(room_id, ["h", "i"])
 | ||
|         )
 | ||
|         self.assertCountEqual(auth_chain_ids, ["k", "j"])
 | ||
| 
 | ||
|         # e gets returned even though include_given is false, but it is in the
 | ||
|         # auth chain of b.
 | ||
|         auth_chain_ids = self.get_success(
 | ||
|             self.store.get_auth_chain_ids(room_id, ["b", "e"])
 | ||
|         )
 | ||
|         self.assertCountEqual(auth_chain_ids, ["e", "f", "g", "h", "i", "j", "k"])
 | ||
| 
 | ||
|         # Test include_given.
 | ||
|         auth_chain_ids = self.get_success(
 | ||
|             self.store.get_auth_chain_ids(room_id, ["i"], include_given=True)
 | ||
|         )
 | ||
|         self.assertCountEqual(auth_chain_ids, ["i", "j"])
 | ||
| 
 | ||
|     @parameterized.expand([(True,), (False,)])
 | ||
|     def test_auth_difference(self, use_chain_cover_index: bool):
 | ||
|         room_id = self._setup_auth_chain(use_chain_cover_index)
 | ||
| 
 | ||
|         # Now actually test that various combinations give the right result:
 | ||
| 
 | ||
|         difference = self.get_success(
 | ||
|             self.store.get_auth_chain_difference(room_id, [{"a"}, {"b"}])
 | ||
|         )
 | ||
|         self.assertSetEqual(difference, {"a", "b"})
 | ||
| 
 | ||
|         difference = self.get_success(
 | ||
|             self.store.get_auth_chain_difference(room_id, [{"a"}, {"b"}, {"c"}])
 | ||
|         )
 | ||
|         self.assertSetEqual(difference, {"a", "b", "c", "e", "f"})
 | ||
| 
 | ||
|         difference = self.get_success(
 | ||
|             self.store.get_auth_chain_difference(room_id, [{"a", "c"}, {"b"}])
 | ||
|         )
 | ||
|         self.assertSetEqual(difference, {"a", "b", "c"})
 | ||
| 
 | ||
|         difference = self.get_success(
 | ||
|             self.store.get_auth_chain_difference(room_id, [{"a", "c"}, {"b", "c"}])
 | ||
|         )
 | ||
|         self.assertSetEqual(difference, {"a", "b"})
 | ||
| 
 | ||
|         difference = self.get_success(
 | ||
|             self.store.get_auth_chain_difference(room_id, [{"a"}, {"b"}, {"d"}])
 | ||
|         )
 | ||
|         self.assertSetEqual(difference, {"a", "b", "d", "e"})
 | ||
| 
 | ||
|         difference = self.get_success(
 | ||
|             self.store.get_auth_chain_difference(room_id, [{"a"}, {"b"}, {"c"}, {"d"}])
 | ||
|         )
 | ||
|         self.assertSetEqual(difference, {"a", "b", "c", "d", "e", "f"})
 | ||
| 
 | ||
|         difference = self.get_success(
 | ||
|             self.store.get_auth_chain_difference(room_id, [{"a"}, {"b"}, {"e"}])
 | ||
|         )
 | ||
|         self.assertSetEqual(difference, {"a", "b"})
 | ||
| 
 | ||
|         difference = self.get_success(
 | ||
|             self.store.get_auth_chain_difference(room_id, [{"a"}])
 | ||
|         )
 | ||
|         self.assertSetEqual(difference, set())
 | ||
| 
 | ||
|     def test_auth_difference_partial_cover(self):
 | ||
|         """Test that we correctly handle rooms where not all events have a chain
 | ||
|         cover calculated. This can happen in some obscure edge cases, including
 | ||
|         during the background update that calculates the chain cover for old
 | ||
|         rooms.
 | ||
|         """
 | ||
| 
 | ||
|         room_id = "@ROOM:local"
 | ||
| 
 | ||
|         # The silly auth graph we use to test the auth difference algorithm,
 | ||
|         # where the top are the most recent events.
 | ||
|         #
 | ||
|         #   A   B
 | ||
|         #    \ /
 | ||
|         #  D  E
 | ||
|         #  \  |
 | ||
|         #   ` F   C
 | ||
|         #     |  /|
 | ||
|         #     G ´ |
 | ||
|         #     | \ |
 | ||
|         #     H   I
 | ||
|         #     |   |
 | ||
|         #     K   J
 | ||
| 
 | ||
|         auth_graph = {
 | ||
|             "a": ["e"],
 | ||
|             "b": ["e"],
 | ||
|             "c": ["g", "i"],
 | ||
|             "d": ["f"],
 | ||
|             "e": ["f"],
 | ||
|             "f": ["g"],
 | ||
|             "g": ["h", "i"],
 | ||
|             "h": ["k"],
 | ||
|             "i": ["j"],
 | ||
|             "k": [],
 | ||
|             "j": [],
 | ||
|         }
 | ||
| 
 | ||
|         depth_map = {
 | ||
|             "a": 7,
 | ||
|             "b": 7,
 | ||
|             "c": 4,
 | ||
|             "d": 6,
 | ||
|             "e": 6,
 | ||
|             "f": 5,
 | ||
|             "g": 3,
 | ||
|             "h": 2,
 | ||
|             "i": 2,
 | ||
|             "k": 1,
 | ||
|             "j": 1,
 | ||
|         }
 | ||
| 
 | ||
|         # We rudely fiddle with the appropriate tables directly, as that's much
 | ||
|         # easier than constructing events properly.
 | ||
| 
 | ||
|         def insert_event(txn):
 | ||
|             # First insert the room and mark it as having a chain cover.
 | ||
|             self.store.db_pool.simple_insert_txn(
 | ||
|                 txn,
 | ||
|                 "rooms",
 | ||
|                 {
 | ||
|                     "room_id": room_id,
 | ||
|                     "creator": "room_creator_user_id",
 | ||
|                     "is_public": True,
 | ||
|                     "room_version": "6",
 | ||
|                     "has_auth_chain_index": True,
 | ||
|                 },
 | ||
|             )
 | ||
| 
 | ||
|             stream_ordering = 0
 | ||
| 
 | ||
|             for event_id in auth_graph:
 | ||
|                 stream_ordering += 1
 | ||
|                 depth = depth_map[event_id]
 | ||
| 
 | ||
|                 self.store.db_pool.simple_insert_txn(
 | ||
|                     txn,
 | ||
|                     table="events",
 | ||
|                     values={
 | ||
|                         "event_id": event_id,
 | ||
|                         "room_id": room_id,
 | ||
|                         "depth": depth,
 | ||
|                         "topological_ordering": depth,
 | ||
|                         "type": "m.test",
 | ||
|                         "processed": True,
 | ||
|                         "outlier": False,
 | ||
|                         "stream_ordering": stream_ordering,
 | ||
|                     },
 | ||
|                 )
 | ||
| 
 | ||
|             # Insert all events apart from 'B'
 | ||
|             self.hs.datastores.persist_events._persist_event_auth_chain_txn(
 | ||
|                 txn,
 | ||
|                 [
 | ||
|                     FakeEvent(event_id, room_id, auth_graph[event_id])
 | ||
|                     for event_id in auth_graph
 | ||
|                     if event_id != "b"
 | ||
|                 ],
 | ||
|             )
 | ||
| 
 | ||
|             # Now we insert the event 'B' without a chain cover, by temporarily
 | ||
|             # pretending the room doesn't have a chain cover.
 | ||
| 
 | ||
|             self.store.db_pool.simple_update_txn(
 | ||
|                 txn,
 | ||
|                 table="rooms",
 | ||
|                 keyvalues={"room_id": room_id},
 | ||
|                 updatevalues={"has_auth_chain_index": False},
 | ||
|             )
 | ||
| 
 | ||
|             self.hs.datastores.persist_events._persist_event_auth_chain_txn(
 | ||
|                 txn,
 | ||
|                 [FakeEvent("b", room_id, auth_graph["b"])],
 | ||
|             )
 | ||
| 
 | ||
|             self.store.db_pool.simple_update_txn(
 | ||
|                 txn,
 | ||
|                 table="rooms",
 | ||
|                 keyvalues={"room_id": room_id},
 | ||
|                 updatevalues={"has_auth_chain_index": True},
 | ||
|             )
 | ||
| 
 | ||
|         self.get_success(
 | ||
|             self.store.db_pool.runInteraction(
 | ||
|                 "insert",
 | ||
|                 insert_event,
 | ||
|             )
 | ||
|         )
 | ||
| 
 | ||
|         # Now actually test that various combinations give the right result:
 | ||
| 
 | ||
|         difference = self.get_success(
 | ||
|             self.store.get_auth_chain_difference(room_id, [{"a"}, {"b"}])
 | ||
|         )
 | ||
|         self.assertSetEqual(difference, {"a", "b"})
 | ||
| 
 | ||
|         difference = self.get_success(
 | ||
|             self.store.get_auth_chain_difference(room_id, [{"a"}, {"b"}, {"c"}])
 | ||
|         )
 | ||
|         self.assertSetEqual(difference, {"a", "b", "c", "e", "f"})
 | ||
| 
 | ||
|         difference = self.get_success(
 | ||
|             self.store.get_auth_chain_difference(room_id, [{"a", "c"}, {"b"}])
 | ||
|         )
 | ||
|         self.assertSetEqual(difference, {"a", "b", "c"})
 | ||
| 
 | ||
|         difference = self.get_success(
 | ||
|             self.store.get_auth_chain_difference(room_id, [{"a", "c"}, {"b", "c"}])
 | ||
|         )
 | ||
|         self.assertSetEqual(difference, {"a", "b"})
 | ||
| 
 | ||
|         difference = self.get_success(
 | ||
|             self.store.get_auth_chain_difference(room_id, [{"a"}, {"b"}, {"d"}])
 | ||
|         )
 | ||
|         self.assertSetEqual(difference, {"a", "b", "d", "e"})
 | ||
| 
 | ||
|         difference = self.get_success(
 | ||
|             self.store.get_auth_chain_difference(room_id, [{"a"}, {"b"}, {"c"}, {"d"}])
 | ||
|         )
 | ||
|         self.assertSetEqual(difference, {"a", "b", "c", "d", "e", "f"})
 | ||
| 
 | ||
|         difference = self.get_success(
 | ||
|             self.store.get_auth_chain_difference(room_id, [{"a"}, {"b"}, {"e"}])
 | ||
|         )
 | ||
|         self.assertSetEqual(difference, {"a", "b"})
 | ||
| 
 | ||
|         difference = self.get_success(
 | ||
|             self.store.get_auth_chain_difference(room_id, [{"a"}])
 | ||
|         )
 | ||
|         self.assertSetEqual(difference, set())
 | ||
| 
 | ||
|     @parameterized.expand(
 | ||
|         [(room_version,) for room_version in KNOWN_ROOM_VERSIONS.values()]
 | ||
|     )
 | ||
|     def test_prune_inbound_federation_queue(self, room_version: RoomVersion):
 | ||
|         """Test that pruning of inbound federation queues work"""
 | ||
| 
 | ||
|         room_id = "some_room_id"
 | ||
| 
 | ||
|         def prev_event_format(prev_event_id: str) -> Union[Tuple[str, dict], str]:
 | ||
|             """Account for differences in prev_events format across room versions"""
 | ||
|             if room_version.event_format == EventFormatVersions.ROOM_V1_V2:
 | ||
|                 return prev_event_id, {}
 | ||
| 
 | ||
|             return prev_event_id
 | ||
| 
 | ||
|         # Insert a bunch of events that all reference the previous one.
 | ||
|         self.get_success(
 | ||
|             self.store.db_pool.simple_insert_many(
 | ||
|                 table="federation_inbound_events_staging",
 | ||
|                 keys=(
 | ||
|                     "origin",
 | ||
|                     "room_id",
 | ||
|                     "received_ts",
 | ||
|                     "event_id",
 | ||
|                     "event_json",
 | ||
|                     "internal_metadata",
 | ||
|                 ),
 | ||
|                 values=[
 | ||
|                     (
 | ||
|                         "some_origin",
 | ||
|                         room_id,
 | ||
|                         0,
 | ||
|                         f"$fake_event_id_{i + 1}",
 | ||
|                         json_encoder.encode(
 | ||
|                             {"prev_events": [prev_event_format(f"$fake_event_id_{i}")]}
 | ||
|                         ),
 | ||
|                         "{}",
 | ||
|                     )
 | ||
|                     for i in range(500)
 | ||
|                 ],
 | ||
|                 desc="test_prune_inbound_federation_queue",
 | ||
|             )
 | ||
|         )
 | ||
| 
 | ||
|         # Calling prune once should return True, i.e. a prune happen. The second
 | ||
|         # time it shouldn't.
 | ||
|         pruned = self.get_success(
 | ||
|             self.store.prune_staged_events_in_room(room_id, room_version)
 | ||
|         )
 | ||
|         self.assertTrue(pruned)
 | ||
| 
 | ||
|         pruned = self.get_success(
 | ||
|             self.store.prune_staged_events_in_room(room_id, room_version)
 | ||
|         )
 | ||
|         self.assertFalse(pruned)
 | ||
| 
 | ||
|         # Assert that we only have a single event left in the queue, and that it
 | ||
|         # is the last one.
 | ||
|         count = self.get_success(
 | ||
|             self.store.db_pool.simple_select_one_onecol(
 | ||
|                 table="federation_inbound_events_staging",
 | ||
|                 keyvalues={"room_id": room_id},
 | ||
|                 retcol="COUNT(*)",
 | ||
|                 desc="test_prune_inbound_federation_queue",
 | ||
|             )
 | ||
|         )
 | ||
|         self.assertEqual(count, 1)
 | ||
| 
 | ||
|         next_staged_event_info = self.get_success(
 | ||
|             self.store.get_next_staged_event_id_for_room(room_id)
 | ||
|         )
 | ||
|         assert next_staged_event_info
 | ||
|         _, event_id = next_staged_event_info
 | ||
|         self.assertEqual(event_id, "$fake_event_id_500")
 | ||
| 
 | ||
|     def _setup_room_for_backfill_tests(self) -> _BackfillSetupInfo:
 | ||
|         """
 | ||
|         Sets up a room with various events and backward extremities to test
 | ||
|         backfill functions against.
 | ||
| 
 | ||
|         Returns:
 | ||
|             _BackfillSetupInfo including the `room_id` to test against and
 | ||
|             `depth_map` of events in the room
 | ||
|         """
 | ||
|         room_id = "!backfill-room-test:some-host"
 | ||
| 
 | ||
|         # The silly graph we use to test grabbing backward extremities,
 | ||
|         # where the top is the oldest events.
 | ||
|         #    1 (oldest)
 | ||
|         #    |
 | ||
|         #    2 ⹁
 | ||
|         #    |  \
 | ||
|         #    |   [b1, b2, b3]
 | ||
|         #    |   |
 | ||
|         #    |   A
 | ||
|         #    |  /
 | ||
|         #    3 {
 | ||
|         #    |  \
 | ||
|         #    |   [b4, b5, b6]
 | ||
|         #    |   |
 | ||
|         #    |   B
 | ||
|         #    |  /
 | ||
|         #    4 ´
 | ||
|         #    |
 | ||
|         #    5 (newest)
 | ||
| 
 | ||
|         event_graph: Dict[str, List[str]] = {
 | ||
|             "1": [],
 | ||
|             "2": ["1"],
 | ||
|             "3": ["2", "A"],
 | ||
|             "4": ["3", "B"],
 | ||
|             "5": ["4"],
 | ||
|             "A": ["b1", "b2", "b3"],
 | ||
|             "b1": ["2"],
 | ||
|             "b2": ["2"],
 | ||
|             "b3": ["2"],
 | ||
|             "B": ["b4", "b5", "b6"],
 | ||
|             "b4": ["3"],
 | ||
|             "b5": ["3"],
 | ||
|             "b6": ["3"],
 | ||
|         }
 | ||
| 
 | ||
|         depth_map: Dict[str, int] = {
 | ||
|             "1": 1,
 | ||
|             "2": 2,
 | ||
|             "b1": 3,
 | ||
|             "b2": 3,
 | ||
|             "b3": 3,
 | ||
|             "A": 4,
 | ||
|             "3": 5,
 | ||
|             "b4": 6,
 | ||
|             "b5": 6,
 | ||
|             "b6": 6,
 | ||
|             "B": 7,
 | ||
|             "4": 8,
 | ||
|             "5": 9,
 | ||
|         }
 | ||
| 
 | ||
|         # The events we have persisted on our server.
 | ||
|         # The rest are events in the room but not backfilled tet.
 | ||
|         our_server_events = {"5", "4", "B", "3", "A"}
 | ||
| 
 | ||
|         complete_event_dict_map: Dict[str, JsonDict] = {}
 | ||
|         stream_ordering = 0
 | ||
|         for (event_id, prev_event_ids) in event_graph.items():
 | ||
|             depth = depth_map[event_id]
 | ||
| 
 | ||
|             complete_event_dict_map[event_id] = {
 | ||
|                 "event_id": event_id,
 | ||
|                 "type": "test_regular_type",
 | ||
|                 "room_id": room_id,
 | ||
|                 "sender": "@sender",
 | ||
|                 "prev_event_ids": prev_event_ids,
 | ||
|                 "auth_event_ids": [],
 | ||
|                 "origin_server_ts": stream_ordering,
 | ||
|                 "depth": depth,
 | ||
|                 "stream_ordering": stream_ordering,
 | ||
|                 "content": {"body": "event" + event_id},
 | ||
|             }
 | ||
| 
 | ||
|             stream_ordering += 1
 | ||
| 
 | ||
|         def populate_db(txn: LoggingTransaction):
 | ||
|             # Insert the room to satisfy the foreign key constraint of
 | ||
|             # `event_failed_pull_attempts`
 | ||
|             self.store.db_pool.simple_insert_txn(
 | ||
|                 txn,
 | ||
|                 "rooms",
 | ||
|                 {
 | ||
|                     "room_id": room_id,
 | ||
|                     "creator": "room_creator_user_id",
 | ||
|                     "is_public": True,
 | ||
|                     "room_version": "6",
 | ||
|                 },
 | ||
|             )
 | ||
| 
 | ||
|             # Insert our server events
 | ||
|             for event_id in our_server_events:
 | ||
|                 event_dict = complete_event_dict_map[event_id]
 | ||
| 
 | ||
|                 self.store.db_pool.simple_insert_txn(
 | ||
|                     txn,
 | ||
|                     table="events",
 | ||
|                     values={
 | ||
|                         "event_id": event_dict.get("event_id"),
 | ||
|                         "type": event_dict.get("type"),
 | ||
|                         "room_id": event_dict.get("room_id"),
 | ||
|                         "depth": event_dict.get("depth"),
 | ||
|                         "topological_ordering": event_dict.get("depth"),
 | ||
|                         "stream_ordering": event_dict.get("stream_ordering"),
 | ||
|                         "processed": True,
 | ||
|                         "outlier": False,
 | ||
|                     },
 | ||
|                 )
 | ||
| 
 | ||
|             # Insert the event edges
 | ||
|             for event_id in our_server_events:
 | ||
|                 for prev_event_id in event_graph[event_id]:
 | ||
|                     self.store.db_pool.simple_insert_txn(
 | ||
|                         txn,
 | ||
|                         table="event_edges",
 | ||
|                         values={
 | ||
|                             "event_id": event_id,
 | ||
|                             "prev_event_id": prev_event_id,
 | ||
|                             "room_id": room_id,
 | ||
|                         },
 | ||
|                     )
 | ||
| 
 | ||
|             # Insert the backward extremities
 | ||
|             prev_events_of_our_events = {
 | ||
|                 prev_event_id
 | ||
|                 for our_server_event in our_server_events
 | ||
|                 for prev_event_id in complete_event_dict_map[our_server_event][
 | ||
|                     "prev_event_ids"
 | ||
|                 ]
 | ||
|             }
 | ||
|             backward_extremities = prev_events_of_our_events - our_server_events
 | ||
|             for backward_extremity in backward_extremities:
 | ||
|                 self.store.db_pool.simple_insert_txn(
 | ||
|                     txn,
 | ||
|                     table="event_backward_extremities",
 | ||
|                     values={
 | ||
|                         "event_id": backward_extremity,
 | ||
|                         "room_id": room_id,
 | ||
|                     },
 | ||
|                 )
 | ||
| 
 | ||
|         self.get_success(
 | ||
|             self.store.db_pool.runInteraction(
 | ||
|                 "_setup_room_for_backfill_tests_populate_db",
 | ||
|                 populate_db,
 | ||
|             )
 | ||
|         )
 | ||
| 
 | ||
|         return _BackfillSetupInfo(room_id=room_id, depth_map=depth_map)
 | ||
| 
 | ||
|     def test_get_backfill_points_in_room(self):
 | ||
|         """
 | ||
|         Test to make sure only backfill points that are older and come before
 | ||
|         the `current_depth` are returned.
 | ||
|         """
 | ||
|         setup_info = self._setup_room_for_backfill_tests()
 | ||
|         room_id = setup_info.room_id
 | ||
|         depth_map = setup_info.depth_map
 | ||
| 
 | ||
|         # Try at "B"
 | ||
|         backfill_points = self.get_success(
 | ||
|             self.store.get_backfill_points_in_room(room_id, depth_map["B"], limit=100)
 | ||
|         )
 | ||
|         backfill_event_ids = [backfill_point[0] for backfill_point in backfill_points]
 | ||
|         self.assertEqual(backfill_event_ids, ["b6", "b5", "b4", "2", "b3", "b2", "b1"])
 | ||
| 
 | ||
|         # Try at "A"
 | ||
|         backfill_points = self.get_success(
 | ||
|             self.store.get_backfill_points_in_room(room_id, depth_map["A"], limit=100)
 | ||
|         )
 | ||
|         backfill_event_ids = [backfill_point[0] for backfill_point in backfill_points]
 | ||
|         # Event "2" has a depth of 2 but is not included here because we only
 | ||
|         # know the approximate depth of 5 from our event "3".
 | ||
|         self.assertListEqual(backfill_event_ids, ["b3", "b2", "b1"])
 | ||
| 
 | ||
|     def test_get_backfill_points_in_room_excludes_events_we_have_attempted(
 | ||
|         self,
 | ||
|     ):
 | ||
|         """
 | ||
|         Test to make sure that events we have attempted to backfill (and within
 | ||
|         backoff timeout duration) do not show up as an event to backfill again.
 | ||
|         """
 | ||
|         setup_info = self._setup_room_for_backfill_tests()
 | ||
|         room_id = setup_info.room_id
 | ||
|         depth_map = setup_info.depth_map
 | ||
| 
 | ||
|         # Record some attempts to backfill these events which will make
 | ||
|         # `get_backfill_points_in_room` exclude them because we
 | ||
|         # haven't passed the backoff interval.
 | ||
|         self.get_success(
 | ||
|             self.store.record_event_failed_pull_attempt(room_id, "b5", "fake cause")
 | ||
|         )
 | ||
|         self.get_success(
 | ||
|             self.store.record_event_failed_pull_attempt(room_id, "b4", "fake cause")
 | ||
|         )
 | ||
|         self.get_success(
 | ||
|             self.store.record_event_failed_pull_attempt(room_id, "b3", "fake cause")
 | ||
|         )
 | ||
|         self.get_success(
 | ||
|             self.store.record_event_failed_pull_attempt(room_id, "b2", "fake cause")
 | ||
|         )
 | ||
| 
 | ||
|         # No time has passed since we attempted to backfill ^
 | ||
| 
 | ||
|         # Try at "B"
 | ||
|         backfill_points = self.get_success(
 | ||
|             self.store.get_backfill_points_in_room(room_id, depth_map["B"], limit=100)
 | ||
|         )
 | ||
|         backfill_event_ids = [backfill_point[0] for backfill_point in backfill_points]
 | ||
|         # Only the backfill points that we didn't record earlier exist here.
 | ||
|         self.assertEqual(backfill_event_ids, ["b6", "2", "b1"])
 | ||
| 
 | ||
|     def test_get_backfill_points_in_room_attempted_event_retry_after_backoff_duration(
 | ||
|         self,
 | ||
|     ):
 | ||
|         """
 | ||
|         Test to make sure after we fake attempt to backfill event "b3" many times,
 | ||
|         we can see retry and see the "b3" again after the backoff timeout duration
 | ||
|         has exceeded.
 | ||
|         """
 | ||
|         setup_info = self._setup_room_for_backfill_tests()
 | ||
|         room_id = setup_info.room_id
 | ||
|         depth_map = setup_info.depth_map
 | ||
| 
 | ||
|         # Record some attempts to backfill these events which will make
 | ||
|         # `get_backfill_points_in_room` exclude them because we
 | ||
|         # haven't passed the backoff interval.
 | ||
|         self.get_success(
 | ||
|             self.store.record_event_failed_pull_attempt(room_id, "b3", "fake cause")
 | ||
|         )
 | ||
|         self.get_success(
 | ||
|             self.store.record_event_failed_pull_attempt(room_id, "b1", "fake cause")
 | ||
|         )
 | ||
|         self.get_success(
 | ||
|             self.store.record_event_failed_pull_attempt(room_id, "b1", "fake cause")
 | ||
|         )
 | ||
|         self.get_success(
 | ||
|             self.store.record_event_failed_pull_attempt(room_id, "b1", "fake cause")
 | ||
|         )
 | ||
|         self.get_success(
 | ||
|             self.store.record_event_failed_pull_attempt(room_id, "b1", "fake cause")
 | ||
|         )
 | ||
| 
 | ||
|         # Now advance time by 2 hours and we should only be able to see "b3"
 | ||
|         # because we have waited long enough for the single attempt (2^1 hours)
 | ||
|         # but we still shouldn't see "b1" because we haven't waited long enough
 | ||
|         # for this many attempts. We didn't do anything to "b2" so it should be
 | ||
|         # visible regardless.
 | ||
|         self.reactor.advance(datetime.timedelta(hours=2).total_seconds())
 | ||
| 
 | ||
|         # Try at "A" and make sure that "b1" is not in the list because we've
 | ||
|         # already attempted many times
 | ||
|         backfill_points = self.get_success(
 | ||
|             self.store.get_backfill_points_in_room(room_id, depth_map["A"], limit=100)
 | ||
|         )
 | ||
|         backfill_event_ids = [backfill_point[0] for backfill_point in backfill_points]
 | ||
|         self.assertEqual(backfill_event_ids, ["b3", "b2"])
 | ||
| 
 | ||
|         # Now advance time by 20 hours (above 2^4 because we made 4 attemps) and
 | ||
|         # see if we can now backfill it
 | ||
|         self.reactor.advance(datetime.timedelta(hours=20).total_seconds())
 | ||
| 
 | ||
|         # Try at "A" again after we advanced enough time and we should see "b3" again
 | ||
|         backfill_points = self.get_success(
 | ||
|             self.store.get_backfill_points_in_room(room_id, depth_map["A"], limit=100)
 | ||
|         )
 | ||
|         backfill_event_ids = [backfill_point[0] for backfill_point in backfill_points]
 | ||
|         self.assertEqual(backfill_event_ids, ["b3", "b2", "b1"])
 | ||
| 
 | ||
|     def test_get_backfill_points_in_room_works_after_many_failed_pull_attempts_that_could_naively_overflow(
 | ||
|         self,
 | ||
|     ) -> None:
 | ||
|         """
 | ||
|         A test that reproduces #13929 (Postgres only).
 | ||
| 
 | ||
|         Test to make sure we can still get backfill points after many failed pull
 | ||
|         attempts that cause us to backoff to the limit. Even if the backoff formula
 | ||
|         would tell us to wait for more seconds than can be expressed in a 32 bit
 | ||
|         signed int.
 | ||
|         """
 | ||
|         setup_info = self._setup_room_for_backfill_tests()
 | ||
|         room_id = setup_info.room_id
 | ||
|         depth_map = setup_info.depth_map
 | ||
| 
 | ||
|         # Pretend that we have tried and failed 10 times to backfill event b1.
 | ||
|         for _ in range(10):
 | ||
|             self.get_success(
 | ||
|                 self.store.record_event_failed_pull_attempt(room_id, "b1", "fake cause")
 | ||
|             )
 | ||
| 
 | ||
|         # If the backoff periods grow without limit:
 | ||
|         # After the first failed attempt, we would have backed off for 1 << 1 = 2 hours.
 | ||
|         # After the second failed attempt we would have backed off for 1 << 2 = 4 hours,
 | ||
|         # so after the 10th failed attempt we should backoff for 1 << 10 == 1024 hours.
 | ||
|         # Wait 1100 hours just so we have a nice round number.
 | ||
|         self.reactor.advance(datetime.timedelta(hours=1100).total_seconds())
 | ||
| 
 | ||
|         # 1024 hours in milliseconds is 1024 * 3600000, which exceeds the largest 32 bit
 | ||
|         # signed integer. The bug we're reproducing is that this overflow causes an
 | ||
|         # error in postgres preventing us from fetching a set of backwards extremities
 | ||
|         # to retry fetching.
 | ||
|         backfill_points = self.get_success(
 | ||
|             self.store.get_backfill_points_in_room(room_id, depth_map["A"], limit=100)
 | ||
|         )
 | ||
| 
 | ||
|         # We should aim to fetch all backoff points: b1's latest backoff period has
 | ||
|         # expired, and we haven't tried the rest.
 | ||
|         backfill_event_ids = [backfill_point[0] for backfill_point in backfill_points]
 | ||
|         self.assertEqual(backfill_event_ids, ["b3", "b2", "b1"])
 | ||
| 
 | ||
|     def _setup_room_for_insertion_backfill_tests(self) -> _BackfillSetupInfo:
 | ||
|         """
 | ||
|         Sets up a room with various insertion event backward extremities to test
 | ||
|         backfill functions against.
 | ||
| 
 | ||
|         Returns:
 | ||
|             _BackfillSetupInfo including the `room_id` to test against and
 | ||
|             `depth_map` of events in the room
 | ||
|         """
 | ||
|         room_id = "!backfill-room-test:some-host"
 | ||
| 
 | ||
|         depth_map: Dict[str, int] = {
 | ||
|             "1": 1,
 | ||
|             "2": 2,
 | ||
|             "insertion_eventA": 3,
 | ||
|             "3": 4,
 | ||
|             "insertion_eventB": 5,
 | ||
|             "4": 6,
 | ||
|             "5": 7,
 | ||
|         }
 | ||
| 
 | ||
|         def populate_db(txn: LoggingTransaction):
 | ||
|             # Insert the room to satisfy the foreign key constraint of
 | ||
|             # `event_failed_pull_attempts`
 | ||
|             self.store.db_pool.simple_insert_txn(
 | ||
|                 txn,
 | ||
|                 "rooms",
 | ||
|                 {
 | ||
|                     "room_id": room_id,
 | ||
|                     "creator": "room_creator_user_id",
 | ||
|                     "is_public": True,
 | ||
|                     "room_version": "6",
 | ||
|                 },
 | ||
|             )
 | ||
| 
 | ||
|             # Insert our server events
 | ||
|             stream_ordering = 0
 | ||
|             for event_id, depth in depth_map.items():
 | ||
|                 self.store.db_pool.simple_insert_txn(
 | ||
|                     txn,
 | ||
|                     table="events",
 | ||
|                     values={
 | ||
|                         "event_id": event_id,
 | ||
|                         "type": EventTypes.MSC2716_INSERTION
 | ||
|                         if event_id.startswith("insertion_event")
 | ||
|                         else "test_regular_type",
 | ||
|                         "room_id": room_id,
 | ||
|                         "depth": depth,
 | ||
|                         "topological_ordering": depth,
 | ||
|                         "stream_ordering": stream_ordering,
 | ||
|                         "processed": True,
 | ||
|                         "outlier": False,
 | ||
|                     },
 | ||
|                 )
 | ||
| 
 | ||
|                 if event_id.startswith("insertion_event"):
 | ||
|                     self.store.db_pool.simple_insert_txn(
 | ||
|                         txn,
 | ||
|                         table="insertion_event_extremities",
 | ||
|                         values={
 | ||
|                             "event_id": event_id,
 | ||
|                             "room_id": room_id,
 | ||
|                         },
 | ||
|                     )
 | ||
| 
 | ||
|                 stream_ordering += 1
 | ||
| 
 | ||
|         self.get_success(
 | ||
|             self.store.db_pool.runInteraction(
 | ||
|                 "_setup_room_for_insertion_backfill_tests_populate_db",
 | ||
|                 populate_db,
 | ||
|             )
 | ||
|         )
 | ||
| 
 | ||
|         return _BackfillSetupInfo(room_id=room_id, depth_map=depth_map)
 | ||
| 
 | ||
|     def test_get_insertion_event_backward_extremities_in_room(self):
 | ||
|         """
 | ||
|         Test to make sure only insertion event backward extremities that are
 | ||
|         older and come before the `current_depth` are returned.
 | ||
|         """
 | ||
|         setup_info = self._setup_room_for_insertion_backfill_tests()
 | ||
|         room_id = setup_info.room_id
 | ||
|         depth_map = setup_info.depth_map
 | ||
| 
 | ||
|         # Try at "insertion_eventB"
 | ||
|         backfill_points = self.get_success(
 | ||
|             self.store.get_insertion_event_backward_extremities_in_room(
 | ||
|                 room_id, depth_map["insertion_eventB"], limit=100
 | ||
|             )
 | ||
|         )
 | ||
|         backfill_event_ids = [backfill_point[0] for backfill_point in backfill_points]
 | ||
|         self.assertEqual(backfill_event_ids, ["insertion_eventB", "insertion_eventA"])
 | ||
| 
 | ||
|         # Try at "insertion_eventA"
 | ||
|         backfill_points = self.get_success(
 | ||
|             self.store.get_insertion_event_backward_extremities_in_room(
 | ||
|                 room_id, depth_map["insertion_eventA"], limit=100
 | ||
|             )
 | ||
|         )
 | ||
|         backfill_event_ids = [backfill_point[0] for backfill_point in backfill_points]
 | ||
|         # Event "2" has a depth of 2 but is not included here because we only
 | ||
|         # know the approximate depth of 5 from our event "3".
 | ||
|         self.assertListEqual(backfill_event_ids, ["insertion_eventA"])
 | ||
| 
 | ||
|     def test_get_insertion_event_backward_extremities_in_room_excludes_events_we_have_attempted(
 | ||
|         self,
 | ||
|     ):
 | ||
|         """
 | ||
|         Test to make sure that insertion events we have attempted to backfill
 | ||
|         (and within backoff timeout duration) do not show up as an event to
 | ||
|         backfill again.
 | ||
|         """
 | ||
|         setup_info = self._setup_room_for_insertion_backfill_tests()
 | ||
|         room_id = setup_info.room_id
 | ||
|         depth_map = setup_info.depth_map
 | ||
| 
 | ||
|         # Record some attempts to backfill these events which will make
 | ||
|         # `get_insertion_event_backward_extremities_in_room` exclude them
 | ||
|         # because we haven't passed the backoff interval.
 | ||
|         self.get_success(
 | ||
|             self.store.record_event_failed_pull_attempt(
 | ||
|                 room_id, "insertion_eventA", "fake cause"
 | ||
|             )
 | ||
|         )
 | ||
| 
 | ||
|         # No time has passed since we attempted to backfill ^
 | ||
| 
 | ||
|         # Try at "insertion_eventB"
 | ||
|         backfill_points = self.get_success(
 | ||
|             self.store.get_insertion_event_backward_extremities_in_room(
 | ||
|                 room_id, depth_map["insertion_eventB"], limit=100
 | ||
|             )
 | ||
|         )
 | ||
|         backfill_event_ids = [backfill_point[0] for backfill_point in backfill_points]
 | ||
|         # Only the backfill points that we didn't record earlier exist here.
 | ||
|         self.assertEqual(backfill_event_ids, ["insertion_eventB"])
 | ||
| 
 | ||
|     def test_get_insertion_event_backward_extremities_in_room_attempted_event_retry_after_backoff_duration(
 | ||
|         self,
 | ||
|     ):
 | ||
|         """
 | ||
|         Test to make sure after we fake attempt to backfill event
 | ||
|         "insertion_eventA" many times, we can see retry and see the
 | ||
|         "insertion_eventA" again after the backoff timeout duration has
 | ||
|         exceeded.
 | ||
|         """
 | ||
|         setup_info = self._setup_room_for_insertion_backfill_tests()
 | ||
|         room_id = setup_info.room_id
 | ||
|         depth_map = setup_info.depth_map
 | ||
| 
 | ||
|         # Record some attempts to backfill these events which will make
 | ||
|         # `get_backfill_points_in_room` exclude them because we
 | ||
|         # haven't passed the backoff interval.
 | ||
|         self.get_success(
 | ||
|             self.store.record_event_failed_pull_attempt(
 | ||
|                 room_id, "insertion_eventB", "fake cause"
 | ||
|             )
 | ||
|         )
 | ||
|         self.get_success(
 | ||
|             self.store.record_event_failed_pull_attempt(
 | ||
|                 room_id, "insertion_eventA", "fake cause"
 | ||
|             )
 | ||
|         )
 | ||
|         self.get_success(
 | ||
|             self.store.record_event_failed_pull_attempt(
 | ||
|                 room_id, "insertion_eventA", "fake cause"
 | ||
|             )
 | ||
|         )
 | ||
|         self.get_success(
 | ||
|             self.store.record_event_failed_pull_attempt(
 | ||
|                 room_id, "insertion_eventA", "fake cause"
 | ||
|             )
 | ||
|         )
 | ||
|         self.get_success(
 | ||
|             self.store.record_event_failed_pull_attempt(
 | ||
|                 room_id, "insertion_eventA", "fake cause"
 | ||
|             )
 | ||
|         )
 | ||
| 
 | ||
|         # Now advance time by 2 hours and we should only be able to see
 | ||
|         # "insertion_eventB" because we have waited long enough for the single
 | ||
|         # attempt (2^1 hours) but we still shouldn't see "insertion_eventA"
 | ||
|         # because we haven't waited long enough for this many attempts.
 | ||
|         self.reactor.advance(datetime.timedelta(hours=2).total_seconds())
 | ||
| 
 | ||
|         # Try at "insertion_eventA" and make sure that "insertion_eventA" is not
 | ||
|         # in the list because we've already attempted many times
 | ||
|         backfill_points = self.get_success(
 | ||
|             self.store.get_insertion_event_backward_extremities_in_room(
 | ||
|                 room_id, depth_map["insertion_eventA"], limit=100
 | ||
|             )
 | ||
|         )
 | ||
|         backfill_event_ids = [backfill_point[0] for backfill_point in backfill_points]
 | ||
|         self.assertEqual(backfill_event_ids, [])
 | ||
| 
 | ||
|         # Now advance time by 20 hours (above 2^4 because we made 4 attemps) and
 | ||
|         # see if we can now backfill it
 | ||
|         self.reactor.advance(datetime.timedelta(hours=20).total_seconds())
 | ||
| 
 | ||
|         # Try at "insertion_eventA" again after we advanced enough time and we
 | ||
|         # should see "insertion_eventA" again
 | ||
|         backfill_points = self.get_success(
 | ||
|             self.store.get_insertion_event_backward_extremities_in_room(
 | ||
|                 room_id, depth_map["insertion_eventA"], limit=100
 | ||
|             )
 | ||
|         )
 | ||
|         backfill_event_ids = [backfill_point[0] for backfill_point in backfill_points]
 | ||
|         self.assertEqual(backfill_event_ids, ["insertion_eventA"])
 | ||
| 
 | ||
|     def test_get_event_ids_to_not_pull_from_backoff(
 | ||
|         self,
 | ||
|     ):
 | ||
|         """
 | ||
|         Test to make sure only event IDs we should backoff from are returned.
 | ||
|         """
 | ||
|         # Create the room
 | ||
|         user_id = self.register_user("alice", "test")
 | ||
|         tok = self.login("alice", "test")
 | ||
|         room_id = self.helper.create_room_as(room_creator=user_id, tok=tok)
 | ||
| 
 | ||
|         self.get_success(
 | ||
|             self.store.record_event_failed_pull_attempt(
 | ||
|                 room_id, "$failed_event_id", "fake cause"
 | ||
|             )
 | ||
|         )
 | ||
| 
 | ||
|         event_ids_to_backoff = self.get_success(
 | ||
|             self.store.get_event_ids_to_not_pull_from_backoff(
 | ||
|                 room_id=room_id, event_ids=["$failed_event_id", "$normal_event_id"]
 | ||
|             )
 | ||
|         )
 | ||
| 
 | ||
|         self.assertEqual(event_ids_to_backoff, ["$failed_event_id"])
 | ||
| 
 | ||
|     def test_get_event_ids_to_not_pull_from_backoff_retry_after_backoff_duration(
 | ||
|         self,
 | ||
|     ):
 | ||
|         """
 | ||
|         Test to make sure no event IDs are returned after the backoff duration has
 | ||
|         elapsed.
 | ||
|         """
 | ||
|         # Create the room
 | ||
|         user_id = self.register_user("alice", "test")
 | ||
|         tok = self.login("alice", "test")
 | ||
|         room_id = self.helper.create_room_as(room_creator=user_id, tok=tok)
 | ||
| 
 | ||
|         self.get_success(
 | ||
|             self.store.record_event_failed_pull_attempt(
 | ||
|                 room_id, "$failed_event_id", "fake cause"
 | ||
|             )
 | ||
|         )
 | ||
| 
 | ||
|         # Now advance time by 2 hours so we wait long enough for the single failed
 | ||
|         # attempt (2^1 hours).
 | ||
|         self.reactor.advance(datetime.timedelta(hours=2).total_seconds())
 | ||
| 
 | ||
|         event_ids_to_backoff = self.get_success(
 | ||
|             self.store.get_event_ids_to_not_pull_from_backoff(
 | ||
|                 room_id=room_id, event_ids=["$failed_event_id", "$normal_event_id"]
 | ||
|             )
 | ||
|         )
 | ||
|         # Since this function only returns events we should backoff from, time has
 | ||
|         # elapsed past the backoff range so there is no events to backoff from.
 | ||
|         self.assertEqual(event_ids_to_backoff, [])
 | ||
| 
 | ||
| 
 | ||
| @attr.s
 | ||
| class FakeEvent:
 | ||
|     event_id = attr.ib()
 | ||
|     room_id = attr.ib()
 | ||
|     auth_events = attr.ib()
 | ||
| 
 | ||
|     type = "foo"
 | ||
|     state_key = "foo"
 | ||
| 
 | ||
|     internal_metadata = _EventInternalMetadata({})
 | ||
| 
 | ||
|     def auth_event_ids(self):
 | ||
|         return self.auth_events
 | ||
| 
 | ||
|     def is_state(self):
 | ||
|         return True
 |