2021-08-26 22:41:44 +02:00
|
|
|
# Copyright 2021 The Matrix.org Foundation C.I.C.
|
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
|
|
|
import logging
|
|
|
|
from http import HTTPStatus
|
|
|
|
from typing import (
|
|
|
|
TYPE_CHECKING,
|
|
|
|
Collection,
|
|
|
|
Container,
|
|
|
|
Dict,
|
|
|
|
Iterable,
|
|
|
|
List,
|
|
|
|
Optional,
|
|
|
|
Sequence,
|
|
|
|
Set,
|
|
|
|
Tuple,
|
|
|
|
)
|
|
|
|
|
|
|
|
from prometheus_client import Counter
|
|
|
|
|
|
|
|
from synapse.api.constants import (
|
|
|
|
EventContentFields,
|
|
|
|
EventTypes,
|
2021-09-06 13:17:16 +02:00
|
|
|
GuestAccess,
|
2021-08-26 22:41:44 +02:00
|
|
|
Membership,
|
|
|
|
RejectedReason,
|
|
|
|
RoomEncryptionAlgorithms,
|
|
|
|
)
|
|
|
|
from synapse.api.errors import (
|
|
|
|
AuthError,
|
|
|
|
Codes,
|
|
|
|
FederationError,
|
|
|
|
HttpResponseException,
|
|
|
|
RequestSendFailed,
|
|
|
|
SynapseError,
|
|
|
|
)
|
|
|
|
from synapse.api.room_versions import KNOWN_ROOM_VERSIONS
|
Split `event_auth.check` into two parts (#10940)
Broadly, the existing `event_auth.check` function has two parts:
* a validation section: checks that the event isn't too big, that it has the rught signatures, etc.
This bit is independent of the rest of the state in the room, and so need only be done once
for each event.
* an auth section: ensures that the event is allowed, given the rest of the state in the room.
This gets done multiple times, against various sets of room state, because it forms part of
the state res algorithm.
Currently, this is implemented with `do_sig_check` and `do_size_check` parameters, but I think
that makes everything hard to follow. Instead, we split the function in two and call each part
separately where it is needed.
2021-09-29 19:59:15 +02:00
|
|
|
from synapse.event_auth import (
|
|
|
|
auth_types_for_event,
|
|
|
|
check_auth_rules_for_event,
|
|
|
|
validate_event_for_room_version,
|
|
|
|
)
|
2021-08-26 22:41:44 +02:00
|
|
|
from synapse.events import EventBase
|
|
|
|
from synapse.events.snapshot import EventContext
|
|
|
|
from synapse.federation.federation_client import InvalidResponseError
|
2021-09-24 12:56:13 +02:00
|
|
|
from synapse.logging.context import nested_logging_context, run_in_background
|
2021-08-26 22:41:44 +02:00
|
|
|
from synapse.logging.utils import log_function
|
|
|
|
from synapse.metrics.background_process_metrics import run_as_background_process
|
|
|
|
from synapse.replication.http.devices import ReplicationUserDevicesResyncRestServlet
|
|
|
|
from synapse.replication.http.federation import (
|
|
|
|
ReplicationFederationSendEventsRestServlet,
|
|
|
|
)
|
|
|
|
from synapse.state import StateResolutionStore
|
|
|
|
from synapse.storage.databases.main.events_worker import EventRedactBehaviour
|
|
|
|
from synapse.types import (
|
|
|
|
MutableStateMap,
|
|
|
|
PersistedEventPosition,
|
|
|
|
RoomStreamToken,
|
|
|
|
StateMap,
|
|
|
|
UserID,
|
|
|
|
get_domain_from_id,
|
|
|
|
)
|
2021-09-28 16:25:07 +02:00
|
|
|
from synapse.util.async_helpers import Linearizer, concurrently_execute
|
2021-08-26 22:41:44 +02:00
|
|
|
from synapse.util.iterutils import batch_iter
|
|
|
|
from synapse.util.retryutils import NotRetryingDestination
|
|
|
|
from synapse.util.stringutils import shortstr
|
|
|
|
|
|
|
|
if TYPE_CHECKING:
|
|
|
|
from synapse.server import HomeServer
|
|
|
|
|
|
|
|
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
soft_failed_event_counter = Counter(
|
|
|
|
"synapse_federation_soft_failed_events_total",
|
|
|
|
"Events received over federation that we marked as soft_failed",
|
|
|
|
)
|
|
|
|
|
|
|
|
|
2021-09-06 15:49:33 +02:00
|
|
|
class FederationEventHandler:
|
2021-08-26 22:41:44 +02:00
|
|
|
"""Handles events that originated from federation.
|
|
|
|
|
|
|
|
Responsible for handing incoming events and passing them on to the rest
|
|
|
|
of the homeserver (including auth and state conflict resolutions)
|
|
|
|
"""
|
|
|
|
|
|
|
|
def __init__(self, hs: "HomeServer"):
|
2021-09-07 12:15:51 +02:00
|
|
|
self._store = hs.get_datastore()
|
|
|
|
self._storage = hs.get_storage()
|
|
|
|
self._state_store = self._storage.state
|
2021-08-26 22:41:44 +02:00
|
|
|
|
2021-09-07 12:15:51 +02:00
|
|
|
self._state_handler = hs.get_state_handler()
|
|
|
|
self._event_creation_handler = hs.get_event_creation_handler()
|
2021-08-26 22:41:44 +02:00
|
|
|
self._event_auth_handler = hs.get_event_auth_handler()
|
|
|
|
self._message_handler = hs.get_message_handler()
|
2021-09-07 12:15:51 +02:00
|
|
|
self._action_generator = hs.get_action_generator()
|
2021-08-26 22:41:44 +02:00
|
|
|
self._state_resolution_handler = hs.get_state_resolution_handler()
|
2021-09-06 15:49:33 +02:00
|
|
|
# avoid a circular dependency by deferring execution here
|
|
|
|
self._get_room_member_handler = hs.get_room_member_handler
|
2021-08-26 22:41:44 +02:00
|
|
|
|
2021-09-07 12:15:51 +02:00
|
|
|
self._federation_client = hs.get_federation_client()
|
|
|
|
self._third_party_event_rules = hs.get_third_party_event_rules()
|
2021-09-06 15:49:33 +02:00
|
|
|
self._notifier = hs.get_notifier()
|
2021-08-26 22:41:44 +02:00
|
|
|
|
2021-09-07 12:15:51 +02:00
|
|
|
self._is_mine_id = hs.is_mine_id
|
2021-09-06 15:49:33 +02:00
|
|
|
self._server_name = hs.hostname
|
2021-08-26 22:41:44 +02:00
|
|
|
self._instance_name = hs.get_instance_name()
|
|
|
|
|
2021-09-07 12:15:51 +02:00
|
|
|
self._config = hs.config
|
2021-08-26 22:41:44 +02:00
|
|
|
self._ephemeral_messages_enabled = hs.config.server.enable_ephemeral_messages
|
|
|
|
|
|
|
|
self._send_events = ReplicationFederationSendEventsRestServlet.make_client(hs)
|
2021-09-13 19:07:12 +02:00
|
|
|
if hs.config.worker.worker_app:
|
2021-08-26 22:41:44 +02:00
|
|
|
self._user_device_resync = (
|
|
|
|
ReplicationUserDevicesResyncRestServlet.make_client(hs)
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
self._device_list_updater = hs.get_device_handler().device_list_updater
|
|
|
|
|
|
|
|
# When joining a room we need to queue any events for that room up.
|
|
|
|
# For each room, a list of (pdu, origin) tuples.
|
|
|
|
# TODO: replace this with something more elegant, probably based around the
|
|
|
|
# federation event staging area.
|
|
|
|
self.room_queues: Dict[str, List[Tuple[EventBase, str]]] = {}
|
|
|
|
|
|
|
|
self._room_pdu_linearizer = Linearizer("fed_room_pdu")
|
|
|
|
|
|
|
|
async def on_receive_pdu(self, origin: str, pdu: EventBase) -> None:
|
|
|
|
"""Process a PDU received via a federation /send/ transaction
|
|
|
|
|
|
|
|
Args:
|
|
|
|
origin: server which initiated the /send/ transaction. Will
|
|
|
|
be used to fetch missing events or state.
|
|
|
|
pdu: received PDU
|
|
|
|
"""
|
|
|
|
|
2021-09-08 11:41:13 +02:00
|
|
|
# We should never see any outliers here.
|
|
|
|
assert not pdu.internal_metadata.outlier
|
|
|
|
|
2021-08-26 22:41:44 +02:00
|
|
|
room_id = pdu.room_id
|
|
|
|
event_id = pdu.event_id
|
|
|
|
|
|
|
|
# We reprocess pdus when we have seen them only as outliers
|
2021-09-07 12:15:51 +02:00
|
|
|
existing = await self._store.get_event(
|
2021-08-26 22:41:44 +02:00
|
|
|
event_id, allow_none=True, allow_rejected=True
|
|
|
|
)
|
|
|
|
|
|
|
|
# FIXME: Currently we fetch an event again when we already have it
|
|
|
|
# if it has been marked as an outlier.
|
|
|
|
if existing:
|
|
|
|
if not existing.internal_metadata.is_outlier():
|
|
|
|
logger.info(
|
|
|
|
"Ignoring received event %s which we have already seen", event_id
|
|
|
|
)
|
|
|
|
return
|
|
|
|
if pdu.internal_metadata.is_outlier():
|
|
|
|
logger.info(
|
|
|
|
"Ignoring received outlier %s which we already have as an outlier",
|
|
|
|
event_id,
|
|
|
|
)
|
|
|
|
return
|
|
|
|
logger.info("De-outliering event %s", event_id)
|
|
|
|
|
|
|
|
# do some initial sanity-checking of the event. In particular, make
|
|
|
|
# sure it doesn't have hundreds of prev_events or auth_events, which
|
|
|
|
# could cause a huge state resolution or cascade of event fetches.
|
|
|
|
try:
|
|
|
|
self._sanity_check_event(pdu)
|
|
|
|
except SynapseError as err:
|
|
|
|
logger.warning("Received event failed sanity checks")
|
|
|
|
raise FederationError("ERROR", err.code, err.msg, affected=pdu.event_id)
|
|
|
|
|
|
|
|
# If we are currently in the process of joining this room, then we
|
|
|
|
# queue up events for later processing.
|
|
|
|
if room_id in self.room_queues:
|
|
|
|
logger.info(
|
|
|
|
"Queuing PDU from %s for now: join in progress",
|
|
|
|
origin,
|
|
|
|
)
|
|
|
|
self.room_queues[room_id].append((pdu, origin))
|
|
|
|
return
|
|
|
|
|
|
|
|
# If we're not in the room just ditch the event entirely. This is
|
|
|
|
# probably an old server that has come back and thinks we're still in
|
|
|
|
# the room (or we've been rejoined to the room by a state reset).
|
|
|
|
#
|
|
|
|
# Note that if we were never in the room then we would have already
|
|
|
|
# dropped the event, since we wouldn't know the room version.
|
|
|
|
is_in_room = await self._event_auth_handler.check_host_in_room(
|
2021-09-06 15:49:33 +02:00
|
|
|
room_id, self._server_name
|
2021-08-26 22:41:44 +02:00
|
|
|
)
|
|
|
|
if not is_in_room:
|
|
|
|
logger.info(
|
|
|
|
"Ignoring PDU from %s as we're not in the room",
|
|
|
|
origin,
|
|
|
|
)
|
|
|
|
return None
|
|
|
|
|
2021-09-08 11:41:13 +02:00
|
|
|
# Try to fetch any missing prev events to fill in gaps in the graph
|
|
|
|
prevs = set(pdu.prev_event_ids())
|
|
|
|
seen = await self._store.have_events_in_timeline(prevs)
|
|
|
|
missing_prevs = prevs - seen
|
2021-08-26 22:41:44 +02:00
|
|
|
|
2021-09-08 11:41:13 +02:00
|
|
|
if missing_prevs:
|
|
|
|
# We only backfill backwards to the min depth.
|
|
|
|
min_depth = await self.get_min_depth_for_context(pdu.room_id)
|
|
|
|
logger.debug("min_depth: %d", min_depth)
|
2021-08-26 22:41:44 +02:00
|
|
|
|
2021-09-08 11:41:13 +02:00
|
|
|
if min_depth is not None and pdu.depth > min_depth:
|
|
|
|
# If we're missing stuff, ensure we only fetch stuff one
|
|
|
|
# at a time.
|
|
|
|
logger.info(
|
|
|
|
"Acquiring room lock to fetch %d missing prev_events: %s",
|
|
|
|
len(missing_prevs),
|
|
|
|
shortstr(missing_prevs),
|
|
|
|
)
|
|
|
|
with (await self._room_pdu_linearizer.queue(pdu.room_id)):
|
2021-08-26 22:41:44 +02:00
|
|
|
logger.info(
|
2021-09-08 11:41:13 +02:00
|
|
|
"Acquired room lock to fetch %d missing prev_events",
|
2021-08-26 22:41:44 +02:00
|
|
|
len(missing_prevs),
|
|
|
|
)
|
2021-09-08 11:41:13 +02:00
|
|
|
|
|
|
|
try:
|
|
|
|
await self._get_missing_events_for_pdu(
|
|
|
|
origin, pdu, prevs, min_depth
|
2021-08-26 22:41:44 +02:00
|
|
|
)
|
2021-09-08 11:41:13 +02:00
|
|
|
except Exception as e:
|
|
|
|
raise Exception(
|
|
|
|
"Error fetching missing prev_events for %s: %s"
|
|
|
|
% (event_id, e)
|
|
|
|
) from e
|
2021-08-26 22:41:44 +02:00
|
|
|
|
2021-09-08 11:41:13 +02:00
|
|
|
# Update the set of things we've seen after trying to
|
|
|
|
# fetch the missing stuff
|
|
|
|
seen = await self._store.have_events_in_timeline(prevs)
|
|
|
|
missing_prevs = prevs - seen
|
|
|
|
|
|
|
|
if not missing_prevs:
|
|
|
|
logger.info("Found all missing prev_events")
|
|
|
|
|
|
|
|
if missing_prevs:
|
|
|
|
# since this event was pushed to us, it is possible for it to
|
|
|
|
# become the only forward-extremity in the room, and we would then
|
|
|
|
# trust its state to be the state for the whole room. This is very
|
|
|
|
# bad. Further, if the event was pushed to us, there is no excuse
|
|
|
|
# for us not to have all the prev_events. (XXX: apart from
|
|
|
|
# min_depth?)
|
|
|
|
#
|
|
|
|
# We therefore reject any such events.
|
|
|
|
logger.warning(
|
|
|
|
"Rejecting: failed to fetch %d prev events: %s",
|
|
|
|
len(missing_prevs),
|
|
|
|
shortstr(missing_prevs),
|
|
|
|
)
|
|
|
|
raise FederationError(
|
|
|
|
"ERROR",
|
|
|
|
403,
|
|
|
|
(
|
|
|
|
"Your server isn't divulging details about prev_events "
|
|
|
|
"referenced in this event."
|
|
|
|
),
|
|
|
|
affected=pdu.event_id,
|
|
|
|
)
|
2021-08-26 22:41:44 +02:00
|
|
|
|
|
|
|
await self._process_received_pdu(origin, pdu, state=None)
|
|
|
|
|
|
|
|
@log_function
|
|
|
|
async def on_send_membership_event(
|
|
|
|
self, origin: str, event: EventBase
|
|
|
|
) -> Tuple[EventBase, EventContext]:
|
|
|
|
"""
|
|
|
|
We have received a join/leave/knock event for a room via send_join/leave/knock.
|
|
|
|
|
|
|
|
Verify that event and send it into the room on the remote homeserver's behalf.
|
|
|
|
|
|
|
|
This is quite similar to on_receive_pdu, with the following principal
|
|
|
|
differences:
|
|
|
|
* only membership events are permitted (and only events with
|
|
|
|
sender==state_key -- ie, no kicks or bans)
|
|
|
|
* *We* send out the event on behalf of the remote server.
|
|
|
|
* We enforce the membership restrictions of restricted rooms.
|
|
|
|
* Rejected events result in an exception rather than being stored.
|
|
|
|
|
|
|
|
There are also other differences, however it is not clear if these are by
|
|
|
|
design or omission. In particular, we do not attempt to backfill any missing
|
|
|
|
prev_events.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
origin: The homeserver of the remote (joining/invited/knocking) user.
|
|
|
|
event: The member event that has been signed by the remote homeserver.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
The event and context of the event after inserting it into the room graph.
|
|
|
|
|
|
|
|
Raises:
|
|
|
|
SynapseError if the event is not accepted into the room
|
|
|
|
"""
|
|
|
|
logger.debug(
|
|
|
|
"on_send_membership_event: Got event: %s, signatures: %s",
|
|
|
|
event.event_id,
|
|
|
|
event.signatures,
|
|
|
|
)
|
|
|
|
|
|
|
|
if get_domain_from_id(event.sender) != origin:
|
|
|
|
logger.info(
|
|
|
|
"Got send_membership request for user %r from different origin %s",
|
|
|
|
event.sender,
|
|
|
|
origin,
|
|
|
|
)
|
|
|
|
raise SynapseError(403, "User not from origin", Codes.FORBIDDEN)
|
|
|
|
|
|
|
|
if event.sender != event.state_key:
|
|
|
|
raise SynapseError(400, "state_key and sender must match", Codes.BAD_JSON)
|
|
|
|
|
|
|
|
assert not event.internal_metadata.outlier
|
|
|
|
|
|
|
|
# Send this event on behalf of the other server.
|
|
|
|
#
|
|
|
|
# The remote server isn't a full participant in the room at this point, so
|
|
|
|
# may not have an up-to-date list of the other homeservers participating in
|
|
|
|
# the room, so we send it on their behalf.
|
|
|
|
event.internal_metadata.send_on_behalf_of = origin
|
|
|
|
|
2021-09-07 12:15:51 +02:00
|
|
|
context = await self._state_handler.compute_event_context(event)
|
2021-08-26 22:41:44 +02:00
|
|
|
context = await self._check_event_auth(origin, event, context)
|
|
|
|
if context.rejected:
|
|
|
|
raise SynapseError(
|
|
|
|
403, f"{event.membership} event was rejected", Codes.FORBIDDEN
|
|
|
|
)
|
|
|
|
|
|
|
|
# for joins, we need to check the restrictions of restricted rooms
|
|
|
|
if event.membership == Membership.JOIN:
|
|
|
|
await self.check_join_restrictions(context, event)
|
|
|
|
|
|
|
|
# for knock events, we run the third-party event rules. It's not entirely clear
|
|
|
|
# why we don't do this for other sorts of membership events.
|
|
|
|
if event.membership == Membership.KNOCK:
|
2021-09-07 12:15:51 +02:00
|
|
|
event_allowed, _ = await self._third_party_event_rules.check_event_allowed(
|
2021-08-26 22:41:44 +02:00
|
|
|
event, context
|
|
|
|
)
|
|
|
|
if not event_allowed:
|
|
|
|
logger.info("Sending of knock %s forbidden by third-party rules", event)
|
|
|
|
raise SynapseError(
|
|
|
|
403, "This event is not allowed in this context", Codes.FORBIDDEN
|
|
|
|
)
|
|
|
|
|
|
|
|
# all looks good, we can persist the event.
|
|
|
|
await self._run_push_actions_and_persist_event(event, context)
|
|
|
|
return event, context
|
|
|
|
|
|
|
|
async def check_join_restrictions(
|
|
|
|
self, context: EventContext, event: EventBase
|
|
|
|
) -> None:
|
|
|
|
"""Check that restrictions in restricted join rules are matched
|
|
|
|
|
|
|
|
Called when we receive a join event via send_join.
|
|
|
|
|
|
|
|
Raises an auth error if the restrictions are not matched.
|
|
|
|
"""
|
|
|
|
prev_state_ids = await context.get_prev_state_ids()
|
|
|
|
|
|
|
|
# Check if the user is already in the room or invited to the room.
|
|
|
|
user_id = event.state_key
|
|
|
|
prev_member_event_id = prev_state_ids.get((EventTypes.Member, user_id), None)
|
|
|
|
prev_member_event = None
|
|
|
|
if prev_member_event_id:
|
2021-09-07 12:15:51 +02:00
|
|
|
prev_member_event = await self._store.get_event(prev_member_event_id)
|
2021-08-26 22:41:44 +02:00
|
|
|
|
|
|
|
# Check if the member should be allowed access via membership in a space.
|
|
|
|
await self._event_auth_handler.check_restricted_join_rules(
|
|
|
|
prev_state_ids,
|
|
|
|
event.room_version,
|
|
|
|
user_id,
|
|
|
|
prev_member_event,
|
|
|
|
)
|
|
|
|
|
|
|
|
@log_function
|
|
|
|
async def backfill(
|
|
|
|
self, dest: str, room_id: str, limit: int, extremities: List[str]
|
|
|
|
) -> None:
|
|
|
|
"""Trigger a backfill request to `dest` for the given `room_id`
|
|
|
|
|
|
|
|
This will attempt to get more events from the remote. If the other side
|
|
|
|
has no new events to offer, this will return an empty list.
|
|
|
|
|
|
|
|
As the events are received, we check their signatures, and also do some
|
|
|
|
sanity-checking on them. If any of the backfilled events are invalid,
|
|
|
|
this method throws a SynapseError.
|
|
|
|
|
|
|
|
We might also raise an InvalidResponseError if the response from the remote
|
|
|
|
server is just bogus.
|
|
|
|
|
|
|
|
TODO: make this more useful to distinguish failures of the remote
|
|
|
|
server from invalid events (there is probably no point in trying to
|
|
|
|
re-fetch invalid events from every other HS in the room.)
|
|
|
|
"""
|
2021-09-06 15:49:33 +02:00
|
|
|
if dest == self._server_name:
|
2021-08-26 22:41:44 +02:00
|
|
|
raise SynapseError(400, "Can't backfill from self.")
|
|
|
|
|
2021-09-07 12:15:51 +02:00
|
|
|
events = await self._federation_client.backfill(
|
2021-08-26 22:41:44 +02:00
|
|
|
dest, room_id, limit=limit, extremities=extremities
|
|
|
|
)
|
|
|
|
|
|
|
|
if not events:
|
|
|
|
return
|
|
|
|
|
|
|
|
# if there are any events in the wrong room, the remote server is buggy and
|
|
|
|
# should not be trusted.
|
|
|
|
for ev in events:
|
|
|
|
if ev.room_id != room_id:
|
|
|
|
raise InvalidResponseError(
|
|
|
|
f"Remote server {dest} returned event {ev.event_id} which is in "
|
|
|
|
f"room {ev.room_id}, when we were backfilling in {room_id}"
|
|
|
|
)
|
|
|
|
|
|
|
|
await self._process_pulled_events(dest, events, backfilled=True)
|
|
|
|
|
|
|
|
async def _get_missing_events_for_pdu(
|
|
|
|
self, origin: str, pdu: EventBase, prevs: Set[str], min_depth: int
|
|
|
|
) -> None:
|
|
|
|
"""
|
|
|
|
Args:
|
|
|
|
origin: Origin of the pdu. Will be called to get the missing events
|
|
|
|
pdu: received pdu
|
|
|
|
prevs: List of event ids which we are missing
|
|
|
|
min_depth: Minimum depth of events to return.
|
|
|
|
"""
|
|
|
|
|
|
|
|
room_id = pdu.room_id
|
|
|
|
event_id = pdu.event_id
|
|
|
|
|
2021-09-07 12:15:51 +02:00
|
|
|
seen = await self._store.have_events_in_timeline(prevs)
|
2021-08-26 22:41:44 +02:00
|
|
|
|
|
|
|
if not prevs - seen:
|
|
|
|
return
|
|
|
|
|
2021-09-07 12:15:51 +02:00
|
|
|
latest_list = await self._store.get_latest_event_ids_in_room(room_id)
|
2021-08-26 22:41:44 +02:00
|
|
|
|
|
|
|
# We add the prev events that we have seen to the latest
|
|
|
|
# list to ensure the remote server doesn't give them to us
|
|
|
|
latest = set(latest_list)
|
|
|
|
latest |= seen
|
|
|
|
|
|
|
|
logger.info(
|
|
|
|
"Requesting missing events between %s and %s",
|
|
|
|
shortstr(latest),
|
|
|
|
event_id,
|
|
|
|
)
|
|
|
|
|
|
|
|
# XXX: we set timeout to 10s to help workaround
|
|
|
|
# https://github.com/matrix-org/synapse/issues/1733.
|
|
|
|
# The reason is to avoid holding the linearizer lock
|
|
|
|
# whilst processing inbound /send transactions, causing
|
|
|
|
# FDs to stack up and block other inbound transactions
|
|
|
|
# which empirically can currently take up to 30 minutes.
|
|
|
|
#
|
|
|
|
# N.B. this explicitly disables retry attempts.
|
|
|
|
#
|
|
|
|
# N.B. this also increases our chances of falling back to
|
|
|
|
# fetching fresh state for the room if the missing event
|
|
|
|
# can't be found, which slightly reduces our security.
|
|
|
|
# it may also increase our DAG extremity count for the room,
|
|
|
|
# causing additional state resolution? See #1760.
|
|
|
|
# However, fetching state doesn't hold the linearizer lock
|
|
|
|
# apparently.
|
|
|
|
#
|
|
|
|
# see https://github.com/matrix-org/synapse/pull/1744
|
|
|
|
#
|
|
|
|
# ----
|
|
|
|
#
|
|
|
|
# Update richvdh 2018/09/18: There are a number of problems with timing this
|
|
|
|
# request out aggressively on the client side:
|
|
|
|
#
|
|
|
|
# - it plays badly with the server-side rate-limiter, which starts tarpitting you
|
|
|
|
# if you send too many requests at once, so you end up with the server carefully
|
|
|
|
# working through the backlog of your requests, which you have already timed
|
|
|
|
# out.
|
|
|
|
#
|
|
|
|
# - for this request in particular, we now (as of
|
|
|
|
# https://github.com/matrix-org/synapse/pull/3456) reject any PDUs where the
|
|
|
|
# server can't produce a plausible-looking set of prev_events - so we becone
|
|
|
|
# much more likely to reject the event.
|
|
|
|
#
|
|
|
|
# - contrary to what it says above, we do *not* fall back to fetching fresh state
|
|
|
|
# for the room if get_missing_events times out. Rather, we give up processing
|
|
|
|
# the PDU whose prevs we are missing, which then makes it much more likely that
|
|
|
|
# we'll end up back here for the *next* PDU in the list, which exacerbates the
|
|
|
|
# problem.
|
|
|
|
#
|
|
|
|
# - the aggressive 10s timeout was introduced to deal with incoming federation
|
|
|
|
# requests taking 8 hours to process. It's not entirely clear why that was going
|
|
|
|
# on; certainly there were other issues causing traffic storms which are now
|
|
|
|
# resolved, and I think in any case we may be more sensible about our locking
|
|
|
|
# now. We're *certainly* more sensible about our logging.
|
|
|
|
#
|
|
|
|
# All that said: Let's try increasing the timeout to 60s and see what happens.
|
|
|
|
|
|
|
|
try:
|
2021-09-07 12:15:51 +02:00
|
|
|
missing_events = await self._federation_client.get_missing_events(
|
2021-08-26 22:41:44 +02:00
|
|
|
origin,
|
|
|
|
room_id,
|
|
|
|
earliest_events_ids=list(latest),
|
|
|
|
latest_events=[pdu],
|
|
|
|
limit=10,
|
|
|
|
min_depth=min_depth,
|
|
|
|
timeout=60000,
|
|
|
|
)
|
|
|
|
except (RequestSendFailed, HttpResponseException, NotRetryingDestination) as e:
|
|
|
|
# We failed to get the missing events, but since we need to handle
|
|
|
|
# the case of `get_missing_events` not returning the necessary
|
|
|
|
# events anyway, it is safe to simply log the error and continue.
|
|
|
|
logger.warning("Failed to get prev_events: %s", e)
|
|
|
|
return
|
|
|
|
|
|
|
|
logger.info("Got %d prev_events", len(missing_events))
|
|
|
|
await self._process_pulled_events(origin, missing_events, backfilled=False)
|
|
|
|
|
|
|
|
async def _process_pulled_events(
|
|
|
|
self, origin: str, events: Iterable[EventBase], backfilled: bool
|
|
|
|
) -> None:
|
|
|
|
"""Process a batch of events we have pulled from a remote server
|
|
|
|
|
|
|
|
Pulls in any events required to auth the events, persists the received events,
|
|
|
|
and notifies clients, if appropriate.
|
|
|
|
|
|
|
|
Assumes the events have already had their signatures and hashes checked.
|
|
|
|
|
|
|
|
Params:
|
|
|
|
origin: The server we received these events from
|
|
|
|
events: The received events.
|
|
|
|
backfilled: True if this is part of a historical batch of events (inhibits
|
|
|
|
notification to clients, and validation of device keys.)
|
|
|
|
"""
|
|
|
|
|
|
|
|
# We want to sort these by depth so we process them and
|
|
|
|
# tell clients about them in order.
|
|
|
|
sorted_events = sorted(events, key=lambda x: x.depth)
|
|
|
|
|
|
|
|
for ev in sorted_events:
|
|
|
|
with nested_logging_context(ev.event_id):
|
|
|
|
await self._process_pulled_event(origin, ev, backfilled=backfilled)
|
|
|
|
|
|
|
|
async def _process_pulled_event(
|
|
|
|
self, origin: str, event: EventBase, backfilled: bool
|
|
|
|
) -> None:
|
|
|
|
"""Process a single event that we have pulled from a remote server
|
|
|
|
|
|
|
|
Pulls in any events required to auth the event, persists the received event,
|
|
|
|
and notifies clients, if appropriate.
|
|
|
|
|
|
|
|
Assumes the event has already had its signatures and hashes checked.
|
|
|
|
|
|
|
|
This is somewhat equivalent to on_receive_pdu, but applies somewhat different
|
|
|
|
logic in the case that we are missing prev_events (in particular, it just
|
|
|
|
requests the state at that point, rather than triggering a get_missing_events) -
|
|
|
|
so is appropriate when we have pulled the event from a remote server, rather
|
|
|
|
than having it pushed to us.
|
|
|
|
|
|
|
|
Params:
|
|
|
|
origin: The server we received this event from
|
|
|
|
events: The received event
|
|
|
|
backfilled: True if this is part of a historical batch of events (inhibits
|
|
|
|
notification to clients, and validation of device keys.)
|
|
|
|
"""
|
|
|
|
logger.info("Processing pulled event %s", event)
|
|
|
|
|
|
|
|
# these should not be outliers.
|
|
|
|
assert not event.internal_metadata.is_outlier()
|
|
|
|
|
|
|
|
event_id = event.event_id
|
|
|
|
|
2021-09-07 12:15:51 +02:00
|
|
|
existing = await self._store.get_event(
|
2021-08-26 22:41:44 +02:00
|
|
|
event_id, allow_none=True, allow_rejected=True
|
|
|
|
)
|
|
|
|
if existing:
|
|
|
|
if not existing.internal_metadata.is_outlier():
|
|
|
|
logger.info(
|
|
|
|
"Ignoring received event %s which we have already seen",
|
|
|
|
event_id,
|
|
|
|
)
|
|
|
|
return
|
|
|
|
logger.info("De-outliering event %s", event_id)
|
|
|
|
|
|
|
|
try:
|
|
|
|
self._sanity_check_event(event)
|
|
|
|
except SynapseError as err:
|
|
|
|
logger.warning("Event %s failed sanity check: %s", event_id, err)
|
|
|
|
return
|
|
|
|
|
|
|
|
try:
|
|
|
|
state = await self._resolve_state_at_missing_prevs(origin, event)
|
|
|
|
await self._process_received_pdu(
|
|
|
|
origin, event, state=state, backfilled=backfilled
|
|
|
|
)
|
|
|
|
except FederationError as e:
|
|
|
|
if e.code == 403:
|
|
|
|
logger.warning("Pulled event %s failed history check.", event_id)
|
|
|
|
else:
|
|
|
|
raise
|
|
|
|
|
|
|
|
async def _resolve_state_at_missing_prevs(
|
|
|
|
self, dest: str, event: EventBase
|
|
|
|
) -> Optional[Iterable[EventBase]]:
|
|
|
|
"""Calculate the state at an event with missing prev_events.
|
|
|
|
|
|
|
|
This is used when we have pulled a batch of events from a remote server, and
|
|
|
|
still don't have all the prev_events.
|
|
|
|
|
|
|
|
If we already have all the prev_events for `event`, this method does nothing.
|
|
|
|
|
|
|
|
Otherwise, the missing prevs become new backwards extremities, and we fall back
|
|
|
|
to asking the remote server for the state after each missing `prev_event`,
|
|
|
|
and resolving across them.
|
|
|
|
|
|
|
|
That's ok provided we then resolve the state against other bits of the DAG
|
|
|
|
before using it - in other words, that the received event `event` is not going
|
|
|
|
to become the only forwards_extremity in the room (which will ensure that you
|
|
|
|
can't just take over a room by sending an event, withholding its prev_events,
|
|
|
|
and declaring yourself to be an admin in the subsequent state request).
|
|
|
|
|
|
|
|
In other words: we should only call this method if `event` has been *pulled*
|
|
|
|
as part of a batch of missing prev events, or similar.
|
|
|
|
|
|
|
|
Params:
|
|
|
|
dest: the remote server to ask for state at the missing prevs. Typically,
|
|
|
|
this will be the server we got `event` from.
|
|
|
|
event: an event to check for missing prevs.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
if we already had all the prev events, `None`. Otherwise, returns a list of
|
|
|
|
the events in the state at `event`.
|
|
|
|
"""
|
|
|
|
room_id = event.room_id
|
|
|
|
event_id = event.event_id
|
|
|
|
|
|
|
|
prevs = set(event.prev_event_ids())
|
2021-09-07 12:15:51 +02:00
|
|
|
seen = await self._store.have_events_in_timeline(prevs)
|
2021-08-26 22:41:44 +02:00
|
|
|
missing_prevs = prevs - seen
|
|
|
|
|
|
|
|
if not missing_prevs:
|
|
|
|
return None
|
|
|
|
|
|
|
|
logger.info(
|
|
|
|
"Event %s is missing prev_events %s: calculating state for a "
|
|
|
|
"backwards extremity",
|
|
|
|
event_id,
|
|
|
|
shortstr(missing_prevs),
|
|
|
|
)
|
|
|
|
# Calculate the state after each of the previous events, and
|
|
|
|
# resolve them to find the correct state at the current event.
|
|
|
|
event_map = {event_id: event}
|
|
|
|
try:
|
|
|
|
# Get the state of the events we know about
|
2021-09-07 12:15:51 +02:00
|
|
|
ours = await self._state_store.get_state_groups_ids(room_id, seen)
|
2021-08-26 22:41:44 +02:00
|
|
|
|
|
|
|
# state_maps is a list of mappings from (type, state_key) to event_id
|
|
|
|
state_maps: List[StateMap[str]] = list(ours.values())
|
|
|
|
|
|
|
|
# we don't need this any more, let's delete it.
|
|
|
|
del ours
|
|
|
|
|
|
|
|
# Ask the remote server for the states we don't
|
|
|
|
# know about
|
|
|
|
for p in missing_prevs:
|
|
|
|
logger.info("Requesting state after missing prev_event %s", p)
|
|
|
|
|
|
|
|
with nested_logging_context(p):
|
|
|
|
# note that if any of the missing prevs share missing state or
|
|
|
|
# auth events, the requests to fetch those events are deduped
|
|
|
|
# by the get_pdu_cache in federation_client.
|
|
|
|
remote_state = await self._get_state_after_missing_prev_event(
|
|
|
|
dest, room_id, p
|
|
|
|
)
|
|
|
|
|
|
|
|
remote_state_map = {
|
|
|
|
(x.type, x.state_key): x.event_id for x in remote_state
|
|
|
|
}
|
|
|
|
state_maps.append(remote_state_map)
|
|
|
|
|
|
|
|
for x in remote_state:
|
|
|
|
event_map[x.event_id] = x
|
|
|
|
|
2021-09-07 12:15:51 +02:00
|
|
|
room_version = await self._store.get_room_version_id(room_id)
|
2021-08-26 22:41:44 +02:00
|
|
|
state_map = await self._state_resolution_handler.resolve_events_with_store(
|
|
|
|
room_id,
|
|
|
|
room_version,
|
|
|
|
state_maps,
|
|
|
|
event_map,
|
2021-09-07 12:15:51 +02:00
|
|
|
state_res_store=StateResolutionStore(self._store),
|
2021-08-26 22:41:44 +02:00
|
|
|
)
|
|
|
|
|
|
|
|
# We need to give _process_received_pdu the actual state events
|
|
|
|
# rather than event ids, so generate that now.
|
|
|
|
|
|
|
|
# First though we need to fetch all the events that are in
|
|
|
|
# state_map, so we can build up the state below.
|
2021-09-07 12:15:51 +02:00
|
|
|
evs = await self._store.get_events(
|
2021-08-26 22:41:44 +02:00
|
|
|
list(state_map.values()),
|
|
|
|
get_prev_content=False,
|
|
|
|
redact_behaviour=EventRedactBehaviour.AS_IS,
|
|
|
|
)
|
|
|
|
event_map.update(evs)
|
|
|
|
|
|
|
|
state = [event_map[e] for e in state_map.values()]
|
|
|
|
except Exception:
|
|
|
|
logger.warning(
|
|
|
|
"Error attempting to resolve state at missing prev_events",
|
|
|
|
exc_info=True,
|
|
|
|
)
|
|
|
|
raise FederationError(
|
|
|
|
"ERROR",
|
|
|
|
403,
|
|
|
|
"We can't get valid state history.",
|
|
|
|
affected=event_id,
|
|
|
|
)
|
|
|
|
return state
|
|
|
|
|
|
|
|
async def _get_state_after_missing_prev_event(
|
|
|
|
self,
|
|
|
|
destination: str,
|
|
|
|
room_id: str,
|
|
|
|
event_id: str,
|
|
|
|
) -> List[EventBase]:
|
|
|
|
"""Requests all of the room state at a given event from a remote homeserver.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
destination: The remote homeserver to query for the state.
|
|
|
|
room_id: The id of the room we're interested in.
|
|
|
|
event_id: The id of the event we want the state at.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
A list of events in the state, including the event itself
|
|
|
|
"""
|
|
|
|
(
|
|
|
|
state_event_ids,
|
|
|
|
auth_event_ids,
|
2021-09-07 12:15:51 +02:00
|
|
|
) = await self._federation_client.get_room_state_ids(
|
2021-08-26 22:41:44 +02:00
|
|
|
destination, room_id, event_id=event_id
|
|
|
|
)
|
|
|
|
|
|
|
|
logger.debug(
|
|
|
|
"state_ids returned %i state events, %i auth events",
|
|
|
|
len(state_event_ids),
|
|
|
|
len(auth_event_ids),
|
|
|
|
)
|
|
|
|
|
|
|
|
# start by just trying to fetch the events from the store
|
|
|
|
desired_events = set(state_event_ids)
|
|
|
|
desired_events.add(event_id)
|
|
|
|
logger.debug("Fetching %i events from cache/store", len(desired_events))
|
2021-09-07 12:15:51 +02:00
|
|
|
fetched_events = await self._store.get_events(
|
2021-08-26 22:41:44 +02:00
|
|
|
desired_events, allow_rejected=True
|
|
|
|
)
|
|
|
|
|
|
|
|
missing_desired_events = desired_events - fetched_events.keys()
|
|
|
|
logger.debug(
|
|
|
|
"We are missing %i events (got %i)",
|
|
|
|
len(missing_desired_events),
|
|
|
|
len(fetched_events),
|
|
|
|
)
|
|
|
|
|
|
|
|
# We probably won't need most of the auth events, so let's just check which
|
|
|
|
# we have for now, rather than thrashing the event cache with them all
|
|
|
|
# unnecessarily.
|
|
|
|
|
|
|
|
# TODO: we probably won't actually need all of the auth events, since we
|
|
|
|
# already have a bunch of the state events. It would be nice if the
|
|
|
|
# federation api gave us a way of finding out which we actually need.
|
|
|
|
|
|
|
|
missing_auth_events = set(auth_event_ids) - fetched_events.keys()
|
|
|
|
missing_auth_events.difference_update(
|
2021-09-07 12:15:51 +02:00
|
|
|
await self._store.have_seen_events(room_id, missing_auth_events)
|
2021-08-26 22:41:44 +02:00
|
|
|
)
|
|
|
|
logger.debug("We are also missing %i auth events", len(missing_auth_events))
|
|
|
|
|
|
|
|
missing_events = missing_desired_events | missing_auth_events
|
|
|
|
logger.debug("Fetching %i events from remote", len(missing_events))
|
|
|
|
await self._get_events_and_persist(
|
2021-09-08 11:37:50 +02:00
|
|
|
destination=destination, room_id=room_id, event_ids=missing_events
|
2021-08-26 22:41:44 +02:00
|
|
|
)
|
|
|
|
|
|
|
|
# we need to make sure we re-load from the database to get the rejected
|
|
|
|
# state correct.
|
|
|
|
fetched_events.update(
|
2021-09-07 12:15:51 +02:00
|
|
|
await self._store.get_events(missing_desired_events, allow_rejected=True)
|
2021-08-26 22:41:44 +02:00
|
|
|
)
|
|
|
|
|
|
|
|
# check for events which were in the wrong room.
|
|
|
|
#
|
|
|
|
# this can happen if a remote server claims that the state or
|
|
|
|
# auth_events at an event in room A are actually events in room B
|
|
|
|
|
|
|
|
bad_events = [
|
|
|
|
(event_id, event.room_id)
|
|
|
|
for event_id, event in fetched_events.items()
|
|
|
|
if event.room_id != room_id
|
|
|
|
]
|
|
|
|
|
|
|
|
for bad_event_id, bad_room_id in bad_events:
|
|
|
|
# This is a bogus situation, but since we may only discover it a long time
|
|
|
|
# after it happened, we try our best to carry on, by just omitting the
|
|
|
|
# bad events from the returned state set.
|
|
|
|
logger.warning(
|
|
|
|
"Remote server %s claims event %s in room %s is an auth/state "
|
|
|
|
"event in room %s",
|
|
|
|
destination,
|
|
|
|
bad_event_id,
|
|
|
|
bad_room_id,
|
|
|
|
room_id,
|
|
|
|
)
|
|
|
|
|
|
|
|
del fetched_events[bad_event_id]
|
|
|
|
|
|
|
|
# if we couldn't get the prev event in question, that's a problem.
|
|
|
|
remote_event = fetched_events.get(event_id)
|
|
|
|
if not remote_event:
|
|
|
|
raise Exception("Unable to get missing prev_event %s" % (event_id,))
|
|
|
|
|
|
|
|
# missing state at that event is a warning, not a blocker
|
|
|
|
# XXX: this doesn't sound right? it means that we'll end up with incomplete
|
|
|
|
# state.
|
|
|
|
failed_to_fetch = desired_events - fetched_events.keys()
|
|
|
|
if failed_to_fetch:
|
|
|
|
logger.warning(
|
|
|
|
"Failed to fetch missing state events for %s %s",
|
|
|
|
event_id,
|
|
|
|
failed_to_fetch,
|
|
|
|
)
|
|
|
|
|
|
|
|
remote_state = [
|
|
|
|
fetched_events[e_id] for e_id in state_event_ids if e_id in fetched_events
|
|
|
|
]
|
|
|
|
|
|
|
|
if remote_event.is_state() and remote_event.rejected_reason is None:
|
|
|
|
remote_state.append(remote_event)
|
|
|
|
|
|
|
|
return remote_state
|
|
|
|
|
|
|
|
async def _process_received_pdu(
|
|
|
|
self,
|
|
|
|
origin: str,
|
|
|
|
event: EventBase,
|
|
|
|
state: Optional[Iterable[EventBase]],
|
|
|
|
backfilled: bool = False,
|
|
|
|
) -> None:
|
2021-09-08 11:41:13 +02:00
|
|
|
"""Called when we have a new non-outlier event.
|
|
|
|
|
|
|
|
This is called when we have a new event to add to the room DAG - either directly
|
|
|
|
via a /send request, retrieved via get_missing_events after a /send request, or
|
|
|
|
backfilled after a client request.
|
|
|
|
|
|
|
|
We need to do auth checks and put it through the StateHandler.
|
2021-08-26 22:41:44 +02:00
|
|
|
|
|
|
|
Args:
|
|
|
|
origin: server sending the event
|
|
|
|
|
|
|
|
event: event to be persisted
|
|
|
|
|
|
|
|
state: Normally None, but if we are handling a gap in the graph
|
|
|
|
(ie, we are missing one or more prev_events), the resolved state at the
|
|
|
|
event
|
|
|
|
|
|
|
|
backfilled: True if this is part of a historical batch of events (inhibits
|
|
|
|
notification to clients, and validation of device keys.)
|
|
|
|
"""
|
|
|
|
logger.debug("Processing event: %s", event)
|
2021-09-08 11:41:13 +02:00
|
|
|
assert not event.internal_metadata.outlier
|
2021-08-26 22:41:44 +02:00
|
|
|
|
|
|
|
try:
|
2021-09-07 12:15:51 +02:00
|
|
|
context = await self._state_handler.compute_event_context(
|
2021-08-26 22:41:44 +02:00
|
|
|
event, old_state=state
|
|
|
|
)
|
2021-09-08 20:03:08 +02:00
|
|
|
context = await self._check_event_auth(
|
|
|
|
origin,
|
|
|
|
event,
|
|
|
|
context,
|
|
|
|
state=state,
|
|
|
|
backfilled=backfilled,
|
2021-08-26 22:41:44 +02:00
|
|
|
)
|
|
|
|
except AuthError as e:
|
|
|
|
raise FederationError("ERROR", e.code, e.msg, affected=event.event_id)
|
|
|
|
|
2021-09-08 20:03:08 +02:00
|
|
|
await self._run_push_actions_and_persist_event(event, context, backfilled)
|
|
|
|
|
2021-08-26 22:41:44 +02:00
|
|
|
if backfilled:
|
|
|
|
return
|
|
|
|
|
|
|
|
# For encrypted messages we check that we know about the sending device,
|
|
|
|
# if we don't then we mark the device cache for that user as stale.
|
|
|
|
if event.type == EventTypes.Encrypted:
|
|
|
|
device_id = event.content.get("device_id")
|
|
|
|
sender_key = event.content.get("sender_key")
|
|
|
|
|
2021-09-07 12:15:51 +02:00
|
|
|
cached_devices = await self._store.get_cached_devices_for_user(event.sender)
|
2021-08-26 22:41:44 +02:00
|
|
|
|
|
|
|
resync = False # Whether we should resync device lists.
|
|
|
|
|
|
|
|
device = None
|
|
|
|
if device_id is not None:
|
|
|
|
device = cached_devices.get(device_id)
|
|
|
|
if device is None:
|
|
|
|
logger.info(
|
|
|
|
"Received event from remote device not in our cache: %s %s",
|
|
|
|
event.sender,
|
|
|
|
device_id,
|
|
|
|
)
|
|
|
|
resync = True
|
|
|
|
|
|
|
|
# We also check if the `sender_key` matches what we expect.
|
|
|
|
if sender_key is not None:
|
|
|
|
# Figure out what sender key we're expecting. If we know the
|
|
|
|
# device and recognize the algorithm then we can work out the
|
|
|
|
# exact key to expect. Otherwise check it matches any key we
|
|
|
|
# have for that device.
|
|
|
|
|
|
|
|
current_keys: Container[str] = []
|
|
|
|
|
|
|
|
if device:
|
|
|
|
keys = device.get("keys", {}).get("keys", {})
|
|
|
|
|
|
|
|
if (
|
|
|
|
event.content.get("algorithm")
|
|
|
|
== RoomEncryptionAlgorithms.MEGOLM_V1_AES_SHA2
|
|
|
|
):
|
|
|
|
# For this algorithm we expect a curve25519 key.
|
|
|
|
key_name = "curve25519:%s" % (device_id,)
|
|
|
|
current_keys = [keys.get(key_name)]
|
|
|
|
else:
|
|
|
|
# We don't know understand the algorithm, so we just
|
|
|
|
# check it matches a key for the device.
|
|
|
|
current_keys = keys.values()
|
|
|
|
elif device_id:
|
|
|
|
# We don't have any keys for the device ID.
|
|
|
|
pass
|
|
|
|
else:
|
|
|
|
# The event didn't include a device ID, so we just look for
|
|
|
|
# keys across all devices.
|
|
|
|
current_keys = [
|
|
|
|
key
|
|
|
|
for device in cached_devices.values()
|
|
|
|
for key in device.get("keys", {}).get("keys", {}).values()
|
|
|
|
]
|
|
|
|
|
|
|
|
# We now check that the sender key matches (one of) the expected
|
|
|
|
# keys.
|
|
|
|
if sender_key not in current_keys:
|
|
|
|
logger.info(
|
|
|
|
"Received event from remote device with unexpected sender key: %s %s: %s",
|
|
|
|
event.sender,
|
|
|
|
device_id or "<no device_id>",
|
|
|
|
sender_key,
|
|
|
|
)
|
|
|
|
resync = True
|
|
|
|
|
|
|
|
if resync:
|
|
|
|
run_as_background_process(
|
|
|
|
"resync_device_due_to_pdu",
|
|
|
|
self._resync_device,
|
|
|
|
event.sender,
|
|
|
|
)
|
|
|
|
|
|
|
|
await self._handle_marker_event(origin, event)
|
|
|
|
|
|
|
|
async def _resync_device(self, sender: str) -> None:
|
|
|
|
"""We have detected that the device list for the given user may be out
|
|
|
|
of sync, so we try and resync them.
|
|
|
|
"""
|
|
|
|
|
|
|
|
try:
|
2021-09-07 12:15:51 +02:00
|
|
|
await self._store.mark_remote_user_device_cache_as_stale(sender)
|
2021-08-26 22:41:44 +02:00
|
|
|
|
|
|
|
# Immediately attempt a resync in the background
|
2021-09-13 19:07:12 +02:00
|
|
|
if self._config.worker.worker_app:
|
2021-08-26 22:41:44 +02:00
|
|
|
await self._user_device_resync(user_id=sender)
|
|
|
|
else:
|
|
|
|
await self._device_list_updater.user_device_resync(sender)
|
|
|
|
except Exception:
|
|
|
|
logger.exception("Failed to resync device for %s", sender)
|
|
|
|
|
2021-09-20 14:56:23 +02:00
|
|
|
async def _handle_marker_event(self, origin: str, marker_event: EventBase) -> None:
|
2021-08-26 22:41:44 +02:00
|
|
|
"""Handles backfilling the insertion event when we receive a marker
|
|
|
|
event that points to one.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
origin: Origin of the event. Will be called to get the insertion event
|
|
|
|
marker_event: The event to process
|
|
|
|
"""
|
|
|
|
|
|
|
|
if marker_event.type != EventTypes.MSC2716_MARKER:
|
|
|
|
# Not a marker event
|
|
|
|
return
|
|
|
|
|
|
|
|
if marker_event.rejected_reason is not None:
|
|
|
|
# Rejected event
|
|
|
|
return
|
|
|
|
|
|
|
|
# Skip processing a marker event if the room version doesn't
|
Allow room creator to send MSC2716 related events in existing room versions (#10566)
* Allow room creator to send MSC2716 related events in existing room versions
Discussed at https://github.com/matrix-org/matrix-doc/pull/2716/#discussion_r682474869
Restoring `get_create_event_for_room_txn` from,
https://github.com/matrix-org/synapse/pull/10245/commits/44bb3f0cf5cb365ef9281554daceeecfb17cc94d
* Add changelog
* Stop people from trying to redact MSC2716 events in unsupported room versions
* Populate rooms.creator column for easy lookup
> From some [out of band discussion](https://matrix.to/#/!UytJQHLQYfvYWsGrGY:jki.re/$p2fKESoFst038x6pOOmsY0C49S2gLKMr0jhNMz_JJz0?via=jki.re&via=matrix.org), my plan is to use `rooms.creator`. But currently, we don't fill in `creator` for remote rooms when a user is invited to a room for example. So we need to add some code to fill in `creator` wherever we add to the `rooms` table. And also add a background update to fill in the rows missing `creator` (we can use the same logic that `get_create_event_for_room_txn` is doing by looking in the state events to get the `creator`).
>
> https://github.com/matrix-org/synapse/pull/10566#issuecomment-901616642
* Remove and switch away from get_create_event_for_room_txn
* Fix no create event being found because no state events persisted yet
* Fix and add tests for rooms creator bg update
* Populate rooms.creator field for easy lookup
Part of https://github.com/matrix-org/synapse/pull/10566
- Fill in creator whenever we insert into the rooms table
- Add background update to backfill any missing creator values
* Add changelog
* Fix usage
* Remove extra delta already included in #10697
* Don't worry about setting creator for invite
* Only iterate over rows missing the creator
See https://github.com/matrix-org/synapse/pull/10697#discussion_r695940898
* Use constant to fetch room creator field
See https://github.com/matrix-org/synapse/pull/10697#discussion_r696803029
* More protection from other random types
See https://github.com/matrix-org/synapse/pull/10697#discussion_r696806853
* Move new background update to end of list
See https://github.com/matrix-org/synapse/pull/10697#discussion_r696814181
* Fix query casing
* Fix ambiguity iterating over cursor instead of list
Fix `psycopg2.ProgrammingError: no results to fetch` error
when tests run with Postgres.
```
SYNAPSE_POSTGRES=1 SYNAPSE_TEST_LOG_LEVEL=INFO python -m twisted.trial tests.storage.databases.main.test_room
```
---
We use `txn.fetchall` because it will return the results as a
list or an empty list when there are no results.
Docs:
> `cursor` objects are iterable, so, instead of calling explicitly fetchone() in a loop, the object itself can be used:
>
> https://www.psycopg.org/docs/cursor.html#cursor-iterable
And I'm guessing iterating over a raw cursor does something weird when there are no results.
---
Test CI failure: https://github.com/matrix-org/synapse/pull/10697/checks?check_run_id=3468916530
```
tests.test_visibility.FilterEventsForServerTestCase.test_large_room
===============================================================================
[FAIL]
Traceback (most recent call last):
File "/home/runner/work/synapse/synapse/tests/storage/databases/main/test_room.py", line 85, in test_background_populate_rooms_creator_column
self.get_success(
File "/home/runner/work/synapse/synapse/tests/unittest.py", line 500, in get_success
return self.successResultOf(d)
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/trial/_synctest.py", line 700, in successResultOf
self.fail(
twisted.trial.unittest.FailTest: Success result expected on <Deferred at 0x7f4022f3eb50 current result: None>, found failure result instead:
Traceback (most recent call last):
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/internet/defer.py", line 701, in errback
self._startRunCallbacks(fail)
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/internet/defer.py", line 764, in _startRunCallbacks
self._runCallbacks()
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/internet/defer.py", line 858, in _runCallbacks
current.result = callback( # type: ignore[misc]
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/internet/defer.py", line 1751, in gotResult
current_context.run(_inlineCallbacks, r, gen, status)
--- <exception caught here> ---
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/internet/defer.py", line 1657, in _inlineCallbacks
result = current_context.run(
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/python/failure.py", line 500, in throwExceptionIntoGenerator
return g.throw(self.type, self.value, self.tb)
File "/home/runner/work/synapse/synapse/synapse/storage/background_updates.py", line 224, in do_next_background_update
await self._do_background_update(desired_duration_ms)
File "/home/runner/work/synapse/synapse/synapse/storage/background_updates.py", line 261, in _do_background_update
items_updated = await update_handler(progress, batch_size)
File "/home/runner/work/synapse/synapse/synapse/storage/databases/main/room.py", line 1399, in _background_populate_rooms_creator_column
end = await self.db_pool.runInteraction(
File "/home/runner/work/synapse/synapse/synapse/storage/database.py", line 686, in runInteraction
result = await self.runWithConnection(
File "/home/runner/work/synapse/synapse/synapse/storage/database.py", line 791, in runWithConnection
return await make_deferred_yieldable(
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/internet/defer.py", line 858, in _runCallbacks
current.result = callback( # type: ignore[misc]
File "/home/runner/work/synapse/synapse/tests/server.py", line 425, in <lambda>
d.addCallback(lambda x: function(*args, **kwargs))
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/enterprise/adbapi.py", line 293, in _runWithConnection
compat.reraise(excValue, excTraceback)
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/python/deprecate.py", line 298, in deprecatedFunction
return function(*args, **kwargs)
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/python/compat.py", line 404, in reraise
raise exception.with_traceback(traceback)
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/enterprise/adbapi.py", line 284, in _runWithConnection
result = func(conn, *args, **kw)
File "/home/runner/work/synapse/synapse/synapse/storage/database.py", line 786, in inner_func
return func(db_conn, *args, **kwargs)
File "/home/runner/work/synapse/synapse/synapse/storage/database.py", line 554, in new_transaction
r = func(cursor, *args, **kwargs)
File "/home/runner/work/synapse/synapse/synapse/storage/databases/main/room.py", line 1375, in _background_populate_rooms_creator_column_txn
for room_id, event_json in txn:
psycopg2.ProgrammingError: no results to fetch
```
* Move code not under the MSC2716 room version underneath an experimental config option
See https://github.com/matrix-org/synapse/pull/10566#issuecomment-906437909
* Add ordering to rooms creator background update
See https://github.com/matrix-org/synapse/pull/10697#discussion_r696815277
* Add comment to better document constant
See https://github.com/matrix-org/synapse/pull/10697#discussion_r699674458
* Use constant field
2021-09-04 07:58:49 +02:00
|
|
|
# support it or the event is not from the room creator.
|
2021-09-07 12:15:51 +02:00
|
|
|
room_version = await self._store.get_room_version(marker_event.room_id)
|
|
|
|
create_event = await self._store.get_create_event_for_room(marker_event.room_id)
|
Allow room creator to send MSC2716 related events in existing room versions (#10566)
* Allow room creator to send MSC2716 related events in existing room versions
Discussed at https://github.com/matrix-org/matrix-doc/pull/2716/#discussion_r682474869
Restoring `get_create_event_for_room_txn` from,
https://github.com/matrix-org/synapse/pull/10245/commits/44bb3f0cf5cb365ef9281554daceeecfb17cc94d
* Add changelog
* Stop people from trying to redact MSC2716 events in unsupported room versions
* Populate rooms.creator column for easy lookup
> From some [out of band discussion](https://matrix.to/#/!UytJQHLQYfvYWsGrGY:jki.re/$p2fKESoFst038x6pOOmsY0C49S2gLKMr0jhNMz_JJz0?via=jki.re&via=matrix.org), my plan is to use `rooms.creator`. But currently, we don't fill in `creator` for remote rooms when a user is invited to a room for example. So we need to add some code to fill in `creator` wherever we add to the `rooms` table. And also add a background update to fill in the rows missing `creator` (we can use the same logic that `get_create_event_for_room_txn` is doing by looking in the state events to get the `creator`).
>
> https://github.com/matrix-org/synapse/pull/10566#issuecomment-901616642
* Remove and switch away from get_create_event_for_room_txn
* Fix no create event being found because no state events persisted yet
* Fix and add tests for rooms creator bg update
* Populate rooms.creator field for easy lookup
Part of https://github.com/matrix-org/synapse/pull/10566
- Fill in creator whenever we insert into the rooms table
- Add background update to backfill any missing creator values
* Add changelog
* Fix usage
* Remove extra delta already included in #10697
* Don't worry about setting creator for invite
* Only iterate over rows missing the creator
See https://github.com/matrix-org/synapse/pull/10697#discussion_r695940898
* Use constant to fetch room creator field
See https://github.com/matrix-org/synapse/pull/10697#discussion_r696803029
* More protection from other random types
See https://github.com/matrix-org/synapse/pull/10697#discussion_r696806853
* Move new background update to end of list
See https://github.com/matrix-org/synapse/pull/10697#discussion_r696814181
* Fix query casing
* Fix ambiguity iterating over cursor instead of list
Fix `psycopg2.ProgrammingError: no results to fetch` error
when tests run with Postgres.
```
SYNAPSE_POSTGRES=1 SYNAPSE_TEST_LOG_LEVEL=INFO python -m twisted.trial tests.storage.databases.main.test_room
```
---
We use `txn.fetchall` because it will return the results as a
list or an empty list when there are no results.
Docs:
> `cursor` objects are iterable, so, instead of calling explicitly fetchone() in a loop, the object itself can be used:
>
> https://www.psycopg.org/docs/cursor.html#cursor-iterable
And I'm guessing iterating over a raw cursor does something weird when there are no results.
---
Test CI failure: https://github.com/matrix-org/synapse/pull/10697/checks?check_run_id=3468916530
```
tests.test_visibility.FilterEventsForServerTestCase.test_large_room
===============================================================================
[FAIL]
Traceback (most recent call last):
File "/home/runner/work/synapse/synapse/tests/storage/databases/main/test_room.py", line 85, in test_background_populate_rooms_creator_column
self.get_success(
File "/home/runner/work/synapse/synapse/tests/unittest.py", line 500, in get_success
return self.successResultOf(d)
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/trial/_synctest.py", line 700, in successResultOf
self.fail(
twisted.trial.unittest.FailTest: Success result expected on <Deferred at 0x7f4022f3eb50 current result: None>, found failure result instead:
Traceback (most recent call last):
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/internet/defer.py", line 701, in errback
self._startRunCallbacks(fail)
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/internet/defer.py", line 764, in _startRunCallbacks
self._runCallbacks()
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/internet/defer.py", line 858, in _runCallbacks
current.result = callback( # type: ignore[misc]
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/internet/defer.py", line 1751, in gotResult
current_context.run(_inlineCallbacks, r, gen, status)
--- <exception caught here> ---
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/internet/defer.py", line 1657, in _inlineCallbacks
result = current_context.run(
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/python/failure.py", line 500, in throwExceptionIntoGenerator
return g.throw(self.type, self.value, self.tb)
File "/home/runner/work/synapse/synapse/synapse/storage/background_updates.py", line 224, in do_next_background_update
await self._do_background_update(desired_duration_ms)
File "/home/runner/work/synapse/synapse/synapse/storage/background_updates.py", line 261, in _do_background_update
items_updated = await update_handler(progress, batch_size)
File "/home/runner/work/synapse/synapse/synapse/storage/databases/main/room.py", line 1399, in _background_populate_rooms_creator_column
end = await self.db_pool.runInteraction(
File "/home/runner/work/synapse/synapse/synapse/storage/database.py", line 686, in runInteraction
result = await self.runWithConnection(
File "/home/runner/work/synapse/synapse/synapse/storage/database.py", line 791, in runWithConnection
return await make_deferred_yieldable(
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/internet/defer.py", line 858, in _runCallbacks
current.result = callback( # type: ignore[misc]
File "/home/runner/work/synapse/synapse/tests/server.py", line 425, in <lambda>
d.addCallback(lambda x: function(*args, **kwargs))
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/enterprise/adbapi.py", line 293, in _runWithConnection
compat.reraise(excValue, excTraceback)
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/python/deprecate.py", line 298, in deprecatedFunction
return function(*args, **kwargs)
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/python/compat.py", line 404, in reraise
raise exception.with_traceback(traceback)
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/enterprise/adbapi.py", line 284, in _runWithConnection
result = func(conn, *args, **kw)
File "/home/runner/work/synapse/synapse/synapse/storage/database.py", line 786, in inner_func
return func(db_conn, *args, **kwargs)
File "/home/runner/work/synapse/synapse/synapse/storage/database.py", line 554, in new_transaction
r = func(cursor, *args, **kwargs)
File "/home/runner/work/synapse/synapse/synapse/storage/databases/main/room.py", line 1375, in _background_populate_rooms_creator_column_txn
for room_id, event_json in txn:
psycopg2.ProgrammingError: no results to fetch
```
* Move code not under the MSC2716 room version underneath an experimental config option
See https://github.com/matrix-org/synapse/pull/10566#issuecomment-906437909
* Add ordering to rooms creator background update
See https://github.com/matrix-org/synapse/pull/10697#discussion_r696815277
* Add comment to better document constant
See https://github.com/matrix-org/synapse/pull/10697#discussion_r699674458
* Use constant field
2021-09-04 07:58:49 +02:00
|
|
|
room_creator = create_event.content.get(EventContentFields.ROOM_CREATOR)
|
|
|
|
if (
|
|
|
|
not room_version.msc2716_historical
|
2021-09-07 12:15:51 +02:00
|
|
|
or not self._config.experimental.msc2716_enabled
|
Allow room creator to send MSC2716 related events in existing room versions (#10566)
* Allow room creator to send MSC2716 related events in existing room versions
Discussed at https://github.com/matrix-org/matrix-doc/pull/2716/#discussion_r682474869
Restoring `get_create_event_for_room_txn` from,
https://github.com/matrix-org/synapse/pull/10245/commits/44bb3f0cf5cb365ef9281554daceeecfb17cc94d
* Add changelog
* Stop people from trying to redact MSC2716 events in unsupported room versions
* Populate rooms.creator column for easy lookup
> From some [out of band discussion](https://matrix.to/#/!UytJQHLQYfvYWsGrGY:jki.re/$p2fKESoFst038x6pOOmsY0C49S2gLKMr0jhNMz_JJz0?via=jki.re&via=matrix.org), my plan is to use `rooms.creator`. But currently, we don't fill in `creator` for remote rooms when a user is invited to a room for example. So we need to add some code to fill in `creator` wherever we add to the `rooms` table. And also add a background update to fill in the rows missing `creator` (we can use the same logic that `get_create_event_for_room_txn` is doing by looking in the state events to get the `creator`).
>
> https://github.com/matrix-org/synapse/pull/10566#issuecomment-901616642
* Remove and switch away from get_create_event_for_room_txn
* Fix no create event being found because no state events persisted yet
* Fix and add tests for rooms creator bg update
* Populate rooms.creator field for easy lookup
Part of https://github.com/matrix-org/synapse/pull/10566
- Fill in creator whenever we insert into the rooms table
- Add background update to backfill any missing creator values
* Add changelog
* Fix usage
* Remove extra delta already included in #10697
* Don't worry about setting creator for invite
* Only iterate over rows missing the creator
See https://github.com/matrix-org/synapse/pull/10697#discussion_r695940898
* Use constant to fetch room creator field
See https://github.com/matrix-org/synapse/pull/10697#discussion_r696803029
* More protection from other random types
See https://github.com/matrix-org/synapse/pull/10697#discussion_r696806853
* Move new background update to end of list
See https://github.com/matrix-org/synapse/pull/10697#discussion_r696814181
* Fix query casing
* Fix ambiguity iterating over cursor instead of list
Fix `psycopg2.ProgrammingError: no results to fetch` error
when tests run with Postgres.
```
SYNAPSE_POSTGRES=1 SYNAPSE_TEST_LOG_LEVEL=INFO python -m twisted.trial tests.storage.databases.main.test_room
```
---
We use `txn.fetchall` because it will return the results as a
list or an empty list when there are no results.
Docs:
> `cursor` objects are iterable, so, instead of calling explicitly fetchone() in a loop, the object itself can be used:
>
> https://www.psycopg.org/docs/cursor.html#cursor-iterable
And I'm guessing iterating over a raw cursor does something weird when there are no results.
---
Test CI failure: https://github.com/matrix-org/synapse/pull/10697/checks?check_run_id=3468916530
```
tests.test_visibility.FilterEventsForServerTestCase.test_large_room
===============================================================================
[FAIL]
Traceback (most recent call last):
File "/home/runner/work/synapse/synapse/tests/storage/databases/main/test_room.py", line 85, in test_background_populate_rooms_creator_column
self.get_success(
File "/home/runner/work/synapse/synapse/tests/unittest.py", line 500, in get_success
return self.successResultOf(d)
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/trial/_synctest.py", line 700, in successResultOf
self.fail(
twisted.trial.unittest.FailTest: Success result expected on <Deferred at 0x7f4022f3eb50 current result: None>, found failure result instead:
Traceback (most recent call last):
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/internet/defer.py", line 701, in errback
self._startRunCallbacks(fail)
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/internet/defer.py", line 764, in _startRunCallbacks
self._runCallbacks()
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/internet/defer.py", line 858, in _runCallbacks
current.result = callback( # type: ignore[misc]
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/internet/defer.py", line 1751, in gotResult
current_context.run(_inlineCallbacks, r, gen, status)
--- <exception caught here> ---
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/internet/defer.py", line 1657, in _inlineCallbacks
result = current_context.run(
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/python/failure.py", line 500, in throwExceptionIntoGenerator
return g.throw(self.type, self.value, self.tb)
File "/home/runner/work/synapse/synapse/synapse/storage/background_updates.py", line 224, in do_next_background_update
await self._do_background_update(desired_duration_ms)
File "/home/runner/work/synapse/synapse/synapse/storage/background_updates.py", line 261, in _do_background_update
items_updated = await update_handler(progress, batch_size)
File "/home/runner/work/synapse/synapse/synapse/storage/databases/main/room.py", line 1399, in _background_populate_rooms_creator_column
end = await self.db_pool.runInteraction(
File "/home/runner/work/synapse/synapse/synapse/storage/database.py", line 686, in runInteraction
result = await self.runWithConnection(
File "/home/runner/work/synapse/synapse/synapse/storage/database.py", line 791, in runWithConnection
return await make_deferred_yieldable(
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/internet/defer.py", line 858, in _runCallbacks
current.result = callback( # type: ignore[misc]
File "/home/runner/work/synapse/synapse/tests/server.py", line 425, in <lambda>
d.addCallback(lambda x: function(*args, **kwargs))
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/enterprise/adbapi.py", line 293, in _runWithConnection
compat.reraise(excValue, excTraceback)
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/python/deprecate.py", line 298, in deprecatedFunction
return function(*args, **kwargs)
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/python/compat.py", line 404, in reraise
raise exception.with_traceback(traceback)
File "/home/runner/work/synapse/synapse/.tox/py/lib/python3.9/site-packages/twisted/enterprise/adbapi.py", line 284, in _runWithConnection
result = func(conn, *args, **kw)
File "/home/runner/work/synapse/synapse/synapse/storage/database.py", line 786, in inner_func
return func(db_conn, *args, **kwargs)
File "/home/runner/work/synapse/synapse/synapse/storage/database.py", line 554, in new_transaction
r = func(cursor, *args, **kwargs)
File "/home/runner/work/synapse/synapse/synapse/storage/databases/main/room.py", line 1375, in _background_populate_rooms_creator_column_txn
for room_id, event_json in txn:
psycopg2.ProgrammingError: no results to fetch
```
* Move code not under the MSC2716 room version underneath an experimental config option
See https://github.com/matrix-org/synapse/pull/10566#issuecomment-906437909
* Add ordering to rooms creator background update
See https://github.com/matrix-org/synapse/pull/10697#discussion_r696815277
* Add comment to better document constant
See https://github.com/matrix-org/synapse/pull/10697#discussion_r699674458
* Use constant field
2021-09-04 07:58:49 +02:00
|
|
|
or marker_event.sender != room_creator
|
|
|
|
):
|
2021-08-26 22:41:44 +02:00
|
|
|
return
|
|
|
|
|
|
|
|
logger.debug("_handle_marker_event: received %s", marker_event)
|
|
|
|
|
|
|
|
insertion_event_id = marker_event.content.get(
|
|
|
|
EventContentFields.MSC2716_MARKER_INSERTION
|
|
|
|
)
|
|
|
|
|
|
|
|
if insertion_event_id is None:
|
|
|
|
# Nothing to retrieve then (invalid marker)
|
|
|
|
return
|
|
|
|
|
|
|
|
logger.debug(
|
|
|
|
"_handle_marker_event: backfilling insertion event %s", insertion_event_id
|
|
|
|
)
|
|
|
|
|
|
|
|
await self._get_events_and_persist(
|
|
|
|
origin,
|
|
|
|
marker_event.room_id,
|
|
|
|
[insertion_event_id],
|
|
|
|
)
|
|
|
|
|
2021-09-07 12:15:51 +02:00
|
|
|
insertion_event = await self._store.get_event(
|
2021-08-26 22:41:44 +02:00
|
|
|
insertion_event_id, allow_none=True
|
|
|
|
)
|
|
|
|
if insertion_event is None:
|
|
|
|
logger.warning(
|
|
|
|
"_handle_marker_event: server %s didn't return insertion event %s for marker %s",
|
|
|
|
origin,
|
|
|
|
insertion_event_id,
|
|
|
|
marker_event.event_id,
|
|
|
|
)
|
|
|
|
return
|
|
|
|
|
|
|
|
logger.debug(
|
|
|
|
"_handle_marker_event: succesfully backfilled insertion event %s from marker event %s",
|
|
|
|
insertion_event,
|
|
|
|
marker_event,
|
|
|
|
)
|
|
|
|
|
2021-09-07 12:15:51 +02:00
|
|
|
await self._store.insert_insertion_extremity(
|
2021-08-26 22:41:44 +02:00
|
|
|
insertion_event_id, marker_event.room_id
|
|
|
|
)
|
|
|
|
|
|
|
|
logger.debug(
|
|
|
|
"_handle_marker_event: insertion extremity added for %s from marker event %s",
|
|
|
|
insertion_event,
|
|
|
|
marker_event,
|
|
|
|
)
|
|
|
|
|
|
|
|
async def _get_events_and_persist(
|
2021-09-08 11:37:50 +02:00
|
|
|
self, destination: str, room_id: str, event_ids: Collection[str]
|
2021-08-26 22:41:44 +02:00
|
|
|
) -> None:
|
|
|
|
"""Fetch the given events from a server, and persist them as outliers.
|
|
|
|
|
|
|
|
This function *does not* recursively get missing auth events of the
|
2021-09-08 11:37:50 +02:00
|
|
|
newly fetched events. Callers must include in the `event_ids` argument
|
2021-08-26 22:41:44 +02:00
|
|
|
any missing events from the auth chain.
|
|
|
|
|
|
|
|
Logs a warning if we can't find the given event.
|
|
|
|
"""
|
|
|
|
|
2021-09-07 12:15:51 +02:00
|
|
|
room_version = await self._store.get_room_version(room_id)
|
2021-08-26 22:41:44 +02:00
|
|
|
|
2021-09-24 12:56:33 +02:00
|
|
|
events: List[EventBase] = []
|
2021-08-26 22:41:44 +02:00
|
|
|
|
2021-09-20 14:56:23 +02:00
|
|
|
async def get_event(event_id: str) -> None:
|
2021-08-26 22:41:44 +02:00
|
|
|
with nested_logging_context(event_id):
|
|
|
|
try:
|
2021-09-07 12:15:51 +02:00
|
|
|
event = await self._federation_client.get_pdu(
|
2021-08-26 22:41:44 +02:00
|
|
|
[destination],
|
|
|
|
event_id,
|
|
|
|
room_version,
|
|
|
|
outlier=True,
|
|
|
|
)
|
|
|
|
if event is None:
|
|
|
|
logger.warning(
|
|
|
|
"Server %s didn't return event %s",
|
|
|
|
destination,
|
|
|
|
event_id,
|
|
|
|
)
|
|
|
|
return
|
2021-09-24 12:56:33 +02:00
|
|
|
events.append(event)
|
2021-08-26 22:41:44 +02:00
|
|
|
|
|
|
|
except Exception as e:
|
|
|
|
logger.warning(
|
|
|
|
"Error fetching missing state/auth event %s: %s %s",
|
|
|
|
event_id,
|
|
|
|
type(e),
|
|
|
|
e,
|
|
|
|
)
|
|
|
|
|
2021-09-08 11:37:50 +02:00
|
|
|
await concurrently_execute(get_event, event_ids, 5)
|
2021-09-24 12:56:33 +02:00
|
|
|
logger.info("Fetched %i events of %i requested", len(events), len(event_ids))
|
|
|
|
await self._auth_and_persist_fetched_events(destination, room_id, events)
|
|
|
|
|
|
|
|
async def _auth_and_persist_fetched_events(
|
|
|
|
self, origin: str, room_id: str, events: Iterable[EventBase]
|
|
|
|
) -> None:
|
|
|
|
"""Persist the events fetched by _get_events_and_persist or _get_remote_auth_chain_for_event
|
|
|
|
|
|
|
|
The events to be persisted must be outliers.
|
|
|
|
|
|
|
|
We first sort the events to make sure that we process each event's auth_events
|
|
|
|
before the event itself, and then auth and persist them.
|
|
|
|
|
|
|
|
Notifies about the events where appropriate.
|
|
|
|
|
|
|
|
Params:
|
|
|
|
origin: where the events came from
|
|
|
|
room_id: the room that the events are meant to be in (though this has
|
|
|
|
not yet been checked)
|
|
|
|
events: the events that have been fetched
|
|
|
|
"""
|
|
|
|
event_map = {event.event_id: event for event in events}
|
2021-08-26 22:41:44 +02:00
|
|
|
|
2021-09-08 11:37:50 +02:00
|
|
|
# XXX: it might be possible to kick this process off in parallel with fetching
|
|
|
|
# the events.
|
|
|
|
while event_map:
|
|
|
|
# build a list of events whose auth events are not in the queue.
|
|
|
|
roots = tuple(
|
|
|
|
ev
|
|
|
|
for ev in event_map.values()
|
|
|
|
if not any(aid in event_map for aid in ev.auth_event_ids())
|
|
|
|
)
|
2021-08-26 22:41:44 +02:00
|
|
|
|
2021-09-08 11:37:50 +02:00
|
|
|
if not roots:
|
|
|
|
# if *none* of the remaining events are ready, that means
|
|
|
|
# we have a loop. This either means a bug in our logic, or that
|
|
|
|
# somebody has managed to create a loop (which requires finding a
|
|
|
|
# hash collision in room v2 and later).
|
|
|
|
logger.warning(
|
|
|
|
"Loop found in auth events while fetching missing state/auth "
|
|
|
|
"events: %s",
|
|
|
|
shortstr(event_map.keys()),
|
|
|
|
)
|
|
|
|
return
|
|
|
|
|
|
|
|
logger.info(
|
|
|
|
"Persisting %i of %i remaining events", len(roots), len(event_map)
|
|
|
|
)
|
|
|
|
|
2021-09-24 12:56:33 +02:00
|
|
|
await self._auth_and_persist_fetched_events_inner(origin, room_id, roots)
|
2021-09-08 11:37:50 +02:00
|
|
|
|
|
|
|
for ev in roots:
|
|
|
|
del event_map[ev.event_id]
|
|
|
|
|
2021-09-24 12:56:33 +02:00
|
|
|
async def _auth_and_persist_fetched_events_inner(
|
2021-09-08 11:37:50 +02:00
|
|
|
self, origin: str, room_id: str, fetched_events: Collection[EventBase]
|
|
|
|
) -> None:
|
2021-09-24 12:56:33 +02:00
|
|
|
"""Helper for _auth_and_persist_fetched_events
|
2021-09-08 11:37:50 +02:00
|
|
|
|
2021-09-24 12:56:33 +02:00
|
|
|
Persists a batch of events where we have (theoretically) already persisted all
|
|
|
|
of their auth events.
|
2021-09-08 11:37:50 +02:00
|
|
|
|
|
|
|
Notifies about the events where appropriate.
|
|
|
|
|
|
|
|
Params:
|
|
|
|
origin: where the events came from
|
|
|
|
room_id: the room that the events are meant to be in (though this has
|
|
|
|
not yet been checked)
|
2021-09-24 12:56:33 +02:00
|
|
|
fetched_events: the events to persist
|
2021-09-08 11:37:50 +02:00
|
|
|
"""
|
|
|
|
# get all the auth events for all the events in this batch. By now, they should
|
|
|
|
# have been persisted.
|
|
|
|
auth_events = {
|
|
|
|
aid for event in fetched_events for aid in event.auth_event_ids()
|
|
|
|
}
|
2021-09-07 12:15:51 +02:00
|
|
|
persisted_events = await self._store.get_events(
|
2021-08-26 22:41:44 +02:00
|
|
|
auth_events,
|
|
|
|
allow_rejected=True,
|
|
|
|
)
|
|
|
|
|
2021-09-28 16:25:07 +02:00
|
|
|
room_version = await self._store.get_room_version_id(room_id)
|
|
|
|
room_version_obj = KNOWN_ROOM_VERSIONS[room_version]
|
|
|
|
|
|
|
|
def prep(event: EventBase) -> Optional[Tuple[EventBase, EventContext]]:
|
2021-08-26 22:41:44 +02:00
|
|
|
with nested_logging_context(suffix=event.event_id):
|
2021-09-24 12:56:13 +02:00
|
|
|
auth = {}
|
|
|
|
for auth_event_id in event.auth_event_ids():
|
|
|
|
ae = persisted_events.get(auth_event_id)
|
2021-09-24 15:01:45 +02:00
|
|
|
if not ae:
|
|
|
|
logger.warning(
|
|
|
|
"Event %s relies on auth_event %s, which could not be found.",
|
|
|
|
event,
|
|
|
|
auth_event_id,
|
|
|
|
)
|
|
|
|
# the fact we can't find the auth event doesn't mean it doesn't
|
|
|
|
# exist, which means it is premature to reject `event`. Instead we
|
|
|
|
# just ignore it for now.
|
|
|
|
return None
|
|
|
|
auth[(ae.type, ae.state_key)] = ae
|
2021-09-24 12:56:13 +02:00
|
|
|
|
|
|
|
context = EventContext.for_outlier()
|
2021-09-28 16:25:07 +02:00
|
|
|
try:
|
Split `event_auth.check` into two parts (#10940)
Broadly, the existing `event_auth.check` function has two parts:
* a validation section: checks that the event isn't too big, that it has the rught signatures, etc.
This bit is independent of the rest of the state in the room, and so need only be done once
for each event.
* an auth section: ensures that the event is allowed, given the rest of the state in the room.
This gets done multiple times, against various sets of room state, because it forms part of
the state res algorithm.
Currently, this is implemented with `do_sig_check` and `do_size_check` parameters, but I think
that makes everything hard to follow. Instead, we split the function in two and call each part
separately where it is needed.
2021-09-29 19:59:15 +02:00
|
|
|
validate_event_for_room_version(room_version_obj, event)
|
|
|
|
check_auth_rules_for_event(room_version_obj, event, auth)
|
2021-09-28 16:25:07 +02:00
|
|
|
except AuthError as e:
|
|
|
|
logger.warning("Rejecting %r because %s", event, e)
|
|
|
|
context.rejected = RejectedReason.AUTH_ERROR
|
|
|
|
|
2021-09-24 12:56:13 +02:00
|
|
|
return event, context
|
2021-08-26 22:41:44 +02:00
|
|
|
|
2021-09-28 16:25:07 +02:00
|
|
|
events_to_persist = (x for x in (prep(event) for event in fetched_events) if x)
|
2021-09-24 15:01:45 +02:00
|
|
|
await self.persist_events_and_notify(room_id, tuple(events_to_persist))
|
2021-08-26 22:41:44 +02:00
|
|
|
|
|
|
|
async def _check_event_auth(
|
|
|
|
self,
|
|
|
|
origin: str,
|
|
|
|
event: EventBase,
|
|
|
|
context: EventContext,
|
|
|
|
state: Optional[Iterable[EventBase]] = None,
|
|
|
|
backfilled: bool = False,
|
|
|
|
) -> EventContext:
|
|
|
|
"""
|
|
|
|
Checks whether an event should be rejected (for failing auth checks).
|
|
|
|
|
|
|
|
Args:
|
|
|
|
origin: The host the event originates from.
|
|
|
|
event: The event itself.
|
|
|
|
context:
|
|
|
|
The event context.
|
|
|
|
|
|
|
|
state:
|
|
|
|
The state events used to check the event for soft-fail. If this is
|
|
|
|
not provided the current state events will be used.
|
|
|
|
|
|
|
|
backfilled: True if the event was backfilled.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
The updated context object.
|
|
|
|
"""
|
2021-09-28 16:25:07 +02:00
|
|
|
# This method should only be used for non-outliers
|
|
|
|
assert not event.internal_metadata.outlier
|
2021-09-08 11:41:13 +02:00
|
|
|
|
2021-10-05 13:50:07 +02:00
|
|
|
# first of all, check that the event itself is valid.
|
2021-09-07 12:15:51 +02:00
|
|
|
room_version = await self._store.get_room_version_id(event.room_id)
|
2021-08-26 22:41:44 +02:00
|
|
|
room_version_obj = KNOWN_ROOM_VERSIONS[room_version]
|
|
|
|
|
2021-10-05 13:50:07 +02:00
|
|
|
try:
|
|
|
|
validate_event_for_room_version(room_version_obj, event)
|
|
|
|
except AuthError as e:
|
|
|
|
logger.warning("While validating received event %r: %s", event, e)
|
|
|
|
# TODO: use a different rejected reason here?
|
|
|
|
context.rejected = RejectedReason.AUTH_ERROR
|
|
|
|
return context
|
|
|
|
|
2021-09-28 16:25:07 +02:00
|
|
|
# calculate what the auth events *should* be, to use as a basis for auth.
|
|
|
|
prev_state_ids = await context.get_prev_state_ids()
|
|
|
|
auth_events_ids = self._event_auth_handler.compute_auth_events(
|
|
|
|
event, prev_state_ids, for_verification=True
|
|
|
|
)
|
|
|
|
auth_events_x = await self._store.get_events(auth_events_ids)
|
|
|
|
calculated_auth_event_map = {
|
|
|
|
(e.type, e.state_key): e for e in auth_events_x.values()
|
|
|
|
}
|
2021-08-26 22:41:44 +02:00
|
|
|
|
|
|
|
try:
|
|
|
|
(
|
|
|
|
context,
|
|
|
|
auth_events_for_auth,
|
|
|
|
) = await self._update_auth_events_and_context_for_auth(
|
2021-09-28 16:25:07 +02:00
|
|
|
origin,
|
|
|
|
event,
|
|
|
|
context,
|
|
|
|
calculated_auth_event_map=calculated_auth_event_map,
|
2021-08-26 22:41:44 +02:00
|
|
|
)
|
|
|
|
except Exception:
|
|
|
|
# We don't really mind if the above fails, so lets not fail
|
|
|
|
# processing if it does. However, it really shouldn't fail so
|
|
|
|
# let's still log as an exception since we'll still want to fix
|
|
|
|
# any bugs.
|
|
|
|
logger.exception(
|
|
|
|
"Failed to double check auth events for %s with remote. "
|
|
|
|
"Ignoring failure and continuing processing of event.",
|
|
|
|
event.event_id,
|
|
|
|
)
|
2021-09-28 16:25:07 +02:00
|
|
|
auth_events_for_auth = calculated_auth_event_map
|
2021-08-26 22:41:44 +02:00
|
|
|
|
|
|
|
try:
|
Split `event_auth.check` into two parts (#10940)
Broadly, the existing `event_auth.check` function has two parts:
* a validation section: checks that the event isn't too big, that it has the rught signatures, etc.
This bit is independent of the rest of the state in the room, and so need only be done once
for each event.
* an auth section: ensures that the event is allowed, given the rest of the state in the room.
This gets done multiple times, against various sets of room state, because it forms part of
the state res algorithm.
Currently, this is implemented with `do_sig_check` and `do_size_check` parameters, but I think
that makes everything hard to follow. Instead, we split the function in two and call each part
separately where it is needed.
2021-09-29 19:59:15 +02:00
|
|
|
check_auth_rules_for_event(room_version_obj, event, auth_events_for_auth)
|
2021-08-26 22:41:44 +02:00
|
|
|
except AuthError as e:
|
|
|
|
logger.warning("Failed auth resolution for %r because %s", event, e)
|
|
|
|
context.rejected = RejectedReason.AUTH_ERROR
|
|
|
|
|
|
|
|
if not context.rejected:
|
|
|
|
await self._check_for_soft_fail(event, state, backfilled, origin=origin)
|
2021-09-06 13:17:16 +02:00
|
|
|
await self._maybe_kick_guest_users(event)
|
2021-08-26 22:41:44 +02:00
|
|
|
|
|
|
|
# If we are going to send this event over federation we precaclculate
|
|
|
|
# the joined hosts.
|
|
|
|
if event.internal_metadata.get_send_on_behalf_of():
|
2021-09-07 12:15:51 +02:00
|
|
|
await self._event_creation_handler.cache_joined_hosts_for_event(
|
2021-08-26 22:41:44 +02:00
|
|
|
event, context
|
|
|
|
)
|
|
|
|
|
|
|
|
return context
|
|
|
|
|
2021-09-06 13:17:16 +02:00
|
|
|
async def _maybe_kick_guest_users(self, event: EventBase) -> None:
|
|
|
|
if event.type != EventTypes.GuestAccess:
|
|
|
|
return
|
|
|
|
|
|
|
|
guest_access = event.content.get(EventContentFields.GUEST_ACCESS)
|
|
|
|
if guest_access == GuestAccess.CAN_JOIN:
|
|
|
|
return
|
|
|
|
|
2021-09-07 12:15:51 +02:00
|
|
|
current_state_map = await self._state_handler.get_current_state(event.room_id)
|
2021-09-06 13:17:16 +02:00
|
|
|
current_state = list(current_state_map.values())
|
2021-09-06 15:49:33 +02:00
|
|
|
await self._get_room_member_handler().kick_guest_users(current_state)
|
2021-09-06 13:17:16 +02:00
|
|
|
|
2021-08-26 22:41:44 +02:00
|
|
|
async def _check_for_soft_fail(
|
|
|
|
self,
|
|
|
|
event: EventBase,
|
|
|
|
state: Optional[Iterable[EventBase]],
|
|
|
|
backfilled: bool,
|
|
|
|
origin: str,
|
|
|
|
) -> None:
|
|
|
|
"""Checks if we should soft fail the event; if so, marks the event as
|
|
|
|
such.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
event
|
|
|
|
state: The state at the event if we don't have all the event's prev events
|
|
|
|
backfilled: Whether the event is from backfill
|
|
|
|
origin: The host the event originates from.
|
|
|
|
"""
|
|
|
|
# For new (non-backfilled and non-outlier) events we check if the event
|
|
|
|
# passes auth based on the current state. If it doesn't then we
|
|
|
|
# "soft-fail" the event.
|
|
|
|
if backfilled or event.internal_metadata.is_outlier():
|
|
|
|
return
|
|
|
|
|
2021-09-07 12:15:51 +02:00
|
|
|
extrem_ids_list = await self._store.get_latest_event_ids_in_room(event.room_id)
|
2021-08-26 22:41:44 +02:00
|
|
|
extrem_ids = set(extrem_ids_list)
|
|
|
|
prev_event_ids = set(event.prev_event_ids())
|
|
|
|
|
|
|
|
if extrem_ids == prev_event_ids:
|
|
|
|
# If they're the same then the current state is the same as the
|
|
|
|
# state at the event, so no point rechecking auth for soft fail.
|
|
|
|
return
|
|
|
|
|
2021-09-07 12:15:51 +02:00
|
|
|
room_version = await self._store.get_room_version_id(event.room_id)
|
2021-08-26 22:41:44 +02:00
|
|
|
room_version_obj = KNOWN_ROOM_VERSIONS[room_version]
|
|
|
|
|
|
|
|
# Calculate the "current state".
|
|
|
|
if state is not None:
|
|
|
|
# If we're explicitly given the state then we won't have all the
|
|
|
|
# prev events, and so we have a gap in the graph. In this case
|
|
|
|
# we want to be a little careful as we might have been down for
|
|
|
|
# a while and have an incorrect view of the current state,
|
|
|
|
# however we still want to do checks as gaps are easy to
|
|
|
|
# maliciously manufacture.
|
|
|
|
#
|
|
|
|
# So we use a "current state" that is actually a state
|
|
|
|
# resolution across the current forward extremities and the
|
|
|
|
# given state at the event. This should correctly handle cases
|
|
|
|
# like bans, especially with state res v2.
|
|
|
|
|
2021-09-07 12:15:51 +02:00
|
|
|
state_sets_d = await self._state_store.get_state_groups(
|
2021-08-26 22:41:44 +02:00
|
|
|
event.room_id, extrem_ids
|
|
|
|
)
|
|
|
|
state_sets: List[Iterable[EventBase]] = list(state_sets_d.values())
|
|
|
|
state_sets.append(state)
|
2021-09-07 12:15:51 +02:00
|
|
|
current_states = await self._state_handler.resolve_events(
|
2021-08-26 22:41:44 +02:00
|
|
|
room_version, state_sets, event
|
|
|
|
)
|
|
|
|
current_state_ids: StateMap[str] = {
|
|
|
|
k: e.event_id for k, e in current_states.items()
|
|
|
|
}
|
|
|
|
else:
|
2021-09-07 12:15:51 +02:00
|
|
|
current_state_ids = await self._state_handler.get_current_state_ids(
|
2021-08-26 22:41:44 +02:00
|
|
|
event.room_id, latest_event_ids=extrem_ids
|
|
|
|
)
|
|
|
|
|
|
|
|
logger.debug(
|
|
|
|
"Doing soft-fail check for %s: state %s",
|
|
|
|
event.event_id,
|
|
|
|
current_state_ids,
|
|
|
|
)
|
|
|
|
|
|
|
|
# Now check if event pass auth against said current state
|
|
|
|
auth_types = auth_types_for_event(room_version_obj, event)
|
|
|
|
current_state_ids_list = [
|
|
|
|
e for k, e in current_state_ids.items() if k in auth_types
|
|
|
|
]
|
|
|
|
|
2021-09-07 12:15:51 +02:00
|
|
|
auth_events_map = await self._store.get_events(current_state_ids_list)
|
2021-08-26 22:41:44 +02:00
|
|
|
current_auth_events = {
|
|
|
|
(e.type, e.state_key): e for e in auth_events_map.values()
|
|
|
|
}
|
|
|
|
|
|
|
|
try:
|
Split `event_auth.check` into two parts (#10940)
Broadly, the existing `event_auth.check` function has two parts:
* a validation section: checks that the event isn't too big, that it has the rught signatures, etc.
This bit is independent of the rest of the state in the room, and so need only be done once
for each event.
* an auth section: ensures that the event is allowed, given the rest of the state in the room.
This gets done multiple times, against various sets of room state, because it forms part of
the state res algorithm.
Currently, this is implemented with `do_sig_check` and `do_size_check` parameters, but I think
that makes everything hard to follow. Instead, we split the function in two and call each part
separately where it is needed.
2021-09-29 19:59:15 +02:00
|
|
|
check_auth_rules_for_event(room_version_obj, event, current_auth_events)
|
2021-08-26 22:41:44 +02:00
|
|
|
except AuthError as e:
|
|
|
|
logger.warning(
|
|
|
|
"Soft-failing %r (from %s) because %s",
|
|
|
|
event,
|
|
|
|
e,
|
|
|
|
origin,
|
|
|
|
extra={
|
|
|
|
"room_id": event.room_id,
|
|
|
|
"mxid": event.sender,
|
|
|
|
"hs": origin,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
soft_failed_event_counter.inc()
|
|
|
|
event.internal_metadata.soft_failed = True
|
|
|
|
|
|
|
|
async def _update_auth_events_and_context_for_auth(
|
|
|
|
self,
|
|
|
|
origin: str,
|
|
|
|
event: EventBase,
|
|
|
|
context: EventContext,
|
2021-09-28 16:25:07 +02:00
|
|
|
calculated_auth_event_map: StateMap[EventBase],
|
2021-08-26 22:41:44 +02:00
|
|
|
) -> Tuple[EventContext, StateMap[EventBase]]:
|
|
|
|
"""Helper for _check_event_auth. See there for docs.
|
|
|
|
|
|
|
|
Checks whether a given event has the expected auth events. If it
|
|
|
|
doesn't then we talk to the remote server to compare state to see if
|
|
|
|
we can come to a consensus (e.g. if one server missed some valid
|
|
|
|
state).
|
|
|
|
|
|
|
|
This attempts to resolve any potential divergence of state between
|
|
|
|
servers, but is not essential and so failures should not block further
|
|
|
|
processing of the event.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
origin:
|
|
|
|
event:
|
|
|
|
context:
|
|
|
|
|
2021-09-28 16:25:07 +02:00
|
|
|
calculated_auth_event_map:
|
|
|
|
Our calculated auth_events based on the state of the room
|
|
|
|
at the event's position in the DAG.
|
2021-08-26 22:41:44 +02:00
|
|
|
|
|
|
|
Returns:
|
|
|
|
updated context, updated auth event map
|
|
|
|
"""
|
2021-09-28 16:25:07 +02:00
|
|
|
assert not event.internal_metadata.outlier
|
|
|
|
|
|
|
|
# take a copy of calculated_auth_event_map before we modify it.
|
|
|
|
auth_events: MutableStateMap[EventBase] = dict(calculated_auth_event_map)
|
2021-08-26 22:41:44 +02:00
|
|
|
|
|
|
|
event_auth_events = set(event.auth_event_ids())
|
|
|
|
|
|
|
|
# missing_auth is the set of the event's auth_events which we don't yet have
|
|
|
|
# in auth_events.
|
|
|
|
missing_auth = event_auth_events.difference(
|
|
|
|
e.event_id for e in auth_events.values()
|
|
|
|
)
|
|
|
|
|
|
|
|
# if we have missing events, we need to fetch those events from somewhere.
|
|
|
|
#
|
|
|
|
# we start by checking if they are in the store, and then try calling /event_auth/.
|
|
|
|
if missing_auth:
|
2021-09-07 12:15:51 +02:00
|
|
|
have_events = await self._store.have_seen_events(
|
|
|
|
event.room_id, missing_auth
|
|
|
|
)
|
2021-08-26 22:41:44 +02:00
|
|
|
logger.debug("Events %s are in the store", have_events)
|
|
|
|
missing_auth.difference_update(have_events)
|
|
|
|
|
2021-10-05 13:50:38 +02:00
|
|
|
# missing_auth is now the set of event_ids which:
|
|
|
|
# a. are listed in event.auth_events, *and*
|
|
|
|
# b. are *not* part of our calculated auth events based on room state, *and*
|
|
|
|
# c. are *not* yet in our database.
|
|
|
|
|
2021-08-26 22:41:44 +02:00
|
|
|
if missing_auth:
|
|
|
|
# If we don't have all the auth events, we need to get them.
|
|
|
|
logger.info("auth_events contains unknown events: %s", missing_auth)
|
|
|
|
try:
|
2021-09-23 18:34:33 +02:00
|
|
|
await self._get_remote_auth_chain_for_event(
|
|
|
|
origin, event.room_id, event.event_id
|
2021-08-26 22:41:44 +02:00
|
|
|
)
|
|
|
|
except Exception:
|
|
|
|
logger.exception("Failed to get auth chain")
|
2021-09-23 18:34:33 +02:00
|
|
|
else:
|
|
|
|
# load any auth events we might have persisted from the database. This
|
|
|
|
# has the side-effect of correctly setting the rejected_reason on them.
|
|
|
|
auth_events.update(
|
|
|
|
{
|
|
|
|
(ae.type, ae.state_key): ae
|
|
|
|
for ae in await self._store.get_events_as_list(
|
|
|
|
missing_auth, allow_rejected=True
|
|
|
|
)
|
|
|
|
}
|
|
|
|
)
|
2021-08-26 22:41:44 +02:00
|
|
|
|
2021-10-05 13:50:38 +02:00
|
|
|
# auth_events now contains
|
|
|
|
# 1. our *calculated* auth events based on the room state, plus:
|
|
|
|
# 2. any events which:
|
|
|
|
# a. are listed in `event.auth_events`, *and*
|
|
|
|
# b. are not part of our calculated auth events, *and*
|
|
|
|
# c. were not in our database before the call to /event_auth
|
|
|
|
# d. have since been added to our database (most likely by /event_auth).
|
|
|
|
|
2021-08-26 22:41:44 +02:00
|
|
|
different_auth = event_auth_events.difference(
|
|
|
|
e.event_id for e in auth_events.values()
|
|
|
|
)
|
|
|
|
|
2021-10-05 13:50:38 +02:00
|
|
|
# different_auth is the set of events which *are* in `event.auth_events`, but
|
|
|
|
# which are *not* in `auth_events`. Comparing with (2.) above, this means
|
|
|
|
# exclusively the set of `event.auth_events` which we already had in our
|
|
|
|
# database before any call to /event_auth.
|
|
|
|
#
|
|
|
|
# I'm reasonably sure that the fact that events returned by /event_auth are
|
|
|
|
# blindly added to auth_events (and hence excluded from different_auth) is a bug
|
|
|
|
# - though it's a very long-standing one (see
|
|
|
|
# https://github.com/matrix-org/synapse/commit/78015948a7febb18e000651f72f8f58830a55b93#diff-0bc92da3d703202f5b9be2d3f845e375f5b1a6bc6ba61705a8af9be1121f5e42R786
|
|
|
|
# from Jan 2015 which seems to add it, though it actually just moves it from
|
|
|
|
# elsewhere (before that, it gets lost in a mess of huge "various bug fixes"
|
|
|
|
# PRs).
|
|
|
|
|
2021-08-26 22:41:44 +02:00
|
|
|
if not different_auth:
|
|
|
|
return context, auth_events
|
|
|
|
|
|
|
|
logger.info(
|
|
|
|
"auth_events refers to events which are not in our calculated auth "
|
|
|
|
"chain: %s",
|
|
|
|
different_auth,
|
|
|
|
)
|
|
|
|
|
|
|
|
# XXX: currently this checks for redactions but I'm not convinced that is
|
|
|
|
# necessary?
|
2021-09-07 12:15:51 +02:00
|
|
|
different_events = await self._store.get_events_as_list(different_auth)
|
2021-08-26 22:41:44 +02:00
|
|
|
|
|
|
|
for d in different_events:
|
|
|
|
if d.room_id != event.room_id:
|
|
|
|
logger.warning(
|
|
|
|
"Event %s refers to auth_event %s which is in a different room",
|
|
|
|
event.event_id,
|
|
|
|
d.event_id,
|
|
|
|
)
|
|
|
|
|
|
|
|
# don't attempt to resolve the claimed auth events against our own
|
|
|
|
# in this case: just use our own auth events.
|
|
|
|
#
|
|
|
|
# XXX: should we reject the event in this case? It feels like we should,
|
|
|
|
# but then shouldn't we also do so if we've failed to fetch any of the
|
|
|
|
# auth events?
|
|
|
|
return context, auth_events
|
|
|
|
|
|
|
|
# now we state-resolve between our own idea of the auth events, and the remote's
|
|
|
|
# idea of them.
|
|
|
|
|
|
|
|
local_state = auth_events.values()
|
|
|
|
remote_auth_events = dict(auth_events)
|
|
|
|
remote_auth_events.update({(d.type, d.state_key): d for d in different_events})
|
|
|
|
remote_state = remote_auth_events.values()
|
|
|
|
|
2021-09-07 12:15:51 +02:00
|
|
|
room_version = await self._store.get_room_version_id(event.room_id)
|
|
|
|
new_state = await self._state_handler.resolve_events(
|
2021-08-26 22:41:44 +02:00
|
|
|
room_version, (local_state, remote_state), event
|
|
|
|
)
|
|
|
|
|
|
|
|
logger.info(
|
|
|
|
"After state res: updating auth_events with new state %s",
|
|
|
|
{
|
|
|
|
(d.type, d.state_key): d.event_id
|
|
|
|
for d in new_state.values()
|
|
|
|
if auth_events.get((d.type, d.state_key)) != d
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
auth_events.update(new_state)
|
|
|
|
|
|
|
|
context = await self._update_context_for_auth_events(
|
|
|
|
event, context, auth_events
|
|
|
|
)
|
|
|
|
|
|
|
|
return context, auth_events
|
|
|
|
|
2021-09-23 18:34:33 +02:00
|
|
|
async def _get_remote_auth_chain_for_event(
|
|
|
|
self, destination: str, room_id: str, event_id: str
|
|
|
|
) -> None:
|
|
|
|
"""If we are missing some of an event's auth events, attempt to request them
|
|
|
|
|
|
|
|
Args:
|
|
|
|
destination: where to fetch the auth tree from
|
|
|
|
room_id: the room in which we are lacking auth events
|
|
|
|
event_id: the event for which we are lacking auth events
|
|
|
|
"""
|
|
|
|
try:
|
2021-09-24 12:56:33 +02:00
|
|
|
remote_event_map = {
|
|
|
|
e.event_id: e
|
|
|
|
for e in await self._federation_client.get_event_auth(
|
|
|
|
destination, room_id, event_id
|
|
|
|
)
|
|
|
|
}
|
2021-09-23 18:34:33 +02:00
|
|
|
except RequestSendFailed as e1:
|
|
|
|
# The other side isn't around or doesn't implement the
|
|
|
|
# endpoint, so lets just bail out.
|
|
|
|
logger.info("Failed to get event auth from remote: %s", e1)
|
|
|
|
return
|
|
|
|
|
2021-09-24 12:56:33 +02:00
|
|
|
logger.info("/event_auth returned %i events", len(remote_event_map))
|
2021-09-23 18:34:33 +02:00
|
|
|
|
2021-09-24 12:56:33 +02:00
|
|
|
# `event` may be returned, but we should not yet process it.
|
|
|
|
remote_event_map.pop(event_id, None)
|
2021-09-23 18:34:33 +02:00
|
|
|
|
2021-09-24 12:56:33 +02:00
|
|
|
# nor should we reprocess any events we have already seen.
|
|
|
|
seen_remotes = await self._store.have_seen_events(
|
|
|
|
room_id, remote_event_map.keys()
|
|
|
|
)
|
|
|
|
for s in seen_remotes:
|
|
|
|
remote_event_map.pop(s, None)
|
2021-09-23 18:34:33 +02:00
|
|
|
|
2021-09-24 12:56:33 +02:00
|
|
|
await self._auth_and_persist_fetched_events(
|
|
|
|
destination, room_id, remote_event_map.values()
|
|
|
|
)
|
2021-09-23 18:34:33 +02:00
|
|
|
|
2021-08-26 22:41:44 +02:00
|
|
|
async def _update_context_for_auth_events(
|
|
|
|
self, event: EventBase, context: EventContext, auth_events: StateMap[EventBase]
|
|
|
|
) -> EventContext:
|
|
|
|
"""Update the state_ids in an event context after auth event resolution,
|
|
|
|
storing the changes as a new state group.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
event: The event we're handling the context for
|
|
|
|
|
|
|
|
context: initial event context
|
|
|
|
|
|
|
|
auth_events: Events to update in the event context.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
new event context
|
|
|
|
"""
|
|
|
|
# exclude the state key of the new event from the current_state in the context.
|
|
|
|
if event.is_state():
|
|
|
|
event_key: Optional[Tuple[str, str]] = (event.type, event.state_key)
|
|
|
|
else:
|
|
|
|
event_key = None
|
|
|
|
state_updates = {
|
|
|
|
k: a.event_id for k, a in auth_events.items() if k != event_key
|
|
|
|
}
|
|
|
|
|
|
|
|
current_state_ids = await context.get_current_state_ids()
|
|
|
|
current_state_ids = dict(current_state_ids) # type: ignore
|
|
|
|
|
|
|
|
current_state_ids.update(state_updates)
|
|
|
|
|
|
|
|
prev_state_ids = await context.get_prev_state_ids()
|
|
|
|
prev_state_ids = dict(prev_state_ids)
|
|
|
|
|
|
|
|
prev_state_ids.update({k: a.event_id for k, a in auth_events.items()})
|
|
|
|
|
|
|
|
# create a new state group as a delta from the existing one.
|
|
|
|
prev_group = context.state_group
|
2021-09-07 12:15:51 +02:00
|
|
|
state_group = await self._state_store.store_state_group(
|
2021-08-26 22:41:44 +02:00
|
|
|
event.event_id,
|
|
|
|
event.room_id,
|
|
|
|
prev_group=prev_group,
|
|
|
|
delta_ids=state_updates,
|
|
|
|
current_state_ids=current_state_ids,
|
|
|
|
)
|
|
|
|
|
|
|
|
return EventContext.with_state(
|
|
|
|
state_group=state_group,
|
|
|
|
state_group_before_event=context.state_group_before_event,
|
|
|
|
current_state_ids=current_state_ids,
|
|
|
|
prev_state_ids=prev_state_ids,
|
|
|
|
prev_group=prev_group,
|
|
|
|
delta_ids=state_updates,
|
|
|
|
)
|
|
|
|
|
|
|
|
async def _run_push_actions_and_persist_event(
|
|
|
|
self, event: EventBase, context: EventContext, backfilled: bool = False
|
2021-09-20 14:56:23 +02:00
|
|
|
) -> None:
|
2021-08-26 22:41:44 +02:00
|
|
|
"""Run the push actions for a received event, and persist it.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
event: The event itself.
|
|
|
|
context: The event context.
|
|
|
|
backfilled: True if the event was backfilled.
|
|
|
|
"""
|
2021-09-08 20:03:08 +02:00
|
|
|
# this method should not be called on outliers (those code paths call
|
|
|
|
# persist_events_and_notify directly.)
|
|
|
|
assert not event.internal_metadata.outlier
|
|
|
|
|
2021-08-26 22:41:44 +02:00
|
|
|
try:
|
|
|
|
if (
|
2021-09-08 20:03:08 +02:00
|
|
|
not backfilled
|
2021-08-26 22:41:44 +02:00
|
|
|
and not context.rejected
|
2021-09-07 12:15:51 +02:00
|
|
|
and (await self._store.get_min_depth(event.room_id)) <= event.depth
|
2021-08-26 22:41:44 +02:00
|
|
|
):
|
2021-09-07 12:15:51 +02:00
|
|
|
await self._action_generator.handle_push_actions_for_event(
|
2021-08-26 22:41:44 +02:00
|
|
|
event, context
|
|
|
|
)
|
|
|
|
|
|
|
|
await self.persist_events_and_notify(
|
|
|
|
event.room_id, [(event, context)], backfilled=backfilled
|
|
|
|
)
|
|
|
|
except Exception:
|
|
|
|
run_in_background(
|
2021-09-07 12:15:51 +02:00
|
|
|
self._store.remove_push_actions_from_staging, event.event_id
|
2021-08-26 22:41:44 +02:00
|
|
|
)
|
|
|
|
raise
|
|
|
|
|
|
|
|
async def persist_events_and_notify(
|
|
|
|
self,
|
|
|
|
room_id: str,
|
|
|
|
event_and_contexts: Sequence[Tuple[EventBase, EventContext]],
|
|
|
|
backfilled: bool = False,
|
|
|
|
) -> int:
|
|
|
|
"""Persists events and tells the notifier/pushers about them, if
|
|
|
|
necessary.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
room_id: The room ID of events being persisted.
|
|
|
|
event_and_contexts: Sequence of events with their associated
|
|
|
|
context that should be persisted. All events must belong to
|
|
|
|
the same room.
|
|
|
|
backfilled: Whether these events are a result of
|
|
|
|
backfilling or not
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
The stream ID after which all events have been persisted.
|
|
|
|
"""
|
|
|
|
if not event_and_contexts:
|
2021-09-07 12:15:51 +02:00
|
|
|
return self._store.get_current_events_token()
|
2021-08-26 22:41:44 +02:00
|
|
|
|
2021-09-07 12:15:51 +02:00
|
|
|
instance = self._config.worker.events_shard_config.get_instance(room_id)
|
2021-08-26 22:41:44 +02:00
|
|
|
if instance != self._instance_name:
|
|
|
|
# Limit the number of events sent over replication. We choose 200
|
|
|
|
# here as that is what we default to in `max_request_body_size(..)`
|
|
|
|
for batch in batch_iter(event_and_contexts, 200):
|
|
|
|
result = await self._send_events(
|
|
|
|
instance_name=instance,
|
2021-09-07 12:15:51 +02:00
|
|
|
store=self._store,
|
2021-08-26 22:41:44 +02:00
|
|
|
room_id=room_id,
|
|
|
|
event_and_contexts=batch,
|
|
|
|
backfilled=backfilled,
|
|
|
|
)
|
|
|
|
return result["max_stream_id"]
|
|
|
|
else:
|
2021-09-07 12:15:51 +02:00
|
|
|
assert self._storage.persistence
|
2021-08-26 22:41:44 +02:00
|
|
|
|
|
|
|
# Note that this returns the events that were persisted, which may not be
|
|
|
|
# the same as were passed in if some were deduplicated due to transaction IDs.
|
2021-09-07 12:15:51 +02:00
|
|
|
events, max_stream_token = await self._storage.persistence.persist_events(
|
2021-08-26 22:41:44 +02:00
|
|
|
event_and_contexts, backfilled=backfilled
|
|
|
|
)
|
|
|
|
|
|
|
|
if self._ephemeral_messages_enabled:
|
|
|
|
for event in events:
|
|
|
|
# If there's an expiry timestamp on the event, schedule its expiry.
|
|
|
|
self._message_handler.maybe_schedule_expiry(event)
|
|
|
|
|
|
|
|
if not backfilled: # Never notify for backfilled events
|
|
|
|
for event in events:
|
|
|
|
await self._notify_persisted_event(event, max_stream_token)
|
|
|
|
|
|
|
|
return max_stream_token.stream
|
|
|
|
|
|
|
|
async def _notify_persisted_event(
|
|
|
|
self, event: EventBase, max_stream_token: RoomStreamToken
|
|
|
|
) -> None:
|
|
|
|
"""Checks to see if notifier/pushers should be notified about the
|
|
|
|
event or not.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
event:
|
|
|
|
max_stream_token: The max_stream_id returned by persist_events
|
|
|
|
"""
|
|
|
|
|
|
|
|
extra_users = []
|
|
|
|
if event.type == EventTypes.Member:
|
|
|
|
target_user_id = event.state_key
|
|
|
|
|
|
|
|
# We notify for memberships if its an invite for one of our
|
|
|
|
# users
|
|
|
|
if event.internal_metadata.is_outlier():
|
|
|
|
if event.membership != Membership.INVITE:
|
2021-09-07 12:15:51 +02:00
|
|
|
if not self._is_mine_id(target_user_id):
|
2021-08-26 22:41:44 +02:00
|
|
|
return
|
|
|
|
|
|
|
|
target_user = UserID.from_string(target_user_id)
|
|
|
|
extra_users.append(target_user)
|
|
|
|
elif event.internal_metadata.is_outlier():
|
|
|
|
return
|
|
|
|
|
|
|
|
# the event has been persisted so it should have a stream ordering.
|
|
|
|
assert event.internal_metadata.stream_ordering
|
|
|
|
|
|
|
|
event_pos = PersistedEventPosition(
|
|
|
|
self._instance_name, event.internal_metadata.stream_ordering
|
|
|
|
)
|
2021-09-06 15:49:33 +02:00
|
|
|
self._notifier.on_new_room_event(
|
2021-08-26 22:41:44 +02:00
|
|
|
event, event_pos, max_stream_token, extra_users=extra_users
|
|
|
|
)
|
|
|
|
|
|
|
|
def _sanity_check_event(self, ev: EventBase) -> None:
|
|
|
|
"""
|
|
|
|
Do some early sanity checks of a received event
|
|
|
|
|
|
|
|
In particular, checks it doesn't have an excessive number of
|
|
|
|
prev_events or auth_events, which could cause a huge state resolution
|
|
|
|
or cascade of event fetches.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
ev: event to be checked
|
|
|
|
|
|
|
|
Raises:
|
|
|
|
SynapseError if the event does not pass muster
|
|
|
|
"""
|
|
|
|
if len(ev.prev_event_ids()) > 20:
|
|
|
|
logger.warning(
|
|
|
|
"Rejecting event %s which has %i prev_events",
|
|
|
|
ev.event_id,
|
|
|
|
len(ev.prev_event_ids()),
|
|
|
|
)
|
|
|
|
raise SynapseError(HTTPStatus.BAD_REQUEST, "Too many prev_events")
|
|
|
|
|
|
|
|
if len(ev.auth_event_ids()) > 10:
|
|
|
|
logger.warning(
|
|
|
|
"Rejecting event %s which has %i auth_events",
|
|
|
|
ev.event_id,
|
|
|
|
len(ev.auth_event_ids()),
|
|
|
|
)
|
|
|
|
raise SynapseError(HTTPStatus.BAD_REQUEST, "Too many auth_events")
|
|
|
|
|
|
|
|
async def get_min_depth_for_context(self, context: str) -> int:
|
2021-09-07 12:15:51 +02:00
|
|
|
return await self._store.get_min_depth(context)
|