1109 lines
		
	
	
		
			43 KiB
		
	
	
	
		
			Python
		
	
	
			
		
		
	
	
			1109 lines
		
	
	
		
			43 KiB
		
	
	
	
		
			Python
		
	
	
| # -*- coding: utf-8 -*-
 | |
| # Copyright 2014-2016 OpenMarket Ltd
 | |
| # Copyright 2017-2018 New Vector Ltd
 | |
| # Copyright 2019 The Matrix.org Foundation C.I.C.
 | |
| #
 | |
| # Licensed under the Apache License, Version 2.0 (the "License");
 | |
| # you may not use this file except in compliance with the License.
 | |
| # You may obtain a copy of the License at
 | |
| #
 | |
| #     http://www.apache.org/licenses/LICENSE-2.0
 | |
| #
 | |
| # Unless required by applicable law or agreed to in writing, software
 | |
| # distributed under the License is distributed on an "AS IS" BASIS,
 | |
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
| # See the License for the specific language governing permissions and
 | |
| # limitations under the License.
 | |
| import logging
 | |
| from typing import Optional
 | |
| 
 | |
| from six import iteritems, itervalues, string_types
 | |
| 
 | |
| from canonicaljson import encode_canonical_json, json
 | |
| 
 | |
| from twisted.internet import defer
 | |
| from twisted.internet.defer import succeed
 | |
| from twisted.internet.interfaces import IDelayedCall
 | |
| 
 | |
| from synapse import event_auth
 | |
| from synapse.api.constants import (
 | |
|     EventContentFields,
 | |
|     EventTypes,
 | |
|     Membership,
 | |
|     RelationTypes,
 | |
|     UserTypes,
 | |
| )
 | |
| from synapse.api.errors import (
 | |
|     AuthError,
 | |
|     Codes,
 | |
|     ConsentNotGivenError,
 | |
|     NotFoundError,
 | |
|     SynapseError,
 | |
| )
 | |
| from synapse.api.room_versions import RoomVersions
 | |
| from synapse.api.urls import ConsentURIBuilder
 | |
| from synapse.events.validator import EventValidator
 | |
| from synapse.logging.context import run_in_background
 | |
| from synapse.metrics.background_process_metrics import run_as_background_process
 | |
| from synapse.replication.http.send_event import ReplicationSendEventRestServlet
 | |
| from synapse.storage.state import StateFilter
 | |
| from synapse.types import RoomAlias, UserID, create_requester
 | |
| from synapse.util.async_helpers import Linearizer
 | |
| from synapse.util.frozenutils import frozendict_json_encoder
 | |
| from synapse.util.metrics import measure_func
 | |
| from synapse.visibility import filter_events_for_client
 | |
| 
 | |
| from ._base import BaseHandler
 | |
| 
 | |
| logger = logging.getLogger(__name__)
 | |
| 
 | |
| 
 | |
| class MessageHandler(object):
 | |
|     """Contains some read only APIs to get state about a room
 | |
|     """
 | |
| 
 | |
|     def __init__(self, hs):
 | |
|         self.auth = hs.get_auth()
 | |
|         self.clock = hs.get_clock()
 | |
|         self.state = hs.get_state_handler()
 | |
|         self.store = hs.get_datastore()
 | |
|         self.storage = hs.get_storage()
 | |
|         self.state_store = self.storage.state
 | |
|         self._event_serializer = hs.get_event_client_serializer()
 | |
|         self._ephemeral_events_enabled = hs.config.enable_ephemeral_messages
 | |
|         self._is_worker_app = bool(hs.config.worker_app)
 | |
| 
 | |
|         # The scheduled call to self._expire_event. None if no call is currently
 | |
|         # scheduled.
 | |
|         self._scheduled_expiry = None  # type: Optional[IDelayedCall]
 | |
| 
 | |
|         if not hs.config.worker_app:
 | |
|             run_as_background_process(
 | |
|                 "_schedule_next_expiry", self._schedule_next_expiry
 | |
|             )
 | |
| 
 | |
|     @defer.inlineCallbacks
 | |
|     def get_room_data(
 | |
|         self, user_id=None, room_id=None, event_type=None, state_key="", is_guest=False
 | |
|     ):
 | |
|         """ Get data from a room.
 | |
| 
 | |
|         Args:
 | |
|             event : The room path event
 | |
|         Returns:
 | |
|             The path data content.
 | |
|         Raises:
 | |
|             SynapseError if something went wrong.
 | |
|         """
 | |
|         (
 | |
|             membership,
 | |
|             membership_event_id,
 | |
|         ) = yield self.auth.check_in_room_or_world_readable(room_id, user_id)
 | |
| 
 | |
|         if membership == Membership.JOIN:
 | |
|             data = yield self.state.get_current_state(room_id, event_type, state_key)
 | |
|         elif membership == Membership.LEAVE:
 | |
|             key = (event_type, state_key)
 | |
|             room_state = yield self.state_store.get_state_for_events(
 | |
|                 [membership_event_id], StateFilter.from_types([key])
 | |
|             )
 | |
|             data = room_state[membership_event_id].get(key)
 | |
| 
 | |
|         return data
 | |
| 
 | |
|     @defer.inlineCallbacks
 | |
|     def get_state_events(
 | |
|         self,
 | |
|         user_id,
 | |
|         room_id,
 | |
|         state_filter=StateFilter.all(),
 | |
|         at_token=None,
 | |
|         is_guest=False,
 | |
|     ):
 | |
|         """Retrieve all state events for a given room. If the user is
 | |
|         joined to the room then return the current state. If the user has
 | |
|         left the room return the state events from when they left. If an explicit
 | |
|         'at' parameter is passed, return the state events as of that event, if
 | |
|         visible.
 | |
| 
 | |
|         Args:
 | |
|             user_id(str): The user requesting state events.
 | |
|             room_id(str): The room ID to get all state events from.
 | |
|             state_filter (StateFilter): The state filter used to fetch state
 | |
|                 from the database.
 | |
|             at_token(StreamToken|None): the stream token of the at which we are requesting
 | |
|                 the stats. If the user is not allowed to view the state as of that
 | |
|                 stream token, we raise a 403 SynapseError. If None, returns the current
 | |
|                 state based on the current_state_events table.
 | |
|             is_guest(bool): whether this user is a guest
 | |
|         Returns:
 | |
|             A list of dicts representing state events. [{}, {}, {}]
 | |
|         Raises:
 | |
|             NotFoundError (404) if the at token does not yield an event
 | |
| 
 | |
|             AuthError (403) if the user doesn't have permission to view
 | |
|             members of this room.
 | |
|         """
 | |
|         if at_token:
 | |
|             # FIXME this claims to get the state at a stream position, but
 | |
|             # get_recent_events_for_room operates by topo ordering. This therefore
 | |
|             # does not reliably give you the state at the given stream position.
 | |
|             # (https://github.com/matrix-org/synapse/issues/3305)
 | |
|             last_events, _ = yield self.store.get_recent_events_for_room(
 | |
|                 room_id, end_token=at_token.room_key, limit=1
 | |
|             )
 | |
| 
 | |
|             if not last_events:
 | |
|                 raise NotFoundError("Can't find event for token %s" % (at_token,))
 | |
| 
 | |
|             visible_events = yield filter_events_for_client(
 | |
|                 self.storage, user_id, last_events, apply_retention_policies=False
 | |
|             )
 | |
| 
 | |
|             event = last_events[0]
 | |
|             if visible_events:
 | |
|                 room_state = yield self.state_store.get_state_for_events(
 | |
|                     [event.event_id], state_filter=state_filter
 | |
|                 )
 | |
|                 room_state = room_state[event.event_id]
 | |
|             else:
 | |
|                 raise AuthError(
 | |
|                     403,
 | |
|                     "User %s not allowed to view events in room %s at token %s"
 | |
|                     % (user_id, room_id, at_token),
 | |
|                 )
 | |
|         else:
 | |
|             (
 | |
|                 membership,
 | |
|                 membership_event_id,
 | |
|             ) = yield self.auth.check_in_room_or_world_readable(room_id, user_id)
 | |
| 
 | |
|             if membership == Membership.JOIN:
 | |
|                 state_ids = yield self.store.get_filtered_current_state_ids(
 | |
|                     room_id, state_filter=state_filter
 | |
|                 )
 | |
|                 room_state = yield self.store.get_events(state_ids.values())
 | |
|             elif membership == Membership.LEAVE:
 | |
|                 room_state = yield self.state_store.get_state_for_events(
 | |
|                     [membership_event_id], state_filter=state_filter
 | |
|                 )
 | |
|                 room_state = room_state[membership_event_id]
 | |
| 
 | |
|         now = self.clock.time_msec()
 | |
|         events = yield self._event_serializer.serialize_events(
 | |
|             room_state.values(),
 | |
|             now,
 | |
|             # We don't bother bundling aggregations in when asked for state
 | |
|             # events, as clients won't use them.
 | |
|             bundle_aggregations=False,
 | |
|         )
 | |
|         return events
 | |
| 
 | |
|     @defer.inlineCallbacks
 | |
|     def get_joined_members(self, requester, room_id):
 | |
|         """Get all the joined members in the room and their profile information.
 | |
| 
 | |
|         If the user has left the room return the state events from when they left.
 | |
| 
 | |
|         Args:
 | |
|             requester(Requester): The user requesting state events.
 | |
|             room_id(str): The room ID to get all state events from.
 | |
|         Returns:
 | |
|             A dict of user_id to profile info
 | |
|         """
 | |
|         user_id = requester.user.to_string()
 | |
|         if not requester.app_service:
 | |
|             # We check AS auth after fetching the room membership, as it
 | |
|             # requires us to pull out all joined members anyway.
 | |
|             membership, _ = yield self.auth.check_in_room_or_world_readable(
 | |
|                 room_id, user_id
 | |
|             )
 | |
|             if membership != Membership.JOIN:
 | |
|                 raise NotImplementedError(
 | |
|                     "Getting joined members after leaving is not implemented"
 | |
|                 )
 | |
| 
 | |
|         users_with_profile = yield self.state.get_current_users_in_room(room_id)
 | |
| 
 | |
|         # If this is an AS, double check that they are allowed to see the members.
 | |
|         # This can either be because the AS user is in the room or because there
 | |
|         # is a user in the room that the AS is "interested in"
 | |
|         if requester.app_service and user_id not in users_with_profile:
 | |
|             for uid in users_with_profile:
 | |
|                 if requester.app_service.is_interested_in_user(uid):
 | |
|                     break
 | |
|             else:
 | |
|                 # Loop fell through, AS has no interested users in room
 | |
|                 raise AuthError(403, "Appservice not in room")
 | |
| 
 | |
|         return {
 | |
|             user_id: {
 | |
|                 "avatar_url": profile.avatar_url,
 | |
|                 "display_name": profile.display_name,
 | |
|             }
 | |
|             for user_id, profile in iteritems(users_with_profile)
 | |
|         }
 | |
| 
 | |
|     def maybe_schedule_expiry(self, event):
 | |
|         """Schedule the expiry of an event if there's not already one scheduled,
 | |
|         or if the one running is for an event that will expire after the provided
 | |
|         timestamp.
 | |
| 
 | |
|         This function needs to invalidate the event cache, which is only possible on
 | |
|         the master process, and therefore needs to be run on there.
 | |
| 
 | |
|         Args:
 | |
|             event (EventBase): The event to schedule the expiry of.
 | |
|         """
 | |
|         assert not self._is_worker_app
 | |
| 
 | |
|         expiry_ts = event.content.get(EventContentFields.SELF_DESTRUCT_AFTER)
 | |
|         if not isinstance(expiry_ts, int) or event.is_state():
 | |
|             return
 | |
| 
 | |
|         # _schedule_expiry_for_event won't actually schedule anything if there's already
 | |
|         # a task scheduled for a timestamp that's sooner than the provided one.
 | |
|         self._schedule_expiry_for_event(event.event_id, expiry_ts)
 | |
| 
 | |
|     @defer.inlineCallbacks
 | |
|     def _schedule_next_expiry(self):
 | |
|         """Retrieve the ID and the expiry timestamp of the next event to be expired,
 | |
|         and schedule an expiry task for it.
 | |
| 
 | |
|         If there's no event left to expire, set _expiry_scheduled to None so that a
 | |
|         future call to save_expiry_ts can schedule a new expiry task.
 | |
|         """
 | |
|         # Try to get the expiry timestamp of the next event to expire.
 | |
|         res = yield self.store.get_next_event_to_expire()
 | |
|         if res:
 | |
|             event_id, expiry_ts = res
 | |
|             self._schedule_expiry_for_event(event_id, expiry_ts)
 | |
| 
 | |
|     def _schedule_expiry_for_event(self, event_id, expiry_ts):
 | |
|         """Schedule an expiry task for the provided event if there's not already one
 | |
|         scheduled at a timestamp that's sooner than the provided one.
 | |
| 
 | |
|         Args:
 | |
|             event_id (str): The ID of the event to expire.
 | |
|             expiry_ts (int): The timestamp at which to expire the event.
 | |
|         """
 | |
|         if self._scheduled_expiry:
 | |
|             # If the provided timestamp refers to a time before the scheduled time of the
 | |
|             # next expiry task, cancel that task and reschedule it for this timestamp.
 | |
|             next_scheduled_expiry_ts = self._scheduled_expiry.getTime() * 1000
 | |
|             if expiry_ts < next_scheduled_expiry_ts:
 | |
|                 self._scheduled_expiry.cancel()
 | |
|             else:
 | |
|                 return
 | |
| 
 | |
|         # Figure out how many seconds we need to wait before expiring the event.
 | |
|         now_ms = self.clock.time_msec()
 | |
|         delay = (expiry_ts - now_ms) / 1000
 | |
| 
 | |
|         # callLater doesn't support negative delays, so trim the delay to 0 if we're
 | |
|         # in that case.
 | |
|         if delay < 0:
 | |
|             delay = 0
 | |
| 
 | |
|         logger.info("Scheduling expiry for event %s in %.3fs", event_id, delay)
 | |
| 
 | |
|         self._scheduled_expiry = self.clock.call_later(
 | |
|             delay,
 | |
|             run_as_background_process,
 | |
|             "_expire_event",
 | |
|             self._expire_event,
 | |
|             event_id,
 | |
|         )
 | |
| 
 | |
|     @defer.inlineCallbacks
 | |
|     def _expire_event(self, event_id):
 | |
|         """Retrieve and expire an event that needs to be expired from the database.
 | |
| 
 | |
|         If the event doesn't exist in the database, log it and delete the expiry date
 | |
|         from the database (so that we don't try to expire it again).
 | |
|         """
 | |
|         assert self._ephemeral_events_enabled
 | |
| 
 | |
|         self._scheduled_expiry = None
 | |
| 
 | |
|         logger.info("Expiring event %s", event_id)
 | |
| 
 | |
|         try:
 | |
|             # Expire the event if we know about it. This function also deletes the expiry
 | |
|             # date from the database in the same database transaction.
 | |
|             yield self.store.expire_event(event_id)
 | |
|         except Exception as e:
 | |
|             logger.error("Could not expire event %s: %r", event_id, e)
 | |
| 
 | |
|         # Schedule the expiry of the next event to expire.
 | |
|         yield self._schedule_next_expiry()
 | |
| 
 | |
| 
 | |
| # The duration (in ms) after which rooms should be removed
 | |
| # `_rooms_to_exclude_from_dummy_event_insertion` (with the effect that we will try
 | |
| # to generate a dummy event for them once more)
 | |
| #
 | |
| _DUMMY_EVENT_ROOM_EXCLUSION_EXPIRY = 7 * 24 * 60 * 60 * 1000
 | |
| 
 | |
| 
 | |
| class EventCreationHandler(object):
 | |
|     def __init__(self, hs):
 | |
|         self.hs = hs
 | |
|         self.auth = hs.get_auth()
 | |
|         self.store = hs.get_datastore()
 | |
|         self.storage = hs.get_storage()
 | |
|         self.state = hs.get_state_handler()
 | |
|         self.clock = hs.get_clock()
 | |
|         self.validator = EventValidator()
 | |
|         self.profile_handler = hs.get_profile_handler()
 | |
|         self.event_builder_factory = hs.get_event_builder_factory()
 | |
|         self.server_name = hs.hostname
 | |
|         self.ratelimiter = hs.get_ratelimiter()
 | |
|         self.notifier = hs.get_notifier()
 | |
|         self.config = hs.config
 | |
|         self.require_membership_for_aliases = hs.config.require_membership_for_aliases
 | |
| 
 | |
|         self.room_invite_state_types = self.hs.config.room_invite_state_types
 | |
| 
 | |
|         self.send_event_to_master = ReplicationSendEventRestServlet.make_client(hs)
 | |
| 
 | |
|         # This is only used to get at ratelimit function, and maybe_kick_guest_users
 | |
|         self.base_handler = BaseHandler(hs)
 | |
| 
 | |
|         self.pusher_pool = hs.get_pusherpool()
 | |
| 
 | |
|         # We arbitrarily limit concurrent event creation for a room to 5.
 | |
|         # This is to stop us from diverging history *too* much.
 | |
|         self.limiter = Linearizer(max_count=5, name="room_event_creation_limit")
 | |
| 
 | |
|         self.action_generator = hs.get_action_generator()
 | |
| 
 | |
|         self.spam_checker = hs.get_spam_checker()
 | |
|         self.third_party_event_rules = hs.get_third_party_event_rules()
 | |
| 
 | |
|         self._block_events_without_consent_error = (
 | |
|             self.config.block_events_without_consent_error
 | |
|         )
 | |
| 
 | |
|         # Rooms which should be excluded from dummy insertion. (For instance,
 | |
|         # those without local users who can send events into the room).
 | |
|         #
 | |
|         # map from room id to time-of-last-attempt.
 | |
|         #
 | |
|         self._rooms_to_exclude_from_dummy_event_insertion = {}  # type: dict[str, int]
 | |
| 
 | |
|         # we need to construct a ConsentURIBuilder here, as it checks that the necessary
 | |
|         # config options, but *only* if we have a configuration for which we are
 | |
|         # going to need it.
 | |
|         if self._block_events_without_consent_error:
 | |
|             self._consent_uri_builder = ConsentURIBuilder(self.config)
 | |
| 
 | |
|         if (
 | |
|             not self.config.worker_app
 | |
|             and self.config.cleanup_extremities_with_dummy_events
 | |
|         ):
 | |
|             self.clock.looping_call(
 | |
|                 lambda: run_as_background_process(
 | |
|                     "send_dummy_events_to_fill_extremities",
 | |
|                     self._send_dummy_events_to_fill_extremities,
 | |
|                 ),
 | |
|                 5 * 60 * 1000,
 | |
|             )
 | |
| 
 | |
|         self._message_handler = hs.get_message_handler()
 | |
| 
 | |
|         self._ephemeral_events_enabled = hs.config.enable_ephemeral_messages
 | |
| 
 | |
|     @defer.inlineCallbacks
 | |
|     def create_event(
 | |
|         self,
 | |
|         requester,
 | |
|         event_dict,
 | |
|         token_id=None,
 | |
|         txn_id=None,
 | |
|         prev_events_and_hashes=None,
 | |
|         require_consent=True,
 | |
|     ):
 | |
|         """
 | |
|         Given a dict from a client, create a new event.
 | |
| 
 | |
|         Creates an FrozenEvent object, filling out auth_events, prev_events,
 | |
|         etc.
 | |
| 
 | |
|         Adds display names to Join membership events.
 | |
| 
 | |
|         Args:
 | |
|             requester
 | |
|             event_dict (dict): An entire event
 | |
|             token_id (str)
 | |
|             txn_id (str)
 | |
| 
 | |
|             prev_events_and_hashes (list[(str, dict[str, str], int)]|None):
 | |
|                 the forward extremities to use as the prev_events for the
 | |
|                 new event. For each event, a tuple of (event_id, hashes, depth)
 | |
|                 where *hashes* is a map from algorithm to hash.
 | |
| 
 | |
|                 If None, they will be requested from the database.
 | |
| 
 | |
|             require_consent (bool): Whether to check if the requester has
 | |
|                 consented to privacy policy.
 | |
|         Raises:
 | |
|             ResourceLimitError if server is blocked to some resource being
 | |
|             exceeded
 | |
|         Returns:
 | |
|             Tuple of created event (FrozenEvent), Context
 | |
|         """
 | |
|         yield self.auth.check_auth_blocking(requester.user.to_string())
 | |
| 
 | |
|         if event_dict["type"] == EventTypes.Create and event_dict["state_key"] == "":
 | |
|             room_version = event_dict["content"]["room_version"]
 | |
|         else:
 | |
|             try:
 | |
|                 room_version = yield self.store.get_room_version(event_dict["room_id"])
 | |
|             except NotFoundError:
 | |
|                 raise AuthError(403, "Unknown room")
 | |
| 
 | |
|         builder = self.event_builder_factory.new(room_version, event_dict)
 | |
| 
 | |
|         self.validator.validate_builder(builder)
 | |
| 
 | |
|         if builder.type == EventTypes.Member:
 | |
|             membership = builder.content.get("membership", None)
 | |
|             target = UserID.from_string(builder.state_key)
 | |
| 
 | |
|             if membership in {Membership.JOIN, Membership.INVITE}:
 | |
|                 # If event doesn't include a display name, add one.
 | |
|                 profile = self.profile_handler
 | |
|                 content = builder.content
 | |
| 
 | |
|                 try:
 | |
|                     if "displayname" not in content:
 | |
|                         content["displayname"] = yield profile.get_displayname(target)
 | |
|                     if "avatar_url" not in content:
 | |
|                         content["avatar_url"] = yield profile.get_avatar_url(target)
 | |
|                 except Exception as e:
 | |
|                     logger.info(
 | |
|                         "Failed to get profile information for %r: %s", target, e
 | |
|                     )
 | |
| 
 | |
|         is_exempt = yield self._is_exempt_from_privacy_policy(builder, requester)
 | |
|         if require_consent and not is_exempt:
 | |
|             yield self.assert_accepted_privacy_policy(requester)
 | |
| 
 | |
|         if token_id is not None:
 | |
|             builder.internal_metadata.token_id = token_id
 | |
| 
 | |
|         if txn_id is not None:
 | |
|             builder.internal_metadata.txn_id = txn_id
 | |
| 
 | |
|         event, context = yield self.create_new_client_event(
 | |
|             builder=builder,
 | |
|             requester=requester,
 | |
|             prev_events_and_hashes=prev_events_and_hashes,
 | |
|         )
 | |
| 
 | |
|         # In an ideal world we wouldn't need the second part of this condition. However,
 | |
|         # this behaviour isn't spec'd yet, meaning we should be able to deactivate this
 | |
|         # behaviour. Another reason is that this code is also evaluated each time a new
 | |
|         # m.room.aliases event is created, which includes hitting a /directory route.
 | |
|         # Therefore not including this condition here would render the similar one in
 | |
|         # synapse.handlers.directory pointless.
 | |
|         if builder.type == EventTypes.Aliases and self.require_membership_for_aliases:
 | |
|             # Ideally we'd do the membership check in event_auth.check(), which
 | |
|             # describes a spec'd algorithm for authenticating events received over
 | |
|             # federation as well as those created locally. As of room v3, aliases events
 | |
|             # can be created by users that are not in the room, therefore we have to
 | |
|             # tolerate them in event_auth.check().
 | |
|             prev_state_ids = yield context.get_prev_state_ids(self.store)
 | |
|             prev_event_id = prev_state_ids.get((EventTypes.Member, event.sender))
 | |
|             prev_event = (
 | |
|                 yield self.store.get_event(prev_event_id, allow_none=True)
 | |
|                 if prev_event_id
 | |
|                 else None
 | |
|             )
 | |
|             if not prev_event or prev_event.membership != Membership.JOIN:
 | |
|                 logger.warning(
 | |
|                     (
 | |
|                         "Attempt to send `m.room.aliases` in room %s by user %s but"
 | |
|                         " membership is %s"
 | |
|                     ),
 | |
|                     event.room_id,
 | |
|                     event.sender,
 | |
|                     prev_event.membership if prev_event else None,
 | |
|                 )
 | |
| 
 | |
|                 raise AuthError(
 | |
|                     403, "You must be in the room to create an alias for it"
 | |
|                 )
 | |
| 
 | |
|         self.validator.validate_new(event, self.config)
 | |
| 
 | |
|         return (event, context)
 | |
| 
 | |
|     def _is_exempt_from_privacy_policy(self, builder, requester):
 | |
|         """"Determine if an event to be sent is exempt from having to consent
 | |
|         to the privacy policy
 | |
| 
 | |
|         Args:
 | |
|             builder (synapse.events.builder.EventBuilder): event being created
 | |
|             requester (Requster): user requesting this event
 | |
| 
 | |
|         Returns:
 | |
|             Deferred[bool]: true if the event can be sent without the user
 | |
|                 consenting
 | |
|         """
 | |
|         # the only thing the user can do is join the server notices room.
 | |
|         if builder.type == EventTypes.Member:
 | |
|             membership = builder.content.get("membership", None)
 | |
|             if membership == Membership.JOIN:
 | |
|                 return self._is_server_notices_room(builder.room_id)
 | |
|             elif membership == Membership.LEAVE:
 | |
|                 # the user is always allowed to leave (but not kick people)
 | |
|                 return builder.state_key == requester.user.to_string()
 | |
|         return succeed(False)
 | |
| 
 | |
|     @defer.inlineCallbacks
 | |
|     def _is_server_notices_room(self, room_id):
 | |
|         if self.config.server_notices_mxid is None:
 | |
|             return False
 | |
|         user_ids = yield self.store.get_users_in_room(room_id)
 | |
|         return self.config.server_notices_mxid in user_ids
 | |
| 
 | |
|     @defer.inlineCallbacks
 | |
|     def assert_accepted_privacy_policy(self, requester):
 | |
|         """Check if a user has accepted the privacy policy
 | |
| 
 | |
|         Called when the given user is about to do something that requires
 | |
|         privacy consent. We see if the user is exempt and otherwise check that
 | |
|         they have given consent. If they have not, a ConsentNotGiven error is
 | |
|         raised.
 | |
| 
 | |
|         Args:
 | |
|             requester (synapse.types.Requester):
 | |
|                 The user making the request
 | |
| 
 | |
|         Returns:
 | |
|             Deferred[None]: returns normally if the user has consented or is
 | |
|                 exempt
 | |
| 
 | |
|         Raises:
 | |
|             ConsentNotGivenError: if the user has not given consent yet
 | |
|         """
 | |
|         if self._block_events_without_consent_error is None:
 | |
|             return
 | |
| 
 | |
|         # exempt AS users from needing consent
 | |
|         if requester.app_service is not None:
 | |
|             return
 | |
| 
 | |
|         user_id = requester.user.to_string()
 | |
| 
 | |
|         # exempt the system notices user
 | |
|         if (
 | |
|             self.config.server_notices_mxid is not None
 | |
|             and user_id == self.config.server_notices_mxid
 | |
|         ):
 | |
|             return
 | |
| 
 | |
|         u = yield self.store.get_user_by_id(user_id)
 | |
|         assert u is not None
 | |
|         if u["user_type"] in (UserTypes.SUPPORT, UserTypes.BOT):
 | |
|             # support and bot users are not required to consent
 | |
|             return
 | |
|         if u["appservice_id"] is not None:
 | |
|             # users registered by an appservice are exempt
 | |
|             return
 | |
|         if u["consent_version"] == self.config.user_consent_version:
 | |
|             return
 | |
| 
 | |
|         consent_uri = self._consent_uri_builder.build_user_consent_uri(
 | |
|             requester.user.localpart
 | |
|         )
 | |
|         msg = self._block_events_without_consent_error % {"consent_uri": consent_uri}
 | |
|         raise ConsentNotGivenError(msg=msg, consent_uri=consent_uri)
 | |
| 
 | |
|     @defer.inlineCallbacks
 | |
|     def send_nonmember_event(self, requester, event, context, ratelimit=True):
 | |
|         """
 | |
|         Persists and notifies local clients and federation of an event.
 | |
| 
 | |
|         Args:
 | |
|             event (FrozenEvent) the event to send.
 | |
|             context (Context) the context of the event.
 | |
|             ratelimit (bool): Whether to rate limit this send.
 | |
|             is_guest (bool): Whether the sender is a guest.
 | |
|         """
 | |
|         if event.type == EventTypes.Member:
 | |
|             raise SynapseError(
 | |
|                 500, "Tried to send member event through non-member codepath"
 | |
|             )
 | |
| 
 | |
|         user = UserID.from_string(event.sender)
 | |
| 
 | |
|         assert self.hs.is_mine(user), "User must be our own: %s" % (user,)
 | |
| 
 | |
|         if event.is_state():
 | |
|             prev_state = yield self.deduplicate_state_event(event, context)
 | |
|             if prev_state is not None:
 | |
|                 logger.info(
 | |
|                     "Not bothering to persist state event %s duplicated by %s",
 | |
|                     event.event_id,
 | |
|                     prev_state.event_id,
 | |
|                 )
 | |
|                 return prev_state
 | |
| 
 | |
|         yield self.handle_new_client_event(
 | |
|             requester=requester, event=event, context=context, ratelimit=ratelimit
 | |
|         )
 | |
| 
 | |
|     @defer.inlineCallbacks
 | |
|     def deduplicate_state_event(self, event, context):
 | |
|         """
 | |
|         Checks whether event is in the latest resolved state in context.
 | |
| 
 | |
|         If so, returns the version of the event in context.
 | |
|         Otherwise, returns None.
 | |
|         """
 | |
|         prev_state_ids = yield context.get_prev_state_ids(self.store)
 | |
|         prev_event_id = prev_state_ids.get((event.type, event.state_key))
 | |
|         if not prev_event_id:
 | |
|             return
 | |
|         prev_event = yield self.store.get_event(prev_event_id, allow_none=True)
 | |
|         if not prev_event:
 | |
|             return
 | |
| 
 | |
|         if prev_event and event.user_id == prev_event.user_id:
 | |
|             prev_content = encode_canonical_json(prev_event.content)
 | |
|             next_content = encode_canonical_json(event.content)
 | |
|             if prev_content == next_content:
 | |
|                 return prev_event
 | |
|         return
 | |
| 
 | |
|     @defer.inlineCallbacks
 | |
|     def create_and_send_nonmember_event(
 | |
|         self, requester, event_dict, ratelimit=True, txn_id=None
 | |
|     ):
 | |
|         """
 | |
|         Creates an event, then sends it.
 | |
| 
 | |
|         See self.create_event and self.send_nonmember_event.
 | |
|         """
 | |
| 
 | |
|         # We limit the number of concurrent event sends in a room so that we
 | |
|         # don't fork the DAG too much. If we don't limit then we can end up in
 | |
|         # a situation where event persistence can't keep up, causing
 | |
|         # extremities to pile up, which in turn leads to state resolution
 | |
|         # taking longer.
 | |
|         with (yield self.limiter.queue(event_dict["room_id"])):
 | |
|             event, context = yield self.create_event(
 | |
|                 requester, event_dict, token_id=requester.access_token_id, txn_id=txn_id
 | |
|             )
 | |
| 
 | |
|             spam_error = self.spam_checker.check_event_for_spam(event)
 | |
|             if spam_error:
 | |
|                 if not isinstance(spam_error, string_types):
 | |
|                     spam_error = "Spam is not permitted here"
 | |
|                 raise SynapseError(403, spam_error, Codes.FORBIDDEN)
 | |
| 
 | |
|             yield self.send_nonmember_event(
 | |
|                 requester, event, context, ratelimit=ratelimit
 | |
|             )
 | |
|         return event
 | |
| 
 | |
|     @measure_func("create_new_client_event")
 | |
|     @defer.inlineCallbacks
 | |
|     def create_new_client_event(
 | |
|         self, builder, requester=None, prev_events_and_hashes=None
 | |
|     ):
 | |
|         """Create a new event for a local client
 | |
| 
 | |
|         Args:
 | |
|             builder (EventBuilder):
 | |
| 
 | |
|             requester (synapse.types.Requester|None):
 | |
| 
 | |
|             prev_events_and_hashes (list[(str, dict[str, str], int)]|None):
 | |
|                 the forward extremities to use as the prev_events for the
 | |
|                 new event. For each event, a tuple of (event_id, hashes, depth)
 | |
|                 where *hashes* is a map from algorithm to hash.
 | |
| 
 | |
|                 If None, they will be requested from the database.
 | |
| 
 | |
|         Returns:
 | |
|             Deferred[(synapse.events.EventBase, synapse.events.snapshot.EventContext)]
 | |
|         """
 | |
| 
 | |
|         if prev_events_and_hashes is not None:
 | |
|             assert len(prev_events_and_hashes) <= 10, (
 | |
|                 "Attempting to create an event with %i prev_events"
 | |
|                 % (len(prev_events_and_hashes),)
 | |
|             )
 | |
|         else:
 | |
|             prev_events_and_hashes = yield self.store.get_prev_events_for_room(
 | |
|                 builder.room_id
 | |
|             )
 | |
| 
 | |
|         prev_events = [
 | |
|             (event_id, prev_hashes)
 | |
|             for event_id, prev_hashes, _ in prev_events_and_hashes
 | |
|         ]
 | |
| 
 | |
|         event = yield builder.build(prev_event_ids=[p for p, _ in prev_events])
 | |
|         context = yield self.state.compute_event_context(event)
 | |
|         if requester:
 | |
|             context.app_service = requester.app_service
 | |
| 
 | |
|         self.validator.validate_new(event, self.config)
 | |
| 
 | |
|         # If this event is an annotation then we check that that the sender
 | |
|         # can't annotate the same way twice (e.g. stops users from liking an
 | |
|         # event multiple times).
 | |
|         relation = event.content.get("m.relates_to", {})
 | |
|         if relation.get("rel_type") == RelationTypes.ANNOTATION:
 | |
|             relates_to = relation["event_id"]
 | |
|             aggregation_key = relation["key"]
 | |
| 
 | |
|             already_exists = yield self.store.has_user_annotated_event(
 | |
|                 relates_to, event.type, aggregation_key, event.sender
 | |
|             )
 | |
|             if already_exists:
 | |
|                 raise SynapseError(400, "Can't send same reaction twice")
 | |
| 
 | |
|         logger.debug("Created event %s", event.event_id)
 | |
| 
 | |
|         return (event, context)
 | |
| 
 | |
|     @measure_func("handle_new_client_event")
 | |
|     @defer.inlineCallbacks
 | |
|     def handle_new_client_event(
 | |
|         self, requester, event, context, ratelimit=True, extra_users=[]
 | |
|     ):
 | |
|         """Processes a new event. This includes checking auth, persisting it,
 | |
|         notifying users, sending to remote servers, etc.
 | |
| 
 | |
|         If called from a worker will hit out to the master process for final
 | |
|         processing.
 | |
| 
 | |
|         Args:
 | |
|             requester (Requester)
 | |
|             event (FrozenEvent)
 | |
|             context (EventContext)
 | |
|             ratelimit (bool)
 | |
|             extra_users (list(UserID)): Any extra users to notify about event
 | |
|         """
 | |
| 
 | |
|         if event.is_state() and (event.type, event.state_key) == (
 | |
|             EventTypes.Create,
 | |
|             "",
 | |
|         ):
 | |
|             room_version = event.content.get("room_version", RoomVersions.V1.identifier)
 | |
|         else:
 | |
|             room_version = yield self.store.get_room_version(event.room_id)
 | |
| 
 | |
|         event_allowed = yield self.third_party_event_rules.check_event_allowed(
 | |
|             event, context
 | |
|         )
 | |
|         if not event_allowed:
 | |
|             raise SynapseError(
 | |
|                 403, "This event is not allowed in this context", Codes.FORBIDDEN
 | |
|             )
 | |
| 
 | |
|         try:
 | |
|             yield self.auth.check_from_context(room_version, event, context)
 | |
|         except AuthError as err:
 | |
|             logger.warning("Denying new event %r because %s", event, err)
 | |
|             raise err
 | |
| 
 | |
|         # Ensure that we can round trip before trying to persist in db
 | |
|         try:
 | |
|             dump = frozendict_json_encoder.encode(event.content)
 | |
|             json.loads(dump)
 | |
|         except Exception:
 | |
|             logger.exception("Failed to encode content: %r", event.content)
 | |
|             raise
 | |
| 
 | |
|         yield self.action_generator.handle_push_actions_for_event(event, context)
 | |
| 
 | |
|         # reraise does not allow inlineCallbacks to preserve the stacktrace, so we
 | |
|         # hack around with a try/finally instead.
 | |
|         success = False
 | |
|         try:
 | |
|             # If we're a worker we need to hit out to the master.
 | |
|             if self.config.worker_app:
 | |
|                 yield self.send_event_to_master(
 | |
|                     event_id=event.event_id,
 | |
|                     store=self.store,
 | |
|                     requester=requester,
 | |
|                     event=event,
 | |
|                     context=context,
 | |
|                     ratelimit=ratelimit,
 | |
|                     extra_users=extra_users,
 | |
|                 )
 | |
|                 success = True
 | |
|                 return
 | |
| 
 | |
|             yield self.persist_and_notify_client_event(
 | |
|                 requester, event, context, ratelimit=ratelimit, extra_users=extra_users
 | |
|             )
 | |
| 
 | |
|             success = True
 | |
|         finally:
 | |
|             if not success:
 | |
|                 # Ensure that we actually remove the entries in the push actions
 | |
|                 # staging area, if we calculated them.
 | |
|                 run_in_background(
 | |
|                     self.store.remove_push_actions_from_staging, event.event_id
 | |
|                 )
 | |
| 
 | |
|     @defer.inlineCallbacks
 | |
|     def persist_and_notify_client_event(
 | |
|         self, requester, event, context, ratelimit=True, extra_users=[]
 | |
|     ):
 | |
|         """Called when we have fully built the event, have already
 | |
|         calculated the push actions for the event, and checked auth.
 | |
| 
 | |
|         This should only be run on master.
 | |
|         """
 | |
|         assert not self.config.worker_app
 | |
| 
 | |
|         if ratelimit:
 | |
|             # We check if this is a room admin redacting an event so that we
 | |
|             # can apply different ratelimiting. We do this by simply checking
 | |
|             # it's not a self-redaction (to avoid having to look up whether the
 | |
|             # user is actually admin or not).
 | |
|             is_admin_redaction = False
 | |
|             if event.type == EventTypes.Redaction:
 | |
|                 original_event = yield self.store.get_event(
 | |
|                     event.redacts,
 | |
|                     check_redacted=False,
 | |
|                     get_prev_content=False,
 | |
|                     allow_rejected=False,
 | |
|                     allow_none=True,
 | |
|                 )
 | |
| 
 | |
|                 is_admin_redaction = (
 | |
|                     original_event and event.sender != original_event.sender
 | |
|                 )
 | |
| 
 | |
|             yield self.base_handler.ratelimit(
 | |
|                 requester, is_admin_redaction=is_admin_redaction
 | |
|             )
 | |
| 
 | |
|         yield self.base_handler.maybe_kick_guest_users(event, context)
 | |
| 
 | |
|         if event.type == EventTypes.CanonicalAlias:
 | |
|             # Check the alias is acually valid (at this time at least)
 | |
|             room_alias_str = event.content.get("alias", None)
 | |
|             if room_alias_str:
 | |
|                 room_alias = RoomAlias.from_string(room_alias_str)
 | |
|                 directory_handler = self.hs.get_handlers().directory_handler
 | |
|                 mapping = yield directory_handler.get_association(room_alias)
 | |
| 
 | |
|                 if mapping["room_id"] != event.room_id:
 | |
|                     raise SynapseError(
 | |
|                         400,
 | |
|                         "Room alias %s does not point to the room" % (room_alias_str,),
 | |
|                     )
 | |
| 
 | |
|         federation_handler = self.hs.get_handlers().federation_handler
 | |
| 
 | |
|         if event.type == EventTypes.Member:
 | |
|             if event.content["membership"] == Membership.INVITE:
 | |
| 
 | |
|                 def is_inviter_member_event(e):
 | |
|                     return e.type == EventTypes.Member and e.sender == event.sender
 | |
| 
 | |
|                 current_state_ids = yield context.get_current_state_ids(self.store)
 | |
| 
 | |
|                 state_to_include_ids = [
 | |
|                     e_id
 | |
|                     for k, e_id in iteritems(current_state_ids)
 | |
|                     if k[0] in self.room_invite_state_types
 | |
|                     or k == (EventTypes.Member, event.sender)
 | |
|                 ]
 | |
| 
 | |
|                 state_to_include = yield self.store.get_events(state_to_include_ids)
 | |
| 
 | |
|                 event.unsigned["invite_room_state"] = [
 | |
|                     {
 | |
|                         "type": e.type,
 | |
|                         "state_key": e.state_key,
 | |
|                         "content": e.content,
 | |
|                         "sender": e.sender,
 | |
|                     }
 | |
|                     for e in itervalues(state_to_include)
 | |
|                 ]
 | |
| 
 | |
|                 invitee = UserID.from_string(event.state_key)
 | |
|                 if not self.hs.is_mine(invitee):
 | |
|                     # TODO: Can we add signature from remote server in a nicer
 | |
|                     # way? If we have been invited by a remote server, we need
 | |
|                     # to get them to sign the event.
 | |
| 
 | |
|                     returned_invite = yield federation_handler.send_invite(
 | |
|                         invitee.domain, event
 | |
|                     )
 | |
| 
 | |
|                     event.unsigned.pop("room_state", None)
 | |
| 
 | |
|                     # TODO: Make sure the signatures actually are correct.
 | |
|                     event.signatures.update(returned_invite.signatures)
 | |
| 
 | |
|         if event.type == EventTypes.Redaction:
 | |
|             original_event = yield self.store.get_event(
 | |
|                 event.redacts,
 | |
|                 check_redacted=False,
 | |
|                 get_prev_content=False,
 | |
|                 allow_rejected=False,
 | |
|                 allow_none=True,
 | |
|             )
 | |
| 
 | |
|             # we can make some additional checks now if we have the original event.
 | |
|             if original_event:
 | |
|                 if original_event.type == EventTypes.Create:
 | |
|                     raise AuthError(403, "Redacting create events is not permitted")
 | |
| 
 | |
|                 if original_event.room_id != event.room_id:
 | |
|                     raise SynapseError(400, "Cannot redact event from a different room")
 | |
| 
 | |
|             prev_state_ids = yield context.get_prev_state_ids(self.store)
 | |
|             auth_events_ids = yield self.auth.compute_auth_events(
 | |
|                 event, prev_state_ids, for_verification=True
 | |
|             )
 | |
|             auth_events = yield self.store.get_events(auth_events_ids)
 | |
|             auth_events = {(e.type, e.state_key): e for e in auth_events.values()}
 | |
|             room_version = yield self.store.get_room_version(event.room_id)
 | |
| 
 | |
|             if event_auth.check_redaction(room_version, event, auth_events=auth_events):
 | |
|                 # this user doesn't have 'redact' rights, so we need to do some more
 | |
|                 # checks on the original event. Let's start by checking the original
 | |
|                 # event exists.
 | |
|                 if not original_event:
 | |
|                     raise NotFoundError("Could not find event %s" % (event.redacts,))
 | |
| 
 | |
|                 if event.user_id != original_event.user_id:
 | |
|                     raise AuthError(403, "You don't have permission to redact events")
 | |
| 
 | |
|                 # all the checks are done.
 | |
|                 event.internal_metadata.recheck_redaction = False
 | |
| 
 | |
|         if event.type == EventTypes.Create:
 | |
|             prev_state_ids = yield context.get_prev_state_ids(self.store)
 | |
|             if prev_state_ids:
 | |
|                 raise AuthError(403, "Changing the room create event is forbidden")
 | |
| 
 | |
|         event_stream_id, max_stream_id = yield self.storage.persistence.persist_event(
 | |
|             event, context=context
 | |
|         )
 | |
| 
 | |
|         if self._ephemeral_events_enabled:
 | |
|             # If there's an expiry timestamp on the event, schedule its expiry.
 | |
|             self._message_handler.maybe_schedule_expiry(event)
 | |
| 
 | |
|         yield self.pusher_pool.on_new_notifications(event_stream_id, max_stream_id)
 | |
| 
 | |
|         def _notify():
 | |
|             try:
 | |
|                 self.notifier.on_new_room_event(
 | |
|                     event, event_stream_id, max_stream_id, extra_users=extra_users
 | |
|                 )
 | |
|             except Exception:
 | |
|                 logger.exception("Error notifying about new room event")
 | |
| 
 | |
|         run_in_background(_notify)
 | |
| 
 | |
|         if event.type == EventTypes.Message:
 | |
|             # We don't want to block sending messages on any presence code. This
 | |
|             # matters as sometimes presence code can take a while.
 | |
|             run_in_background(self._bump_active_time, requester.user)
 | |
| 
 | |
|     @defer.inlineCallbacks
 | |
|     def _bump_active_time(self, user):
 | |
|         try:
 | |
|             presence = self.hs.get_presence_handler()
 | |
|             yield presence.bump_presence_active_time(user)
 | |
|         except Exception:
 | |
|             logger.exception("Error bumping presence active time")
 | |
| 
 | |
|     @defer.inlineCallbacks
 | |
|     def _send_dummy_events_to_fill_extremities(self):
 | |
|         """Background task to send dummy events into rooms that have a large
 | |
|         number of extremities
 | |
|         """
 | |
|         self._expire_rooms_to_exclude_from_dummy_event_insertion()
 | |
|         room_ids = yield self.store.get_rooms_with_many_extremities(
 | |
|             min_count=10,
 | |
|             limit=5,
 | |
|             room_id_filter=self._rooms_to_exclude_from_dummy_event_insertion.keys(),
 | |
|         )
 | |
| 
 | |
|         for room_id in room_ids:
 | |
|             # For each room we need to find a joined member we can use to send
 | |
|             # the dummy event with.
 | |
| 
 | |
|             prev_events_and_hashes = yield self.store.get_prev_events_for_room(room_id)
 | |
| 
 | |
|             latest_event_ids = (event_id for (event_id, _, _) in prev_events_and_hashes)
 | |
| 
 | |
|             members = yield self.state.get_current_users_in_room(
 | |
|                 room_id, latest_event_ids=latest_event_ids
 | |
|             )
 | |
|             dummy_event_sent = False
 | |
|             for user_id in members:
 | |
|                 if not self.hs.is_mine_id(user_id):
 | |
|                     continue
 | |
|                 requester = create_requester(user_id)
 | |
|                 try:
 | |
|                     event, context = yield self.create_event(
 | |
|                         requester,
 | |
|                         {
 | |
|                             "type": "org.matrix.dummy_event",
 | |
|                             "content": {},
 | |
|                             "room_id": room_id,
 | |
|                             "sender": user_id,
 | |
|                         },
 | |
|                         prev_events_and_hashes=prev_events_and_hashes,
 | |
|                     )
 | |
| 
 | |
|                     event.internal_metadata.proactively_send = False
 | |
| 
 | |
|                     yield self.send_nonmember_event(
 | |
|                         requester, event, context, ratelimit=False
 | |
|                     )
 | |
|                     dummy_event_sent = True
 | |
|                     break
 | |
|                 except ConsentNotGivenError:
 | |
|                     logger.info(
 | |
|                         "Failed to send dummy event into room %s for user %s due to "
 | |
|                         "lack of consent. Will try another user" % (room_id, user_id)
 | |
|                     )
 | |
|                 except AuthError:
 | |
|                     logger.info(
 | |
|                         "Failed to send dummy event into room %s for user %s due to "
 | |
|                         "lack of power. Will try another user" % (room_id, user_id)
 | |
|                     )
 | |
| 
 | |
|             if not dummy_event_sent:
 | |
|                 # Did not find a valid user in the room, so remove from future attempts
 | |
|                 # Exclusion is time limited, so the room will be rechecked in the future
 | |
|                 # dependent on _DUMMY_EVENT_ROOM_EXCLUSION_EXPIRY
 | |
|                 logger.info(
 | |
|                     "Failed to send dummy event into room %s. Will exclude it from "
 | |
|                     "future attempts until cache expires" % (room_id,)
 | |
|                 )
 | |
|                 now = self.clock.time_msec()
 | |
|                 self._rooms_to_exclude_from_dummy_event_insertion[room_id] = now
 | |
| 
 | |
|     def _expire_rooms_to_exclude_from_dummy_event_insertion(self):
 | |
|         expire_before = self.clock.time_msec() - _DUMMY_EVENT_ROOM_EXCLUSION_EXPIRY
 | |
|         to_expire = set()
 | |
|         for room_id, time in self._rooms_to_exclude_from_dummy_event_insertion.items():
 | |
|             if time < expire_before:
 | |
|                 to_expire.add(room_id)
 | |
|         for room_id in to_expire:
 | |
|             logger.debug(
 | |
|                 "Expiring room id %s from dummy event insertion exclusion cache",
 | |
|                 room_id,
 | |
|             )
 | |
|             del self._rooms_to_exclude_from_dummy_event_insertion[room_id]
 |