2014-08-27 18:59:36 +02:00
|
|
|
# -*- coding: utf-8 -*-
|
2016-01-05 19:12:37 +01:00
|
|
|
# Copyright 2014 - 2016 OpenMarket Ltd
|
2018-02-06 17:40:38 +01:00
|
|
|
# Copyright 2017 - 2018 New Vector Ltd
|
2014-08-27 18:59:36 +02:00
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
2018-04-27 12:40:06 +02:00
|
|
|
import logging
|
|
|
|
|
2018-07-09 08:09:20 +02:00
|
|
|
from six import iteritems, itervalues, string_types
|
|
|
|
|
|
|
|
from canonicaljson import encode_canonical_json, json
|
|
|
|
|
2018-06-25 15:08:28 +02:00
|
|
|
from twisted.internet import defer
|
2018-05-22 09:56:52 +02:00
|
|
|
from twisted.internet.defer import succeed
|
2014-08-27 18:59:36 +02:00
|
|
|
|
2018-07-09 08:09:20 +02:00
|
|
|
from synapse.api.constants import MAX_DEPTH, EventTypes, Membership
|
2018-08-15 17:35:22 +02:00
|
|
|
from synapse.api.errors import (
|
|
|
|
AuthError,
|
|
|
|
Codes,
|
|
|
|
ConsentNotGivenError,
|
|
|
|
NotFoundError,
|
|
|
|
SynapseError,
|
|
|
|
)
|
2018-05-22 09:56:52 +02:00
|
|
|
from synapse.api.urls import ConsentURIBuilder
|
2016-05-11 10:09:20 +02:00
|
|
|
from synapse.crypto.event_signing import add_hashes_and_signatures
|
2015-01-26 17:11:28 +01:00
|
|
|
from synapse.events.utils import serialize_event
|
2014-12-10 18:59:47 +01:00
|
|
|
from synapse.events.validator import EventValidator
|
2018-07-31 14:53:54 +02:00
|
|
|
from synapse.replication.http.send_event import ReplicationSendEventRestServlet
|
2018-07-20 16:32:23 +02:00
|
|
|
from synapse.types import RoomAlias, UserID
|
2018-08-10 15:50:21 +02:00
|
|
|
from synapse.util.async_helpers import Linearizer
|
2018-07-09 08:09:20 +02:00
|
|
|
from synapse.util.frozenutils import frozendict_json_encoder
|
2018-04-27 15:32:08 +02:00
|
|
|
from synapse.util.logcontext import run_in_background
|
2016-08-23 16:23:39 +02:00
|
|
|
from synapse.util.metrics import measure_func
|
2018-08-15 17:35:22 +02:00
|
|
|
from synapse.visibility import filter_events_for_client
|
2014-12-10 18:59:47 +01:00
|
|
|
|
2014-09-05 22:35:56 +02:00
|
|
|
from ._base import BaseHandler
|
2014-08-27 18:59:36 +02:00
|
|
|
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
2018-07-18 16:22:02 +02:00
|
|
|
class MessageHandler(object):
|
|
|
|
"""Contains some read only APIs to get state about a room
|
|
|
|
"""
|
2014-08-27 18:59:36 +02:00
|
|
|
|
|
|
|
def __init__(self, hs):
|
2018-07-18 16:22:02 +02:00
|
|
|
self.auth = hs.get_auth()
|
|
|
|
self.clock = hs.get_clock()
|
2015-02-09 18:41:29 +01:00
|
|
|
self.state = hs.get_state_handler()
|
2018-07-18 16:22:02 +02:00
|
|
|
self.store = hs.get_datastore()
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def get_room_data(self, user_id=None, room_id=None,
|
|
|
|
event_type=None, state_key="", is_guest=False):
|
|
|
|
""" Get data from a room.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
event : The room path event
|
|
|
|
Returns:
|
|
|
|
The path data content.
|
|
|
|
Raises:
|
|
|
|
SynapseError if something went wrong.
|
|
|
|
"""
|
2018-07-20 16:30:59 +02:00
|
|
|
membership, membership_event_id = yield self.auth.check_in_room_or_world_readable(
|
2018-07-18 16:22:02 +02:00
|
|
|
room_id, user_id
|
|
|
|
)
|
|
|
|
|
|
|
|
if membership == Membership.JOIN:
|
|
|
|
data = yield self.state.get_current_state(
|
|
|
|
room_id, event_type, state_key
|
|
|
|
)
|
|
|
|
elif membership == Membership.LEAVE:
|
|
|
|
key = (event_type, state_key)
|
|
|
|
room_state = yield self.store.get_state_for_events(
|
|
|
|
[membership_event_id], [key]
|
|
|
|
)
|
|
|
|
data = room_state[membership_event_id].get(key)
|
|
|
|
|
|
|
|
defer.returnValue(data)
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
2018-08-15 17:35:22 +02:00
|
|
|
def get_state_events(
|
|
|
|
self, user_id, room_id, types=None, filtered_types=None,
|
|
|
|
at_token=None, is_guest=False,
|
|
|
|
):
|
2018-07-18 16:22:02 +02:00
|
|
|
"""Retrieve all state events for a given room. If the user is
|
|
|
|
joined to the room then return the current state. If the user has
|
2018-08-15 17:35:22 +02:00
|
|
|
left the room return the state events from when they left. If an explicit
|
|
|
|
'at' parameter is passed, return the state events as of that event, if
|
|
|
|
visible.
|
2018-07-18 16:22:02 +02:00
|
|
|
|
|
|
|
Args:
|
|
|
|
user_id(str): The user requesting state events.
|
|
|
|
room_id(str): The room ID to get all state events from.
|
2018-08-15 17:35:22 +02:00
|
|
|
types(list[(str, str|None)]|None): List of (type, state_key) tuples
|
|
|
|
which are used to filter the state fetched. If `state_key` is None,
|
|
|
|
all events are returned of the given type.
|
|
|
|
May be None, which matches any key.
|
|
|
|
filtered_types(list[str]|None): Only apply filtering via `types` to this
|
|
|
|
list of event types. Other types of events are returned unfiltered.
|
|
|
|
If None, `types` filtering is applied to all events.
|
|
|
|
at_token(StreamToken|None): the stream token of the at which we are requesting
|
|
|
|
the stats. If the user is not allowed to view the state as of that
|
|
|
|
stream token, we raise a 403 SynapseError. If None, returns the current
|
|
|
|
state based on the current_state_events table.
|
|
|
|
is_guest(bool): whether this user is a guest
|
2018-07-18 16:22:02 +02:00
|
|
|
Returns:
|
|
|
|
A list of dicts representing state events. [{}, {}, {}]
|
2018-08-15 17:35:22 +02:00
|
|
|
Raises:
|
|
|
|
NotFoundError (404) if the at token does not yield an event
|
|
|
|
|
|
|
|
AuthError (403) if the user doesn't have permission to view
|
|
|
|
members of this room.
|
2018-07-18 16:22:02 +02:00
|
|
|
"""
|
2018-08-15 17:35:22 +02:00
|
|
|
if at_token:
|
|
|
|
# FIXME this claims to get the state at a stream position, but
|
|
|
|
# get_recent_events_for_room operates by topo ordering. This therefore
|
|
|
|
# does not reliably give you the state at the given stream position.
|
|
|
|
# (https://github.com/matrix-org/synapse/issues/3305)
|
|
|
|
last_events, _ = yield self.store.get_recent_events_for_room(
|
|
|
|
room_id, end_token=at_token.room_key, limit=1,
|
|
|
|
)
|
2018-07-18 16:22:02 +02:00
|
|
|
|
2018-08-15 17:35:22 +02:00
|
|
|
if not last_events:
|
|
|
|
raise NotFoundError("Can't find event for token %s" % (at_token, ))
|
|
|
|
|
|
|
|
visible_events = yield filter_events_for_client(
|
|
|
|
self.store, user_id, last_events,
|
2018-07-18 16:22:02 +02:00
|
|
|
)
|
2018-08-15 17:35:22 +02:00
|
|
|
|
|
|
|
event = last_events[0]
|
|
|
|
if visible_events:
|
|
|
|
room_state = yield self.store.get_state_for_events(
|
|
|
|
[event.event_id], types, filtered_types=filtered_types,
|
|
|
|
)
|
|
|
|
room_state = room_state[event.event_id]
|
|
|
|
else:
|
|
|
|
raise AuthError(
|
|
|
|
403,
|
|
|
|
"User %s not allowed to view events in room %s at token %s" % (
|
|
|
|
user_id, room_id, at_token,
|
|
|
|
)
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
membership, membership_event_id = (
|
|
|
|
yield self.auth.check_in_room_or_world_readable(
|
|
|
|
room_id, user_id,
|
|
|
|
)
|
|
|
|
)
|
|
|
|
|
|
|
|
if membership == Membership.JOIN:
|
|
|
|
state_ids = yield self.store.get_filtered_current_state_ids(
|
|
|
|
room_id, types, filtered_types=filtered_types,
|
|
|
|
)
|
|
|
|
room_state = yield self.store.get_events(state_ids.values())
|
|
|
|
elif membership == Membership.LEAVE:
|
|
|
|
room_state = yield self.store.get_state_for_events(
|
|
|
|
[membership_event_id], types, filtered_types=filtered_types,
|
|
|
|
)
|
|
|
|
room_state = room_state[membership_event_id]
|
2018-07-18 16:22:02 +02:00
|
|
|
|
|
|
|
now = self.clock.time_msec()
|
|
|
|
defer.returnValue(
|
|
|
|
[serialize_event(c, now) for c in room_state.values()]
|
|
|
|
)
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def get_joined_members(self, requester, room_id):
|
|
|
|
"""Get all the joined members in the room and their profile information.
|
|
|
|
|
|
|
|
If the user has left the room return the state events from when they left.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
requester(Requester): The user requesting state events.
|
|
|
|
room_id(str): The room ID to get all state events from.
|
|
|
|
Returns:
|
|
|
|
A dict of user_id to profile info
|
|
|
|
"""
|
|
|
|
user_id = requester.user.to_string()
|
|
|
|
if not requester.app_service:
|
|
|
|
# We check AS auth after fetching the room membership, as it
|
|
|
|
# requires us to pull out all joined members anyway.
|
2018-07-20 16:30:59 +02:00
|
|
|
membership, _ = yield self.auth.check_in_room_or_world_readable(
|
2018-07-18 16:22:02 +02:00
|
|
|
room_id, user_id
|
|
|
|
)
|
|
|
|
if membership != Membership.JOIN:
|
|
|
|
raise NotImplementedError(
|
|
|
|
"Getting joined members after leaving is not implemented"
|
|
|
|
)
|
|
|
|
|
|
|
|
users_with_profile = yield self.state.get_current_user_in_room(room_id)
|
|
|
|
|
|
|
|
# If this is an AS, double check that they are allowed to see the members.
|
|
|
|
# This can either be because the AS user is in the room or because there
|
|
|
|
# is a user in the room that the AS is "interested in"
|
|
|
|
if requester.app_service and user_id not in users_with_profile:
|
|
|
|
for uid in users_with_profile:
|
|
|
|
if requester.app_service.is_interested_in_user(uid):
|
|
|
|
break
|
|
|
|
else:
|
|
|
|
# Loop fell through, AS has no interested users in room
|
|
|
|
raise AuthError(403, "Appservice not in room")
|
|
|
|
|
|
|
|
defer.returnValue({
|
|
|
|
user_id: {
|
|
|
|
"avatar_url": profile.avatar_url,
|
|
|
|
"display_name": profile.display_name,
|
|
|
|
}
|
|
|
|
for user_id, profile in iteritems(users_with_profile)
|
|
|
|
})
|
|
|
|
|
|
|
|
|
2018-01-15 17:51:53 +01:00
|
|
|
class EventCreationHandler(object):
|
|
|
|
def __init__(self, hs):
|
|
|
|
self.hs = hs
|
|
|
|
self.auth = hs.get_auth()
|
|
|
|
self.store = hs.get_datastore()
|
|
|
|
self.state = hs.get_state_handler()
|
|
|
|
self.clock = hs.get_clock()
|
|
|
|
self.validator = EventValidator()
|
|
|
|
self.profile_handler = hs.get_profile_handler()
|
|
|
|
self.event_builder_factory = hs.get_event_builder_factory()
|
|
|
|
self.server_name = hs.hostname
|
|
|
|
self.ratelimiter = hs.get_ratelimiter()
|
|
|
|
self.notifier = hs.get_notifier()
|
2018-02-05 18:22:16 +01:00
|
|
|
self.config = hs.config
|
|
|
|
|
2018-07-31 14:53:54 +02:00
|
|
|
self.send_event_to_master = ReplicationSendEventRestServlet.make_client(hs)
|
2018-01-15 17:51:53 +01:00
|
|
|
|
|
|
|
# This is only used to get at ratelimit function, and maybe_kick_guest_users
|
|
|
|
self.base_handler = BaseHandler(hs)
|
|
|
|
|
|
|
|
self.pusher_pool = hs.get_pusherpool()
|
|
|
|
|
|
|
|
# We arbitrarily limit concurrent event creation for a room to 5.
|
|
|
|
# This is to stop us from diverging history *too* much.
|
2018-07-20 14:11:43 +02:00
|
|
|
self.limiter = Linearizer(max_count=5, name="room_event_creation_limit")
|
2018-01-15 17:51:53 +01:00
|
|
|
|
|
|
|
self.action_generator = hs.get_action_generator()
|
|
|
|
|
|
|
|
self.spam_checker = hs.get_spam_checker()
|
|
|
|
|
2018-05-22 09:56:52 +02:00
|
|
|
if self.config.block_events_without_consent_error is not None:
|
|
|
|
self._consent_uri_builder = ConsentURIBuilder(self.config)
|
|
|
|
|
2014-12-04 16:50:01 +01:00
|
|
|
@defer.inlineCallbacks
|
2017-05-02 12:36:11 +02:00
|
|
|
def create_event(self, requester, event_dict, token_id=None, txn_id=None,
|
2018-04-16 19:41:37 +02:00
|
|
|
prev_events_and_hashes=None):
|
2016-01-15 17:27:26 +01:00
|
|
|
"""
|
|
|
|
Given a dict from a client, create a new event.
|
2014-12-15 18:01:12 +01:00
|
|
|
|
|
|
|
Creates an FrozenEvent object, filling out auth_events, prev_events,
|
|
|
|
etc.
|
|
|
|
|
|
|
|
Adds display names to Join membership events.
|
|
|
|
|
|
|
|
Args:
|
2017-05-02 12:36:11 +02:00
|
|
|
requester
|
2014-12-15 18:01:12 +01:00
|
|
|
event_dict (dict): An entire event
|
2016-04-01 18:39:32 +02:00
|
|
|
token_id (str)
|
|
|
|
txn_id (str)
|
2018-04-16 19:41:37 +02:00
|
|
|
|
|
|
|
prev_events_and_hashes (list[(str, dict[str, str], int)]|None):
|
|
|
|
the forward extremities to use as the prev_events for the
|
|
|
|
new event. For each event, a tuple of (event_id, hashes, depth)
|
|
|
|
where *hashes* is a map from algorithm to hash.
|
|
|
|
|
|
|
|
If None, they will be requested from the database.
|
2018-08-16 22:25:16 +02:00
|
|
|
Raises:
|
|
|
|
ResourceLimitError if server is blocked to some resource being
|
|
|
|
exceeded
|
2016-01-15 17:27:26 +01:00
|
|
|
Returns:
|
|
|
|
Tuple of created event (FrozenEvent), Context
|
2014-12-15 18:01:12 +01:00
|
|
|
"""
|
2018-08-16 22:25:16 +02:00
|
|
|
yield self.auth.check_auth_blocking(requester.user.to_string())
|
|
|
|
|
2014-12-04 16:50:01 +01:00
|
|
|
builder = self.event_builder_factory.new(event_dict)
|
|
|
|
|
2018-04-09 13:07:39 +02:00
|
|
|
self.validator.validate_new(builder)
|
|
|
|
|
|
|
|
if builder.type == EventTypes.Member:
|
|
|
|
membership = builder.content.get("membership", None)
|
|
|
|
target = UserID.from_string(builder.state_key)
|
|
|
|
|
|
|
|
if membership in {Membership.JOIN, Membership.INVITE}:
|
|
|
|
# If event doesn't include a display name, add one.
|
|
|
|
profile = self.profile_handler
|
|
|
|
content = builder.content
|
|
|
|
|
|
|
|
try:
|
|
|
|
if "displayname" not in content:
|
|
|
|
content["displayname"] = yield profile.get_displayname(target)
|
|
|
|
if "avatar_url" not in content:
|
|
|
|
content["avatar_url"] = yield profile.get_avatar_url(target)
|
|
|
|
except Exception as e:
|
|
|
|
logger.info(
|
|
|
|
"Failed to get profile information for %r: %s",
|
|
|
|
target, e
|
|
|
|
)
|
|
|
|
|
2018-06-25 18:56:10 +02:00
|
|
|
is_exempt = yield self._is_exempt_from_privacy_policy(builder, requester)
|
2018-05-22 09:56:52 +02:00
|
|
|
if not is_exempt:
|
|
|
|
yield self.assert_accepted_privacy_policy(requester)
|
|
|
|
|
2018-04-09 13:07:39 +02:00
|
|
|
if token_id is not None:
|
|
|
|
builder.internal_metadata.token_id = token_id
|
|
|
|
|
|
|
|
if txn_id is not None:
|
|
|
|
builder.internal_metadata.txn_id = txn_id
|
|
|
|
|
|
|
|
event, context = yield self.create_new_client_event(
|
|
|
|
builder=builder,
|
|
|
|
requester=requester,
|
2018-04-16 19:41:37 +02:00
|
|
|
prev_events_and_hashes=prev_events_and_hashes,
|
2018-04-09 13:07:39 +02:00
|
|
|
)
|
2015-01-28 17:58:23 +01:00
|
|
|
|
2016-01-15 17:27:26 +01:00
|
|
|
defer.returnValue((event, context))
|
|
|
|
|
2018-06-25 18:56:10 +02:00
|
|
|
def _is_exempt_from_privacy_policy(self, builder, requester):
|
2018-05-22 09:56:52 +02:00
|
|
|
""""Determine if an event to be sent is exempt from having to consent
|
|
|
|
to the privacy policy
|
|
|
|
|
|
|
|
Args:
|
|
|
|
builder (synapse.events.builder.EventBuilder): event being created
|
2018-06-25 18:56:10 +02:00
|
|
|
requester (Requster): user requesting this event
|
2018-05-22 09:56:52 +02:00
|
|
|
|
|
|
|
Returns:
|
|
|
|
Deferred[bool]: true if the event can be sent without the user
|
|
|
|
consenting
|
|
|
|
"""
|
|
|
|
# the only thing the user can do is join the server notices room.
|
|
|
|
if builder.type == EventTypes.Member:
|
|
|
|
membership = builder.content.get("membership", None)
|
|
|
|
if membership == Membership.JOIN:
|
|
|
|
return self._is_server_notices_room(builder.room_id)
|
2018-06-25 18:56:10 +02:00
|
|
|
elif membership == Membership.LEAVE:
|
|
|
|
# the user is always allowed to leave (but not kick people)
|
|
|
|
return builder.state_key == requester.user.to_string()
|
2018-05-22 09:56:52 +02:00
|
|
|
return succeed(False)
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def _is_server_notices_room(self, room_id):
|
|
|
|
if self.config.server_notices_mxid is None:
|
|
|
|
defer.returnValue(False)
|
|
|
|
user_ids = yield self.store.get_users_in_room(room_id)
|
|
|
|
defer.returnValue(self.config.server_notices_mxid in user_ids)
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def assert_accepted_privacy_policy(self, requester):
|
|
|
|
"""Check if a user has accepted the privacy policy
|
|
|
|
|
|
|
|
Called when the given user is about to do something that requires
|
|
|
|
privacy consent. We see if the user is exempt and otherwise check that
|
|
|
|
they have given consent. If they have not, a ConsentNotGiven error is
|
|
|
|
raised.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
requester (synapse.types.Requester):
|
|
|
|
The user making the request
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
Deferred[None]: returns normally if the user has consented or is
|
|
|
|
exempt
|
|
|
|
|
|
|
|
Raises:
|
|
|
|
ConsentNotGivenError: if the user has not given consent yet
|
|
|
|
"""
|
|
|
|
if self.config.block_events_without_consent_error is None:
|
|
|
|
return
|
|
|
|
|
|
|
|
# exempt AS users from needing consent
|
|
|
|
if requester.app_service is not None:
|
|
|
|
return
|
|
|
|
|
|
|
|
user_id = requester.user.to_string()
|
|
|
|
|
|
|
|
# exempt the system notices user
|
|
|
|
if (
|
|
|
|
self.config.server_notices_mxid is not None and
|
|
|
|
user_id == self.config.server_notices_mxid
|
|
|
|
):
|
|
|
|
return
|
|
|
|
|
|
|
|
u = yield self.store.get_user_by_id(user_id)
|
|
|
|
assert u is not None
|
2018-05-29 20:54:32 +02:00
|
|
|
if u["appservice_id"] is not None:
|
|
|
|
# users registered by an appservice are exempt
|
|
|
|
return
|
2018-05-22 09:56:52 +02:00
|
|
|
if u["consent_version"] == self.config.user_consent_version:
|
|
|
|
return
|
|
|
|
|
2018-05-23 16:28:23 +02:00
|
|
|
consent_uri = self._consent_uri_builder.build_user_consent_uri(
|
|
|
|
requester.user.localpart,
|
|
|
|
)
|
2018-05-23 16:24:31 +02:00
|
|
|
msg = self.config.block_events_without_consent_error % {
|
|
|
|
'consent_uri': consent_uri,
|
|
|
|
}
|
2018-05-22 09:56:52 +02:00
|
|
|
raise ConsentNotGivenError(
|
2018-05-23 16:24:31 +02:00
|
|
|
msg=msg,
|
2018-05-22 09:56:52 +02:00
|
|
|
consent_uri=consent_uri,
|
|
|
|
)
|
|
|
|
|
2016-01-15 17:27:26 +01:00
|
|
|
@defer.inlineCallbacks
|
2016-03-03 17:43:42 +01:00
|
|
|
def send_nonmember_event(self, requester, event, context, ratelimit=True):
|
2016-01-15 17:27:26 +01:00
|
|
|
"""
|
|
|
|
Persists and notifies local clients and federation of an event.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
event (FrozenEvent) the event to send.
|
|
|
|
context (Context) the context of the event.
|
|
|
|
ratelimit (bool): Whether to rate limit this send.
|
|
|
|
is_guest (bool): Whether the sender is a guest.
|
|
|
|
"""
|
2016-02-15 19:21:30 +01:00
|
|
|
if event.type == EventTypes.Member:
|
|
|
|
raise SynapseError(
|
|
|
|
500,
|
2016-02-18 12:02:14 +01:00
|
|
|
"Tried to send member event through non-member codepath"
|
2016-02-15 19:21:30 +01:00
|
|
|
)
|
|
|
|
|
2016-01-15 17:27:26 +01:00
|
|
|
user = UserID.from_string(event.sender)
|
|
|
|
|
|
|
|
assert self.hs.is_mine(user), "User must be our own: %s" % (user,)
|
|
|
|
|
2015-12-02 16:50:50 +01:00
|
|
|
if event.is_state():
|
2016-08-25 18:32:22 +02:00
|
|
|
prev_state = yield self.deduplicate_state_event(event, context)
|
2016-02-15 19:21:30 +01:00
|
|
|
if prev_state is not None:
|
|
|
|
defer.returnValue(prev_state)
|
2015-12-02 16:50:50 +01:00
|
|
|
|
2016-02-15 19:21:30 +01:00
|
|
|
yield self.handle_new_client_event(
|
2016-03-03 17:43:42 +01:00
|
|
|
requester=requester,
|
2016-02-15 19:21:30 +01:00
|
|
|
event=event,
|
|
|
|
context=context,
|
|
|
|
ratelimit=ratelimit,
|
|
|
|
)
|
2014-12-04 16:50:01 +01:00
|
|
|
|
2016-08-25 18:32:22 +02:00
|
|
|
@defer.inlineCallbacks
|
2016-02-15 19:21:30 +01:00
|
|
|
def deduplicate_state_event(self, event, context):
|
2016-02-16 15:25:23 +01:00
|
|
|
"""
|
|
|
|
Checks whether event is in the latest resolved state in context.
|
|
|
|
|
|
|
|
If so, returns the version of the event in context.
|
|
|
|
Otherwise, returns None.
|
|
|
|
"""
|
2018-07-23 14:00:22 +02:00
|
|
|
prev_state_ids = yield context.get_prev_state_ids(self.store)
|
|
|
|
prev_event_id = prev_state_ids.get((event.type, event.state_key))
|
2016-08-25 18:32:22 +02:00
|
|
|
prev_event = yield self.store.get_event(prev_event_id, allow_none=True)
|
|
|
|
if not prev_event:
|
|
|
|
return
|
|
|
|
|
2016-02-16 15:25:23 +01:00
|
|
|
if prev_event and event.user_id == prev_event.user_id:
|
|
|
|
prev_content = encode_canonical_json(prev_event.content)
|
2016-02-15 19:21:30 +01:00
|
|
|
next_content = encode_canonical_json(event.content)
|
|
|
|
if prev_content == next_content:
|
2016-08-25 18:32:22 +02:00
|
|
|
defer.returnValue(prev_event)
|
|
|
|
return
|
2016-02-15 19:21:30 +01:00
|
|
|
|
2016-01-15 17:27:26 +01:00
|
|
|
@defer.inlineCallbacks
|
2016-02-15 19:21:30 +01:00
|
|
|
def create_and_send_nonmember_event(
|
|
|
|
self,
|
2016-03-03 17:43:42 +01:00
|
|
|
requester,
|
2016-02-15 19:21:30 +01:00
|
|
|
event_dict,
|
|
|
|
ratelimit=True,
|
|
|
|
txn_id=None
|
|
|
|
):
|
2016-01-15 17:27:26 +01:00
|
|
|
"""
|
|
|
|
Creates an event, then sends it.
|
|
|
|
|
2016-02-15 19:21:30 +01:00
|
|
|
See self.create_event and self.send_nonmember_event.
|
2016-01-15 17:27:26 +01:00
|
|
|
"""
|
2017-09-19 13:20:11 +02:00
|
|
|
|
2018-04-10 15:00:24 +02:00
|
|
|
# We limit the number of concurrent event sends in a room so that we
|
|
|
|
# don't fork the DAG too much. If we don't limit then we can end up in
|
|
|
|
# a situation where event persistence can't keep up, causing
|
|
|
|
# extremities to pile up, which in turn leads to state resolution
|
|
|
|
# taking longer.
|
2018-04-09 13:07:39 +02:00
|
|
|
with (yield self.limiter.queue(event_dict["room_id"])):
|
|
|
|
event, context = yield self.create_event(
|
|
|
|
requester,
|
|
|
|
event_dict,
|
|
|
|
token_id=requester.access_token_id,
|
|
|
|
txn_id=txn_id
|
2017-09-19 13:20:11 +02:00
|
|
|
)
|
|
|
|
|
2018-04-09 13:07:39 +02:00
|
|
|
spam_error = self.spam_checker.check_event_for_spam(event)
|
|
|
|
if spam_error:
|
2018-05-05 22:47:18 +02:00
|
|
|
if not isinstance(spam_error, string_types):
|
2018-04-09 13:07:39 +02:00
|
|
|
spam_error = "Spam is not permitted here"
|
|
|
|
raise SynapseError(
|
|
|
|
403, spam_error, Codes.FORBIDDEN
|
|
|
|
)
|
|
|
|
|
|
|
|
yield self.send_nonmember_event(
|
|
|
|
requester,
|
|
|
|
event,
|
|
|
|
context,
|
|
|
|
ratelimit=ratelimit,
|
|
|
|
)
|
2014-12-04 16:50:01 +01:00
|
|
|
defer.returnValue(event)
|
|
|
|
|
2018-02-06 17:31:50 +01:00
|
|
|
@measure_func("create_new_client_event")
|
2016-05-11 10:09:20 +02:00
|
|
|
@defer.inlineCallbacks
|
2018-04-16 19:41:37 +02:00
|
|
|
def create_new_client_event(self, builder, requester=None,
|
|
|
|
prev_events_and_hashes=None):
|
|
|
|
"""Create a new event for a local client
|
|
|
|
|
|
|
|
Args:
|
|
|
|
builder (EventBuilder):
|
|
|
|
|
|
|
|
requester (synapse.types.Requester|None):
|
|
|
|
|
|
|
|
prev_events_and_hashes (list[(str, dict[str, str], int)]|None):
|
|
|
|
the forward extremities to use as the prev_events for the
|
|
|
|
new event. For each event, a tuple of (event_id, hashes, depth)
|
|
|
|
where *hashes* is a map from algorithm to hash.
|
|
|
|
|
|
|
|
If None, they will be requested from the database.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
Deferred[(synapse.events.EventBase, synapse.events.snapshot.EventContext)]
|
|
|
|
"""
|
|
|
|
|
|
|
|
if prev_events_and_hashes is not None:
|
|
|
|
assert len(prev_events_and_hashes) <= 10, \
|
|
|
|
"Attempting to create an event with %i prev_events" % (
|
|
|
|
len(prev_events_and_hashes),
|
2016-05-11 10:09:20 +02:00
|
|
|
)
|
2018-04-16 19:41:37 +02:00
|
|
|
else:
|
|
|
|
prev_events_and_hashes = \
|
|
|
|
yield self.store.get_prev_events_for_room(builder.room_id)
|
2016-05-11 10:09:20 +02:00
|
|
|
|
2018-04-16 19:41:37 +02:00
|
|
|
if prev_events_and_hashes:
|
|
|
|
depth = max([d for _, _, d in prev_events_and_hashes]) + 1
|
2018-05-01 17:19:39 +02:00
|
|
|
# we cap depth of generated events, to ensure that they are not
|
|
|
|
# rejected by other servers (and so that they can be persisted in
|
|
|
|
# the db)
|
|
|
|
depth = min(depth, MAX_DEPTH)
|
2018-04-16 19:41:37 +02:00
|
|
|
else:
|
|
|
|
depth = 1
|
2016-05-11 10:09:20 +02:00
|
|
|
|
2018-04-16 19:41:37 +02:00
|
|
|
prev_events = [
|
|
|
|
(event_id, prev_hashes)
|
|
|
|
for event_id, prev_hashes, _ in prev_events_and_hashes
|
|
|
|
]
|
2016-05-11 10:09:20 +02:00
|
|
|
|
|
|
|
builder.prev_events = prev_events
|
|
|
|
builder.depth = depth
|
|
|
|
|
2018-01-15 17:51:53 +01:00
|
|
|
context = yield self.state.compute_event_context(builder)
|
2017-05-02 12:36:11 +02:00
|
|
|
if requester:
|
|
|
|
context.app_service = requester.app_service
|
2016-05-11 10:09:20 +02:00
|
|
|
|
|
|
|
if builder.is_state():
|
|
|
|
builder.prev_state = yield self.store.add_event_hashes(
|
|
|
|
context.prev_state_events
|
|
|
|
)
|
|
|
|
|
|
|
|
yield self.auth.add_auth_events(builder, context)
|
|
|
|
|
2016-05-11 13:20:57 +02:00
|
|
|
signing_key = self.hs.config.signing_key[0]
|
2016-05-11 10:09:20 +02:00
|
|
|
add_hashes_and_signatures(
|
2016-05-11 13:20:57 +02:00
|
|
|
builder, self.server_name, signing_key
|
2016-05-11 10:09:20 +02:00
|
|
|
)
|
|
|
|
|
|
|
|
event = builder.build()
|
|
|
|
|
|
|
|
logger.debug(
|
2018-07-23 14:00:22 +02:00
|
|
|
"Created event %s",
|
|
|
|
event.event_id,
|
2016-05-11 10:09:20 +02:00
|
|
|
)
|
|
|
|
|
|
|
|
defer.returnValue(
|
|
|
|
(event, context,)
|
|
|
|
)
|
|
|
|
|
2016-08-23 16:23:39 +02:00
|
|
|
@measure_func("handle_new_client_event")
|
2016-05-11 10:09:20 +02:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def handle_new_client_event(
|
|
|
|
self,
|
|
|
|
requester,
|
|
|
|
event,
|
|
|
|
context,
|
|
|
|
ratelimit=True,
|
2018-02-15 17:30:10 +01:00
|
|
|
extra_users=[],
|
2016-05-11 10:09:20 +02:00
|
|
|
):
|
2018-02-15 17:30:10 +01:00
|
|
|
"""Processes a new event. This includes checking auth, persisting it,
|
|
|
|
notifying users, sending to remote servers, etc.
|
|
|
|
|
|
|
|
If called from a worker will hit out to the master process for final
|
|
|
|
processing.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
requester (Requester)
|
|
|
|
event (FrozenEvent)
|
|
|
|
context (EventContext)
|
|
|
|
ratelimit (bool)
|
2018-03-01 18:39:58 +01:00
|
|
|
extra_users (list(UserID)): Any extra users to notify about event
|
2018-02-15 17:30:10 +01:00
|
|
|
"""
|
|
|
|
|
2018-03-01 11:18:33 +01:00
|
|
|
try:
|
|
|
|
yield self.auth.check_from_context(event, context)
|
|
|
|
except AuthError as err:
|
|
|
|
logger.warn("Denying new event %r because %s", event, err)
|
|
|
|
raise err
|
|
|
|
|
|
|
|
# Ensure that we can round trip before trying to persist in db
|
|
|
|
try:
|
2018-03-29 23:57:28 +02:00
|
|
|
dump = frozendict_json_encoder.encode(event.content)
|
2018-06-28 15:49:57 +02:00
|
|
|
json.loads(dump)
|
2018-03-01 11:18:33 +01:00
|
|
|
except Exception:
|
|
|
|
logger.exception("Failed to encode content: %r", event.content)
|
|
|
|
raise
|
|
|
|
|
2018-02-15 17:30:10 +01:00
|
|
|
yield self.action_generator.handle_push_actions_for_event(
|
|
|
|
event, context
|
|
|
|
)
|
|
|
|
|
2018-10-01 19:48:51 +02:00
|
|
|
# reraise does not allow inlineCallbacks to preserve the stacktrace, so we
|
|
|
|
# hack around with a try/finally instead.
|
|
|
|
success = False
|
2018-02-15 17:30:10 +01:00
|
|
|
try:
|
|
|
|
# If we're a worker we need to hit out to the master.
|
|
|
|
if self.config.worker_app:
|
2018-07-31 14:53:54 +02:00
|
|
|
yield self.send_event_to_master(
|
|
|
|
event_id=event.event_id,
|
2018-07-23 17:28:00 +02:00
|
|
|
store=self.store,
|
2018-02-15 17:30:10 +01:00
|
|
|
requester=requester,
|
|
|
|
event=event,
|
|
|
|
context=context,
|
2018-03-01 11:08:28 +01:00
|
|
|
ratelimit=ratelimit,
|
|
|
|
extra_users=extra_users,
|
2018-02-15 17:30:10 +01:00
|
|
|
)
|
2018-10-01 19:48:51 +02:00
|
|
|
success = True
|
2018-02-15 17:30:10 +01:00
|
|
|
return
|
|
|
|
|
|
|
|
yield self.persist_and_notify_client_event(
|
|
|
|
requester,
|
|
|
|
event,
|
|
|
|
context,
|
|
|
|
ratelimit=ratelimit,
|
|
|
|
extra_users=extra_users,
|
2018-02-05 18:22:16 +01:00
|
|
|
)
|
2018-04-27 12:40:06 +02:00
|
|
|
|
2018-10-01 19:48:51 +02:00
|
|
|
success = True
|
|
|
|
finally:
|
|
|
|
if not success:
|
|
|
|
# Ensure that we actually remove the entries in the push actions
|
|
|
|
# staging area, if we calculated them.
|
|
|
|
run_in_background(
|
|
|
|
self.store.remove_push_actions_from_staging,
|
|
|
|
event.event_id,
|
|
|
|
)
|
2018-02-15 17:30:10 +01:00
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def persist_and_notify_client_event(
|
|
|
|
self,
|
|
|
|
requester,
|
|
|
|
event,
|
|
|
|
context,
|
|
|
|
ratelimit=True,
|
|
|
|
extra_users=[],
|
|
|
|
):
|
2018-03-01 11:18:33 +01:00
|
|
|
"""Called when we have fully built the event, have already
|
|
|
|
calculated the push actions for the event, and checked auth.
|
2018-03-01 11:05:27 +01:00
|
|
|
|
|
|
|
This should only be run on master.
|
2018-02-15 17:30:10 +01:00
|
|
|
"""
|
|
|
|
assert not self.config.worker_app
|
2018-02-05 18:22:16 +01:00
|
|
|
|
2016-05-11 10:09:20 +02:00
|
|
|
if ratelimit:
|
2018-01-15 17:51:53 +01:00
|
|
|
yield self.base_handler.ratelimit(requester)
|
2016-05-11 10:09:20 +02:00
|
|
|
|
2018-01-15 17:51:53 +01:00
|
|
|
yield self.base_handler.maybe_kick_guest_users(event, context)
|
2016-05-11 10:09:20 +02:00
|
|
|
|
|
|
|
if event.type == EventTypes.CanonicalAlias:
|
|
|
|
# Check the alias is acually valid (at this time at least)
|
|
|
|
room_alias_str = event.content.get("alias", None)
|
|
|
|
if room_alias_str:
|
|
|
|
room_alias = RoomAlias.from_string(room_alias_str)
|
|
|
|
directory_handler = self.hs.get_handlers().directory_handler
|
|
|
|
mapping = yield directory_handler.get_association(room_alias)
|
|
|
|
|
|
|
|
if mapping["room_id"] != event.room_id:
|
|
|
|
raise SynapseError(
|
|
|
|
400,
|
|
|
|
"Room alias %s does not point to the room" % (
|
|
|
|
room_alias_str,
|
|
|
|
)
|
|
|
|
)
|
|
|
|
|
|
|
|
federation_handler = self.hs.get_handlers().federation_handler
|
|
|
|
|
|
|
|
if event.type == EventTypes.Member:
|
|
|
|
if event.content["membership"] == Membership.INVITE:
|
|
|
|
def is_inviter_member_event(e):
|
|
|
|
return (
|
|
|
|
e.type == EventTypes.Member and
|
|
|
|
e.sender == event.sender
|
|
|
|
)
|
|
|
|
|
2018-07-23 14:00:22 +02:00
|
|
|
current_state_ids = yield context.get_current_state_ids(self.store)
|
|
|
|
|
2016-08-25 18:32:22 +02:00
|
|
|
state_to_include_ids = [
|
|
|
|
e_id
|
2018-07-23 14:00:22 +02:00
|
|
|
for k, e_id in iteritems(current_state_ids)
|
2016-08-25 18:32:22 +02:00
|
|
|
if k[0] in self.hs.config.room_invite_state_types
|
2017-04-26 17:18:08 +02:00
|
|
|
or k == (EventTypes.Member, event.sender)
|
2016-08-25 18:32:22 +02:00
|
|
|
]
|
|
|
|
|
|
|
|
state_to_include = yield self.store.get_events(state_to_include_ids)
|
|
|
|
|
2016-05-11 10:09:20 +02:00
|
|
|
event.unsigned["invite_room_state"] = [
|
|
|
|
{
|
|
|
|
"type": e.type,
|
|
|
|
"state_key": e.state_key,
|
|
|
|
"content": e.content,
|
|
|
|
"sender": e.sender,
|
|
|
|
}
|
2018-05-05 22:47:18 +02:00
|
|
|
for e in itervalues(state_to_include)
|
2016-05-11 10:09:20 +02:00
|
|
|
]
|
|
|
|
|
|
|
|
invitee = UserID.from_string(event.state_key)
|
|
|
|
if not self.hs.is_mine(invitee):
|
|
|
|
# TODO: Can we add signature from remote server in a nicer
|
|
|
|
# way? If we have been invited by a remote server, we need
|
|
|
|
# to get them to sign the event.
|
|
|
|
|
|
|
|
returned_invite = yield federation_handler.send_invite(
|
|
|
|
invitee.domain,
|
|
|
|
event,
|
|
|
|
)
|
|
|
|
|
|
|
|
event.unsigned.pop("room_state", None)
|
|
|
|
|
|
|
|
# TODO: Make sure the signatures actually are correct.
|
|
|
|
event.signatures.update(
|
|
|
|
returned_invite.signatures
|
|
|
|
)
|
|
|
|
|
|
|
|
if event.type == EventTypes.Redaction:
|
2018-07-23 14:00:22 +02:00
|
|
|
prev_state_ids = yield context.get_prev_state_ids(self.store)
|
2016-08-25 18:32:22 +02:00
|
|
|
auth_events_ids = yield self.auth.compute_auth_events(
|
2018-07-23 14:00:22 +02:00
|
|
|
event, prev_state_ids, for_verification=True,
|
2016-08-25 18:32:22 +02:00
|
|
|
)
|
|
|
|
auth_events = yield self.store.get_events(auth_events_ids)
|
|
|
|
auth_events = {
|
|
|
|
(e.type, e.state_key): e for e in auth_events.values()
|
|
|
|
}
|
|
|
|
if self.auth.check_redaction(event, auth_events=auth_events):
|
2016-05-11 10:09:20 +02:00
|
|
|
original_event = yield self.store.get_event(
|
|
|
|
event.redacts,
|
|
|
|
check_redacted=False,
|
|
|
|
get_prev_content=False,
|
|
|
|
allow_rejected=False,
|
|
|
|
allow_none=False
|
|
|
|
)
|
|
|
|
if event.user_id != original_event.user_id:
|
|
|
|
raise AuthError(
|
|
|
|
403,
|
|
|
|
"You don't have permission to redact events"
|
|
|
|
)
|
|
|
|
|
2018-07-23 14:00:22 +02:00
|
|
|
if event.type == EventTypes.Create:
|
|
|
|
prev_state_ids = yield context.get_prev_state_ids(self.store)
|
|
|
|
if prev_state_ids:
|
|
|
|
raise AuthError(
|
|
|
|
403,
|
|
|
|
"Changing the room create event is forbidden",
|
|
|
|
)
|
2016-05-11 10:09:20 +02:00
|
|
|
|
2018-02-15 17:30:10 +01:00
|
|
|
(event_stream_id, max_stream_id) = yield self.store.persist_event(
|
|
|
|
event, context=context
|
2016-05-11 10:09:20 +02:00
|
|
|
)
|
|
|
|
|
2018-08-17 01:32:39 +02:00
|
|
|
self.pusher_pool.on_new_notifications(
|
|
|
|
event_stream_id, max_stream_id,
|
2016-05-11 10:09:20 +02:00
|
|
|
)
|
|
|
|
|
2016-06-02 14:02:33 +02:00
|
|
|
def _notify():
|
2018-04-27 12:07:40 +02:00
|
|
|
try:
|
|
|
|
self.notifier.on_new_room_event(
|
|
|
|
event, event_stream_id, max_stream_id,
|
|
|
|
extra_users=extra_users
|
|
|
|
)
|
|
|
|
except Exception:
|
|
|
|
logger.exception("Error notifying about new room event")
|
2016-05-11 10:09:20 +02:00
|
|
|
|
2018-04-27 12:29:27 +02:00
|
|
|
run_in_background(_notify)
|
2018-02-06 18:27:00 +01:00
|
|
|
|
|
|
|
if event.type == EventTypes.Message:
|
|
|
|
# We don't want to block sending messages on any presence code. This
|
|
|
|
# matters as sometimes presence code can take a while.
|
2018-04-27 12:07:40 +02:00
|
|
|
run_in_background(self._bump_active_time, requester.user)
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def _bump_active_time(self, user):
|
|
|
|
try:
|
|
|
|
presence = self.hs.get_presence_handler()
|
|
|
|
yield presence.bump_presence_active_time(user)
|
|
|
|
except Exception:
|
|
|
|
logger.exception("Error bumping presence active time")
|