2016-07-15 14:19:07 +02:00
|
|
|
# -*- coding: utf-8 -*-
|
|
|
|
# Copyright 2016 OpenMarket Ltd
|
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
2018-07-09 08:09:20 +02:00
|
|
|
import logging
|
|
|
|
|
|
|
|
from six import iteritems, itervalues
|
|
|
|
|
|
|
|
from twisted.internet import defer
|
|
|
|
|
2016-07-20 18:58:44 +02:00
|
|
|
from synapse.api import errors
|
2017-02-01 18:33:16 +01:00
|
|
|
from synapse.api.constants import EventTypes
|
2018-01-22 19:11:18 +01:00
|
|
|
from synapse.api.errors import FederationDeniedError
|
2018-07-09 08:09:20 +02:00
|
|
|
from synapse.types import RoomStreamToken, get_domain_from_id
|
2016-07-15 14:19:07 +02:00
|
|
|
from synapse.util import stringutils
|
2017-01-26 17:06:54 +01:00
|
|
|
from synapse.util.async import Linearizer
|
2017-03-03 17:02:53 +01:00
|
|
|
from synapse.util.caches.expiringcache import ExpiringCache
|
2017-02-02 19:36:17 +01:00
|
|
|
from synapse.util.metrics import measure_func
|
2018-07-09 08:09:20 +02:00
|
|
|
from synapse.util.retryutils import NotRetryingDestination
|
2016-07-15 14:19:07 +02:00
|
|
|
|
2018-07-09 08:09:20 +02:00
|
|
|
from ._base import BaseHandler
|
2018-04-28 13:19:12 +02:00
|
|
|
|
2016-07-15 14:19:07 +02:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
|
|
|
class DeviceHandler(BaseHandler):
|
|
|
|
def __init__(self, hs):
|
|
|
|
super(DeviceHandler, self).__init__(hs)
|
|
|
|
|
2017-01-26 17:06:54 +01:00
|
|
|
self.hs = hs
|
2017-01-25 15:27:27 +01:00
|
|
|
self.state = hs.get_state_handler()
|
2017-11-01 11:29:34 +01:00
|
|
|
self._auth_handler = hs.get_auth_handler()
|
2017-01-26 17:06:54 +01:00
|
|
|
self.federation_sender = hs.get_federation_sender()
|
2017-03-01 15:12:11 +01:00
|
|
|
|
|
|
|
self._edu_updater = DeviceListEduUpdater(hs, self)
|
2017-01-26 17:06:54 +01:00
|
|
|
|
2018-03-12 17:17:08 +01:00
|
|
|
federation_registry = hs.get_federation_registry()
|
|
|
|
|
|
|
|
federation_registry.register_edu_handler(
|
2017-03-01 15:12:11 +01:00
|
|
|
"m.device_list_update", self._edu_updater.incoming_device_list_update,
|
2017-01-26 17:06:54 +01:00
|
|
|
)
|
2018-03-12 17:17:08 +01:00
|
|
|
federation_registry.register_query_handler(
|
2017-01-26 17:06:54 +01:00
|
|
|
"user_devices", self.on_federation_query_user_devices,
|
|
|
|
)
|
2017-01-25 15:27:27 +01:00
|
|
|
|
2017-01-26 17:39:33 +01:00
|
|
|
hs.get_distributor().observe("user_left_room", self.user_left_room)
|
|
|
|
|
2016-07-15 14:19:07 +02:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def check_device_registered(self, user_id, device_id,
|
2016-08-03 15:30:06 +02:00
|
|
|
initial_device_display_name=None):
|
2016-07-15 14:19:07 +02:00
|
|
|
"""
|
|
|
|
If the given device has not been registered, register it with the
|
|
|
|
supplied display name.
|
|
|
|
|
|
|
|
If no device_id is supplied, we make one up.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
user_id (str): @user:id
|
|
|
|
device_id (str | None): device id supplied by client
|
|
|
|
initial_device_display_name (str | None): device display name from
|
|
|
|
client
|
|
|
|
Returns:
|
|
|
|
str: device id (generated if none was supplied)
|
|
|
|
"""
|
|
|
|
if device_id is not None:
|
2017-01-25 15:27:27 +01:00
|
|
|
new_device = yield self.store.store_device(
|
2016-07-15 14:19:07 +02:00
|
|
|
user_id=user_id,
|
|
|
|
device_id=device_id,
|
|
|
|
initial_device_display_name=initial_device_display_name,
|
|
|
|
)
|
2017-01-25 15:27:27 +01:00
|
|
|
if new_device:
|
2017-01-26 17:06:54 +01:00
|
|
|
yield self.notify_device_update(user_id, [device_id])
|
2016-07-15 14:19:07 +02:00
|
|
|
defer.returnValue(device_id)
|
|
|
|
|
|
|
|
# if the device id is not specified, we'll autogen one, but loop a few
|
|
|
|
# times in case of a clash.
|
|
|
|
attempts = 0
|
|
|
|
while attempts < 5:
|
2017-01-25 15:27:27 +01:00
|
|
|
device_id = stringutils.random_string(10).upper()
|
|
|
|
new_device = yield self.store.store_device(
|
|
|
|
user_id=user_id,
|
|
|
|
device_id=device_id,
|
|
|
|
initial_device_display_name=initial_device_display_name,
|
|
|
|
)
|
|
|
|
if new_device:
|
2017-01-26 17:06:54 +01:00
|
|
|
yield self.notify_device_update(user_id, [device_id])
|
2016-07-15 14:19:07 +02:00
|
|
|
defer.returnValue(device_id)
|
2017-01-25 15:27:27 +01:00
|
|
|
attempts += 1
|
2016-07-15 14:19:07 +02:00
|
|
|
|
2016-07-20 18:58:44 +02:00
|
|
|
raise errors.StoreError(500, "Couldn't generate a device ID.")
|
2016-07-20 17:34:00 +02:00
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def get_devices_by_user(self, user_id):
|
|
|
|
"""
|
|
|
|
Retrieve the given user's devices
|
|
|
|
|
|
|
|
Args:
|
|
|
|
user_id (str):
|
|
|
|
Returns:
|
2016-07-21 16:55:13 +02:00
|
|
|
defer.Deferred: list[dict[str, X]]: info on each device
|
2016-07-20 17:34:00 +02:00
|
|
|
"""
|
|
|
|
|
2016-07-21 16:55:13 +02:00
|
|
|
device_map = yield self.store.get_devices_by_user(user_id)
|
2016-07-20 17:34:00 +02:00
|
|
|
|
|
|
|
ips = yield self.store.get_last_client_ip_by_device(
|
2017-06-27 15:46:12 +02:00
|
|
|
user_id, device_id=None
|
2016-07-20 17:34:00 +02:00
|
|
|
)
|
|
|
|
|
2018-05-31 11:03:47 +02:00
|
|
|
devices = list(device_map.values())
|
2016-07-21 16:55:13 +02:00
|
|
|
for device in devices:
|
2016-07-20 18:58:44 +02:00
|
|
|
_update_device_from_client_ips(device, ips)
|
2016-07-20 17:34:00 +02:00
|
|
|
|
|
|
|
defer.returnValue(devices)
|
2016-07-20 18:58:44 +02:00
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def get_device(self, user_id, device_id):
|
|
|
|
""" Retrieve the given device
|
|
|
|
|
|
|
|
Args:
|
|
|
|
user_id (str):
|
2016-07-22 15:52:53 +02:00
|
|
|
device_id (str):
|
2016-07-20 18:58:44 +02:00
|
|
|
|
|
|
|
Returns:
|
|
|
|
defer.Deferred: dict[str, X]: info on the device
|
|
|
|
Raises:
|
|
|
|
errors.NotFoundError: if the device was not found
|
|
|
|
"""
|
|
|
|
try:
|
|
|
|
device = yield self.store.get_device(user_id, device_id)
|
2016-07-21 14:15:15 +02:00
|
|
|
except errors.StoreError:
|
2016-07-20 18:58:44 +02:00
|
|
|
raise errors.NotFoundError
|
|
|
|
ips = yield self.store.get_last_client_ip_by_device(
|
2017-06-27 15:46:12 +02:00
|
|
|
user_id, device_id,
|
2016-07-20 18:58:44 +02:00
|
|
|
)
|
|
|
|
_update_device_from_client_ips(device, ips)
|
|
|
|
defer.returnValue(device)
|
|
|
|
|
2016-07-22 15:52:53 +02:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def delete_device(self, user_id, device_id):
|
|
|
|
""" Delete the given device
|
|
|
|
|
|
|
|
Args:
|
|
|
|
user_id (str):
|
|
|
|
device_id (str):
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
defer.Deferred:
|
|
|
|
"""
|
|
|
|
|
|
|
|
try:
|
|
|
|
yield self.store.delete_device(user_id, device_id)
|
2018-03-15 16:11:17 +01:00
|
|
|
except errors.StoreError as e:
|
2016-07-22 15:52:53 +02:00
|
|
|
if e.code == 404:
|
|
|
|
# no match
|
|
|
|
pass
|
|
|
|
else:
|
|
|
|
raise
|
|
|
|
|
2017-11-01 11:29:34 +01:00
|
|
|
yield self._auth_handler.delete_access_tokens_for_user(
|
2016-07-26 12:09:47 +02:00
|
|
|
user_id, device_id=device_id,
|
|
|
|
)
|
2016-07-22 15:52:53 +02:00
|
|
|
|
2016-07-27 13:18:03 +02:00
|
|
|
yield self.store.delete_e2e_keys_by_device(
|
|
|
|
user_id=user_id, device_id=device_id
|
|
|
|
)
|
|
|
|
|
2017-01-26 17:06:54 +01:00
|
|
|
yield self.notify_device_update(user_id, [device_id])
|
2017-01-25 15:27:27 +01:00
|
|
|
|
2017-11-29 16:44:59 +01:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def delete_all_devices_for_user(self, user_id, except_device_id=None):
|
|
|
|
"""Delete all of the user's devices
|
|
|
|
|
|
|
|
Args:
|
|
|
|
user_id (str):
|
|
|
|
except_device_id (str|None): optional device id which should not
|
|
|
|
be deleted
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
defer.Deferred:
|
|
|
|
"""
|
|
|
|
device_map = yield self.store.get_devices_by_user(user_id)
|
2018-05-31 11:03:47 +02:00
|
|
|
device_ids = list(device_map)
|
2017-11-29 16:44:59 +01:00
|
|
|
if except_device_id is not None:
|
|
|
|
device_ids = [d for d in device_ids if d != except_device_id]
|
|
|
|
yield self.delete_devices(user_id, device_ids)
|
|
|
|
|
2017-03-13 18:53:23 +01:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def delete_devices(self, user_id, device_ids):
|
|
|
|
""" Delete several devices
|
|
|
|
|
|
|
|
Args:
|
|
|
|
user_id (str):
|
2017-11-29 16:44:59 +01:00
|
|
|
device_ids (List[str]): The list of device IDs to delete
|
2017-03-13 18:53:23 +01:00
|
|
|
|
|
|
|
Returns:
|
|
|
|
defer.Deferred:
|
|
|
|
"""
|
|
|
|
|
|
|
|
try:
|
|
|
|
yield self.store.delete_devices(user_id, device_ids)
|
2018-03-15 16:11:17 +01:00
|
|
|
except errors.StoreError as e:
|
2017-03-13 18:53:23 +01:00
|
|
|
if e.code == 404:
|
|
|
|
# no match
|
|
|
|
pass
|
|
|
|
else:
|
|
|
|
raise
|
|
|
|
|
|
|
|
# Delete access tokens and e2e keys for each device. Not optimised as it is not
|
|
|
|
# considered as part of a critical path.
|
|
|
|
for device_id in device_ids:
|
2017-11-01 11:29:34 +01:00
|
|
|
yield self._auth_handler.delete_access_tokens_for_user(
|
2017-03-13 18:53:23 +01:00
|
|
|
user_id, device_id=device_id,
|
|
|
|
)
|
|
|
|
yield self.store.delete_e2e_keys_by_device(
|
|
|
|
user_id=user_id, device_id=device_id
|
|
|
|
)
|
|
|
|
|
|
|
|
yield self.notify_device_update(user_id, device_ids)
|
|
|
|
|
2016-07-25 18:51:24 +02:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def update_device(self, user_id, device_id, content):
|
|
|
|
""" Update the given device
|
|
|
|
|
|
|
|
Args:
|
|
|
|
user_id (str):
|
|
|
|
device_id (str):
|
|
|
|
content (dict): body of update request
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
defer.Deferred:
|
|
|
|
"""
|
|
|
|
|
|
|
|
try:
|
|
|
|
yield self.store.update_device(
|
|
|
|
user_id,
|
|
|
|
device_id,
|
|
|
|
new_display_name=content.get("display_name")
|
|
|
|
)
|
2017-01-26 17:06:54 +01:00
|
|
|
yield self.notify_device_update(user_id, [device_id])
|
2018-03-15 16:11:17 +01:00
|
|
|
except errors.StoreError as e:
|
2016-07-25 18:51:24 +02:00
|
|
|
if e.code == 404:
|
|
|
|
raise errors.NotFoundError()
|
|
|
|
else:
|
|
|
|
raise
|
2016-07-22 15:52:53 +02:00
|
|
|
|
2017-02-02 19:36:17 +01:00
|
|
|
@measure_func("notify_device_update")
|
2017-01-25 15:27:27 +01:00
|
|
|
@defer.inlineCallbacks
|
2017-01-26 17:06:54 +01:00
|
|
|
def notify_device_update(self, user_id, device_ids):
|
2017-01-26 17:30:37 +01:00
|
|
|
"""Notify that a user's device(s) has changed. Pokes the notifier, and
|
|
|
|
remote servers if the user is local.
|
|
|
|
"""
|
2017-02-02 16:25:00 +01:00
|
|
|
users_who_share_room = yield self.store.get_users_who_share_room_with_user(
|
|
|
|
user_id
|
|
|
|
)
|
2017-01-25 15:27:27 +01:00
|
|
|
|
|
|
|
hosts = set()
|
2017-01-26 17:06:54 +01:00
|
|
|
if self.hs.is_mine_id(user_id):
|
2017-02-02 16:25:00 +01:00
|
|
|
hosts.update(get_domain_from_id(u) for u in users_who_share_room)
|
2017-01-26 17:06:54 +01:00
|
|
|
hosts.discard(self.server_name)
|
2017-01-25 15:27:27 +01:00
|
|
|
|
|
|
|
position = yield self.store.add_device_change_to_streams(
|
2017-01-26 17:06:54 +01:00
|
|
|
user_id, device_ids, list(hosts)
|
2017-01-25 15:27:27 +01:00
|
|
|
)
|
|
|
|
|
2017-03-16 12:51:46 +01:00
|
|
|
room_ids = yield self.store.get_rooms_for_user(user_id)
|
2017-02-02 16:25:00 +01:00
|
|
|
|
2017-01-25 15:27:27 +01:00
|
|
|
yield self.notifier.on_new_event(
|
|
|
|
"device_list_key", position, rooms=room_ids,
|
|
|
|
)
|
|
|
|
|
2017-01-26 17:30:37 +01:00
|
|
|
if hosts:
|
|
|
|
logger.info("Sending device list update notif to: %r", hosts)
|
|
|
|
for host in hosts:
|
|
|
|
self.federation_sender.send_device_messages(host)
|
2017-01-25 15:27:27 +01:00
|
|
|
|
2017-02-02 19:36:17 +01:00
|
|
|
@measure_func("device.get_user_ids_changed")
|
2017-02-01 11:30:03 +01:00
|
|
|
@defer.inlineCallbacks
|
2017-02-01 18:33:16 +01:00
|
|
|
def get_user_ids_changed(self, user_id, from_token):
|
2017-02-01 18:52:57 +01:00
|
|
|
"""Get list of users that have had the devices updated, or have newly
|
|
|
|
joined a room, that `user_id` may be interested in.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
user_id (str)
|
|
|
|
from_token (StreamToken)
|
|
|
|
"""
|
2017-09-13 17:28:08 +02:00
|
|
|
now_token = yield self.hs.get_event_sources().get_current_token()
|
|
|
|
|
2017-03-16 12:51:46 +01:00
|
|
|
room_ids = yield self.store.get_rooms_for_user(user_id)
|
2017-02-01 11:30:03 +01:00
|
|
|
|
2017-02-01 18:33:16 +01:00
|
|
|
# First we check if any devices have changed
|
2017-02-01 11:30:03 +01:00
|
|
|
changed = yield self.store.get_user_whose_devices_changed(
|
2017-02-01 18:33:16 +01:00
|
|
|
from_token.device_list_key
|
2017-02-01 11:30:03 +01:00
|
|
|
)
|
2017-02-01 18:33:16 +01:00
|
|
|
|
|
|
|
# Then work out if any users have since joined
|
|
|
|
rooms_changed = self.store.get_rooms_that_changed(room_ids, from_token.room_key)
|
|
|
|
|
2017-09-13 17:28:08 +02:00
|
|
|
member_events = yield self.store.get_membership_changes_for_user(
|
|
|
|
user_id, from_token.room_key, now_token.room_key
|
|
|
|
)
|
|
|
|
rooms_changed.update(event.room_id for event in member_events)
|
|
|
|
|
2017-02-14 14:59:50 +01:00
|
|
|
stream_ordering = RoomStreamToken.parse_stream_token(
|
2017-09-13 17:28:08 +02:00
|
|
|
from_token.room_key
|
|
|
|
).stream
|
2017-02-14 14:59:50 +01:00
|
|
|
|
2017-02-01 18:33:16 +01:00
|
|
|
possibly_changed = set(changed)
|
2017-09-18 16:38:22 +02:00
|
|
|
possibly_left = set()
|
2017-02-01 18:33:16 +01:00
|
|
|
for room_id in rooms_changed:
|
2017-09-18 16:38:22 +02:00
|
|
|
current_state_ids = yield self.store.get_current_state_ids(room_id)
|
|
|
|
|
2017-09-13 17:28:08 +02:00
|
|
|
# The user may have left the room
|
|
|
|
# TODO: Check if they actually did or if we were just invited.
|
|
|
|
if room_id not in room_ids:
|
2018-04-28 13:19:12 +02:00
|
|
|
for key, event_id in iteritems(current_state_ids):
|
2017-09-18 16:38:22 +02:00
|
|
|
etype, state_key = key
|
|
|
|
if etype != EventTypes.Member:
|
|
|
|
continue
|
|
|
|
possibly_left.add(state_key)
|
2017-09-13 17:28:08 +02:00
|
|
|
continue
|
|
|
|
|
2017-02-14 14:59:50 +01:00
|
|
|
# Fetch the current state at the time.
|
2017-02-02 15:27:19 +01:00
|
|
|
try:
|
|
|
|
event_ids = yield self.store.get_forward_extremeties_for_room(
|
|
|
|
room_id, stream_ordering=stream_ordering
|
|
|
|
)
|
2017-02-14 14:59:50 +01:00
|
|
|
except errors.StoreError:
|
|
|
|
# we have purged the stream_ordering index since the stream
|
|
|
|
# ordering: treat it the same as a new room
|
|
|
|
event_ids = []
|
2017-02-01 18:33:16 +01:00
|
|
|
|
2017-02-14 14:59:50 +01:00
|
|
|
# special-case for an empty prev state: include all members
|
|
|
|
# in the changed list
|
|
|
|
if not event_ids:
|
2018-04-28 13:19:12 +02:00
|
|
|
for key, event_id in iteritems(current_state_ids):
|
2017-02-14 14:59:50 +01:00
|
|
|
etype, state_key = key
|
|
|
|
if etype != EventTypes.Member:
|
|
|
|
continue
|
|
|
|
possibly_changed.add(state_key)
|
|
|
|
continue
|
|
|
|
|
2017-09-13 17:28:08 +02:00
|
|
|
current_member_id = current_state_ids.get((EventTypes.Member, user_id))
|
|
|
|
if not current_member_id:
|
|
|
|
continue
|
|
|
|
|
2017-02-14 14:59:50 +01:00
|
|
|
# mapping from event_id -> state_dict
|
|
|
|
prev_state_ids = yield self.store.get_state_ids_for_events(event_ids)
|
|
|
|
|
2017-09-13 17:28:08 +02:00
|
|
|
# Check if we've joined the room? If so we just blindly add all the users to
|
|
|
|
# the "possibly changed" users.
|
2018-04-28 13:19:12 +02:00
|
|
|
for state_dict in itervalues(prev_state_ids):
|
2017-09-13 17:28:08 +02:00
|
|
|
member_event = state_dict.get((EventTypes.Member, user_id), None)
|
|
|
|
if not member_event or member_event != current_member_id:
|
2018-04-28 13:19:12 +02:00
|
|
|
for key, event_id in iteritems(current_state_ids):
|
2017-09-13 17:28:08 +02:00
|
|
|
etype, state_key = key
|
|
|
|
if etype != EventTypes.Member:
|
|
|
|
continue
|
2017-09-15 10:56:54 +02:00
|
|
|
possibly_changed.add(state_key)
|
2017-09-13 17:28:08 +02:00
|
|
|
break
|
|
|
|
|
2017-02-01 18:33:16 +01:00
|
|
|
# If there has been any change in membership, include them in the
|
|
|
|
# possibly changed list. We'll check if they are joined below,
|
|
|
|
# and we're not toooo worried about spuriously adding users.
|
2018-04-28 13:19:12 +02:00
|
|
|
for key, event_id in iteritems(current_state_ids):
|
2017-02-01 18:33:16 +01:00
|
|
|
etype, state_key = key
|
2017-02-14 14:59:50 +01:00
|
|
|
if etype != EventTypes.Member:
|
|
|
|
continue
|
|
|
|
|
|
|
|
# check if this member has changed since any of the extremities
|
|
|
|
# at the stream_ordering, and add them to the list if so.
|
2018-04-28 13:19:12 +02:00
|
|
|
for state_dict in itervalues(prev_state_ids):
|
2017-02-14 14:59:50 +01:00
|
|
|
prev_event_id = state_dict.get(key, None)
|
2017-02-01 18:33:16 +01:00
|
|
|
if not prev_event_id or prev_event_id != event_id:
|
2017-09-18 16:38:22 +02:00
|
|
|
if state_key != user_id:
|
|
|
|
possibly_changed.add(state_key)
|
2017-02-14 14:59:50 +01:00
|
|
|
break
|
2017-02-01 18:33:16 +01:00
|
|
|
|
2017-09-18 16:38:22 +02:00
|
|
|
if possibly_changed or possibly_left:
|
2017-09-08 15:44:36 +02:00
|
|
|
users_who_share_room = yield self.store.get_users_who_share_room_with_user(
|
|
|
|
user_id
|
|
|
|
)
|
2017-02-01 11:30:03 +01:00
|
|
|
|
2017-09-08 15:44:36 +02:00
|
|
|
# Take the intersection of the users whose devices may have changed
|
|
|
|
# and those that actually still share a room with the user
|
|
|
|
possibly_joined = possibly_changed & users_who_share_room
|
2017-09-18 16:38:22 +02:00
|
|
|
possibly_left = (possibly_changed | possibly_left) - users_who_share_room
|
2017-09-08 15:44:36 +02:00
|
|
|
else:
|
|
|
|
possibly_joined = []
|
|
|
|
possibly_left = []
|
|
|
|
|
|
|
|
defer.returnValue({
|
|
|
|
"changed": list(possibly_joined),
|
|
|
|
"left": list(possibly_left),
|
|
|
|
})
|
2017-02-01 11:30:03 +01:00
|
|
|
|
2017-01-26 17:06:54 +01:00
|
|
|
@defer.inlineCallbacks
|
2017-03-01 15:12:11 +01:00
|
|
|
def on_federation_query_user_devices(self, user_id):
|
|
|
|
stream_id, devices = yield self.store.get_devices_with_keys_by_user(user_id)
|
|
|
|
defer.returnValue({
|
|
|
|
"user_id": user_id,
|
|
|
|
"stream_id": stream_id,
|
|
|
|
"devices": devices,
|
|
|
|
})
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def user_left_room(self, user, room_id):
|
|
|
|
user_id = user.to_string()
|
2017-03-16 12:51:46 +01:00
|
|
|
room_ids = yield self.store.get_rooms_for_user(user_id)
|
|
|
|
if not room_ids:
|
2017-03-01 15:12:11 +01:00
|
|
|
# We no longer share rooms with this user, so we'll no longer
|
|
|
|
# receive device updates. Mark this in DB.
|
|
|
|
yield self.store.mark_remote_user_device_list_as_unsubscribed(user_id)
|
|
|
|
|
|
|
|
|
|
|
|
def _update_device_from_client_ips(device, client_ips):
|
|
|
|
ip = client_ips.get((device["user_id"], device["device_id"]), {})
|
|
|
|
device.update({
|
|
|
|
"last_seen_ts": ip.get("last_seen"),
|
|
|
|
"last_seen_ip": ip.get("ip"),
|
|
|
|
})
|
|
|
|
|
|
|
|
|
|
|
|
class DeviceListEduUpdater(object):
|
|
|
|
"Handles incoming device list updates from federation and updates the DB"
|
|
|
|
|
|
|
|
def __init__(self, hs, device_handler):
|
|
|
|
self.store = hs.get_datastore()
|
2018-03-13 14:26:52 +01:00
|
|
|
self.federation = hs.get_federation_client()
|
2017-03-01 15:12:11 +01:00
|
|
|
self.clock = hs.get_clock()
|
|
|
|
self.device_handler = device_handler
|
|
|
|
|
2017-03-03 16:31:57 +01:00
|
|
|
self._remote_edu_linearizer = Linearizer(name="remote_device_list")
|
2017-03-01 15:12:11 +01:00
|
|
|
|
|
|
|
# user_id -> list of updates waiting to be handled.
|
|
|
|
self._pending_updates = {}
|
|
|
|
|
|
|
|
# Recently seen stream ids. We don't bother keeping these in the DB,
|
|
|
|
# but they're useful to have them about to reduce the number of spurious
|
|
|
|
# resyncs.
|
2017-03-03 17:02:53 +01:00
|
|
|
self._seen_updates = ExpiringCache(
|
|
|
|
cache_name="device_update_edu",
|
|
|
|
clock=self.clock,
|
|
|
|
max_len=10000,
|
|
|
|
expiry_ms=30 * 60 * 1000,
|
|
|
|
iterable=True,
|
|
|
|
)
|
2017-03-01 15:12:11 +01:00
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def incoming_device_list_update(self, origin, edu_content):
|
|
|
|
"""Called on incoming device list update from federation. Responsible
|
|
|
|
for parsing the EDU and adding to pending updates list.
|
|
|
|
"""
|
|
|
|
|
|
|
|
user_id = edu_content.pop("user_id")
|
|
|
|
device_id = edu_content.pop("device_id")
|
|
|
|
stream_id = str(edu_content.pop("stream_id")) # They may come as ints
|
|
|
|
prev_ids = edu_content.pop("prev_id", [])
|
|
|
|
prev_ids = [str(p) for p in prev_ids] # They may come as ints
|
2017-01-26 17:06:54 +01:00
|
|
|
|
|
|
|
if get_domain_from_id(user_id) != origin:
|
|
|
|
# TODO: Raise?
|
2017-01-26 17:30:37 +01:00
|
|
|
logger.warning("Got device list update edu for %r from %r", user_id, origin)
|
2017-01-26 17:06:54 +01:00
|
|
|
return
|
|
|
|
|
2017-03-16 12:51:46 +01:00
|
|
|
room_ids = yield self.store.get_rooms_for_user(user_id)
|
|
|
|
if not room_ids:
|
2017-01-26 17:39:33 +01:00
|
|
|
# We don't share any rooms with this user. Ignore update, as we
|
|
|
|
# probably won't get any further updates.
|
|
|
|
return
|
2017-01-26 17:06:54 +01:00
|
|
|
|
2017-03-01 15:12:11 +01:00
|
|
|
self._pending_updates.setdefault(user_id, []).append(
|
|
|
|
(device_id, stream_id, prev_ids, edu_content)
|
|
|
|
)
|
|
|
|
|
|
|
|
yield self._handle_device_updates(user_id)
|
|
|
|
|
|
|
|
@measure_func("_incoming_device_list_update")
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def _handle_device_updates(self, user_id):
|
|
|
|
"Actually handle pending updates."
|
|
|
|
|
2017-03-03 16:31:57 +01:00
|
|
|
with (yield self._remote_edu_linearizer.queue(user_id)):
|
2017-03-01 15:12:11 +01:00
|
|
|
pending_updates = self._pending_updates.pop(user_id, [])
|
|
|
|
if not pending_updates:
|
|
|
|
# This can happen since we batch updates
|
|
|
|
return
|
|
|
|
|
2017-05-05 11:59:32 +02:00
|
|
|
# Given a list of updates we check if we need to resync. This
|
|
|
|
# happens if we've missed updates.
|
2017-03-01 15:12:11 +01:00
|
|
|
resync = yield self._need_to_do_resync(user_id, pending_updates)
|
2017-01-26 17:06:54 +01:00
|
|
|
|
|
|
|
if resync:
|
2017-01-26 17:30:37 +01:00
|
|
|
# Fetch all devices for the user.
|
2017-03-01 15:12:11 +01:00
|
|
|
origin = get_domain_from_id(user_id)
|
2017-05-05 11:34:10 +02:00
|
|
|
try:
|
|
|
|
result = yield self.federation.query_user_devices(origin, user_id)
|
|
|
|
except NotRetryingDestination:
|
2017-05-05 11:59:32 +02:00
|
|
|
# TODO: Remember that we are now out of sync and try again
|
|
|
|
# later
|
2017-05-05 11:34:10 +02:00
|
|
|
logger.warn(
|
|
|
|
"Failed to handle device list update for %s,"
|
|
|
|
" we're not retrying the remote",
|
|
|
|
user_id,
|
|
|
|
)
|
2017-05-05 15:01:17 +02:00
|
|
|
# We abort on exceptions rather than accepting the update
|
|
|
|
# as otherwise synapse will 'forget' that its device list
|
|
|
|
# is out of date. If we bail then we will retry the resync
|
|
|
|
# next time we get a device list update for this user_id.
|
|
|
|
# This makes it more likely that the device lists will
|
|
|
|
# eventually become consistent.
|
2017-05-05 11:34:10 +02:00
|
|
|
return
|
2018-01-22 19:11:18 +01:00
|
|
|
except FederationDeniedError as e:
|
|
|
|
logger.info(e)
|
|
|
|
return
|
2017-05-05 11:34:10 +02:00
|
|
|
except Exception:
|
2017-05-05 11:59:32 +02:00
|
|
|
# TODO: Remember that we are now out of sync and try again
|
|
|
|
# later
|
2017-05-05 11:34:10 +02:00
|
|
|
logger.exception(
|
|
|
|
"Failed to handle device list update for %s", user_id
|
|
|
|
)
|
|
|
|
return
|
|
|
|
|
2017-01-26 17:06:54 +01:00
|
|
|
stream_id = result["stream_id"]
|
|
|
|
devices = result["devices"]
|
|
|
|
yield self.store.update_remote_device_list_cache(
|
|
|
|
user_id, devices, stream_id,
|
|
|
|
)
|
|
|
|
device_ids = [device["device_id"] for device in devices]
|
2017-03-01 15:12:11 +01:00
|
|
|
yield self.device_handler.notify_device_update(user_id, device_ids)
|
2017-01-26 17:06:54 +01:00
|
|
|
else:
|
2017-01-26 17:30:37 +01:00
|
|
|
# Simply update the single device, since we know that is the only
|
2018-07-10 18:58:09 +02:00
|
|
|
# change (because of the single prev_id matching the current cache)
|
2017-03-01 15:12:11 +01:00
|
|
|
for device_id, stream_id, prev_ids, content in pending_updates:
|
|
|
|
yield self.store.update_remote_device_list_cache_entry(
|
|
|
|
user_id, device_id, content, stream_id,
|
|
|
|
)
|
|
|
|
|
|
|
|
yield self.device_handler.notify_device_update(
|
|
|
|
user_id, [device_id for device_id, _, _, _ in pending_updates]
|
2017-01-26 17:06:54 +01:00
|
|
|
)
|
|
|
|
|
2017-03-01 15:12:11 +01:00
|
|
|
self._seen_updates.setdefault(user_id, set()).update(
|
2017-03-03 17:02:53 +01:00
|
|
|
stream_id for _, stream_id, _, _ in pending_updates
|
2017-03-01 15:12:11 +01:00
|
|
|
)
|
2017-01-26 17:06:54 +01:00
|
|
|
|
2017-01-26 17:39:33 +01:00
|
|
|
@defer.inlineCallbacks
|
2017-03-01 15:12:11 +01:00
|
|
|
def _need_to_do_resync(self, user_id, updates):
|
|
|
|
"""Given a list of updates for a user figure out if we need to do a full
|
|
|
|
resync, or whether we have enough data that we can just apply the delta.
|
|
|
|
"""
|
|
|
|
seen_updates = self._seen_updates.get(user_id, set())
|
2017-01-26 17:39:33 +01:00
|
|
|
|
2017-03-01 15:12:11 +01:00
|
|
|
extremity = yield self.store.get_device_list_last_stream_id_for_remote(
|
|
|
|
user_id
|
|
|
|
)
|
2016-07-20 18:58:44 +02:00
|
|
|
|
2017-03-01 15:12:11 +01:00
|
|
|
stream_id_in_updates = set() # stream_ids in updates list
|
|
|
|
for _, stream_id, prev_ids, _ in updates:
|
|
|
|
if not prev_ids:
|
|
|
|
# We always do a resync if there are no previous IDs
|
|
|
|
defer.returnValue(True)
|
|
|
|
|
|
|
|
for prev_id in prev_ids:
|
|
|
|
if prev_id == extremity:
|
|
|
|
continue
|
|
|
|
elif prev_id in seen_updates:
|
|
|
|
continue
|
|
|
|
elif prev_id in stream_id_in_updates:
|
|
|
|
continue
|
|
|
|
else:
|
|
|
|
defer.returnValue(True)
|
|
|
|
|
|
|
|
stream_id_in_updates.add(stream_id)
|
|
|
|
|
|
|
|
defer.returnValue(False)
|