1032 lines
		
	
	
		
			40 KiB
		
	
	
	
		
			Python
		
	
	
			
		
		
	
	
			1032 lines
		
	
	
		
			40 KiB
		
	
	
	
		
			Python
		
	
	
#!/usr/bin/env python
 | 
						|
# -*- coding: utf-8 -*-
 | 
						|
# Copyright 2016 OpenMarket Ltd
 | 
						|
# Copyright 2020 The Matrix.org Foundation C.I.C.
 | 
						|
#
 | 
						|
# Licensed under the Apache License, Version 2.0 (the "License");
 | 
						|
# you may not use this file except in compliance with the License.
 | 
						|
# You may obtain a copy of the License at
 | 
						|
#
 | 
						|
#     http://www.apache.org/licenses/LICENSE-2.0
 | 
						|
#
 | 
						|
# Unless required by applicable law or agreed to in writing, software
 | 
						|
# distributed under the License is distributed on an "AS IS" BASIS,
 | 
						|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | 
						|
# See the License for the specific language governing permissions and
 | 
						|
# limitations under the License.
 | 
						|
import contextlib
 | 
						|
import logging
 | 
						|
import sys
 | 
						|
from typing import Dict, Iterable, Optional, Set
 | 
						|
 | 
						|
from typing_extensions import ContextManager
 | 
						|
 | 
						|
from twisted.internet import defer, reactor
 | 
						|
 | 
						|
import synapse
 | 
						|
import synapse.events
 | 
						|
from synapse.api.errors import HttpResponseException, RequestSendFailed, SynapseError
 | 
						|
from synapse.api.urls import (
 | 
						|
    CLIENT_API_PREFIX,
 | 
						|
    FEDERATION_PREFIX,
 | 
						|
    LEGACY_MEDIA_PREFIX,
 | 
						|
    MEDIA_PREFIX,
 | 
						|
    SERVER_KEY_V2_PREFIX,
 | 
						|
)
 | 
						|
from synapse.app import _base
 | 
						|
from synapse.config._base import ConfigError
 | 
						|
from synapse.config.homeserver import HomeServerConfig
 | 
						|
from synapse.config.logger import setup_logging
 | 
						|
from synapse.config.server import ListenerConfig
 | 
						|
from synapse.federation import send_queue
 | 
						|
from synapse.federation.transport.server import TransportLayerServer
 | 
						|
from synapse.handlers.presence import (
 | 
						|
    BasePresenceHandler,
 | 
						|
    PresenceState,
 | 
						|
    get_interested_parties,
 | 
						|
)
 | 
						|
from synapse.http.server import JsonResource, OptionsResource
 | 
						|
from synapse.http.servlet import RestServlet, parse_json_object_from_request
 | 
						|
from synapse.http.site import SynapseSite
 | 
						|
from synapse.logging.context import LoggingContext
 | 
						|
from synapse.metrics import METRICS_PREFIX, MetricsResource, RegistryProxy
 | 
						|
from synapse.metrics.background_process_metrics import run_as_background_process
 | 
						|
from synapse.replication.http import REPLICATION_PREFIX, ReplicationRestResource
 | 
						|
from synapse.replication.http.presence import (
 | 
						|
    ReplicationBumpPresenceActiveTime,
 | 
						|
    ReplicationPresenceSetState,
 | 
						|
)
 | 
						|
from synapse.replication.slave.storage._base import BaseSlavedStore
 | 
						|
from synapse.replication.slave.storage.account_data import SlavedAccountDataStore
 | 
						|
from synapse.replication.slave.storage.appservice import SlavedApplicationServiceStore
 | 
						|
from synapse.replication.slave.storage.client_ips import SlavedClientIpStore
 | 
						|
from synapse.replication.slave.storage.deviceinbox import SlavedDeviceInboxStore
 | 
						|
from synapse.replication.slave.storage.devices import SlavedDeviceStore
 | 
						|
from synapse.replication.slave.storage.directory import DirectoryStore
 | 
						|
from synapse.replication.slave.storage.events import SlavedEventStore
 | 
						|
from synapse.replication.slave.storage.filtering import SlavedFilteringStore
 | 
						|
from synapse.replication.slave.storage.groups import SlavedGroupServerStore
 | 
						|
from synapse.replication.slave.storage.keys import SlavedKeyStore
 | 
						|
from synapse.replication.slave.storage.presence import SlavedPresenceStore
 | 
						|
from synapse.replication.slave.storage.profile import SlavedProfileStore
 | 
						|
from synapse.replication.slave.storage.push_rule import SlavedPushRuleStore
 | 
						|
from synapse.replication.slave.storage.pushers import SlavedPusherStore
 | 
						|
from synapse.replication.slave.storage.receipts import SlavedReceiptsStore
 | 
						|
from synapse.replication.slave.storage.registration import SlavedRegistrationStore
 | 
						|
from synapse.replication.slave.storage.room import RoomStore
 | 
						|
from synapse.replication.slave.storage.transactions import SlavedTransactionStore
 | 
						|
from synapse.replication.tcp.client import ReplicationDataHandler
 | 
						|
from synapse.replication.tcp.commands import ClearUserSyncsCommand
 | 
						|
from synapse.replication.tcp.streams import (
 | 
						|
    AccountDataStream,
 | 
						|
    DeviceListsStream,
 | 
						|
    GroupServerStream,
 | 
						|
    PresenceStream,
 | 
						|
    PushersStream,
 | 
						|
    PushRulesStream,
 | 
						|
    ReceiptsStream,
 | 
						|
    TagAccountDataStream,
 | 
						|
    ToDeviceStream,
 | 
						|
    TypingStream,
 | 
						|
)
 | 
						|
from synapse.rest.admin import register_servlets_for_media_repo
 | 
						|
from synapse.rest.client.v1 import events
 | 
						|
from synapse.rest.client.v1.initial_sync import InitialSyncRestServlet
 | 
						|
from synapse.rest.client.v1.login import LoginRestServlet
 | 
						|
from synapse.rest.client.v1.profile import (
 | 
						|
    ProfileAvatarURLRestServlet,
 | 
						|
    ProfileDisplaynameRestServlet,
 | 
						|
    ProfileRestServlet,
 | 
						|
)
 | 
						|
from synapse.rest.client.v1.push_rule import PushRuleRestServlet
 | 
						|
from synapse.rest.client.v1.room import (
 | 
						|
    JoinedRoomMemberListRestServlet,
 | 
						|
    JoinRoomAliasServlet,
 | 
						|
    PublicRoomListRestServlet,
 | 
						|
    RoomEventContextServlet,
 | 
						|
    RoomInitialSyncRestServlet,
 | 
						|
    RoomMemberListRestServlet,
 | 
						|
    RoomMembershipRestServlet,
 | 
						|
    RoomMessageListRestServlet,
 | 
						|
    RoomSendEventRestServlet,
 | 
						|
    RoomStateEventRestServlet,
 | 
						|
    RoomStateRestServlet,
 | 
						|
)
 | 
						|
from synapse.rest.client.v1.voip import VoipRestServlet
 | 
						|
from synapse.rest.client.v2_alpha import groups, sync, user_directory
 | 
						|
from synapse.rest.client.v2_alpha._base import client_patterns
 | 
						|
from synapse.rest.client.v2_alpha.account import ThreepidRestServlet
 | 
						|
from synapse.rest.client.v2_alpha.account_data import (
 | 
						|
    AccountDataServlet,
 | 
						|
    RoomAccountDataServlet,
 | 
						|
)
 | 
						|
from synapse.rest.client.v2_alpha.keys import KeyChangesServlet, KeyQueryServlet
 | 
						|
from synapse.rest.client.v2_alpha.register import RegisterRestServlet
 | 
						|
from synapse.rest.client.versions import VersionsRestServlet
 | 
						|
from synapse.rest.key.v2 import KeyApiV2Resource
 | 
						|
from synapse.server import HomeServer
 | 
						|
from synapse.storage.data_stores.main.censor_events import CensorEventsStore
 | 
						|
from synapse.storage.data_stores.main.media_repository import MediaRepositoryStore
 | 
						|
from synapse.storage.data_stores.main.monthly_active_users import (
 | 
						|
    MonthlyActiveUsersWorkerStore,
 | 
						|
)
 | 
						|
from synapse.storage.data_stores.main.presence import UserPresenceState
 | 
						|
from synapse.storage.data_stores.main.search import SearchWorkerStore
 | 
						|
from synapse.storage.data_stores.main.ui_auth import UIAuthWorkerStore
 | 
						|
from synapse.storage.data_stores.main.user_directory import UserDirectoryStore
 | 
						|
from synapse.types import ReadReceipt
 | 
						|
from synapse.util.async_helpers import Linearizer
 | 
						|
from synapse.util.httpresourcetree import create_resource_tree
 | 
						|
from synapse.util.manhole import manhole
 | 
						|
from synapse.util.versionstring import get_version_string
 | 
						|
 | 
						|
logger = logging.getLogger("synapse.app.generic_worker")
 | 
						|
 | 
						|
 | 
						|
class PresenceStatusStubServlet(RestServlet):
 | 
						|
    """If presence is disabled this servlet can be used to stub out setting
 | 
						|
    presence status.
 | 
						|
    """
 | 
						|
 | 
						|
    PATTERNS = client_patterns("/presence/(?P<user_id>[^/]*)/status")
 | 
						|
 | 
						|
    def __init__(self, hs):
 | 
						|
        super(PresenceStatusStubServlet, self).__init__()
 | 
						|
        self.auth = hs.get_auth()
 | 
						|
 | 
						|
    async def on_GET(self, request, user_id):
 | 
						|
        await self.auth.get_user_by_req(request)
 | 
						|
        return 200, {"presence": "offline"}
 | 
						|
 | 
						|
    async def on_PUT(self, request, user_id):
 | 
						|
        await self.auth.get_user_by_req(request)
 | 
						|
        return 200, {}
 | 
						|
 | 
						|
 | 
						|
class KeyUploadServlet(RestServlet):
 | 
						|
    """An implementation of the `KeyUploadServlet` that responds to read only
 | 
						|
    requests, but otherwise proxies through to the master instance.
 | 
						|
    """
 | 
						|
 | 
						|
    PATTERNS = client_patterns("/keys/upload(/(?P<device_id>[^/]+))?$")
 | 
						|
 | 
						|
    def __init__(self, hs):
 | 
						|
        """
 | 
						|
        Args:
 | 
						|
            hs (synapse.server.HomeServer): server
 | 
						|
        """
 | 
						|
        super(KeyUploadServlet, self).__init__()
 | 
						|
        self.auth = hs.get_auth()
 | 
						|
        self.store = hs.get_datastore()
 | 
						|
        self.http_client = hs.get_simple_http_client()
 | 
						|
        self.main_uri = hs.config.worker_main_http_uri
 | 
						|
 | 
						|
    async def on_POST(self, request, device_id):
 | 
						|
        requester = await self.auth.get_user_by_req(request, allow_guest=True)
 | 
						|
        user_id = requester.user.to_string()
 | 
						|
        body = parse_json_object_from_request(request)
 | 
						|
 | 
						|
        if device_id is not None:
 | 
						|
            # passing the device_id here is deprecated; however, we allow it
 | 
						|
            # for now for compatibility with older clients.
 | 
						|
            if requester.device_id is not None and device_id != requester.device_id:
 | 
						|
                logger.warning(
 | 
						|
                    "Client uploading keys for a different device "
 | 
						|
                    "(logged in as %s, uploading for %s)",
 | 
						|
                    requester.device_id,
 | 
						|
                    device_id,
 | 
						|
                )
 | 
						|
        else:
 | 
						|
            device_id = requester.device_id
 | 
						|
 | 
						|
        if device_id is None:
 | 
						|
            raise SynapseError(
 | 
						|
                400, "To upload keys, you must pass device_id when authenticating"
 | 
						|
            )
 | 
						|
 | 
						|
        if body:
 | 
						|
            # They're actually trying to upload something, proxy to main synapse.
 | 
						|
            # Pass through the auth headers, if any, in case the access token
 | 
						|
            # is there.
 | 
						|
            auth_headers = request.requestHeaders.getRawHeaders(b"Authorization", [])
 | 
						|
            headers = {"Authorization": auth_headers}
 | 
						|
            try:
 | 
						|
                result = await self.http_client.post_json_get_json(
 | 
						|
                    self.main_uri + request.uri.decode("ascii"), body, headers=headers
 | 
						|
                )
 | 
						|
            except HttpResponseException as e:
 | 
						|
                raise e.to_synapse_error() from e
 | 
						|
            except RequestSendFailed as e:
 | 
						|
                raise SynapseError(502, "Failed to talk to master") from e
 | 
						|
 | 
						|
            return 200, result
 | 
						|
        else:
 | 
						|
            # Just interested in counts.
 | 
						|
            result = await self.store.count_e2e_one_time_keys(user_id, device_id)
 | 
						|
            return 200, {"one_time_key_counts": result}
 | 
						|
 | 
						|
 | 
						|
class _NullContextManager(ContextManager[None]):
 | 
						|
    """A context manager which does nothing."""
 | 
						|
 | 
						|
    def __exit__(self, exc_type, exc_val, exc_tb):
 | 
						|
        pass
 | 
						|
 | 
						|
 | 
						|
UPDATE_SYNCING_USERS_MS = 10 * 1000
 | 
						|
 | 
						|
 | 
						|
class GenericWorkerPresence(BasePresenceHandler):
 | 
						|
    def __init__(self, hs):
 | 
						|
        super().__init__(hs)
 | 
						|
        self.hs = hs
 | 
						|
        self.is_mine_id = hs.is_mine_id
 | 
						|
        self.http_client = hs.get_simple_http_client()
 | 
						|
 | 
						|
        self._presence_enabled = hs.config.use_presence
 | 
						|
 | 
						|
        # The number of ongoing syncs on this process, by user id.
 | 
						|
        # Empty if _presence_enabled is false.
 | 
						|
        self._user_to_num_current_syncs = {}  # type: Dict[str, int]
 | 
						|
 | 
						|
        self.notifier = hs.get_notifier()
 | 
						|
        self.instance_id = hs.get_instance_id()
 | 
						|
 | 
						|
        # user_id -> last_sync_ms. Lists the users that have stopped syncing
 | 
						|
        # but we haven't notified the master of that yet
 | 
						|
        self.users_going_offline = {}
 | 
						|
 | 
						|
        self._bump_active_client = ReplicationBumpPresenceActiveTime.make_client(hs)
 | 
						|
        self._set_state_client = ReplicationPresenceSetState.make_client(hs)
 | 
						|
 | 
						|
        self._send_stop_syncing_loop = self.clock.looping_call(
 | 
						|
            self.send_stop_syncing, UPDATE_SYNCING_USERS_MS
 | 
						|
        )
 | 
						|
 | 
						|
        hs.get_reactor().addSystemEventTrigger(
 | 
						|
            "before",
 | 
						|
            "shutdown",
 | 
						|
            run_as_background_process,
 | 
						|
            "generic_presence.on_shutdown",
 | 
						|
            self._on_shutdown,
 | 
						|
        )
 | 
						|
 | 
						|
    def _on_shutdown(self):
 | 
						|
        if self._presence_enabled:
 | 
						|
            self.hs.get_tcp_replication().send_command(
 | 
						|
                ClearUserSyncsCommand(self.instance_id)
 | 
						|
            )
 | 
						|
 | 
						|
    def send_user_sync(self, user_id, is_syncing, last_sync_ms):
 | 
						|
        if self._presence_enabled:
 | 
						|
            self.hs.get_tcp_replication().send_user_sync(
 | 
						|
                self.instance_id, user_id, is_syncing, last_sync_ms
 | 
						|
            )
 | 
						|
 | 
						|
    def mark_as_coming_online(self, user_id):
 | 
						|
        """A user has started syncing. Send a UserSync to the master, unless they
 | 
						|
        had recently stopped syncing.
 | 
						|
 | 
						|
        Args:
 | 
						|
            user_id (str)
 | 
						|
        """
 | 
						|
        going_offline = self.users_going_offline.pop(user_id, None)
 | 
						|
        if not going_offline:
 | 
						|
            # Safe to skip because we haven't yet told the master they were offline
 | 
						|
            self.send_user_sync(user_id, True, self.clock.time_msec())
 | 
						|
 | 
						|
    def mark_as_going_offline(self, user_id):
 | 
						|
        """A user has stopped syncing. We wait before notifying the master as
 | 
						|
        its likely they'll come back soon. This allows us to avoid sending
 | 
						|
        a stopped syncing immediately followed by a started syncing notification
 | 
						|
        to the master
 | 
						|
 | 
						|
        Args:
 | 
						|
            user_id (str)
 | 
						|
        """
 | 
						|
        self.users_going_offline[user_id] = self.clock.time_msec()
 | 
						|
 | 
						|
    def send_stop_syncing(self):
 | 
						|
        """Check if there are any users who have stopped syncing a while ago
 | 
						|
        and haven't come back yet. If there are poke the master about them.
 | 
						|
        """
 | 
						|
        now = self.clock.time_msec()
 | 
						|
        for user_id, last_sync_ms in list(self.users_going_offline.items()):
 | 
						|
            if now - last_sync_ms > UPDATE_SYNCING_USERS_MS:
 | 
						|
                self.users_going_offline.pop(user_id, None)
 | 
						|
                self.send_user_sync(user_id, False, last_sync_ms)
 | 
						|
 | 
						|
    async def user_syncing(
 | 
						|
        self, user_id: str, affect_presence: bool
 | 
						|
    ) -> ContextManager[None]:
 | 
						|
        """Record that a user is syncing.
 | 
						|
 | 
						|
        Called by the sync and events servlets to record that a user has connected to
 | 
						|
        this worker and is waiting for some events.
 | 
						|
        """
 | 
						|
        if not affect_presence or not self._presence_enabled:
 | 
						|
            return _NullContextManager()
 | 
						|
 | 
						|
        curr_sync = self._user_to_num_current_syncs.get(user_id, 0)
 | 
						|
        self._user_to_num_current_syncs[user_id] = curr_sync + 1
 | 
						|
 | 
						|
        # If we went from no in flight sync to some, notify replication
 | 
						|
        if self._user_to_num_current_syncs[user_id] == 1:
 | 
						|
            self.mark_as_coming_online(user_id)
 | 
						|
 | 
						|
        def _end():
 | 
						|
            # We check that the user_id is in user_to_num_current_syncs because
 | 
						|
            # user_to_num_current_syncs may have been cleared if we are
 | 
						|
            # shutting down.
 | 
						|
            if user_id in self._user_to_num_current_syncs:
 | 
						|
                self._user_to_num_current_syncs[user_id] -= 1
 | 
						|
 | 
						|
                # If we went from one in flight sync to non, notify replication
 | 
						|
                if self._user_to_num_current_syncs[user_id] == 0:
 | 
						|
                    self.mark_as_going_offline(user_id)
 | 
						|
 | 
						|
        @contextlib.contextmanager
 | 
						|
        def _user_syncing():
 | 
						|
            try:
 | 
						|
                yield
 | 
						|
            finally:
 | 
						|
                _end()
 | 
						|
 | 
						|
        return _user_syncing()
 | 
						|
 | 
						|
    @defer.inlineCallbacks
 | 
						|
    def notify_from_replication(self, states, stream_id):
 | 
						|
        parties = yield get_interested_parties(self.store, states)
 | 
						|
        room_ids_to_states, users_to_states = parties
 | 
						|
 | 
						|
        self.notifier.on_new_event(
 | 
						|
            "presence_key",
 | 
						|
            stream_id,
 | 
						|
            rooms=room_ids_to_states.keys(),
 | 
						|
            users=users_to_states.keys(),
 | 
						|
        )
 | 
						|
 | 
						|
    @defer.inlineCallbacks
 | 
						|
    def process_replication_rows(self, token, rows):
 | 
						|
        states = [
 | 
						|
            UserPresenceState(
 | 
						|
                row.user_id,
 | 
						|
                row.state,
 | 
						|
                row.last_active_ts,
 | 
						|
                row.last_federation_update_ts,
 | 
						|
                row.last_user_sync_ts,
 | 
						|
                row.status_msg,
 | 
						|
                row.currently_active,
 | 
						|
            )
 | 
						|
            for row in rows
 | 
						|
        ]
 | 
						|
 | 
						|
        for state in states:
 | 
						|
            self.user_to_current_state[state.user_id] = state
 | 
						|
 | 
						|
        stream_id = token
 | 
						|
        yield self.notify_from_replication(states, stream_id)
 | 
						|
 | 
						|
    def get_currently_syncing_users_for_replication(self) -> Iterable[str]:
 | 
						|
        return [
 | 
						|
            user_id
 | 
						|
            for user_id, count in self._user_to_num_current_syncs.items()
 | 
						|
            if count > 0
 | 
						|
        ]
 | 
						|
 | 
						|
    async def set_state(self, target_user, state, ignore_status_msg=False):
 | 
						|
        """Set the presence state of the user.
 | 
						|
        """
 | 
						|
        presence = state["presence"]
 | 
						|
 | 
						|
        valid_presence = (
 | 
						|
            PresenceState.ONLINE,
 | 
						|
            PresenceState.UNAVAILABLE,
 | 
						|
            PresenceState.OFFLINE,
 | 
						|
        )
 | 
						|
        if presence not in valid_presence:
 | 
						|
            raise SynapseError(400, "Invalid presence state")
 | 
						|
 | 
						|
        user_id = target_user.to_string()
 | 
						|
 | 
						|
        # If presence is disabled, no-op
 | 
						|
        if not self.hs.config.use_presence:
 | 
						|
            return
 | 
						|
 | 
						|
        # Proxy request to master
 | 
						|
        await self._set_state_client(
 | 
						|
            user_id=user_id, state=state, ignore_status_msg=ignore_status_msg
 | 
						|
        )
 | 
						|
 | 
						|
    async def bump_presence_active_time(self, user):
 | 
						|
        """We've seen the user do something that indicates they're interacting
 | 
						|
        with the app.
 | 
						|
        """
 | 
						|
        # If presence is disabled, no-op
 | 
						|
        if not self.hs.config.use_presence:
 | 
						|
            return
 | 
						|
 | 
						|
        # Proxy request to master
 | 
						|
        user_id = user.to_string()
 | 
						|
        await self._bump_active_client(user_id=user_id)
 | 
						|
 | 
						|
 | 
						|
class GenericWorkerTyping(object):
 | 
						|
    def __init__(self, hs):
 | 
						|
        self._latest_room_serial = 0
 | 
						|
        self._reset()
 | 
						|
 | 
						|
    def _reset(self):
 | 
						|
        """
 | 
						|
        Reset the typing handler's data caches.
 | 
						|
        """
 | 
						|
        # map room IDs to serial numbers
 | 
						|
        self._room_serials = {}
 | 
						|
        # map room IDs to sets of users currently typing
 | 
						|
        self._room_typing = {}
 | 
						|
 | 
						|
    def process_replication_rows(self, token, rows):
 | 
						|
        if self._latest_room_serial > token:
 | 
						|
            # The master has gone backwards. To prevent inconsistent data, just
 | 
						|
            # clear everything.
 | 
						|
            self._reset()
 | 
						|
 | 
						|
        # Set the latest serial token to whatever the server gave us.
 | 
						|
        self._latest_room_serial = token
 | 
						|
 | 
						|
        for row in rows:
 | 
						|
            self._room_serials[row.room_id] = token
 | 
						|
            self._room_typing[row.room_id] = row.user_ids
 | 
						|
 | 
						|
    def get_current_token(self) -> int:
 | 
						|
        return self._latest_room_serial
 | 
						|
 | 
						|
 | 
						|
class GenericWorkerSlavedStore(
 | 
						|
    # FIXME(#3714): We need to add UserDirectoryStore as we write directly
 | 
						|
    # rather than going via the correct worker.
 | 
						|
    UserDirectoryStore,
 | 
						|
    UIAuthWorkerStore,
 | 
						|
    SlavedDeviceInboxStore,
 | 
						|
    SlavedDeviceStore,
 | 
						|
    SlavedReceiptsStore,
 | 
						|
    SlavedPushRuleStore,
 | 
						|
    SlavedGroupServerStore,
 | 
						|
    SlavedAccountDataStore,
 | 
						|
    SlavedPusherStore,
 | 
						|
    CensorEventsStore,
 | 
						|
    SlavedEventStore,
 | 
						|
    SlavedKeyStore,
 | 
						|
    RoomStore,
 | 
						|
    DirectoryStore,
 | 
						|
    SlavedApplicationServiceStore,
 | 
						|
    SlavedRegistrationStore,
 | 
						|
    SlavedTransactionStore,
 | 
						|
    SlavedProfileStore,
 | 
						|
    SlavedClientIpStore,
 | 
						|
    SlavedPresenceStore,
 | 
						|
    SlavedFilteringStore,
 | 
						|
    MonthlyActiveUsersWorkerStore,
 | 
						|
    MediaRepositoryStore,
 | 
						|
    SearchWorkerStore,
 | 
						|
    BaseSlavedStore,
 | 
						|
):
 | 
						|
    def __init__(self, database, db_conn, hs):
 | 
						|
        super(GenericWorkerSlavedStore, self).__init__(database, db_conn, hs)
 | 
						|
 | 
						|
        # We pull out the current federation stream position now so that we
 | 
						|
        # always have a known value for the federation position in memory so
 | 
						|
        # that we don't have to bounce via a deferred once when we start the
 | 
						|
        # replication streams.
 | 
						|
        self.federation_out_pos_startup = self._get_federation_out_pos(db_conn)
 | 
						|
 | 
						|
    def _get_federation_out_pos(self, db_conn):
 | 
						|
        sql = "SELECT stream_id FROM federation_stream_position WHERE type = ?"
 | 
						|
        sql = self.database_engine.convert_param_style(sql)
 | 
						|
 | 
						|
        txn = db_conn.cursor()
 | 
						|
        txn.execute(sql, ("federation",))
 | 
						|
        rows = txn.fetchall()
 | 
						|
        txn.close()
 | 
						|
 | 
						|
        return rows[0][0] if rows else -1
 | 
						|
 | 
						|
 | 
						|
class GenericWorkerServer(HomeServer):
 | 
						|
    DATASTORE_CLASS = GenericWorkerSlavedStore
 | 
						|
 | 
						|
    def _listen_http(self, listener_config: ListenerConfig):
 | 
						|
        port = listener_config.port
 | 
						|
        bind_addresses = listener_config.bind_addresses
 | 
						|
 | 
						|
        assert listener_config.http_options is not None
 | 
						|
 | 
						|
        site_tag = listener_config.http_options.tag
 | 
						|
        if site_tag is None:
 | 
						|
            site_tag = port
 | 
						|
        resources = {}
 | 
						|
        for res in listener_config.http_options.resources:
 | 
						|
            for name in res.names:
 | 
						|
                if name == "metrics":
 | 
						|
                    resources[METRICS_PREFIX] = MetricsResource(RegistryProxy)
 | 
						|
                elif name == "client":
 | 
						|
                    resource = JsonResource(self, canonical_json=False)
 | 
						|
 | 
						|
                    PublicRoomListRestServlet(self).register(resource)
 | 
						|
                    RoomMemberListRestServlet(self).register(resource)
 | 
						|
                    JoinedRoomMemberListRestServlet(self).register(resource)
 | 
						|
                    RoomStateRestServlet(self).register(resource)
 | 
						|
                    RoomEventContextServlet(self).register(resource)
 | 
						|
                    RoomMessageListRestServlet(self).register(resource)
 | 
						|
                    RegisterRestServlet(self).register(resource)
 | 
						|
                    LoginRestServlet(self).register(resource)
 | 
						|
                    ThreepidRestServlet(self).register(resource)
 | 
						|
                    KeyQueryServlet(self).register(resource)
 | 
						|
                    KeyChangesServlet(self).register(resource)
 | 
						|
                    VoipRestServlet(self).register(resource)
 | 
						|
                    PushRuleRestServlet(self).register(resource)
 | 
						|
                    VersionsRestServlet(self).register(resource)
 | 
						|
                    RoomSendEventRestServlet(self).register(resource)
 | 
						|
                    RoomMembershipRestServlet(self).register(resource)
 | 
						|
                    RoomStateEventRestServlet(self).register(resource)
 | 
						|
                    JoinRoomAliasServlet(self).register(resource)
 | 
						|
                    ProfileAvatarURLRestServlet(self).register(resource)
 | 
						|
                    ProfileDisplaynameRestServlet(self).register(resource)
 | 
						|
                    ProfileRestServlet(self).register(resource)
 | 
						|
                    KeyUploadServlet(self).register(resource)
 | 
						|
                    AccountDataServlet(self).register(resource)
 | 
						|
                    RoomAccountDataServlet(self).register(resource)
 | 
						|
 | 
						|
                    sync.register_servlets(self, resource)
 | 
						|
                    events.register_servlets(self, resource)
 | 
						|
                    InitialSyncRestServlet(self).register(resource)
 | 
						|
                    RoomInitialSyncRestServlet(self).register(resource)
 | 
						|
 | 
						|
                    user_directory.register_servlets(self, resource)
 | 
						|
 | 
						|
                    # If presence is disabled, use the stub servlet that does
 | 
						|
                    # not allow sending presence
 | 
						|
                    if not self.config.use_presence:
 | 
						|
                        PresenceStatusStubServlet(self).register(resource)
 | 
						|
 | 
						|
                    groups.register_servlets(self, resource)
 | 
						|
 | 
						|
                    resources.update({CLIENT_API_PREFIX: resource})
 | 
						|
                elif name == "federation":
 | 
						|
                    resources.update({FEDERATION_PREFIX: TransportLayerServer(self)})
 | 
						|
                elif name == "media":
 | 
						|
                    if self.config.can_load_media_repo:
 | 
						|
                        media_repo = self.get_media_repository_resource()
 | 
						|
 | 
						|
                        # We need to serve the admin servlets for media on the
 | 
						|
                        # worker.
 | 
						|
                        admin_resource = JsonResource(self, canonical_json=False)
 | 
						|
                        register_servlets_for_media_repo(self, admin_resource)
 | 
						|
 | 
						|
                        resources.update(
 | 
						|
                            {
 | 
						|
                                MEDIA_PREFIX: media_repo,
 | 
						|
                                LEGACY_MEDIA_PREFIX: media_repo,
 | 
						|
                                "/_synapse/admin": admin_resource,
 | 
						|
                            }
 | 
						|
                        )
 | 
						|
                    else:
 | 
						|
                        logger.warning(
 | 
						|
                            "A 'media' listener is configured but the media"
 | 
						|
                            " repository is disabled. Ignoring."
 | 
						|
                        )
 | 
						|
 | 
						|
                if name == "openid" and "federation" not in res.names:
 | 
						|
                    # Only load the openid resource separately if federation resource
 | 
						|
                    # is not specified since federation resource includes openid
 | 
						|
                    # resource.
 | 
						|
                    resources.update(
 | 
						|
                        {
 | 
						|
                            FEDERATION_PREFIX: TransportLayerServer(
 | 
						|
                                self, servlet_groups=["openid"]
 | 
						|
                            )
 | 
						|
                        }
 | 
						|
                    )
 | 
						|
 | 
						|
                if name in ["keys", "federation"]:
 | 
						|
                    resources[SERVER_KEY_V2_PREFIX] = KeyApiV2Resource(self)
 | 
						|
 | 
						|
                if name == "replication":
 | 
						|
                    resources[REPLICATION_PREFIX] = ReplicationRestResource(self)
 | 
						|
 | 
						|
        root_resource = create_resource_tree(resources, OptionsResource())
 | 
						|
 | 
						|
        _base.listen_tcp(
 | 
						|
            bind_addresses,
 | 
						|
            port,
 | 
						|
            SynapseSite(
 | 
						|
                "synapse.access.http.%s" % (site_tag,),
 | 
						|
                site_tag,
 | 
						|
                listener_config,
 | 
						|
                root_resource,
 | 
						|
                self.version_string,
 | 
						|
            ),
 | 
						|
            reactor=self.get_reactor(),
 | 
						|
        )
 | 
						|
 | 
						|
        logger.info("Synapse worker now listening on port %d", port)
 | 
						|
 | 
						|
    def start_listening(self, listeners: Iterable[ListenerConfig]):
 | 
						|
        for listener in listeners:
 | 
						|
            if listener.type == "http":
 | 
						|
                self._listen_http(listener)
 | 
						|
            elif listener.type == "manhole":
 | 
						|
                _base.listen_tcp(
 | 
						|
                    listener.bind_addresses,
 | 
						|
                    listener.port,
 | 
						|
                    manhole(
 | 
						|
                        username="matrix", password="rabbithole", globals={"hs": self}
 | 
						|
                    ),
 | 
						|
                )
 | 
						|
            elif listener.type == "metrics":
 | 
						|
                if not self.get_config().enable_metrics:
 | 
						|
                    logger.warning(
 | 
						|
                        (
 | 
						|
                            "Metrics listener configured, but "
 | 
						|
                            "enable_metrics is not True!"
 | 
						|
                        )
 | 
						|
                    )
 | 
						|
                else:
 | 
						|
                    _base.listen_metrics(listener.bind_addresses, listener.port)
 | 
						|
            else:
 | 
						|
                logger.warning("Unsupported listener type: %s", listener.type)
 | 
						|
 | 
						|
        self.get_tcp_replication().start_replication(self)
 | 
						|
 | 
						|
    def remove_pusher(self, app_id, push_key, user_id):
 | 
						|
        self.get_tcp_replication().send_remove_pusher(app_id, push_key, user_id)
 | 
						|
 | 
						|
    def build_replication_data_handler(self):
 | 
						|
        return GenericWorkerReplicationHandler(self)
 | 
						|
 | 
						|
    def build_presence_handler(self):
 | 
						|
        return GenericWorkerPresence(self)
 | 
						|
 | 
						|
    def build_typing_handler(self):
 | 
						|
        return GenericWorkerTyping(self)
 | 
						|
 | 
						|
 | 
						|
class GenericWorkerReplicationHandler(ReplicationDataHandler):
 | 
						|
    def __init__(self, hs):
 | 
						|
        super(GenericWorkerReplicationHandler, self).__init__(hs)
 | 
						|
 | 
						|
        self.store = hs.get_datastore()
 | 
						|
        self.typing_handler = hs.get_typing_handler()
 | 
						|
        self.presence_handler = hs.get_presence_handler()  # type: GenericWorkerPresence
 | 
						|
        self.notifier = hs.get_notifier()
 | 
						|
 | 
						|
        self.notify_pushers = hs.config.start_pushers
 | 
						|
        self.pusher_pool = hs.get_pusherpool()
 | 
						|
 | 
						|
        self.send_handler = None  # type: Optional[FederationSenderHandler]
 | 
						|
        if hs.config.send_federation:
 | 
						|
            self.send_handler = FederationSenderHandler(hs)
 | 
						|
 | 
						|
    async def on_rdata(self, stream_name, instance_name, token, rows):
 | 
						|
        await super().on_rdata(stream_name, instance_name, token, rows)
 | 
						|
        await self._process_and_notify(stream_name, instance_name, token, rows)
 | 
						|
 | 
						|
    async def _process_and_notify(self, stream_name, instance_name, token, rows):
 | 
						|
        try:
 | 
						|
            if self.send_handler:
 | 
						|
                await self.send_handler.process_replication_rows(
 | 
						|
                    stream_name, token, rows
 | 
						|
                )
 | 
						|
 | 
						|
            if stream_name == PushRulesStream.NAME:
 | 
						|
                self.notifier.on_new_event(
 | 
						|
                    "push_rules_key", token, users=[row.user_id for row in rows]
 | 
						|
                )
 | 
						|
            elif stream_name in (AccountDataStream.NAME, TagAccountDataStream.NAME):
 | 
						|
                self.notifier.on_new_event(
 | 
						|
                    "account_data_key", token, users=[row.user_id for row in rows]
 | 
						|
                )
 | 
						|
            elif stream_name == ReceiptsStream.NAME:
 | 
						|
                self.notifier.on_new_event(
 | 
						|
                    "receipt_key", token, rooms=[row.room_id for row in rows]
 | 
						|
                )
 | 
						|
                await self.pusher_pool.on_new_receipts(
 | 
						|
                    token, token, {row.room_id for row in rows}
 | 
						|
                )
 | 
						|
            elif stream_name == TypingStream.NAME:
 | 
						|
                self.typing_handler.process_replication_rows(token, rows)
 | 
						|
                self.notifier.on_new_event(
 | 
						|
                    "typing_key", token, rooms=[row.room_id for row in rows]
 | 
						|
                )
 | 
						|
            elif stream_name == ToDeviceStream.NAME:
 | 
						|
                entities = [row.entity for row in rows if row.entity.startswith("@")]
 | 
						|
                if entities:
 | 
						|
                    self.notifier.on_new_event("to_device_key", token, users=entities)
 | 
						|
            elif stream_name == DeviceListsStream.NAME:
 | 
						|
                all_room_ids = set()  # type: Set[str]
 | 
						|
                for row in rows:
 | 
						|
                    if row.entity.startswith("@"):
 | 
						|
                        room_ids = await self.store.get_rooms_for_user(row.entity)
 | 
						|
                        all_room_ids.update(room_ids)
 | 
						|
                self.notifier.on_new_event("device_list_key", token, rooms=all_room_ids)
 | 
						|
            elif stream_name == PresenceStream.NAME:
 | 
						|
                await self.presence_handler.process_replication_rows(token, rows)
 | 
						|
            elif stream_name == GroupServerStream.NAME:
 | 
						|
                self.notifier.on_new_event(
 | 
						|
                    "groups_key", token, users=[row.user_id for row in rows]
 | 
						|
                )
 | 
						|
            elif stream_name == PushersStream.NAME:
 | 
						|
                for row in rows:
 | 
						|
                    if row.deleted:
 | 
						|
                        self.stop_pusher(row.user_id, row.app_id, row.pushkey)
 | 
						|
                    else:
 | 
						|
                        await self.start_pusher(row.user_id, row.app_id, row.pushkey)
 | 
						|
        except Exception:
 | 
						|
            logger.exception("Error processing replication")
 | 
						|
 | 
						|
    async def on_position(self, stream_name: str, instance_name: str, token: int):
 | 
						|
        await super().on_position(stream_name, instance_name, token)
 | 
						|
        # Also call on_rdata to ensure that stream positions are properly reset.
 | 
						|
        await self.on_rdata(stream_name, instance_name, token, [])
 | 
						|
 | 
						|
    def stop_pusher(self, user_id, app_id, pushkey):
 | 
						|
        if not self.notify_pushers:
 | 
						|
            return
 | 
						|
 | 
						|
        key = "%s:%s" % (app_id, pushkey)
 | 
						|
        pushers_for_user = self.pusher_pool.pushers.get(user_id, {})
 | 
						|
        pusher = pushers_for_user.pop(key, None)
 | 
						|
        if pusher is None:
 | 
						|
            return
 | 
						|
        logger.info("Stopping pusher %r / %r", user_id, key)
 | 
						|
        pusher.on_stop()
 | 
						|
 | 
						|
    async def start_pusher(self, user_id, app_id, pushkey):
 | 
						|
        if not self.notify_pushers:
 | 
						|
            return
 | 
						|
 | 
						|
        key = "%s:%s" % (app_id, pushkey)
 | 
						|
        logger.info("Starting pusher %r / %r", user_id, key)
 | 
						|
        return await self.pusher_pool.start_pusher_by_id(app_id, pushkey, user_id)
 | 
						|
 | 
						|
    def on_remote_server_up(self, server: str):
 | 
						|
        """Called when get a new REMOTE_SERVER_UP command."""
 | 
						|
 | 
						|
        # Let's wake up the transaction queue for the server in case we have
 | 
						|
        # pending stuff to send to it.
 | 
						|
        if self.send_handler:
 | 
						|
            self.send_handler.wake_destination(server)
 | 
						|
 | 
						|
 | 
						|
class FederationSenderHandler(object):
 | 
						|
    """Processes the fedration replication stream
 | 
						|
 | 
						|
    This class is only instantiate on the worker responsible for sending outbound
 | 
						|
    federation transactions. It receives rows from the replication stream and forwards
 | 
						|
    the appropriate entries to the FederationSender class.
 | 
						|
    """
 | 
						|
 | 
						|
    def __init__(self, hs: GenericWorkerServer):
 | 
						|
        self.store = hs.get_datastore()
 | 
						|
        self._is_mine_id = hs.is_mine_id
 | 
						|
        self.federation_sender = hs.get_federation_sender()
 | 
						|
        self._hs = hs
 | 
						|
 | 
						|
        # if the worker is restarted, we want to pick up where we left off in
 | 
						|
        # the replication stream, so load the position from the database.
 | 
						|
        #
 | 
						|
        # XXX is this actually worthwhile? Whenever the master is restarted, we'll
 | 
						|
        # drop some rows anyway (which is mostly fine because we're only dropping
 | 
						|
        # typing and presence notifications). If the replication stream is
 | 
						|
        # unreliable, why do we do all this hoop-jumping to store the position in the
 | 
						|
        # database? See also https://github.com/matrix-org/synapse/issues/7535.
 | 
						|
        #
 | 
						|
        self.federation_position = self.store.federation_out_pos_startup
 | 
						|
 | 
						|
        self._fed_position_linearizer = Linearizer(name="_fed_position_linearizer")
 | 
						|
        self._last_ack = self.federation_position
 | 
						|
 | 
						|
    def on_start(self):
 | 
						|
        # There may be some events that are persisted but haven't been sent,
 | 
						|
        # so send them now.
 | 
						|
        self.federation_sender.notify_new_events(
 | 
						|
            self.store.get_room_max_stream_ordering()
 | 
						|
        )
 | 
						|
 | 
						|
    def wake_destination(self, server: str):
 | 
						|
        self.federation_sender.wake_destination(server)
 | 
						|
 | 
						|
    async def process_replication_rows(self, stream_name, token, rows):
 | 
						|
        # The federation stream contains things that we want to send out, e.g.
 | 
						|
        # presence, typing, etc.
 | 
						|
        if stream_name == "federation":
 | 
						|
            send_queue.process_rows_for_federation(self.federation_sender, rows)
 | 
						|
            await self.update_token(token)
 | 
						|
 | 
						|
        # We also need to poke the federation sender when new events happen
 | 
						|
        elif stream_name == "events":
 | 
						|
            self.federation_sender.notify_new_events(token)
 | 
						|
 | 
						|
        # ... and when new receipts happen
 | 
						|
        elif stream_name == ReceiptsStream.NAME:
 | 
						|
            await self._on_new_receipts(rows)
 | 
						|
 | 
						|
        # ... as well as device updates and messages
 | 
						|
        elif stream_name == DeviceListsStream.NAME:
 | 
						|
            # The entities are either user IDs (starting with '@') whose devices
 | 
						|
            # have changed, or remote servers that we need to tell about
 | 
						|
            # changes.
 | 
						|
            hosts = {row.entity for row in rows if not row.entity.startswith("@")}
 | 
						|
            for host in hosts:
 | 
						|
                self.federation_sender.send_device_messages(host)
 | 
						|
 | 
						|
        elif stream_name == ToDeviceStream.NAME:
 | 
						|
            # The to_device stream includes stuff to be pushed to both local
 | 
						|
            # clients and remote servers, so we ignore entities that start with
 | 
						|
            # '@' (since they'll be local users rather than destinations).
 | 
						|
            hosts = {row.entity for row in rows if not row.entity.startswith("@")}
 | 
						|
            for host in hosts:
 | 
						|
                self.federation_sender.send_device_messages(host)
 | 
						|
 | 
						|
    async def _on_new_receipts(self, rows):
 | 
						|
        """
 | 
						|
        Args:
 | 
						|
            rows (Iterable[synapse.replication.tcp.streams.ReceiptsStream.ReceiptsStreamRow]):
 | 
						|
                new receipts to be processed
 | 
						|
        """
 | 
						|
        for receipt in rows:
 | 
						|
            # we only want to send on receipts for our own users
 | 
						|
            if not self._is_mine_id(receipt.user_id):
 | 
						|
                continue
 | 
						|
            receipt_info = ReadReceipt(
 | 
						|
                receipt.room_id,
 | 
						|
                receipt.receipt_type,
 | 
						|
                receipt.user_id,
 | 
						|
                [receipt.event_id],
 | 
						|
                receipt.data,
 | 
						|
            )
 | 
						|
            await self.federation_sender.send_read_receipt(receipt_info)
 | 
						|
 | 
						|
    async def update_token(self, token):
 | 
						|
        """Update the record of where we have processed to in the federation stream.
 | 
						|
 | 
						|
        Called after we have processed a an update received over replication. Sends
 | 
						|
        a FEDERATION_ACK back to the master, and stores the token that we have processed
 | 
						|
         in `federation_stream_position` so that we can restart where we left off.
 | 
						|
        """
 | 
						|
        self.federation_position = token
 | 
						|
 | 
						|
        # We save and send the ACK to master asynchronously, so we don't block
 | 
						|
        # processing on persistence. We don't need to do this operation for
 | 
						|
        # every single RDATA we receive, we just need to do it periodically.
 | 
						|
 | 
						|
        if self._fed_position_linearizer.is_queued(None):
 | 
						|
            # There is already a task queued up to save and send the token, so
 | 
						|
            # no need to queue up another task.
 | 
						|
            return
 | 
						|
 | 
						|
        run_as_background_process("_save_and_send_ack", self._save_and_send_ack)
 | 
						|
 | 
						|
    async def _save_and_send_ack(self):
 | 
						|
        """Save the current federation position in the database and send an ACK
 | 
						|
        to master with where we're up to.
 | 
						|
        """
 | 
						|
        try:
 | 
						|
            # We linearize here to ensure we don't have races updating the token
 | 
						|
            #
 | 
						|
            # XXX this appears to be redundant, since the ReplicationCommandHandler
 | 
						|
            # has a linearizer which ensures that we only process one line of
 | 
						|
            # replication data at a time. Should we remove it, or is it doing useful
 | 
						|
            # service for robustness? Or could we replace it with an assertion that
 | 
						|
            # we're not being re-entered?
 | 
						|
 | 
						|
            with (await self._fed_position_linearizer.queue(None)):
 | 
						|
                # We persist and ack the same position, so we take a copy of it
 | 
						|
                # here as otherwise it can get modified from underneath us.
 | 
						|
                current_position = self.federation_position
 | 
						|
 | 
						|
                await self.store.update_federation_out_pos(
 | 
						|
                    "federation", current_position
 | 
						|
                )
 | 
						|
 | 
						|
                # We ACK this token over replication so that the master can drop
 | 
						|
                # its in memory queues
 | 
						|
                self._hs.get_tcp_replication().send_federation_ack(current_position)
 | 
						|
                self._last_ack = current_position
 | 
						|
        except Exception:
 | 
						|
            logger.exception("Error updating federation stream position")
 | 
						|
 | 
						|
 | 
						|
def start(config_options):
 | 
						|
    try:
 | 
						|
        config = HomeServerConfig.load_config("Synapse worker", config_options)
 | 
						|
    except ConfigError as e:
 | 
						|
        sys.stderr.write("\n" + str(e) + "\n")
 | 
						|
        sys.exit(1)
 | 
						|
 | 
						|
    # For backwards compatibility let any of the old app names.
 | 
						|
    assert config.worker_app in (
 | 
						|
        "synapse.app.appservice",
 | 
						|
        "synapse.app.client_reader",
 | 
						|
        "synapse.app.event_creator",
 | 
						|
        "synapse.app.federation_reader",
 | 
						|
        "synapse.app.federation_sender",
 | 
						|
        "synapse.app.frontend_proxy",
 | 
						|
        "synapse.app.generic_worker",
 | 
						|
        "synapse.app.media_repository",
 | 
						|
        "synapse.app.pusher",
 | 
						|
        "synapse.app.synchrotron",
 | 
						|
        "synapse.app.user_dir",
 | 
						|
    )
 | 
						|
 | 
						|
    if config.worker_app == "synapse.app.appservice":
 | 
						|
        if config.notify_appservices:
 | 
						|
            sys.stderr.write(
 | 
						|
                "\nThe appservices must be disabled in the main synapse process"
 | 
						|
                "\nbefore they can be run in a separate worker."
 | 
						|
                "\nPlease add ``notify_appservices: false`` to the main config"
 | 
						|
                "\n"
 | 
						|
            )
 | 
						|
            sys.exit(1)
 | 
						|
 | 
						|
        # Force the appservice to start since they will be disabled in the main config
 | 
						|
        config.notify_appservices = True
 | 
						|
    else:
 | 
						|
        # For other worker types we force this to off.
 | 
						|
        config.notify_appservices = False
 | 
						|
 | 
						|
    if config.worker_app == "synapse.app.pusher":
 | 
						|
        if config.start_pushers:
 | 
						|
            sys.stderr.write(
 | 
						|
                "\nThe pushers must be disabled in the main synapse process"
 | 
						|
                "\nbefore they can be run in a separate worker."
 | 
						|
                "\nPlease add ``start_pushers: false`` to the main config"
 | 
						|
                "\n"
 | 
						|
            )
 | 
						|
            sys.exit(1)
 | 
						|
 | 
						|
        # Force the pushers to start since they will be disabled in the main config
 | 
						|
        config.start_pushers = True
 | 
						|
    else:
 | 
						|
        # For other worker types we force this to off.
 | 
						|
        config.start_pushers = False
 | 
						|
 | 
						|
    if config.worker_app == "synapse.app.user_dir":
 | 
						|
        if config.update_user_directory:
 | 
						|
            sys.stderr.write(
 | 
						|
                "\nThe update_user_directory must be disabled in the main synapse process"
 | 
						|
                "\nbefore they can be run in a separate worker."
 | 
						|
                "\nPlease add ``update_user_directory: false`` to the main config"
 | 
						|
                "\n"
 | 
						|
            )
 | 
						|
            sys.exit(1)
 | 
						|
 | 
						|
        # Force the pushers to start since they will be disabled in the main config
 | 
						|
        config.update_user_directory = True
 | 
						|
    else:
 | 
						|
        # For other worker types we force this to off.
 | 
						|
        config.update_user_directory = False
 | 
						|
 | 
						|
    if config.worker_app == "synapse.app.federation_sender":
 | 
						|
        if config.send_federation:
 | 
						|
            sys.stderr.write(
 | 
						|
                "\nThe send_federation must be disabled in the main synapse process"
 | 
						|
                "\nbefore they can be run in a separate worker."
 | 
						|
                "\nPlease add ``send_federation: false`` to the main config"
 | 
						|
                "\n"
 | 
						|
            )
 | 
						|
            sys.exit(1)
 | 
						|
 | 
						|
        # Force the pushers to start since they will be disabled in the main config
 | 
						|
        config.send_federation = True
 | 
						|
    else:
 | 
						|
        # For other worker types we force this to off.
 | 
						|
        config.send_federation = False
 | 
						|
 | 
						|
    synapse.events.USE_FROZEN_DICTS = config.use_frozen_dicts
 | 
						|
 | 
						|
    hs = GenericWorkerServer(
 | 
						|
        config.server_name,
 | 
						|
        config=config,
 | 
						|
        version_string="Synapse/" + get_version_string(synapse),
 | 
						|
    )
 | 
						|
 | 
						|
    setup_logging(hs, config, use_worker_options=True)
 | 
						|
 | 
						|
    hs.setup()
 | 
						|
 | 
						|
    # Ensure the replication streamer is always started in case we write to any
 | 
						|
    # streams. Will no-op if no streams can be written to by this worker.
 | 
						|
    hs.get_replication_streamer()
 | 
						|
 | 
						|
    reactor.addSystemEventTrigger(
 | 
						|
        "before", "startup", _base.start, hs, config.worker_listeners
 | 
						|
    )
 | 
						|
 | 
						|
    _base.start_worker_reactor("synapse-generic-worker", config)
 | 
						|
 | 
						|
 | 
						|
if __name__ == "__main__":
 | 
						|
    with LoggingContext("main"):
 | 
						|
        start(sys.argv[1:])
 |