482 lines
19 KiB
Python
482 lines
19 KiB
Python
# Copyright 2016 OpenMarket Ltd
|
|
# Copyright 2020 The Matrix.org Foundation C.I.C.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
import logging
|
|
import sys
|
|
from typing import Dict, List, Optional, Tuple
|
|
|
|
from matrix_common.versionstring import get_distribution_version_string
|
|
|
|
from twisted.internet import address
|
|
from twisted.web.resource import Resource
|
|
|
|
import synapse
|
|
import synapse.events
|
|
from synapse.api.errors import HttpResponseException, RequestSendFailed, SynapseError
|
|
from synapse.api.urls import (
|
|
CLIENT_API_PREFIX,
|
|
FEDERATION_PREFIX,
|
|
LEGACY_MEDIA_PREFIX,
|
|
MEDIA_R0_PREFIX,
|
|
MEDIA_V3_PREFIX,
|
|
SERVER_KEY_V2_PREFIX,
|
|
)
|
|
from synapse.app import _base
|
|
from synapse.app._base import (
|
|
handle_startup_exception,
|
|
max_request_body_size,
|
|
redirect_stdio_to_logs,
|
|
register_start,
|
|
)
|
|
from synapse.config._base import ConfigError
|
|
from synapse.config.homeserver import HomeServerConfig
|
|
from synapse.config.logger import setup_logging
|
|
from synapse.config.server import ListenerConfig
|
|
from synapse.federation.transport.server import TransportLayerServer
|
|
from synapse.http.server import JsonResource, OptionsResource
|
|
from synapse.http.servlet import RestServlet, parse_json_object_from_request
|
|
from synapse.http.site import SynapseRequest, SynapseSite
|
|
from synapse.logging.context import LoggingContext
|
|
from synapse.metrics import METRICS_PREFIX, MetricsResource, RegistryProxy
|
|
from synapse.replication.http import REPLICATION_PREFIX, ReplicationRestResource
|
|
from synapse.replication.slave.storage._base import BaseSlavedStore
|
|
from synapse.replication.slave.storage.account_data import SlavedAccountDataStore
|
|
from synapse.replication.slave.storage.appservice import SlavedApplicationServiceStore
|
|
from synapse.replication.slave.storage.deviceinbox import SlavedDeviceInboxStore
|
|
from synapse.replication.slave.storage.devices import SlavedDeviceStore
|
|
from synapse.replication.slave.storage.directory import DirectoryStore
|
|
from synapse.replication.slave.storage.events import SlavedEventStore
|
|
from synapse.replication.slave.storage.filtering import SlavedFilteringStore
|
|
from synapse.replication.slave.storage.groups import SlavedGroupServerStore
|
|
from synapse.replication.slave.storage.keys import SlavedKeyStore
|
|
from synapse.replication.slave.storage.profile import SlavedProfileStore
|
|
from synapse.replication.slave.storage.push_rule import SlavedPushRuleStore
|
|
from synapse.replication.slave.storage.pushers import SlavedPusherStore
|
|
from synapse.replication.slave.storage.receipts import SlavedReceiptsStore
|
|
from synapse.replication.slave.storage.registration import SlavedRegistrationStore
|
|
from synapse.rest.admin import register_servlets_for_media_repo
|
|
from synapse.rest.client import (
|
|
account_data,
|
|
events,
|
|
initial_sync,
|
|
login,
|
|
presence,
|
|
profile,
|
|
push_rule,
|
|
read_marker,
|
|
receipts,
|
|
room,
|
|
room_batch,
|
|
room_keys,
|
|
sendtodevice,
|
|
sync,
|
|
tags,
|
|
user_directory,
|
|
versions,
|
|
voip,
|
|
)
|
|
from synapse.rest.client._base import client_patterns
|
|
from synapse.rest.client.account import ThreepidRestServlet, WhoamiRestServlet
|
|
from synapse.rest.client.devices import DevicesRestServlet
|
|
from synapse.rest.client.keys import (
|
|
KeyChangesServlet,
|
|
KeyQueryServlet,
|
|
OneTimeKeyServlet,
|
|
)
|
|
from synapse.rest.client.register import (
|
|
RegisterRestServlet,
|
|
RegistrationTokenValidityRestServlet,
|
|
)
|
|
from synapse.rest.health import HealthResource
|
|
from synapse.rest.key.v2 import KeyApiV2Resource
|
|
from synapse.rest.synapse.client import build_synapse_client_resource_tree
|
|
from synapse.rest.well_known import well_known_resource
|
|
from synapse.server import HomeServer
|
|
from synapse.storage.databases.main.censor_events import CensorEventsStore
|
|
from synapse.storage.databases.main.client_ips import ClientIpWorkerStore
|
|
from synapse.storage.databases.main.e2e_room_keys import EndToEndRoomKeyStore
|
|
from synapse.storage.databases.main.lock import LockStore
|
|
from synapse.storage.databases.main.media_repository import MediaRepositoryStore
|
|
from synapse.storage.databases.main.metrics import ServerMetricsStore
|
|
from synapse.storage.databases.main.monthly_active_users import (
|
|
MonthlyActiveUsersWorkerStore,
|
|
)
|
|
from synapse.storage.databases.main.presence import PresenceStore
|
|
from synapse.storage.databases.main.room import RoomWorkerStore
|
|
from synapse.storage.databases.main.room_batch import RoomBatchStore
|
|
from synapse.storage.databases.main.search import SearchStore
|
|
from synapse.storage.databases.main.session import SessionStore
|
|
from synapse.storage.databases.main.stats import StatsStore
|
|
from synapse.storage.databases.main.transactions import TransactionWorkerStore
|
|
from synapse.storage.databases.main.ui_auth import UIAuthWorkerStore
|
|
from synapse.storage.databases.main.user_directory import UserDirectoryStore
|
|
from synapse.types import JsonDict
|
|
from synapse.util.httpresourcetree import create_resource_tree
|
|
|
|
logger = logging.getLogger("synapse.app.generic_worker")
|
|
|
|
|
|
class KeyUploadServlet(RestServlet):
|
|
"""An implementation of the `KeyUploadServlet` that responds to read only
|
|
requests, but otherwise proxies through to the master instance.
|
|
"""
|
|
|
|
PATTERNS = client_patterns("/keys/upload(/(?P<device_id>[^/]+))?$")
|
|
|
|
def __init__(self, hs: HomeServer):
|
|
"""
|
|
Args:
|
|
hs: server
|
|
"""
|
|
super().__init__()
|
|
self.auth = hs.get_auth()
|
|
self.store = hs.get_datastores().main
|
|
self.http_client = hs.get_simple_http_client()
|
|
self.main_uri = hs.config.worker.worker_main_http_uri
|
|
|
|
async def on_POST(
|
|
self, request: SynapseRequest, device_id: Optional[str]
|
|
) -> Tuple[int, JsonDict]:
|
|
requester = await self.auth.get_user_by_req(request, allow_guest=True)
|
|
user_id = requester.user.to_string()
|
|
body = parse_json_object_from_request(request)
|
|
|
|
if device_id is not None:
|
|
# passing the device_id here is deprecated; however, we allow it
|
|
# for now for compatibility with older clients.
|
|
if requester.device_id is not None and device_id != requester.device_id:
|
|
logger.warning(
|
|
"Client uploading keys for a different device "
|
|
"(logged in as %s, uploading for %s)",
|
|
requester.device_id,
|
|
device_id,
|
|
)
|
|
else:
|
|
device_id = requester.device_id
|
|
|
|
if device_id is None:
|
|
raise SynapseError(
|
|
400, "To upload keys, you must pass device_id when authenticating"
|
|
)
|
|
|
|
if body:
|
|
# They're actually trying to upload something, proxy to main synapse.
|
|
|
|
# Proxy headers from the original request, such as the auth headers
|
|
# (in case the access token is there) and the original IP /
|
|
# User-Agent of the request.
|
|
headers = {
|
|
header: request.requestHeaders.getRawHeaders(header, [])
|
|
for header in (b"Authorization", b"User-Agent")
|
|
}
|
|
# Add the previous hop to the X-Forwarded-For header.
|
|
x_forwarded_for = request.requestHeaders.getRawHeaders(
|
|
b"X-Forwarded-For", []
|
|
)
|
|
# we use request.client here, since we want the previous hop, not the
|
|
# original client (as returned by request.getClientAddress()).
|
|
if isinstance(request.client, (address.IPv4Address, address.IPv6Address)):
|
|
previous_host = request.client.host.encode("ascii")
|
|
# If the header exists, add to the comma-separated list of the first
|
|
# instance of the header. Otherwise, generate a new header.
|
|
if x_forwarded_for:
|
|
x_forwarded_for = [x_forwarded_for[0] + b", " + previous_host]
|
|
x_forwarded_for.extend(x_forwarded_for[1:])
|
|
else:
|
|
x_forwarded_for = [previous_host]
|
|
headers[b"X-Forwarded-For"] = x_forwarded_for
|
|
|
|
# Replicate the original X-Forwarded-Proto header. Note that
|
|
# XForwardedForRequest overrides isSecure() to give us the original protocol
|
|
# used by the client, as opposed to the protocol used by our upstream proxy
|
|
# - which is what we want here.
|
|
headers[b"X-Forwarded-Proto"] = [
|
|
b"https" if request.isSecure() else b"http"
|
|
]
|
|
|
|
try:
|
|
result = await self.http_client.post_json_get_json(
|
|
self.main_uri + request.uri.decode("ascii"), body, headers=headers
|
|
)
|
|
except HttpResponseException as e:
|
|
raise e.to_synapse_error() from e
|
|
except RequestSendFailed as e:
|
|
raise SynapseError(502, "Failed to talk to master") from e
|
|
|
|
return 200, result
|
|
else:
|
|
# Just interested in counts.
|
|
result = await self.store.count_e2e_one_time_keys(user_id, device_id)
|
|
return 200, {"one_time_key_counts": result}
|
|
|
|
|
|
class GenericWorkerSlavedStore(
|
|
# FIXME(#3714): We need to add UserDirectoryStore as we write directly
|
|
# rather than going via the correct worker.
|
|
UserDirectoryStore,
|
|
StatsStore,
|
|
UIAuthWorkerStore,
|
|
EndToEndRoomKeyStore,
|
|
PresenceStore,
|
|
SlavedDeviceInboxStore,
|
|
SlavedDeviceStore,
|
|
SlavedReceiptsStore,
|
|
SlavedPushRuleStore,
|
|
SlavedGroupServerStore,
|
|
SlavedAccountDataStore,
|
|
SlavedPusherStore,
|
|
CensorEventsStore,
|
|
ClientIpWorkerStore,
|
|
SlavedEventStore,
|
|
SlavedKeyStore,
|
|
RoomWorkerStore,
|
|
RoomBatchStore,
|
|
DirectoryStore,
|
|
SlavedApplicationServiceStore,
|
|
SlavedRegistrationStore,
|
|
SlavedProfileStore,
|
|
SlavedFilteringStore,
|
|
MonthlyActiveUsersWorkerStore,
|
|
MediaRepositoryStore,
|
|
ServerMetricsStore,
|
|
SearchStore,
|
|
TransactionWorkerStore,
|
|
LockStore,
|
|
SessionStore,
|
|
BaseSlavedStore,
|
|
):
|
|
# Properties that multiple storage classes define. Tell mypy what the
|
|
# expected type is.
|
|
server_name: str
|
|
config: HomeServerConfig
|
|
|
|
|
|
class GenericWorkerServer(HomeServer):
|
|
DATASTORE_CLASS = GenericWorkerSlavedStore # type: ignore
|
|
|
|
def _listen_http(self, listener_config: ListenerConfig) -> None:
|
|
port = listener_config.port
|
|
bind_addresses = listener_config.bind_addresses
|
|
|
|
assert listener_config.http_options is not None
|
|
|
|
site_tag = listener_config.http_options.tag
|
|
if site_tag is None:
|
|
site_tag = str(port)
|
|
|
|
# We always include a health resource.
|
|
resources: Dict[str, Resource] = {"/health": HealthResource()}
|
|
|
|
for res in listener_config.http_options.resources:
|
|
for name in res.names:
|
|
if name == "metrics":
|
|
resources[METRICS_PREFIX] = MetricsResource(RegistryProxy)
|
|
elif name == "client":
|
|
resource = JsonResource(self, canonical_json=False)
|
|
|
|
RegisterRestServlet(self).register(resource)
|
|
RegistrationTokenValidityRestServlet(self).register(resource)
|
|
login.register_servlets(self, resource)
|
|
ThreepidRestServlet(self).register(resource)
|
|
WhoamiRestServlet(self).register(resource)
|
|
DevicesRestServlet(self).register(resource)
|
|
|
|
# Read-only
|
|
KeyUploadServlet(self).register(resource)
|
|
KeyQueryServlet(self).register(resource)
|
|
KeyChangesServlet(self).register(resource)
|
|
OneTimeKeyServlet(self).register(resource)
|
|
|
|
voip.register_servlets(self, resource)
|
|
push_rule.register_servlets(self, resource)
|
|
versions.register_servlets(self, resource)
|
|
|
|
profile.register_servlets(self, resource)
|
|
|
|
sync.register_servlets(self, resource)
|
|
events.register_servlets(self, resource)
|
|
room.register_servlets(self, resource, is_worker=True)
|
|
room.register_deprecated_servlets(self, resource)
|
|
initial_sync.register_servlets(self, resource)
|
|
room_batch.register_servlets(self, resource)
|
|
room_keys.register_servlets(self, resource)
|
|
tags.register_servlets(self, resource)
|
|
account_data.register_servlets(self, resource)
|
|
receipts.register_servlets(self, resource)
|
|
read_marker.register_servlets(self, resource)
|
|
|
|
sendtodevice.register_servlets(self, resource)
|
|
|
|
user_directory.register_servlets(self, resource)
|
|
|
|
presence.register_servlets(self, resource)
|
|
|
|
resources.update({CLIENT_API_PREFIX: resource})
|
|
|
|
resources.update(build_synapse_client_resource_tree(self))
|
|
resources.update({"/.well-known": well_known_resource(self)})
|
|
|
|
elif name == "federation":
|
|
resources.update({FEDERATION_PREFIX: TransportLayerServer(self)})
|
|
elif name == "media":
|
|
if self.config.media.can_load_media_repo:
|
|
media_repo = self.get_media_repository_resource()
|
|
|
|
# We need to serve the admin servlets for media on the
|
|
# worker.
|
|
admin_resource = JsonResource(self, canonical_json=False)
|
|
register_servlets_for_media_repo(self, admin_resource)
|
|
|
|
resources.update(
|
|
{
|
|
MEDIA_R0_PREFIX: media_repo,
|
|
MEDIA_V3_PREFIX: media_repo,
|
|
LEGACY_MEDIA_PREFIX: media_repo,
|
|
"/_synapse/admin": admin_resource,
|
|
}
|
|
)
|
|
else:
|
|
logger.warning(
|
|
"A 'media' listener is configured but the media"
|
|
" repository is disabled. Ignoring."
|
|
)
|
|
|
|
if name == "openid" and "federation" not in res.names:
|
|
# Only load the openid resource separately if federation resource
|
|
# is not specified since federation resource includes openid
|
|
# resource.
|
|
resources.update(
|
|
{
|
|
FEDERATION_PREFIX: TransportLayerServer(
|
|
self, servlet_groups=["openid"]
|
|
)
|
|
}
|
|
)
|
|
|
|
if name in ["keys", "federation"]:
|
|
resources[SERVER_KEY_V2_PREFIX] = KeyApiV2Resource(self)
|
|
|
|
if name == "replication":
|
|
resources[REPLICATION_PREFIX] = ReplicationRestResource(self)
|
|
|
|
# Attach additional resources registered by modules.
|
|
resources.update(self._module_web_resources)
|
|
self._module_web_resources_consumed = True
|
|
|
|
root_resource = create_resource_tree(resources, OptionsResource())
|
|
|
|
_base.listen_tcp(
|
|
bind_addresses,
|
|
port,
|
|
SynapseSite(
|
|
"synapse.access.http.%s" % (site_tag,),
|
|
site_tag,
|
|
listener_config,
|
|
root_resource,
|
|
self.version_string,
|
|
max_request_body_size=max_request_body_size(self.config),
|
|
reactor=self.get_reactor(),
|
|
),
|
|
reactor=self.get_reactor(),
|
|
)
|
|
|
|
logger.info("Synapse worker now listening on port %d", port)
|
|
|
|
def start_listening(self) -> None:
|
|
for listener in self.config.worker.worker_listeners:
|
|
if listener.type == "http":
|
|
self._listen_http(listener)
|
|
elif listener.type == "manhole":
|
|
_base.listen_manhole(
|
|
listener.bind_addresses,
|
|
listener.port,
|
|
manhole_settings=self.config.server.manhole_settings,
|
|
manhole_globals={"hs": self},
|
|
)
|
|
elif listener.type == "metrics":
|
|
if not self.config.metrics.enable_metrics:
|
|
logger.warning(
|
|
"Metrics listener configured, but "
|
|
"enable_metrics is not True!"
|
|
)
|
|
else:
|
|
_base.listen_metrics(listener.bind_addresses, listener.port)
|
|
else:
|
|
logger.warning("Unsupported listener type: %s", listener.type)
|
|
|
|
self.get_replication_command_handler().start_replication(self)
|
|
|
|
|
|
def start(config_options: List[str]) -> None:
|
|
try:
|
|
config = HomeServerConfig.load_config("Synapse worker", config_options)
|
|
except ConfigError as e:
|
|
sys.stderr.write("\n" + str(e) + "\n")
|
|
sys.exit(1)
|
|
|
|
# For backwards compatibility let any of the old app names.
|
|
assert config.worker.worker_app in (
|
|
"synapse.app.appservice",
|
|
"synapse.app.client_reader",
|
|
"synapse.app.event_creator",
|
|
"synapse.app.federation_reader",
|
|
"synapse.app.federation_sender",
|
|
"synapse.app.frontend_proxy",
|
|
"synapse.app.generic_worker",
|
|
"synapse.app.media_repository",
|
|
"synapse.app.pusher",
|
|
"synapse.app.synchrotron",
|
|
"synapse.app.user_dir",
|
|
)
|
|
|
|
synapse.events.USE_FROZEN_DICTS = config.server.use_frozen_dicts
|
|
synapse.util.caches.TRACK_MEMORY_USAGE = config.caches.track_memory_usage
|
|
|
|
if config.server.gc_seconds:
|
|
synapse.metrics.MIN_TIME_BETWEEN_GCS = config.server.gc_seconds
|
|
|
|
hs = GenericWorkerServer(
|
|
config.server.server_name,
|
|
config=config,
|
|
version_string="Synapse/" + get_distribution_version_string("matrix-synapse"),
|
|
)
|
|
|
|
setup_logging(hs, config, use_worker_options=True)
|
|
|
|
try:
|
|
hs.setup()
|
|
|
|
# Ensure the replication streamer is always started in case we write to any
|
|
# streams. Will no-op if no streams can be written to by this worker.
|
|
hs.get_replication_streamer()
|
|
except Exception as e:
|
|
handle_startup_exception(e)
|
|
|
|
register_start(_base.start, hs)
|
|
|
|
# redirect stdio to the logs, if configured.
|
|
if not hs.config.logging.no_redirect_stdio:
|
|
redirect_stdio_to_logs()
|
|
|
|
_base.start_worker_reactor("synapse-generic-worker", config)
|
|
|
|
|
|
def main() -> None:
|
|
with LoggingContext("main"):
|
|
start(sys.argv[1:])
|
|
|
|
|
|
if __name__ == "__main__":
|
|
main()
|