2019-01-22 11:59:27 +01:00
|
|
|
# Copyright 2014-2016 OpenMarket Ltd
|
|
|
|
# Copyright 2019 New Vector Ltd
|
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
|
|
|
import logging
|
2019-01-22 00:27:57 +01:00
|
|
|
import random
|
2019-01-22 11:59:27 +01:00
|
|
|
import time
|
2020-07-23 13:05:57 +02:00
|
|
|
from typing import List
|
2019-01-22 11:59:27 +01:00
|
|
|
|
|
|
|
import attr
|
|
|
|
|
|
|
|
from twisted.internet.error import ConnectError
|
|
|
|
from twisted.names import client, dns
|
|
|
|
from twisted.names.error import DNSNameError, DomainError
|
|
|
|
|
2019-07-03 16:07:04 +02:00
|
|
|
from synapse.logging.context import make_deferred_yieldable
|
2019-01-22 11:59:27 +01:00
|
|
|
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
SERVER_CACHE = {}
|
|
|
|
|
|
|
|
|
2019-08-08 16:30:04 +02:00
|
|
|
@attr.s(slots=True, frozen=True)
|
2020-09-04 12:54:56 +02:00
|
|
|
class Server:
|
2019-01-22 11:59:27 +01:00
|
|
|
"""
|
|
|
|
Our record of an individual server which can be tried to reach a destination.
|
|
|
|
|
|
|
|
Attributes:
|
|
|
|
host (bytes): target hostname
|
|
|
|
port (int):
|
|
|
|
priority (int):
|
|
|
|
weight (int):
|
|
|
|
expires (int): when the cache should expire this record - in *seconds* since
|
|
|
|
the epoch
|
|
|
|
"""
|
2019-06-20 11:32:02 +02:00
|
|
|
|
2019-01-22 11:59:27 +01:00
|
|
|
host = attr.ib()
|
|
|
|
port = attr.ib()
|
|
|
|
priority = attr.ib(default=0)
|
|
|
|
weight = attr.ib(default=0)
|
|
|
|
expires = attr.ib(default=0)
|
|
|
|
|
|
|
|
|
2019-08-08 16:30:04 +02:00
|
|
|
def _sort_server_list(server_list):
|
|
|
|
"""Given a list of SRV records sort them into priority order and shuffle
|
|
|
|
each priority with the given weight.
|
|
|
|
"""
|
|
|
|
priority_map = {}
|
|
|
|
|
|
|
|
for server in server_list:
|
|
|
|
priority_map.setdefault(server.priority, []).append(server)
|
|
|
|
|
|
|
|
results = []
|
|
|
|
for priority in sorted(priority_map):
|
2019-08-20 12:49:44 +02:00
|
|
|
servers = priority_map[priority]
|
2019-08-08 16:30:04 +02:00
|
|
|
|
2019-08-27 14:56:42 +02:00
|
|
|
# This algorithms roughly follows the algorithm described in RFC2782,
|
|
|
|
# changed to remove an off-by-one error.
|
2019-08-20 12:49:44 +02:00
|
|
|
#
|
2019-08-27 14:56:42 +02:00
|
|
|
# N.B. Weights can be zero, which means that they should be picked
|
|
|
|
# rarely.
|
2019-08-20 12:49:44 +02:00
|
|
|
|
|
|
|
total_weight = sum(s.weight for s in servers)
|
2019-08-27 14:56:42 +02:00
|
|
|
|
|
|
|
# Total weight can become zero if there are only zero weight servers
|
|
|
|
# left, which we handle by just shuffling and appending to the results.
|
|
|
|
while servers and total_weight:
|
|
|
|
target_weight = random.randint(1, total_weight)
|
2019-08-08 16:30:04 +02:00
|
|
|
|
|
|
|
for s in servers:
|
|
|
|
target_weight -= s.weight
|
|
|
|
|
|
|
|
if target_weight <= 0:
|
|
|
|
break
|
|
|
|
|
|
|
|
results.append(s)
|
|
|
|
servers.remove(s)
|
2019-08-20 12:49:44 +02:00
|
|
|
total_weight -= s.weight
|
2019-08-08 16:30:04 +02:00
|
|
|
|
2019-08-27 14:56:42 +02:00
|
|
|
if servers:
|
|
|
|
random.shuffle(servers)
|
|
|
|
results.extend(servers)
|
|
|
|
|
2019-08-08 16:30:04 +02:00
|
|
|
return results
|
|
|
|
|
|
|
|
|
2020-09-04 12:54:56 +02:00
|
|
|
class SrvResolver:
|
2019-01-22 18:42:26 +01:00
|
|
|
"""Interface to the dns client to do SRV lookups, with result caching.
|
2019-01-22 11:59:27 +01:00
|
|
|
|
|
|
|
The default resolver in twisted.names doesn't do any caching (it has a CacheResolver,
|
|
|
|
but the cache never gets populated), so we add our own caching layer here.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
dns_client (twisted.internet.interfaces.IResolver): twisted resolver impl
|
|
|
|
cache (dict): cache object
|
2019-01-22 18:42:26 +01:00
|
|
|
get_time (callable): clock implementation. Should return seconds since the epoch
|
2019-01-22 11:59:27 +01:00
|
|
|
"""
|
2019-06-20 11:32:02 +02:00
|
|
|
|
2019-01-22 18:42:26 +01:00
|
|
|
def __init__(self, dns_client=client, cache=SERVER_CACHE, get_time=time.time):
|
|
|
|
self._dns_client = dns_client
|
|
|
|
self._cache = cache
|
|
|
|
self._get_time = get_time
|
|
|
|
|
2020-07-23 13:05:57 +02:00
|
|
|
async def resolve_service(self, service_name: bytes) -> List[Server]:
|
2019-01-22 18:42:26 +01:00
|
|
|
"""Look up a SRV record
|
|
|
|
|
|
|
|
Args:
|
|
|
|
service_name (bytes): record to look up
|
|
|
|
|
|
|
|
Returns:
|
2020-07-23 13:05:57 +02:00
|
|
|
a list of the SRV records, or an empty list if none found
|
2019-01-22 18:42:26 +01:00
|
|
|
"""
|
|
|
|
now = int(self._get_time())
|
|
|
|
|
|
|
|
if not isinstance(service_name, bytes):
|
|
|
|
raise TypeError("%r is not a byte string" % (service_name,))
|
|
|
|
|
|
|
|
cache_entry = self._cache.get(service_name, None)
|
2019-01-22 11:59:27 +01:00
|
|
|
if cache_entry:
|
2019-01-22 18:42:26 +01:00
|
|
|
if all(s.expires > now for s in cache_entry):
|
|
|
|
servers = list(cache_entry)
|
2019-08-08 16:30:04 +02:00
|
|
|
return _sort_server_list(servers)
|
2019-01-22 18:42:26 +01:00
|
|
|
|
|
|
|
try:
|
2020-07-23 13:05:57 +02:00
|
|
|
answers, _, _ = await make_deferred_yieldable(
|
2019-06-20 11:32:02 +02:00
|
|
|
self._dns_client.lookupService(service_name)
|
2019-01-22 11:59:27 +01:00
|
|
|
)
|
2019-01-22 18:42:26 +01:00
|
|
|
except DNSNameError:
|
|
|
|
# TODO: cache this. We can get the SOA out of the exception, and use
|
|
|
|
# the negative-TTL value.
|
2019-07-23 15:00:55 +02:00
|
|
|
return []
|
2019-01-22 18:42:26 +01:00
|
|
|
except DomainError as e:
|
|
|
|
# We failed to resolve the name (other than a NameError)
|
|
|
|
# Try something in the cache, else rereaise
|
|
|
|
cache_entry = self._cache.get(service_name, None)
|
|
|
|
if cache_entry:
|
2019-10-31 11:23:24 +01:00
|
|
|
logger.warning(
|
2019-06-20 11:32:02 +02:00
|
|
|
"Failed to resolve %r, falling back to cache. %r", service_name, e
|
2019-01-22 18:42:26 +01:00
|
|
|
)
|
2019-07-23 15:00:55 +02:00
|
|
|
return list(cache_entry)
|
2019-01-22 18:42:26 +01:00
|
|
|
else:
|
|
|
|
raise e
|
|
|
|
|
2019-06-20 11:32:02 +02:00
|
|
|
if (
|
|
|
|
len(answers) == 1
|
|
|
|
and answers[0].type == dns.SRV
|
|
|
|
and answers[0].payload
|
|
|
|
and answers[0].payload.target == dns.Name(b".")
|
|
|
|
):
|
2019-01-22 18:42:26 +01:00
|
|
|
raise ConnectError("Service %s unavailable" % service_name)
|
|
|
|
|
|
|
|
servers = []
|
|
|
|
|
|
|
|
for answer in answers:
|
|
|
|
if answer.type != dns.SRV or not answer.payload:
|
|
|
|
continue
|
|
|
|
|
|
|
|
payload = answer.payload
|
|
|
|
|
2019-06-20 11:32:02 +02:00
|
|
|
servers.append(
|
|
|
|
Server(
|
|
|
|
host=payload.target.name,
|
|
|
|
port=payload.port,
|
|
|
|
priority=payload.priority,
|
|
|
|
weight=payload.weight,
|
|
|
|
expires=now + answer.ttl,
|
|
|
|
)
|
|
|
|
)
|
2019-01-22 18:42:26 +01:00
|
|
|
|
|
|
|
self._cache[service_name] = list(servers)
|
2019-08-08 16:30:04 +02:00
|
|
|
return _sort_server_list(servers)
|