2018-07-18 11:50:33 +02:00
|
|
|
# Copyright 2018 New Vector Ltd
|
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
2018-09-20 18:05:34 +02:00
|
|
|
import logging
|
2018-08-20 18:16:58 +02:00
|
|
|
import threading
|
2019-09-24 15:16:16 +02:00
|
|
|
from functools import wraps
|
2021-04-21 11:03:31 +02:00
|
|
|
from typing import TYPE_CHECKING, Dict, Optional, Set, Union
|
2018-07-18 11:50:33 +02:00
|
|
|
|
2020-05-29 14:25:32 +02:00
|
|
|
from prometheus_client.core import REGISTRY, Counter, Gauge
|
2018-07-18 11:50:33 +02:00
|
|
|
|
|
|
|
from twisted.internet import defer
|
|
|
|
|
2019-07-03 16:07:04 +02:00
|
|
|
from synapse.logging.context import LoggingContext, PreserveLoggingContext
|
2020-10-26 10:30:19 +01:00
|
|
|
from synapse.logging.opentracing import noop_context_manager, start_active_span
|
2020-12-11 20:05:15 +01:00
|
|
|
from synapse.util.async_helpers import maybe_awaitable
|
2018-07-18 11:50:33 +02:00
|
|
|
|
2020-05-29 14:25:32 +02:00
|
|
|
if TYPE_CHECKING:
|
|
|
|
import resource
|
|
|
|
|
|
|
|
|
2018-09-20 18:05:34 +02:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
2018-07-18 11:50:33 +02:00
|
|
|
_background_process_start_count = Counter(
|
|
|
|
"synapse_background_process_start_count",
|
|
|
|
"Number of background processes started",
|
|
|
|
["name"],
|
|
|
|
)
|
|
|
|
|
2020-05-29 14:25:32 +02:00
|
|
|
_background_process_in_flight_count = Gauge(
|
|
|
|
"synapse_background_process_in_flight_count",
|
|
|
|
"Number of background processes in flight",
|
|
|
|
labelnames=["name"],
|
|
|
|
)
|
|
|
|
|
2018-07-18 11:50:33 +02:00
|
|
|
# we set registry=None in all of these to stop them getting registered with
|
|
|
|
# the default registry. Instead we collect them all via the CustomCollector,
|
|
|
|
# which ensures that we can update them before they are collected.
|
|
|
|
#
|
|
|
|
_background_process_ru_utime = Counter(
|
|
|
|
"synapse_background_process_ru_utime_seconds",
|
|
|
|
"User CPU time used by background processes, in seconds",
|
|
|
|
["name"],
|
|
|
|
registry=None,
|
|
|
|
)
|
|
|
|
|
|
|
|
_background_process_ru_stime = Counter(
|
|
|
|
"synapse_background_process_ru_stime_seconds",
|
|
|
|
"System CPU time used by background processes, in seconds",
|
|
|
|
["name"],
|
|
|
|
registry=None,
|
|
|
|
)
|
|
|
|
|
|
|
|
_background_process_db_txn_count = Counter(
|
|
|
|
"synapse_background_process_db_txn_count",
|
|
|
|
"Number of database transactions done by background processes",
|
|
|
|
["name"],
|
|
|
|
registry=None,
|
|
|
|
)
|
|
|
|
|
|
|
|
_background_process_db_txn_duration = Counter(
|
|
|
|
"synapse_background_process_db_txn_duration_seconds",
|
2019-06-20 11:32:02 +02:00
|
|
|
(
|
|
|
|
"Seconds spent by background processes waiting for database "
|
|
|
|
"transactions, excluding scheduling time"
|
|
|
|
),
|
2018-07-18 11:50:33 +02:00
|
|
|
["name"],
|
|
|
|
registry=None,
|
|
|
|
)
|
|
|
|
|
|
|
|
_background_process_db_sched_duration = Counter(
|
|
|
|
"synapse_background_process_db_sched_duration_seconds",
|
|
|
|
"Seconds spent by background processes waiting for database connections",
|
|
|
|
["name"],
|
|
|
|
registry=None,
|
|
|
|
)
|
|
|
|
|
|
|
|
# map from description to a counter, so that we can name our logcontexts
|
|
|
|
# incrementally. (It actually duplicates _background_process_start_count, but
|
|
|
|
# it's much simpler to do so than to try to combine them.)
|
2020-03-19 11:00:24 +01:00
|
|
|
_background_process_counts = {} # type: Dict[str, int]
|
2018-07-18 11:50:33 +02:00
|
|
|
|
2020-05-29 14:25:32 +02:00
|
|
|
# Set of all running background processes that became active active since the
|
|
|
|
# last time metrics were scraped (i.e. background processes that performed some
|
|
|
|
# work since the last scrape.)
|
2018-07-18 11:50:33 +02:00
|
|
|
#
|
2020-05-29 14:25:32 +02:00
|
|
|
# We do it like this to handle the case where we have a large number of
|
|
|
|
# background processes stacking up behind a lock or linearizer, where we then
|
|
|
|
# only need to iterate over and update metrics for the process that have
|
|
|
|
# actually been active and can ignore the idle ones.
|
|
|
|
_background_processes_active_since_last_scrape = set() # type: Set[_BackgroundProcess]
|
2018-07-18 11:50:33 +02:00
|
|
|
|
2020-05-29 14:25:32 +02:00
|
|
|
# A lock that covers the above set and dict
|
2018-08-20 18:16:58 +02:00
|
|
|
_bg_metrics_lock = threading.Lock()
|
|
|
|
|
2018-07-18 11:50:33 +02:00
|
|
|
|
2020-09-04 12:54:56 +02:00
|
|
|
class _Collector:
|
2018-07-18 11:50:33 +02:00
|
|
|
"""A custom metrics collector for the background process metrics.
|
|
|
|
|
|
|
|
Ensures that all of the metrics are up-to-date with any in-flight processes
|
|
|
|
before they are returned.
|
|
|
|
"""
|
2019-06-20 11:32:02 +02:00
|
|
|
|
2018-07-18 11:50:33 +02:00
|
|
|
def collect(self):
|
2020-05-29 14:25:32 +02:00
|
|
|
global _background_processes_active_since_last_scrape
|
2018-07-18 11:50:33 +02:00
|
|
|
|
2020-05-29 14:25:32 +02:00
|
|
|
# We swap out the _background_processes set with an empty one so that
|
|
|
|
# we can safely iterate over the set without holding the lock.
|
2018-08-20 18:16:58 +02:00
|
|
|
with _bg_metrics_lock:
|
2020-05-29 14:25:32 +02:00
|
|
|
_background_processes_copy = _background_processes_active_since_last_scrape
|
|
|
|
_background_processes_active_since_last_scrape = set()
|
2018-08-20 18:16:58 +02:00
|
|
|
|
2020-05-29 14:25:32 +02:00
|
|
|
for process in _background_processes_copy:
|
|
|
|
process.update_metrics()
|
2018-07-18 11:50:33 +02:00
|
|
|
|
|
|
|
# now we need to run collect() over each of the static Counters, and
|
|
|
|
# yield each metric they return.
|
|
|
|
for m in (
|
2019-06-20 11:32:02 +02:00
|
|
|
_background_process_ru_utime,
|
|
|
|
_background_process_ru_stime,
|
|
|
|
_background_process_db_txn_count,
|
|
|
|
_background_process_db_txn_duration,
|
|
|
|
_background_process_db_sched_duration,
|
2018-07-18 11:50:33 +02:00
|
|
|
):
|
|
|
|
for r in m.collect():
|
|
|
|
yield r
|
|
|
|
|
|
|
|
|
|
|
|
REGISTRY.register(_Collector())
|
|
|
|
|
|
|
|
|
2020-09-04 12:54:56 +02:00
|
|
|
class _BackgroundProcess:
|
2018-07-18 11:50:33 +02:00
|
|
|
def __init__(self, desc, ctx):
|
|
|
|
self.desc = desc
|
|
|
|
self._context = ctx
|
|
|
|
self._reported_stats = None
|
|
|
|
|
|
|
|
def update_metrics(self):
|
|
|
|
"""Updates the metrics with values from this process."""
|
|
|
|
new_stats = self._context.get_resource_usage()
|
|
|
|
if self._reported_stats is None:
|
|
|
|
diff = new_stats
|
|
|
|
else:
|
|
|
|
diff = new_stats - self._reported_stats
|
|
|
|
self._reported_stats = new_stats
|
|
|
|
|
|
|
|
_background_process_ru_utime.labels(self.desc).inc(diff.ru_utime)
|
|
|
|
_background_process_ru_stime.labels(self.desc).inc(diff.ru_stime)
|
2019-06-20 11:32:02 +02:00
|
|
|
_background_process_db_txn_count.labels(self.desc).inc(diff.db_txn_count)
|
2018-07-18 11:50:33 +02:00
|
|
|
_background_process_db_txn_duration.labels(self.desc).inc(
|
2019-06-20 11:32:02 +02:00
|
|
|
diff.db_txn_duration_sec
|
2018-07-18 11:50:33 +02:00
|
|
|
)
|
|
|
|
_background_process_db_sched_duration.labels(self.desc).inc(
|
2019-06-20 11:32:02 +02:00
|
|
|
diff.db_sched_duration_sec
|
2018-07-18 11:50:33 +02:00
|
|
|
)
|
|
|
|
|
|
|
|
|
2020-10-26 10:30:19 +01:00
|
|
|
def run_as_background_process(desc: str, func, *args, bg_start_span=True, **kwargs):
|
2018-07-18 11:50:33 +02:00
|
|
|
"""Run the given function in its own logcontext, with resource metrics
|
|
|
|
|
|
|
|
This should be used to wrap processes which are fired off to run in the
|
|
|
|
background, instead of being associated with a particular request.
|
|
|
|
|
2018-07-26 12:44:26 +02:00
|
|
|
It returns a Deferred which completes when the function completes, but it doesn't
|
|
|
|
follow the synapse logcontext rules, which makes it appropriate for passing to
|
|
|
|
clock.looping_call and friends (or for firing-and-forgetting in the middle of a
|
2020-08-19 13:09:24 +02:00
|
|
|
normal synapse async function).
|
2018-07-26 12:44:26 +02:00
|
|
|
|
2018-07-18 11:50:33 +02:00
|
|
|
Args:
|
2020-08-06 14:20:42 +02:00
|
|
|
desc: a description for this background process type
|
2019-09-25 19:17:39 +02:00
|
|
|
func: a function, which may return a Deferred or a coroutine
|
2020-10-26 10:30:19 +01:00
|
|
|
bg_start_span: Whether to start an opentracing span. Defaults to True.
|
|
|
|
Should only be disabled for processes that will not log to or tag
|
|
|
|
a span.
|
2018-07-18 11:50:33 +02:00
|
|
|
args: positional args for func
|
|
|
|
kwargs: keyword args for func
|
|
|
|
|
2018-07-26 12:44:26 +02:00
|
|
|
Returns: Deferred which returns the result of func, but note that it does not
|
|
|
|
follow the synapse logcontext rules.
|
2018-07-18 11:50:33 +02:00
|
|
|
"""
|
2019-06-20 11:32:02 +02:00
|
|
|
|
2020-08-06 14:20:42 +02:00
|
|
|
async def run():
|
2018-08-20 18:16:58 +02:00
|
|
|
with _bg_metrics_lock:
|
|
|
|
count = _background_process_counts.get(desc, 0)
|
|
|
|
_background_process_counts[desc] = count + 1
|
|
|
|
|
2018-07-18 11:50:33 +02:00
|
|
|
_background_process_start_count.labels(desc).inc()
|
2020-05-29 14:25:32 +02:00
|
|
|
_background_process_in_flight_count.labels(desc).inc()
|
2018-07-18 11:50:33 +02:00
|
|
|
|
2021-04-21 11:03:31 +02:00
|
|
|
with BackgroundProcessLoggingContext(desc, count) as context:
|
2018-07-18 11:50:33 +02:00
|
|
|
try:
|
2020-10-26 10:30:19 +01:00
|
|
|
ctx = noop_context_manager()
|
|
|
|
if bg_start_span:
|
2021-04-08 14:01:14 +02:00
|
|
|
ctx = start_active_span(desc, tags={"request_id": str(context)})
|
2020-10-26 10:30:19 +01:00
|
|
|
with ctx:
|
2020-12-11 20:05:15 +01:00
|
|
|
return await maybe_awaitable(func(*args, **kwargs))
|
2018-09-20 17:15:21 +02:00
|
|
|
except Exception:
|
2020-08-06 14:20:42 +02:00
|
|
|
logger.exception(
|
2021-02-16 23:32:34 +01:00
|
|
|
"Background process '%s' threw an exception",
|
|
|
|
desc,
|
2020-07-09 14:01:33 +02:00
|
|
|
)
|
2018-07-18 11:50:33 +02:00
|
|
|
finally:
|
2020-05-29 14:25:32 +02:00
|
|
|
_background_process_in_flight_count.labels(desc).dec()
|
2018-07-18 11:50:33 +02:00
|
|
|
|
|
|
|
with PreserveLoggingContext():
|
2020-08-06 14:20:42 +02:00
|
|
|
# Note that we return a Deferred here so that it can be used in a
|
|
|
|
# looping_call and other places that expect a Deferred.
|
|
|
|
return defer.ensureDeferred(run())
|
2019-09-24 15:16:16 +02:00
|
|
|
|
|
|
|
|
|
|
|
def wrap_as_background_process(desc):
|
|
|
|
"""Decorator that wraps a function that gets called as a background
|
|
|
|
process.
|
|
|
|
|
|
|
|
Equivalent of calling the function with `run_as_background_process`
|
|
|
|
"""
|
|
|
|
|
|
|
|
def wrap_as_background_process_inner(func):
|
|
|
|
@wraps(func)
|
|
|
|
def wrap_as_background_process_inner_2(*args, **kwargs):
|
|
|
|
return run_as_background_process(desc, func, *args, **kwargs)
|
|
|
|
|
|
|
|
return wrap_as_background_process_inner_2
|
|
|
|
|
|
|
|
return wrap_as_background_process_inner
|
2020-05-29 14:25:32 +02:00
|
|
|
|
|
|
|
|
|
|
|
class BackgroundProcessLoggingContext(LoggingContext):
|
|
|
|
"""A logging context that tracks in flight metrics for background
|
|
|
|
processes.
|
|
|
|
"""
|
|
|
|
|
2021-04-20 15:19:00 +02:00
|
|
|
__slots__ = ["_proc"]
|
2020-05-29 14:25:32 +02:00
|
|
|
|
2021-04-21 11:03:31 +02:00
|
|
|
def __init__(self, name: str, instance_id: Optional[Union[int, str]] = None):
|
|
|
|
"""
|
|
|
|
|
|
|
|
Args:
|
|
|
|
name: The name of the background process. Each distinct `name` gets a
|
|
|
|
separate prometheus time series.
|
|
|
|
|
|
|
|
instance_id: an identifer to add to `name` to distinguish this instance of
|
|
|
|
the named background process in the logs. If this is `None`, one is
|
|
|
|
made up based on id(self).
|
|
|
|
"""
|
|
|
|
if instance_id is None:
|
|
|
|
instance_id = id(self)
|
|
|
|
super().__init__("%s-%s" % (name, instance_id))
|
2020-05-29 14:25:32 +02:00
|
|
|
self._proc = _BackgroundProcess(name, self)
|
|
|
|
|
|
|
|
def start(self, rusage: "Optional[resource._RUsage]"):
|
2021-02-16 23:32:34 +01:00
|
|
|
"""Log context has started running (again)."""
|
2020-05-29 14:25:32 +02:00
|
|
|
|
|
|
|
super().start(rusage)
|
|
|
|
|
|
|
|
# We've become active again so we make sure we're in the list of active
|
|
|
|
# procs. (Note that "start" here means we've become active, as opposed
|
|
|
|
# to starting for the first time.)
|
|
|
|
with _bg_metrics_lock:
|
|
|
|
_background_processes_active_since_last_scrape.add(self._proc)
|
|
|
|
|
|
|
|
def __exit__(self, type, value, traceback) -> None:
|
2021-02-16 23:32:34 +01:00
|
|
|
"""Log context has finished."""
|
2020-05-29 14:25:32 +02:00
|
|
|
|
|
|
|
super().__exit__(type, value, traceback)
|
|
|
|
|
2020-10-23 18:38:40 +02:00
|
|
|
# The background process has finished. We explicitly remove and manually
|
2020-05-29 14:25:32 +02:00
|
|
|
# update the metrics here so that if nothing is scraping metrics the set
|
|
|
|
# doesn't infinitely grow.
|
|
|
|
with _bg_metrics_lock:
|
|
|
|
_background_processes_active_since_last_scrape.discard(self._proc)
|
|
|
|
|
|
|
|
self._proc.update_metrics()
|