2015-01-06 08:21:39 -05:00
|
|
|
#
|
2023-11-21 15:29:58 -05:00
|
|
|
# This file is licensed under the Affero General Public License (AGPL) version 3.
|
|
|
|
#
|
2024-01-23 06:26:48 -05:00
|
|
|
# Copyright 2019 The Matrix.org Foundation C.I.C.
|
|
|
|
# Copyright 2014-2016 OpenMarket Ltd
|
2023-11-21 15:29:58 -05:00
|
|
|
# Copyright (C) 2023 New Vector, Ltd
|
|
|
|
#
|
|
|
|
# This program is free software: you can redistribute it and/or modify
|
|
|
|
# it under the terms of the GNU Affero General Public License as
|
|
|
|
# published by the Free Software Foundation, either version 3 of the
|
|
|
|
# License, or (at your option) any later version.
|
|
|
|
#
|
|
|
|
# See the GNU Affero General Public License for more details:
|
|
|
|
# <https://www.gnu.org/licenses/agpl-3.0.html>.
|
|
|
|
#
|
|
|
|
# Originally licensed under the Apache License, Version 2.0:
|
|
|
|
# <http://www.apache.org/licenses/LICENSE-2.0>.
|
|
|
|
#
|
|
|
|
# [This file includes modifications made by New Vector Limited]
|
2015-01-06 08:21:39 -05:00
|
|
|
#
|
|
|
|
#
|
|
|
|
|
2017-03-17 11:11:26 -04:00
|
|
|
""" Thread-local-alike tracking of log contexts within synapse
|
|
|
|
|
|
|
|
This module provides objects and utilities for tracking contexts through
|
|
|
|
synapse code, so that log lines can include a request identifier, and so that
|
|
|
|
CPU and database activity can be accounted for against the request that caused
|
|
|
|
them.
|
|
|
|
|
|
|
|
See doc/log_contexts.rst for details on how this works.
|
|
|
|
"""
|
2014-10-29 21:21:33 -04:00
|
|
|
import logging
|
2018-07-09 02:09:20 -04:00
|
|
|
import threading
|
2021-07-22 07:00:16 -04:00
|
|
|
import typing
|
2020-05-05 09:21:34 -04:00
|
|
|
import warnings
|
2021-12-14 12:35:28 -05:00
|
|
|
from types import TracebackType
|
|
|
|
from typing import (
|
|
|
|
TYPE_CHECKING,
|
|
|
|
Awaitable,
|
|
|
|
Callable,
|
|
|
|
Optional,
|
|
|
|
Tuple,
|
|
|
|
Type,
|
|
|
|
TypeVar,
|
|
|
|
Union,
|
|
|
|
overload,
|
|
|
|
)
|
2020-03-07 12:57:26 -05:00
|
|
|
|
2021-04-08 08:01:14 -04:00
|
|
|
import attr
|
2022-03-08 10:58:14 -05:00
|
|
|
from typing_extensions import Literal, ParamSpec
|
2018-07-09 02:09:20 -04:00
|
|
|
|
2018-10-23 08:12:32 -04:00
|
|
|
from twisted.internet import defer, threads
|
2021-12-14 12:35:28 -05:00
|
|
|
from twisted.python.threadpool import ThreadPool
|
2014-10-29 21:21:33 -04:00
|
|
|
|
2020-03-07 12:57:26 -05:00
|
|
|
if TYPE_CHECKING:
|
|
|
|
from synapse.logging.scopecontextmanager import _LogContextScope
|
2021-12-14 12:35:28 -05:00
|
|
|
from synapse.types import ISynapseReactor
|
2020-03-07 12:57:26 -05:00
|
|
|
|
2014-11-20 12:10:37 -05:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
2015-12-03 16:03:01 -05:00
|
|
|
try:
|
|
|
|
import resource
|
2015-12-04 06:53:38 -05:00
|
|
|
|
|
|
|
# Python doesn't ship with a definition of RUSAGE_THREAD but it's defined
|
|
|
|
# to be 1 on linux so we hard code it.
|
2015-12-03 16:03:01 -05:00
|
|
|
RUSAGE_THREAD = 1
|
2015-12-04 06:53:38 -05:00
|
|
|
|
|
|
|
# If the system doesn't support RUSAGE_THREAD then this should throw an
|
|
|
|
# exception.
|
2015-12-03 16:03:01 -05:00
|
|
|
resource.getrusage(RUSAGE_THREAD)
|
2015-12-04 06:34:05 -05:00
|
|
|
|
2019-09-18 16:55:37 -04:00
|
|
|
is_thread_resource_usage_supported = True
|
|
|
|
|
2021-10-08 09:49:41 -04:00
|
|
|
def get_thread_resource_usage() -> "Optional[resource.struct_rusage]":
|
2015-12-03 16:03:01 -05:00
|
|
|
return resource.getrusage(RUSAGE_THREAD)
|
2019-06-20 05:32:02 -04:00
|
|
|
|
2017-10-23 10:52:32 -04:00
|
|
|
except Exception:
|
2015-12-04 06:53:38 -05:00
|
|
|
# If the system doesn't support resource.getrusage(RUSAGE_THREAD) then we
|
2019-09-18 16:55:37 -04:00
|
|
|
# won't track resource usage.
|
|
|
|
is_thread_resource_usage_supported = False
|
|
|
|
|
2021-10-08 09:49:41 -04:00
|
|
|
def get_thread_resource_usage() -> "Optional[resource.struct_rusage]":
|
2015-12-03 16:03:01 -05:00
|
|
|
return None
|
|
|
|
|
2014-10-30 06:13:46 -04:00
|
|
|
|
2020-09-28 12:58:33 -04:00
|
|
|
# a hook which can be set during testing to assert that we aren't abusing logcontexts.
|
2021-12-14 12:35:28 -05:00
|
|
|
def logcontext_error(msg: str) -> None:
|
2020-09-28 12:58:33 -04:00
|
|
|
logger.warning(msg)
|
|
|
|
|
|
|
|
|
2019-07-03 08:40:45 -04:00
|
|
|
# get an id for the current thread.
|
|
|
|
#
|
|
|
|
# threading.get_ident doesn't actually return an OS-level tid, and annoyingly,
|
|
|
|
# on Linux it actually returns the same value either side of a fork() call. However
|
|
|
|
# we only fork in one place, so it's not worth the hoop-jumping to get a real tid.
|
|
|
|
#
|
|
|
|
get_thread_id = threading.get_ident
|
|
|
|
|
|
|
|
|
2020-09-04 06:54:56 -04:00
|
|
|
class ContextResourceUsage:
|
2018-07-10 08:56:07 -04:00
|
|
|
"""Object for tracking the resources used by a log context
|
|
|
|
|
|
|
|
Attributes:
|
|
|
|
ru_utime (float): user CPU time (in seconds)
|
|
|
|
ru_stime (float): system CPU time (in seconds)
|
|
|
|
db_txn_count (int): number of database transactions done
|
|
|
|
db_sched_duration_sec (float): amount of time spent waiting for a
|
|
|
|
database connection
|
|
|
|
db_txn_duration_sec (float): amount of time spent doing database
|
|
|
|
transactions (excluding scheduling time)
|
|
|
|
evt_db_fetch_count (int): number of events requested from the database
|
|
|
|
"""
|
|
|
|
|
|
|
|
__slots__ = [
|
2019-06-20 05:32:02 -04:00
|
|
|
"ru_stime",
|
|
|
|
"ru_utime",
|
|
|
|
"db_txn_count",
|
|
|
|
"db_txn_duration_sec",
|
|
|
|
"db_sched_duration_sec",
|
2018-07-10 08:56:07 -04:00
|
|
|
"evt_db_fetch_count",
|
|
|
|
]
|
|
|
|
|
2020-03-07 12:57:26 -05:00
|
|
|
def __init__(self, copy_from: "Optional[ContextResourceUsage]" = None) -> None:
|
2018-07-10 08:56:07 -04:00
|
|
|
"""Create a new ContextResourceUsage
|
|
|
|
|
|
|
|
Args:
|
2022-11-16 10:25:24 -05:00
|
|
|
copy_from: if not None, an object to copy stats from
|
2018-07-10 08:56:07 -04:00
|
|
|
"""
|
|
|
|
if copy_from is None:
|
|
|
|
self.reset()
|
|
|
|
else:
|
2020-03-07 12:57:26 -05:00
|
|
|
# FIXME: mypy can't infer the types set via reset() above, so specify explicitly for now
|
2021-07-15 06:02:43 -04:00
|
|
|
self.ru_utime: float = copy_from.ru_utime
|
|
|
|
self.ru_stime: float = copy_from.ru_stime
|
|
|
|
self.db_txn_count: int = copy_from.db_txn_count
|
2018-07-10 08:56:07 -04:00
|
|
|
|
2021-07-15 06:02:43 -04:00
|
|
|
self.db_txn_duration_sec: float = copy_from.db_txn_duration_sec
|
|
|
|
self.db_sched_duration_sec: float = copy_from.db_sched_duration_sec
|
|
|
|
self.evt_db_fetch_count: int = copy_from.evt_db_fetch_count
|
2018-07-10 08:56:07 -04:00
|
|
|
|
2020-03-07 12:57:26 -05:00
|
|
|
def copy(self) -> "ContextResourceUsage":
|
2018-07-10 08:56:07 -04:00
|
|
|
return ContextResourceUsage(copy_from=self)
|
|
|
|
|
2020-03-07 12:57:26 -05:00
|
|
|
def reset(self) -> None:
|
2019-06-20 05:32:02 -04:00
|
|
|
self.ru_stime = 0.0
|
|
|
|
self.ru_utime = 0.0
|
2018-07-10 08:56:07 -04:00
|
|
|
self.db_txn_count = 0
|
|
|
|
|
2020-03-07 12:57:26 -05:00
|
|
|
self.db_txn_duration_sec = 0.0
|
|
|
|
self.db_sched_duration_sec = 0.0
|
2018-07-10 08:56:07 -04:00
|
|
|
self.evt_db_fetch_count = 0
|
|
|
|
|
2020-03-07 12:57:26 -05:00
|
|
|
def __repr__(self) -> str:
|
2019-06-20 05:32:02 -04:00
|
|
|
return (
|
|
|
|
"<ContextResourceUsage ru_stime='%r', ru_utime='%r', "
|
|
|
|
"db_txn_count='%r', db_txn_duration_sec='%r', "
|
|
|
|
"db_sched_duration_sec='%r', evt_db_fetch_count='%r'>"
|
|
|
|
) % (
|
|
|
|
self.ru_stime,
|
|
|
|
self.ru_utime,
|
|
|
|
self.db_txn_count,
|
|
|
|
self.db_txn_duration_sec,
|
|
|
|
self.db_sched_duration_sec,
|
|
|
|
self.evt_db_fetch_count,
|
|
|
|
)
|
2018-07-19 06:58:18 -04:00
|
|
|
|
2020-03-07 12:57:26 -05:00
|
|
|
def __iadd__(self, other: "ContextResourceUsage") -> "ContextResourceUsage":
|
2018-07-10 08:56:07 -04:00
|
|
|
"""Add another ContextResourceUsage's stats to this one's.
|
|
|
|
|
|
|
|
Args:
|
2022-11-16 10:25:24 -05:00
|
|
|
other: the other resource usage object
|
2018-07-10 08:56:07 -04:00
|
|
|
"""
|
|
|
|
self.ru_utime += other.ru_utime
|
|
|
|
self.ru_stime += other.ru_stime
|
|
|
|
self.db_txn_count += other.db_txn_count
|
|
|
|
self.db_txn_duration_sec += other.db_txn_duration_sec
|
|
|
|
self.db_sched_duration_sec += other.db_sched_duration_sec
|
|
|
|
self.evt_db_fetch_count += other.evt_db_fetch_count
|
|
|
|
return self
|
|
|
|
|
2020-03-07 12:57:26 -05:00
|
|
|
def __isub__(self, other: "ContextResourceUsage") -> "ContextResourceUsage":
|
2018-07-10 08:56:07 -04:00
|
|
|
self.ru_utime -= other.ru_utime
|
|
|
|
self.ru_stime -= other.ru_stime
|
|
|
|
self.db_txn_count -= other.db_txn_count
|
|
|
|
self.db_txn_duration_sec -= other.db_txn_duration_sec
|
|
|
|
self.db_sched_duration_sec -= other.db_sched_duration_sec
|
|
|
|
self.evt_db_fetch_count -= other.evt_db_fetch_count
|
|
|
|
return self
|
|
|
|
|
2020-03-07 12:57:26 -05:00
|
|
|
def __add__(self, other: "ContextResourceUsage") -> "ContextResourceUsage":
|
2018-07-10 08:56:07 -04:00
|
|
|
res = ContextResourceUsage(copy_from=self)
|
|
|
|
res += other
|
|
|
|
return res
|
|
|
|
|
2020-03-07 12:57:26 -05:00
|
|
|
def __sub__(self, other: "ContextResourceUsage") -> "ContextResourceUsage":
|
2018-07-10 08:56:07 -04:00
|
|
|
res = ContextResourceUsage(copy_from=self)
|
|
|
|
res -= other
|
|
|
|
return res
|
|
|
|
|
|
|
|
|
2022-01-13 08:49:28 -05:00
|
|
|
@attr.s(slots=True, auto_attribs=True)
|
2021-04-08 08:01:14 -04:00
|
|
|
class ContextRequest:
|
|
|
|
"""
|
|
|
|
A bundle of attributes from the SynapseRequest object.
|
|
|
|
|
|
|
|
This exists to:
|
|
|
|
|
|
|
|
* Avoid a cycle between LoggingContext and SynapseRequest.
|
|
|
|
* Be a single variable that can be passed from parent LoggingContexts to
|
|
|
|
their children.
|
|
|
|
"""
|
|
|
|
|
2022-01-13 08:49:28 -05:00
|
|
|
request_id: str
|
|
|
|
ip_address: str
|
|
|
|
site_tag: str
|
|
|
|
requester: Optional[str]
|
|
|
|
authenticated_entity: Optional[str]
|
|
|
|
method: str
|
|
|
|
url: str
|
|
|
|
protocol: str
|
|
|
|
user_agent: str
|
2021-04-08 08:01:14 -04:00
|
|
|
|
|
|
|
|
2020-03-24 10:45:33 -04:00
|
|
|
LoggingContextOrSentinel = Union["LoggingContext", "_Sentinel"]
|
|
|
|
|
|
|
|
|
2020-09-04 06:54:56 -04:00
|
|
|
class _Sentinel:
|
2020-03-24 10:45:33 -04:00
|
|
|
"""Sentinel to represent the root context"""
|
|
|
|
|
|
|
|
__slots__ = ["previous_context", "finished", "request", "scope", "tag"]
|
|
|
|
|
|
|
|
def __init__(self) -> None:
|
|
|
|
# Minimal set for compatibility with LoggingContext
|
|
|
|
self.previous_context = None
|
|
|
|
self.finished = False
|
|
|
|
self.request = None
|
|
|
|
self.scope = None
|
|
|
|
self.tag = None
|
|
|
|
|
2021-10-28 09:14:42 -04:00
|
|
|
def __str__(self) -> str:
|
2020-03-24 10:45:33 -04:00
|
|
|
return "sentinel"
|
|
|
|
|
2021-12-14 12:35:28 -05:00
|
|
|
def start(self, rusage: "Optional[resource.struct_rusage]") -> None:
|
2020-03-24 10:45:33 -04:00
|
|
|
pass
|
|
|
|
|
2021-12-14 12:35:28 -05:00
|
|
|
def stop(self, rusage: "Optional[resource.struct_rusage]") -> None:
|
2020-03-24 10:45:33 -04:00
|
|
|
pass
|
|
|
|
|
2021-12-14 12:35:28 -05:00
|
|
|
def add_database_transaction(self, duration_sec: float) -> None:
|
2020-03-24 10:45:33 -04:00
|
|
|
pass
|
|
|
|
|
2021-12-14 12:35:28 -05:00
|
|
|
def add_database_scheduled(self, sched_sec: float) -> None:
|
2020-03-24 10:45:33 -04:00
|
|
|
pass
|
|
|
|
|
2021-12-14 12:35:28 -05:00
|
|
|
def record_event_fetch(self, event_count: int) -> None:
|
2020-03-24 10:45:33 -04:00
|
|
|
pass
|
|
|
|
|
2021-10-28 09:14:42 -04:00
|
|
|
def __bool__(self) -> Literal[False]:
|
2020-03-24 10:45:33 -04:00
|
|
|
return False
|
|
|
|
|
|
|
|
|
|
|
|
SENTINEL_CONTEXT = _Sentinel()
|
2020-03-07 12:57:26 -05:00
|
|
|
|
|
|
|
|
2020-09-04 06:54:56 -04:00
|
|
|
class LoggingContext:
|
2014-10-30 06:13:46 -04:00
|
|
|
"""Additional context for log formatting. Contexts are scoped within a
|
2016-02-10 06:29:21 -05:00
|
|
|
"with" block.
|
2018-01-11 17:40:51 -05:00
|
|
|
|
2018-07-10 11:12:36 -04:00
|
|
|
If a parent is given when creating a new context, then:
|
|
|
|
- logging fields are copied from the parent to the new context on entry
|
|
|
|
- when the new context exits, the cpu usage stats are copied from the
|
|
|
|
child to the parent
|
|
|
|
|
2014-10-30 06:13:46 -04:00
|
|
|
Args:
|
2021-04-21 11:39:34 -04:00
|
|
|
name: Name for the context for logging. If this is omitted, it is
|
|
|
|
inherited from the parent context.
|
2018-07-10 11:12:36 -04:00
|
|
|
parent_context (LoggingContext|None): The parent of the new context
|
2014-10-30 06:13:46 -04:00
|
|
|
"""
|
|
|
|
|
2015-12-03 16:03:01 -05:00
|
|
|
__slots__ = [
|
2019-06-20 05:32:02 -04:00
|
|
|
"previous_context",
|
|
|
|
"name",
|
|
|
|
"parent_context",
|
2018-07-10 08:56:07 -04:00
|
|
|
"_resource_usage",
|
2018-05-22 06:16:07 -04:00
|
|
|
"usage_start",
|
2019-06-20 05:32:02 -04:00
|
|
|
"main_thread",
|
2020-03-24 10:45:33 -04:00
|
|
|
"finished",
|
2019-06-20 05:32:02 -04:00
|
|
|
"request",
|
|
|
|
"tag",
|
2019-07-11 05:36:03 -04:00
|
|
|
"scope",
|
2015-12-03 16:03:01 -05:00
|
|
|
]
|
2014-10-29 21:21:33 -04:00
|
|
|
|
2021-01-05 08:06:55 -05:00
|
|
|
def __init__(
|
|
|
|
self,
|
2021-04-21 11:39:34 -04:00
|
|
|
name: Optional[str] = None,
|
2021-01-05 08:06:55 -05:00
|
|
|
parent_context: "Optional[LoggingContext]" = None,
|
2021-04-08 08:01:14 -04:00
|
|
|
request: Optional[ContextRequest] = None,
|
2021-01-05 08:06:55 -05:00
|
|
|
) -> None:
|
2020-03-24 10:45:33 -04:00
|
|
|
self.previous_context = current_context()
|
2018-01-11 13:17:54 -05:00
|
|
|
|
2018-07-10 08:56:07 -04:00
|
|
|
# track the resources used by this context so far
|
|
|
|
self._resource_usage = ContextResourceUsage()
|
2018-06-21 01:15:03 -04:00
|
|
|
|
2020-03-24 10:45:33 -04:00
|
|
|
# The thread resource usage when the logcontext became active. None
|
|
|
|
# if the context is not currently active.
|
2021-10-08 09:49:41 -04:00
|
|
|
self.usage_start: Optional[resource.struct_rusage] = None
|
2018-05-22 06:16:07 -04:00
|
|
|
|
2019-07-03 08:40:45 -04:00
|
|
|
self.main_thread = get_thread_id()
|
2018-01-11 17:40:51 -05:00
|
|
|
self.request = None
|
2016-02-03 08:51:25 -05:00
|
|
|
self.tag = ""
|
2021-07-15 06:02:43 -04:00
|
|
|
self.scope: Optional["_LogContextScope"] = None
|
2014-10-29 21:21:33 -04:00
|
|
|
|
2020-03-24 10:45:33 -04:00
|
|
|
# keep track of whether we have hit the __exit__ block for this context
|
|
|
|
# (suggesting that the the thing that created the context thinks it should
|
|
|
|
# be finished, and that re-activating it would suggest an error).
|
|
|
|
self.finished = False
|
|
|
|
|
2018-07-10 11:12:36 -04:00
|
|
|
self.parent_context = parent_context
|
|
|
|
|
2018-09-27 06:25:34 -04:00
|
|
|
if self.parent_context is not None:
|
2021-04-08 08:01:14 -04:00
|
|
|
# we track the current request_id
|
|
|
|
self.request = self.parent_context.request
|
|
|
|
|
|
|
|
# we also track the current scope:
|
|
|
|
self.scope = self.parent_context.scope
|
2018-09-27 06:25:34 -04:00
|
|
|
|
|
|
|
if request is not None:
|
|
|
|
# the request param overrides the request from the parent context
|
|
|
|
self.request = request
|
|
|
|
|
2021-04-21 11:39:34 -04:00
|
|
|
# if we don't have a `name`, but do have a parent context, use its name.
|
|
|
|
if self.parent_context and name is None:
|
|
|
|
name = str(self.parent_context)
|
|
|
|
if name is None:
|
|
|
|
raise ValueError(
|
|
|
|
"LoggingContext must be given either a name or a parent context"
|
|
|
|
)
|
|
|
|
self.name = name
|
|
|
|
|
2020-03-07 12:57:26 -05:00
|
|
|
def __str__(self) -> str:
|
2021-04-20 09:19:00 -04:00
|
|
|
return self.name
|
2014-10-29 21:21:33 -04:00
|
|
|
|
2020-05-05 09:21:34 -04:00
|
|
|
@classmethod
|
|
|
|
def current_context(cls) -> LoggingContextOrSentinel:
|
|
|
|
"""Get the current logging context from thread local storage
|
|
|
|
|
|
|
|
This exists for backwards compatibility. ``current_context()`` should be
|
|
|
|
called directly.
|
|
|
|
|
|
|
|
Returns:
|
2022-11-16 10:25:24 -05:00
|
|
|
The current logging context
|
2020-05-05 09:21:34 -04:00
|
|
|
"""
|
|
|
|
warnings.warn(
|
|
|
|
"synapse.logging.context.LoggingContext.current_context() is deprecated "
|
|
|
|
"in favor of synapse.logging.context.current_context().",
|
|
|
|
DeprecationWarning,
|
|
|
|
stacklevel=2,
|
|
|
|
)
|
|
|
|
return current_context()
|
|
|
|
|
|
|
|
@classmethod
|
|
|
|
def set_current_context(
|
|
|
|
cls, context: LoggingContextOrSentinel
|
|
|
|
) -> LoggingContextOrSentinel:
|
|
|
|
"""Set the current logging context in thread local storage
|
|
|
|
|
|
|
|
This exists for backwards compatibility. ``set_current_context()`` should be
|
|
|
|
called directly.
|
|
|
|
|
|
|
|
Args:
|
2022-11-16 10:25:24 -05:00
|
|
|
context: The context to activate.
|
|
|
|
|
2020-05-05 09:21:34 -04:00
|
|
|
Returns:
|
|
|
|
The context that was previously active
|
|
|
|
"""
|
|
|
|
warnings.warn(
|
|
|
|
"synapse.logging.context.LoggingContext.set_current_context() is deprecated "
|
|
|
|
"in favor of synapse.logging.context.set_current_context().",
|
|
|
|
DeprecationWarning,
|
|
|
|
stacklevel=2,
|
|
|
|
)
|
|
|
|
return set_current_context(context)
|
|
|
|
|
2020-03-07 12:57:26 -05:00
|
|
|
def __enter__(self) -> "LoggingContext":
|
2014-10-30 06:13:46 -04:00
|
|
|
"""Enters this logging context into thread local storage"""
|
2020-03-24 10:45:33 -04:00
|
|
|
old_context = set_current_context(self)
|
2016-02-10 06:25:19 -05:00
|
|
|
if self.previous_context != old_context:
|
2020-09-28 12:58:33 -04:00
|
|
|
logcontext_error(
|
|
|
|
"Expected previous context %r, found %r"
|
2021-02-16 17:32:34 -05:00
|
|
|
% (
|
|
|
|
self.previous_context,
|
|
|
|
old_context,
|
|
|
|
)
|
2016-02-10 06:23:32 -05:00
|
|
|
)
|
2014-10-29 21:21:33 -04:00
|
|
|
return self
|
|
|
|
|
2021-12-14 12:35:28 -05:00
|
|
|
def __exit__(
|
|
|
|
self,
|
|
|
|
type: Optional[Type[BaseException]],
|
|
|
|
value: Optional[BaseException],
|
|
|
|
traceback: Optional[TracebackType],
|
|
|
|
) -> None:
|
2014-10-30 06:13:46 -04:00
|
|
|
"""Restore the logging context in thread local storage to the state it
|
|
|
|
was before this context was entered.
|
|
|
|
Returns:
|
2018-05-22 06:16:07 -04:00
|
|
|
None to avoid suppressing any exceptions that were thrown.
|
2014-10-30 06:13:46 -04:00
|
|
|
"""
|
2020-03-24 10:45:33 -04:00
|
|
|
current = set_current_context(self.previous_context)
|
2015-12-07 05:51:18 -05:00
|
|
|
if current is not self:
|
2020-03-24 10:45:33 -04:00
|
|
|
if current is SENTINEL_CONTEXT:
|
2020-09-28 12:58:33 -04:00
|
|
|
logcontext_error("Expected logging context %s was lost" % (self,))
|
2014-11-25 05:57:31 -05:00
|
|
|
else:
|
2020-09-28 12:58:33 -04:00
|
|
|
logcontext_error(
|
|
|
|
"Expected logging context %s but found %s" % (self, current)
|
2014-11-25 05:57:31 -05:00
|
|
|
)
|
2018-07-10 11:12:36 -04:00
|
|
|
|
2020-03-24 10:45:33 -04:00
|
|
|
# the fact that we are here suggests that the caller thinks that everything
|
|
|
|
# is done and dusted for this logcontext, and further activity will not get
|
|
|
|
# recorded against the correct metrics.
|
|
|
|
self.finished = True
|
2018-07-10 11:12:36 -04:00
|
|
|
|
2021-10-08 09:49:41 -04:00
|
|
|
def start(self, rusage: "Optional[resource.struct_rusage]") -> None:
|
2020-04-03 08:21:30 -04:00
|
|
|
"""
|
|
|
|
Record that this logcontext is currently running.
|
|
|
|
|
|
|
|
This should not be called directly: use set_current_context
|
|
|
|
|
|
|
|
Args:
|
|
|
|
rusage: the resources used by the current thread, at the point of
|
|
|
|
switching to this logcontext. May be None if this platform doesn't
|
|
|
|
support getrusuage.
|
|
|
|
"""
|
2019-07-03 08:40:45 -04:00
|
|
|
if get_thread_id() != self.main_thread:
|
2020-09-28 12:58:33 -04:00
|
|
|
logcontext_error("Started logcontext %s on different thread" % (self,))
|
2015-12-03 16:03:01 -05:00
|
|
|
return
|
|
|
|
|
2020-03-24 10:45:33 -04:00
|
|
|
if self.finished:
|
2020-09-28 12:58:33 -04:00
|
|
|
logcontext_error("Re-starting finished log context %s" % (self,))
|
2020-03-24 10:45:33 -04:00
|
|
|
|
2018-05-22 06:16:07 -04:00
|
|
|
# If we haven't already started record the thread resource usage so
|
|
|
|
# far
|
2020-03-24 10:45:33 -04:00
|
|
|
if self.usage_start:
|
2020-09-28 12:58:33 -04:00
|
|
|
logcontext_error("Re-starting already-active log context %s" % (self,))
|
2020-03-24 10:45:33 -04:00
|
|
|
else:
|
2020-04-03 08:21:30 -04:00
|
|
|
self.usage_start = rusage
|
2015-12-03 16:03:01 -05:00
|
|
|
|
2021-10-08 09:49:41 -04:00
|
|
|
def stop(self, rusage: "Optional[resource.struct_rusage]") -> None:
|
2020-04-03 08:21:30 -04:00
|
|
|
"""
|
|
|
|
Record that this logcontext is no longer running.
|
|
|
|
|
|
|
|
This should not be called directly: use set_current_context
|
|
|
|
|
|
|
|
Args:
|
|
|
|
rusage: the resources used by the current thread, at the point of
|
|
|
|
switching away from this logcontext. May be None if this platform
|
|
|
|
doesn't support getrusuage.
|
|
|
|
"""
|
|
|
|
|
|
|
|
try:
|
|
|
|
if get_thread_id() != self.main_thread:
|
2020-09-28 12:58:33 -04:00
|
|
|
logcontext_error("Stopped logcontext %s on different thread" % (self,))
|
2020-04-03 08:21:30 -04:00
|
|
|
return
|
|
|
|
|
|
|
|
if not rusage:
|
|
|
|
return
|
2015-12-03 16:03:01 -05:00
|
|
|
|
2020-04-03 08:21:30 -04:00
|
|
|
# Record the cpu used since we started
|
|
|
|
if not self.usage_start:
|
2020-09-28 12:58:33 -04:00
|
|
|
logcontext_error(
|
|
|
|
"Called stop on logcontext %s without recording a start rusage"
|
|
|
|
% (self,)
|
2019-09-18 16:55:37 -04:00
|
|
|
)
|
2020-04-03 08:21:30 -04:00
|
|
|
return
|
2018-07-10 08:56:07 -04:00
|
|
|
|
2020-04-03 08:21:30 -04:00
|
|
|
utime_delta, stime_delta = self._get_cputime(rusage)
|
2020-05-06 16:43:39 -04:00
|
|
|
self.add_cputime(utime_delta, stime_delta)
|
2020-04-03 08:21:30 -04:00
|
|
|
finally:
|
|
|
|
self.usage_start = None
|
2020-03-24 10:45:33 -04:00
|
|
|
|
2020-03-07 12:57:26 -05:00
|
|
|
def get_resource_usage(self) -> ContextResourceUsage:
|
2018-07-10 08:56:07 -04:00
|
|
|
"""Get resources used by this logcontext so far.
|
2018-05-22 06:16:07 -04:00
|
|
|
|
|
|
|
Returns:
|
2022-11-16 10:25:24 -05:00
|
|
|
A *copy* of the object tracking resource usage so far
|
2018-05-22 06:16:07 -04:00
|
|
|
"""
|
2018-07-10 08:56:07 -04:00
|
|
|
# we always return a copy, for consistency
|
|
|
|
res = self._resource_usage.copy()
|
2015-12-03 16:03:01 -05:00
|
|
|
|
2018-05-22 06:16:07 -04:00
|
|
|
# If we are on the correct thread and we're currently running then we
|
|
|
|
# can include resource usage so far.
|
2019-07-03 08:40:45 -04:00
|
|
|
is_main_thread = get_thread_id() == self.main_thread
|
2020-03-24 10:45:33 -04:00
|
|
|
if self.usage_start and is_main_thread:
|
2020-04-03 08:21:30 -04:00
|
|
|
rusage = get_thread_resource_usage()
|
|
|
|
assert rusage is not None
|
|
|
|
utime_delta, stime_delta = self._get_cputime(rusage)
|
2019-06-19 15:55:53 -04:00
|
|
|
res.ru_utime += utime_delta
|
|
|
|
res.ru_stime += stime_delta
|
2015-12-03 16:03:01 -05:00
|
|
|
|
2018-07-10 08:56:07 -04:00
|
|
|
return res
|
2015-12-03 16:03:01 -05:00
|
|
|
|
2021-10-08 09:49:41 -04:00
|
|
|
def _get_cputime(self, current: "resource.struct_rusage") -> Tuple[float, float]:
|
2020-04-03 08:21:30 -04:00
|
|
|
"""Get the cpu usage time between start() and the given rusage
|
|
|
|
|
|
|
|
Args:
|
|
|
|
rusage: the current resource usage
|
2019-06-19 15:55:53 -04:00
|
|
|
|
|
|
|
Returns: Tuple[float, float]: seconds in user mode, seconds in system mode
|
|
|
|
"""
|
2020-03-07 12:57:26 -05:00
|
|
|
assert self.usage_start is not None
|
|
|
|
|
2019-06-19 15:55:53 -04:00
|
|
|
utime_delta = current.ru_utime - self.usage_start.ru_utime
|
|
|
|
stime_delta = current.ru_stime - self.usage_start.ru_stime
|
|
|
|
|
|
|
|
# sanity check
|
|
|
|
if utime_delta < 0:
|
2019-06-24 05:01:16 -04:00
|
|
|
logger.error(
|
|
|
|
"utime went backwards! %f < %f",
|
|
|
|
current.ru_utime,
|
|
|
|
self.usage_start.ru_utime,
|
|
|
|
)
|
|
|
|
utime_delta = 0
|
2019-06-19 15:55:53 -04:00
|
|
|
|
|
|
|
if stime_delta < 0:
|
2019-06-24 05:01:16 -04:00
|
|
|
logger.error(
|
|
|
|
"stime went backwards! %f < %f",
|
|
|
|
current.ru_stime,
|
|
|
|
self.usage_start.ru_stime,
|
|
|
|
)
|
|
|
|
stime_delta = 0
|
2019-06-19 15:55:53 -04:00
|
|
|
|
|
|
|
return utime_delta, stime_delta
|
|
|
|
|
2020-05-06 16:43:39 -04:00
|
|
|
def add_cputime(self, utime_delta: float, stime_delta: float) -> None:
|
|
|
|
"""Update the CPU time usage of this context (and any parents, recursively).
|
|
|
|
|
|
|
|
Args:
|
|
|
|
utime_delta: additional user time, in seconds, spent in this context.
|
|
|
|
stime_delta: additional system time, in seconds, spent in this context.
|
|
|
|
"""
|
|
|
|
self._resource_usage.ru_utime += utime_delta
|
|
|
|
self._resource_usage.ru_stime += stime_delta
|
|
|
|
if self.parent_context:
|
|
|
|
self.parent_context.add_cputime(utime_delta, stime_delta)
|
|
|
|
|
2020-03-07 12:57:26 -05:00
|
|
|
def add_database_transaction(self, duration_sec: float) -> None:
|
2020-05-06 16:43:39 -04:00
|
|
|
"""Record the use of a database transaction and the length of time it took.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
duration_sec: The number of seconds the database transaction took.
|
|
|
|
"""
|
2019-06-19 15:55:53 -04:00
|
|
|
if duration_sec < 0:
|
2019-06-24 05:01:16 -04:00
|
|
|
raise ValueError("DB txn time can only be non-negative")
|
2018-07-10 08:56:07 -04:00
|
|
|
self._resource_usage.db_txn_count += 1
|
|
|
|
self._resource_usage.db_txn_duration_sec += duration_sec
|
2020-05-06 16:43:39 -04:00
|
|
|
if self.parent_context:
|
|
|
|
self.parent_context.add_database_transaction(duration_sec)
|
2015-12-07 12:56:11 -05:00
|
|
|
|
2020-03-07 12:57:26 -05:00
|
|
|
def add_database_scheduled(self, sched_sec: float) -> None:
|
2018-01-11 19:27:14 -05:00
|
|
|
"""Record a use of the database pool
|
|
|
|
|
|
|
|
Args:
|
2020-05-06 16:43:39 -04:00
|
|
|
sched_sec: number of seconds it took us to get a connection
|
2018-01-11 19:27:14 -05:00
|
|
|
"""
|
2019-06-19 15:55:53 -04:00
|
|
|
if sched_sec < 0:
|
2019-06-24 05:01:16 -04:00
|
|
|
raise ValueError("DB scheduling time can only be non-negative")
|
2018-07-10 08:56:07 -04:00
|
|
|
self._resource_usage.db_sched_duration_sec += sched_sec
|
2020-05-06 16:43:39 -04:00
|
|
|
if self.parent_context:
|
|
|
|
self.parent_context.add_database_scheduled(sched_sec)
|
2018-01-11 19:27:14 -05:00
|
|
|
|
2020-03-07 12:57:26 -05:00
|
|
|
def record_event_fetch(self, event_count: int) -> None:
|
2018-06-22 05:42:28 -04:00
|
|
|
"""Record a number of events being fetched from the db
|
|
|
|
|
|
|
|
Args:
|
2020-05-06 16:43:39 -04:00
|
|
|
event_count: number of events being fetched
|
2018-06-22 05:42:28 -04:00
|
|
|
"""
|
2018-07-10 08:56:07 -04:00
|
|
|
self._resource_usage.evt_db_fetch_count += event_count
|
2020-05-06 16:43:39 -04:00
|
|
|
if self.parent_context:
|
|
|
|
self.parent_context.record_event_fetch(event_count)
|
2018-06-22 05:42:28 -04:00
|
|
|
|
2014-10-29 21:21:33 -04:00
|
|
|
|
|
|
|
class LoggingContextFilter(logging.Filter):
|
2014-10-30 06:13:46 -04:00
|
|
|
"""Logging filter that adds values from the current logging context to each
|
|
|
|
record.
|
|
|
|
"""
|
2019-06-20 05:32:02 -04:00
|
|
|
|
2020-12-14 14:19:47 -05:00
|
|
|
def __init__(self, request: str = ""):
|
|
|
|
self._default_request = request
|
2014-10-29 21:21:33 -04:00
|
|
|
|
2021-01-05 08:06:55 -05:00
|
|
|
def filter(self, record: logging.LogRecord) -> Literal[True]:
|
2014-10-30 06:13:46 -04:00
|
|
|
"""Add each fields from the logging contexts to the record.
|
|
|
|
Returns:
|
|
|
|
True to include the record in the log output.
|
|
|
|
"""
|
2020-03-24 10:45:33 -04:00
|
|
|
context = current_context()
|
2022-09-30 12:36:28 -04:00
|
|
|
record.request = self._default_request
|
2018-08-20 13:20:07 -04:00
|
|
|
|
|
|
|
# context should never be None, but if it somehow ends up being, then
|
|
|
|
# we end up in a death spiral of infinite loops, so let's check, for
|
|
|
|
# robustness' sake.
|
|
|
|
if context is not None:
|
2021-04-08 08:01:14 -04:00
|
|
|
# Logging is interested in the request ID. Note that for backwards
|
|
|
|
# compatibility this is stored as the "request" on the record.
|
2022-09-30 12:36:28 -04:00
|
|
|
record.request = str(context)
|
2021-04-08 08:01:14 -04:00
|
|
|
|
|
|
|
# Add some data from the HTTP request.
|
|
|
|
request = context.request
|
|
|
|
if request is None:
|
|
|
|
return True
|
|
|
|
|
2022-09-30 12:36:28 -04:00
|
|
|
record.ip_address = request.ip_address
|
|
|
|
record.site_tag = request.site_tag
|
|
|
|
record.requester = request.requester
|
|
|
|
record.authenticated_entity = request.authenticated_entity
|
|
|
|
record.method = request.method
|
|
|
|
record.url = request.url
|
|
|
|
record.protocol = request.protocol
|
|
|
|
record.user_agent = request.user_agent
|
2018-08-20 13:20:07 -04:00
|
|
|
|
2014-10-29 21:21:33 -04:00
|
|
|
return True
|
|
|
|
|
|
|
|
|
2020-07-21 19:40:27 -04:00
|
|
|
class PreserveLoggingContext:
|
|
|
|
"""Context manager which replaces the logging context
|
2014-10-30 06:13:46 -04:00
|
|
|
|
2021-02-16 17:32:34 -05:00
|
|
|
The previous logging context is restored on exit."""
|
2020-07-21 19:40:27 -04:00
|
|
|
|
|
|
|
__slots__ = ["_old_context", "_new_context"]
|
2015-12-03 16:03:01 -05:00
|
|
|
|
2020-03-24 10:45:33 -04:00
|
|
|
def __init__(
|
|
|
|
self, new_context: LoggingContextOrSentinel = SENTINEL_CONTEXT
|
|
|
|
) -> None:
|
2020-07-21 19:40:27 -04:00
|
|
|
self._new_context = new_context
|
2014-10-30 06:13:46 -04:00
|
|
|
|
2020-03-07 12:57:26 -05:00
|
|
|
def __enter__(self) -> None:
|
2020-07-21 19:40:27 -04:00
|
|
|
self._old_context = set_current_context(self._new_context)
|
2016-02-04 05:22:44 -05:00
|
|
|
|
2021-12-14 12:35:28 -05:00
|
|
|
def __exit__(
|
|
|
|
self,
|
|
|
|
type: Optional[Type[BaseException]],
|
|
|
|
value: Optional[BaseException],
|
|
|
|
traceback: Optional[TracebackType],
|
|
|
|
) -> None:
|
2020-07-21 19:40:27 -04:00
|
|
|
context = set_current_context(self._old_context)
|
2016-02-04 05:22:44 -05:00
|
|
|
|
2020-07-21 19:40:27 -04:00
|
|
|
if context != self._new_context:
|
2020-03-24 10:45:33 -04:00
|
|
|
if not context:
|
2020-09-28 12:58:33 -04:00
|
|
|
logcontext_error(
|
|
|
|
"Expected logging context %s was lost" % (self._new_context,)
|
2020-07-21 19:40:27 -04:00
|
|
|
)
|
2019-05-29 14:27:50 -04:00
|
|
|
else:
|
2020-09-28 12:58:33 -04:00
|
|
|
logcontext_error(
|
|
|
|
"Expected logging context %s but found %s"
|
2021-02-16 17:32:34 -05:00
|
|
|
% (
|
|
|
|
self._new_context,
|
|
|
|
context,
|
|
|
|
)
|
2019-05-29 14:27:50 -04:00
|
|
|
)
|
2016-02-04 05:22:44 -05:00
|
|
|
|
2020-03-24 10:45:33 -04:00
|
|
|
|
|
|
|
_thread_local = threading.local()
|
|
|
|
_thread_local.current_context = SENTINEL_CONTEXT
|
|
|
|
|
|
|
|
|
|
|
|
def current_context() -> LoggingContextOrSentinel:
|
|
|
|
"""Get the current logging context from thread local storage"""
|
|
|
|
return getattr(_thread_local, "current_context", SENTINEL_CONTEXT)
|
|
|
|
|
|
|
|
|
|
|
|
def set_current_context(context: LoggingContextOrSentinel) -> LoggingContextOrSentinel:
|
|
|
|
"""Set the current logging context in thread local storage
|
|
|
|
Args:
|
2022-11-16 10:25:24 -05:00
|
|
|
context: The context to activate.
|
|
|
|
|
2020-03-24 10:45:33 -04:00
|
|
|
Returns:
|
|
|
|
The context that was previously active
|
|
|
|
"""
|
2020-03-31 10:18:41 -04:00
|
|
|
# everything blows up if we allow current_context to be set to None, so sanity-check
|
|
|
|
# that now.
|
|
|
|
if context is None:
|
|
|
|
raise TypeError("'context' argument may not be None")
|
|
|
|
|
2020-03-24 10:45:33 -04:00
|
|
|
current = current_context()
|
|
|
|
|
|
|
|
if current is not context:
|
2020-04-03 08:21:30 -04:00
|
|
|
rusage = get_thread_resource_usage()
|
|
|
|
current.stop(rusage)
|
2020-03-24 10:45:33 -04:00
|
|
|
_thread_local.current_context = context
|
2020-04-03 08:21:30 -04:00
|
|
|
context.start(rusage)
|
|
|
|
|
2020-03-24 10:45:33 -04:00
|
|
|
return current
|
2015-05-08 14:53:34 -04:00
|
|
|
|
2015-05-08 11:32:18 -04:00
|
|
|
|
2021-01-05 08:06:55 -05:00
|
|
|
def nested_logging_context(suffix: str) -> LoggingContext:
|
2018-09-27 06:25:34 -04:00
|
|
|
"""Creates a new logging context as a child of another.
|
|
|
|
|
2021-04-08 08:01:14 -04:00
|
|
|
The nested logging context will have a 'name' made up of the parent context's
|
|
|
|
name, plus the given suffix.
|
2018-09-27 06:25:34 -04:00
|
|
|
|
|
|
|
CPU/db usage stats will be added to the parent context's on exit.
|
|
|
|
|
|
|
|
Normal usage looks like:
|
|
|
|
|
|
|
|
with nested_logging_context(suffix):
|
|
|
|
# ... do stuff
|
|
|
|
|
|
|
|
Args:
|
2021-04-08 08:01:14 -04:00
|
|
|
suffix: suffix to add to the parent context's 'name'.
|
2018-09-27 06:25:34 -04:00
|
|
|
|
|
|
|
Returns:
|
2022-11-16 10:25:24 -05:00
|
|
|
A new logging context.
|
2018-09-27 06:25:34 -04:00
|
|
|
"""
|
2021-01-05 08:06:55 -05:00
|
|
|
curr_context = current_context()
|
|
|
|
if not curr_context:
|
|
|
|
logger.warning(
|
|
|
|
"Starting nested logging context from sentinel context: metrics will be lost"
|
|
|
|
)
|
|
|
|
parent_context = None
|
2020-03-07 12:57:26 -05:00
|
|
|
else:
|
2021-01-05 08:06:55 -05:00
|
|
|
assert isinstance(curr_context, LoggingContext)
|
|
|
|
parent_context = curr_context
|
2021-04-20 09:19:00 -04:00
|
|
|
prefix = str(curr_context)
|
2021-04-08 08:01:14 -04:00
|
|
|
return LoggingContext(
|
|
|
|
prefix + "-" + suffix,
|
|
|
|
parent_context=parent_context,
|
|
|
|
)
|
2018-09-27 06:25:34 -04:00
|
|
|
|
|
|
|
|
2022-03-08 10:58:14 -05:00
|
|
|
P = ParamSpec("P")
|
2021-12-14 12:35:28 -05:00
|
|
|
R = TypeVar("R")
|
|
|
|
|
|
|
|
|
2022-04-27 09:58:26 -04:00
|
|
|
async def _unwrap_awaitable(awaitable: Awaitable[R]) -> R:
|
|
|
|
"""Unwraps an arbitrary awaitable by awaiting it."""
|
|
|
|
return await awaitable
|
|
|
|
|
|
|
|
|
2021-12-14 12:35:28 -05:00
|
|
|
@overload
|
2023-09-05 07:14:14 -04:00
|
|
|
def preserve_fn(
|
2022-03-08 10:58:14 -05:00
|
|
|
f: Callable[P, Awaitable[R]],
|
|
|
|
) -> Callable[P, "defer.Deferred[R]"]:
|
2021-12-14 12:35:28 -05:00
|
|
|
# The `type: ignore[misc]` above suppresses
|
|
|
|
# "Overloaded function signatures 1 and 2 overlap with incompatible return types"
|
|
|
|
...
|
|
|
|
|
|
|
|
|
|
|
|
@overload
|
2024-03-13 12:46:44 -04:00
|
|
|
def preserve_fn(f: Callable[P, R]) -> Callable[P, "defer.Deferred[R]"]: ...
|
2021-12-14 12:35:28 -05:00
|
|
|
|
|
|
|
|
|
|
|
def preserve_fn(
|
|
|
|
f: Union[
|
2022-03-08 10:58:14 -05:00
|
|
|
Callable[P, R],
|
|
|
|
Callable[P, Awaitable[R]],
|
2021-12-14 12:35:28 -05:00
|
|
|
]
|
2022-03-08 10:58:14 -05:00
|
|
|
) -> Callable[P, "defer.Deferred[R]"]:
|
2018-03-07 14:59:24 -05:00
|
|
|
"""Function decorator which wraps the function with run_in_background"""
|
2019-06-20 05:32:02 -04:00
|
|
|
|
2022-03-08 10:58:14 -05:00
|
|
|
def g(*args: P.args, **kwargs: P.kwargs) -> "defer.Deferred[R]":
|
2018-03-07 14:59:24 -05:00
|
|
|
return run_in_background(f, *args, **kwargs)
|
2019-06-20 05:32:02 -04:00
|
|
|
|
2018-03-07 14:59:24 -05:00
|
|
|
return g
|
|
|
|
|
|
|
|
|
2021-12-14 12:35:28 -05:00
|
|
|
@overload
|
2023-09-05 07:14:14 -04:00
|
|
|
def run_in_background(
|
2022-03-08 10:58:14 -05:00
|
|
|
f: Callable[P, Awaitable[R]], *args: P.args, **kwargs: P.kwargs
|
2021-12-14 12:35:28 -05:00
|
|
|
) -> "defer.Deferred[R]":
|
|
|
|
# The `type: ignore[misc]` above suppresses
|
|
|
|
# "Overloaded function signatures 1 and 2 overlap with incompatible return types"
|
|
|
|
...
|
|
|
|
|
|
|
|
|
|
|
|
@overload
|
|
|
|
def run_in_background(
|
2022-03-08 10:58:14 -05:00
|
|
|
f: Callable[P, R], *args: P.args, **kwargs: P.kwargs
|
2024-03-13 12:46:44 -04:00
|
|
|
) -> "defer.Deferred[R]": ...
|
2021-12-14 12:35:28 -05:00
|
|
|
|
|
|
|
|
2024-03-13 13:05:57 -04:00
|
|
|
def run_in_background(
|
2021-12-14 12:35:28 -05:00
|
|
|
f: Union[
|
2022-03-08 10:58:14 -05:00
|
|
|
Callable[P, R],
|
|
|
|
Callable[P, Awaitable[R]],
|
2021-12-14 12:35:28 -05:00
|
|
|
],
|
2022-03-08 10:58:14 -05:00
|
|
|
*args: P.args,
|
|
|
|
**kwargs: P.kwargs,
|
2021-12-14 12:35:28 -05:00
|
|
|
) -> "defer.Deferred[R]":
|
2018-03-07 14:59:24 -05:00
|
|
|
"""Calls a function, ensuring that the current context is restored after
|
2017-03-17 16:56:54 -04:00
|
|
|
return from the function, and that the sentinel context is set once the
|
2018-05-02 06:46:23 -04:00
|
|
|
deferred returned by the function completes.
|
2017-03-14 22:21:07 -04:00
|
|
|
|
2019-07-02 14:01:28 -04:00
|
|
|
Useful for wrapping functions that return a deferred or coroutine, which you don't
|
|
|
|
yield or await on (for instance because you want to pass it to
|
|
|
|
deferred.gatherResults()).
|
2018-04-27 06:07:40 -04:00
|
|
|
|
2020-01-11 08:00:24 -05:00
|
|
|
If f returns a Coroutine object, it will be wrapped into a Deferred (which will have
|
|
|
|
the side effect of executing the coroutine).
|
|
|
|
|
2018-04-27 06:07:40 -04:00
|
|
|
Note that if you completely discard the result, you should make sure that
|
|
|
|
`f` doesn't raise any deferred exceptions, otherwise a scary-looking
|
|
|
|
CRITICAL error about an unhandled error will be logged without much
|
|
|
|
indication about where it came from.
|
2017-03-14 22:21:07 -04:00
|
|
|
"""
|
2020-03-24 10:45:33 -04:00
|
|
|
current = current_context()
|
2018-04-27 07:17:13 -04:00
|
|
|
try:
|
|
|
|
res = f(*args, **kwargs)
|
2021-03-24 09:34:30 -04:00
|
|
|
except Exception:
|
2018-04-27 07:17:13 -04:00
|
|
|
# the assumption here is that the caller doesn't want to be disturbed
|
|
|
|
# by synchronous exceptions, so let's turn them into Failures.
|
|
|
|
return defer.fail()
|
|
|
|
|
2022-04-27 09:58:26 -04:00
|
|
|
# `res` may be a coroutine, `Deferred`, some other kind of awaitable, or a plain
|
|
|
|
# value. Convert it to a `Deferred`.
|
2023-08-24 10:53:07 -04:00
|
|
|
d: "defer.Deferred[R]"
|
2021-07-22 07:00:16 -04:00
|
|
|
if isinstance(res, typing.Coroutine):
|
2022-04-27 09:58:26 -04:00
|
|
|
# Wrap the coroutine in a `Deferred`.
|
2023-08-24 10:53:07 -04:00
|
|
|
d = defer.ensureDeferred(res)
|
2022-04-27 09:58:26 -04:00
|
|
|
elif isinstance(res, defer.Deferred):
|
2023-08-24 10:53:07 -04:00
|
|
|
d = res
|
2022-04-27 09:58:26 -04:00
|
|
|
elif isinstance(res, Awaitable):
|
|
|
|
# `res` is probably some kind of completed awaitable, such as a `DoneAwaitable`
|
|
|
|
# or `Future` from `make_awaitable`.
|
2023-08-24 10:53:07 -04:00
|
|
|
d = defer.ensureDeferred(_unwrap_awaitable(res))
|
2022-04-27 09:58:26 -04:00
|
|
|
else:
|
|
|
|
# `res` is a plain value. Wrap it in a `Deferred`.
|
2023-08-24 10:53:07 -04:00
|
|
|
d = defer.succeed(res)
|
2018-05-02 06:46:23 -04:00
|
|
|
|
2023-08-24 10:53:07 -04:00
|
|
|
if d.called and not d.paused:
|
2018-05-02 06:46:23 -04:00
|
|
|
# The function should have maintained the logcontext, so we can
|
|
|
|
# optimise out the messing about
|
2023-08-24 10:53:07 -04:00
|
|
|
return d
|
2018-05-02 06:46:23 -04:00
|
|
|
|
|
|
|
# The function may have reset the context before returning, so
|
|
|
|
# we need to restore it now.
|
2020-03-24 10:45:33 -04:00
|
|
|
ctx = set_current_context(current)
|
2018-05-02 06:46:23 -04:00
|
|
|
|
|
|
|
# The original context will be restored when the deferred
|
|
|
|
# completes, but there is nothing waiting for it, so it will
|
|
|
|
# get leaked into the reactor or some other function which
|
|
|
|
# wasn't expecting it. We therefore need to reset the context
|
|
|
|
# here.
|
|
|
|
#
|
|
|
|
# (If this feels asymmetric, consider it this way: we are
|
|
|
|
# effectively forking a new thread of execution. We are
|
|
|
|
# probably currently within a ``with LoggingContext()`` block,
|
|
|
|
# which is supposed to have a single entry and exit point. But
|
|
|
|
# by spawning off another deferred, we are effectively
|
|
|
|
# adding a new exit point.)
|
2023-08-24 10:53:07 -04:00
|
|
|
d.addBoth(_set_context_cb, ctx)
|
|
|
|
return d
|
2016-02-04 05:22:44 -05:00
|
|
|
|
|
|
|
|
2021-12-14 12:35:28 -05:00
|
|
|
T = TypeVar("T")
|
|
|
|
|
2017-03-30 08:22:24 -04:00
|
|
|
|
2021-12-14 12:35:28 -05:00
|
|
|
def make_deferred_yieldable(deferred: "defer.Deferred[T]") -> "defer.Deferred[T]":
|
|
|
|
"""Given a deferred, make it follow the Synapse logcontext rules:
|
|
|
|
|
|
|
|
If the deferred has completed, essentially does nothing (just returns another
|
|
|
|
completed deferred with the result/failure).
|
2017-03-30 08:22:24 -04:00
|
|
|
|
|
|
|
If the deferred has not yet completed, resets the logcontext before
|
|
|
|
returning a deferred. Then, when the deferred completes, restores the
|
|
|
|
current logcontext before running callbacks/errbacks.
|
|
|
|
|
2018-04-27 06:29:27 -04:00
|
|
|
(This is more-or-less the opposite operation to run_in_background.)
|
2017-03-30 08:22:24 -04:00
|
|
|
"""
|
2018-05-02 06:46:23 -04:00
|
|
|
if deferred.called and not deferred.paused:
|
|
|
|
# it looks like this deferred is ready to run any callbacks we give it
|
|
|
|
# immediately. We may as well optimise out the logcontext faffery.
|
|
|
|
return deferred
|
|
|
|
|
|
|
|
# ok, we can't be sure that a yield won't block, so let's reset the
|
|
|
|
# logcontext, and add a callback to the deferred to restore it.
|
2020-03-24 10:45:33 -04:00
|
|
|
prev_context = set_current_context(SENTINEL_CONTEXT)
|
2018-05-02 06:46:23 -04:00
|
|
|
deferred.addBoth(_set_context_cb, prev_context)
|
2018-03-01 07:19:09 -05:00
|
|
|
return deferred
|
|
|
|
|
|
|
|
|
2020-03-07 12:57:26 -05:00
|
|
|
ResultT = TypeVar("ResultT")
|
|
|
|
|
|
|
|
|
2023-08-24 10:53:07 -04:00
|
|
|
def _set_context_cb(result: ResultT, context: LoggingContextOrSentinel) -> ResultT:
|
2018-03-01 07:19:09 -05:00
|
|
|
"""A callback function which just sets the logging context"""
|
2020-03-24 10:45:33 -04:00
|
|
|
set_current_context(context)
|
2018-03-01 07:19:09 -05:00
|
|
|
return result
|
2017-03-30 08:22:24 -04:00
|
|
|
|
|
|
|
|
2021-12-14 12:35:28 -05:00
|
|
|
def defer_to_thread(
|
2022-03-08 10:58:14 -05:00
|
|
|
reactor: "ISynapseReactor", f: Callable[P, R], *args: P.args, **kwargs: P.kwargs
|
2021-12-14 12:35:28 -05:00
|
|
|
) -> "defer.Deferred[R]":
|
2018-10-23 08:12:32 -04:00
|
|
|
"""
|
|
|
|
Calls the function `f` using a thread from the reactor's default threadpool and
|
|
|
|
returns the result as a Deferred.
|
|
|
|
|
|
|
|
Creates a new logcontext for `f`, which is created as a child of the current
|
|
|
|
logcontext (so its CPU usage metrics will get attributed to the current
|
|
|
|
logcontext). `f` should preserve the logcontext it is given.
|
|
|
|
|
|
|
|
The result deferred follows the Synapse logcontext rules: you should `yield`
|
|
|
|
on it.
|
|
|
|
|
|
|
|
Args:
|
2022-11-16 10:25:24 -05:00
|
|
|
reactor: The reactor in whose main thread the Deferred will be invoked,
|
|
|
|
and whose threadpool we should use for the function.
|
2018-10-23 08:12:32 -04:00
|
|
|
|
|
|
|
Normally this will be hs.get_reactor().
|
|
|
|
|
2022-11-16 10:25:24 -05:00
|
|
|
f: The function to call.
|
2016-02-04 05:22:44 -05:00
|
|
|
|
2018-10-23 08:12:32 -04:00
|
|
|
args: positional arguments to pass to f.
|
2016-02-04 05:22:44 -05:00
|
|
|
|
2018-10-23 08:12:32 -04:00
|
|
|
kwargs: keyword arguments to pass to f.
|
2016-02-04 05:22:44 -05:00
|
|
|
|
2018-10-23 08:12:32 -04:00
|
|
|
Returns:
|
2022-11-16 10:25:24 -05:00
|
|
|
A Deferred which fires a callback with the result of `f`, or an
|
2018-10-23 08:12:32 -04:00
|
|
|
errback if `f` throws an exception.
|
2016-02-04 05:22:44 -05:00
|
|
|
"""
|
2018-10-23 08:12:32 -04:00
|
|
|
return defer_to_threadpool(reactor, reactor.getThreadPool(), f, *args, **kwargs)
|
2016-02-04 05:22:44 -05:00
|
|
|
|
2018-10-23 08:12:32 -04:00
|
|
|
|
2021-12-14 12:35:28 -05:00
|
|
|
def defer_to_threadpool(
|
|
|
|
reactor: "ISynapseReactor",
|
|
|
|
threadpool: ThreadPool,
|
2022-03-08 10:58:14 -05:00
|
|
|
f: Callable[P, R],
|
|
|
|
*args: P.args,
|
|
|
|
**kwargs: P.kwargs,
|
2021-12-14 12:35:28 -05:00
|
|
|
) -> "defer.Deferred[R]":
|
2018-10-23 08:12:32 -04:00
|
|
|
"""
|
|
|
|
A wrapper for twisted.internet.threads.deferToThreadpool, which handles
|
|
|
|
logcontexts correctly.
|
|
|
|
|
|
|
|
Calls the function `f` using a thread from the given threadpool and returns
|
|
|
|
the result as a Deferred.
|
|
|
|
|
|
|
|
Creates a new logcontext for `f`, which is created as a child of the current
|
|
|
|
logcontext (so its CPU usage metrics will get attributed to the current
|
|
|
|
logcontext). `f` should preserve the logcontext it is given.
|
|
|
|
|
|
|
|
The result deferred follows the Synapse logcontext rules: you should `yield`
|
|
|
|
on it.
|
|
|
|
|
|
|
|
Args:
|
2022-11-16 10:25:24 -05:00
|
|
|
reactor: The reactor in whose main thread the Deferred will be invoked.
|
|
|
|
Normally this will be hs.get_reactor().
|
2018-10-23 08:12:32 -04:00
|
|
|
|
2022-11-16 10:25:24 -05:00
|
|
|
threadpool: The threadpool to use for running `f`. Normally this will be
|
|
|
|
hs.get_reactor().getThreadPool().
|
2018-10-23 08:12:32 -04:00
|
|
|
|
2022-11-16 10:25:24 -05:00
|
|
|
f: The function to call.
|
2018-10-23 08:12:32 -04:00
|
|
|
|
|
|
|
args: positional arguments to pass to f.
|
|
|
|
|
|
|
|
kwargs: keyword arguments to pass to f.
|
|
|
|
|
|
|
|
Returns:
|
2022-11-16 10:25:24 -05:00
|
|
|
A Deferred which fires a callback with the result of `f`, or an
|
2018-10-23 08:12:32 -04:00
|
|
|
errback if `f` throws an exception.
|
|
|
|
"""
|
2021-01-05 08:06:55 -05:00
|
|
|
curr_context = current_context()
|
|
|
|
if not curr_context:
|
|
|
|
logger.warning(
|
|
|
|
"Calling defer_to_threadpool from sentinel context: metrics will be lost"
|
|
|
|
)
|
|
|
|
parent_context = None
|
|
|
|
else:
|
|
|
|
assert isinstance(curr_context, LoggingContext)
|
|
|
|
parent_context = curr_context
|
2018-10-23 08:12:32 -04:00
|
|
|
|
2021-12-14 12:35:28 -05:00
|
|
|
def g() -> R:
|
2021-04-20 09:19:00 -04:00
|
|
|
with LoggingContext(str(curr_context), parent_context=parent_context):
|
2018-10-23 08:12:32 -04:00
|
|
|
return f(*args, **kwargs)
|
|
|
|
|
2019-06-20 05:32:02 -04:00
|
|
|
return make_deferred_yieldable(threads.deferToThreadPool(reactor, threadpool, g))
|