mirror of
https://mau.dev/maunium/synapse.git
synced 2024-10-01 01:36:05 -04:00
Remove usage of deprecated logger.warn method from codebase (#6271)
Replace every instance of `logger.warn` with `logger.warning` as the former is deprecated.
This commit is contained in:
parent
0467f33584
commit
54fef094b3
1
changelog.d/6271.misc
Normal file
1
changelog.d/6271.misc
Normal file
@ -0,0 +1 @@
|
|||||||
|
Replace every instance of `logger.warn` method with `logger.warning` as the former is deprecated.
|
@ -72,7 +72,7 @@ def move_media(origin_server, file_id, src_paths, dest_paths):
|
|||||||
# check that the original exists
|
# check that the original exists
|
||||||
original_file = src_paths.remote_media_filepath(origin_server, file_id)
|
original_file = src_paths.remote_media_filepath(origin_server, file_id)
|
||||||
if not os.path.exists(original_file):
|
if not os.path.exists(original_file):
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Original for %s/%s (%s) does not exist",
|
"Original for %s/%s (%s) does not exist",
|
||||||
origin_server,
|
origin_server,
|
||||||
file_id,
|
file_id,
|
||||||
|
@ -157,7 +157,7 @@ class Store(
|
|||||||
)
|
)
|
||||||
except self.database_engine.module.DatabaseError as e:
|
except self.database_engine.module.DatabaseError as e:
|
||||||
if self.database_engine.is_deadlock(e):
|
if self.database_engine.is_deadlock(e):
|
||||||
logger.warn("[TXN DEADLOCK] {%s} %d/%d", desc, i, N)
|
logger.warning("[TXN DEADLOCK] {%s} %d/%d", desc, i, N)
|
||||||
if i < N:
|
if i < N:
|
||||||
i += 1
|
i += 1
|
||||||
conn.rollback()
|
conn.rollback()
|
||||||
@ -432,7 +432,7 @@ class Porter(object):
|
|||||||
for row in rows:
|
for row in rows:
|
||||||
d = dict(zip(headers, row))
|
d = dict(zip(headers, row))
|
||||||
if "\0" in d['value']:
|
if "\0" in d['value']:
|
||||||
logger.warn('dropping search row %s', d)
|
logger.warning('dropping search row %s', d)
|
||||||
else:
|
else:
|
||||||
rows_dict.append(d)
|
rows_dict.append(d)
|
||||||
|
|
||||||
@ -647,7 +647,7 @@ class Porter(object):
|
|||||||
if isinstance(col, bytes):
|
if isinstance(col, bytes):
|
||||||
return bytearray(col)
|
return bytearray(col)
|
||||||
elif isinstance(col, string_types) and "\0" in col:
|
elif isinstance(col, string_types) and "\0" in col:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"DROPPING ROW: NUL value in table %s col %s: %r",
|
"DROPPING ROW: NUL value in table %s col %s: %r",
|
||||||
table,
|
table,
|
||||||
headers[j],
|
headers[j],
|
||||||
|
@ -497,7 +497,7 @@ class Auth(object):
|
|||||||
token = self.get_access_token_from_request(request)
|
token = self.get_access_token_from_request(request)
|
||||||
service = self.store.get_app_service_by_token(token)
|
service = self.store.get_app_service_by_token(token)
|
||||||
if not service:
|
if not service:
|
||||||
logger.warn("Unrecognised appservice access token.")
|
logger.warning("Unrecognised appservice access token.")
|
||||||
raise InvalidClientTokenError()
|
raise InvalidClientTokenError()
|
||||||
request.authenticated_entity = service.sender
|
request.authenticated_entity = service.sender
|
||||||
return defer.succeed(service)
|
return defer.succeed(service)
|
||||||
|
@ -44,6 +44,8 @@ def check_bind_error(e, address, bind_addresses):
|
|||||||
bind_addresses (list): Addresses on which the service listens.
|
bind_addresses (list): Addresses on which the service listens.
|
||||||
"""
|
"""
|
||||||
if address == "0.0.0.0" and "::" in bind_addresses:
|
if address == "0.0.0.0" and "::" in bind_addresses:
|
||||||
logger.warn("Failed to listen on 0.0.0.0, continuing because listening on [::]")
|
logger.warning(
|
||||||
|
"Failed to listen on 0.0.0.0, continuing because listening on [::]"
|
||||||
|
)
|
||||||
else:
|
else:
|
||||||
raise e
|
raise e
|
||||||
|
@ -94,7 +94,7 @@ class AppserviceServer(HomeServer):
|
|||||||
)
|
)
|
||||||
elif listener["type"] == "metrics":
|
elif listener["type"] == "metrics":
|
||||||
if not self.get_config().enable_metrics:
|
if not self.get_config().enable_metrics:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
(
|
(
|
||||||
"Metrics listener configured, but "
|
"Metrics listener configured, but "
|
||||||
"enable_metrics is not True!"
|
"enable_metrics is not True!"
|
||||||
@ -103,7 +103,7 @@ class AppserviceServer(HomeServer):
|
|||||||
else:
|
else:
|
||||||
_base.listen_metrics(listener["bind_addresses"], listener["port"])
|
_base.listen_metrics(listener["bind_addresses"], listener["port"])
|
||||||
else:
|
else:
|
||||||
logger.warn("Unrecognized listener type: %s", listener["type"])
|
logger.warning("Unrecognized listener type: %s", listener["type"])
|
||||||
|
|
||||||
self.get_tcp_replication().start_replication(self)
|
self.get_tcp_replication().start_replication(self)
|
||||||
|
|
||||||
|
@ -153,7 +153,7 @@ class ClientReaderServer(HomeServer):
|
|||||||
)
|
)
|
||||||
elif listener["type"] == "metrics":
|
elif listener["type"] == "metrics":
|
||||||
if not self.get_config().enable_metrics:
|
if not self.get_config().enable_metrics:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
(
|
(
|
||||||
"Metrics listener configured, but "
|
"Metrics listener configured, but "
|
||||||
"enable_metrics is not True!"
|
"enable_metrics is not True!"
|
||||||
@ -162,7 +162,7 @@ class ClientReaderServer(HomeServer):
|
|||||||
else:
|
else:
|
||||||
_base.listen_metrics(listener["bind_addresses"], listener["port"])
|
_base.listen_metrics(listener["bind_addresses"], listener["port"])
|
||||||
else:
|
else:
|
||||||
logger.warn("Unrecognized listener type: %s", listener["type"])
|
logger.warning("Unrecognized listener type: %s", listener["type"])
|
||||||
|
|
||||||
self.get_tcp_replication().start_replication(self)
|
self.get_tcp_replication().start_replication(self)
|
||||||
|
|
||||||
|
@ -147,7 +147,7 @@ class EventCreatorServer(HomeServer):
|
|||||||
)
|
)
|
||||||
elif listener["type"] == "metrics":
|
elif listener["type"] == "metrics":
|
||||||
if not self.get_config().enable_metrics:
|
if not self.get_config().enable_metrics:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
(
|
(
|
||||||
"Metrics listener configured, but "
|
"Metrics listener configured, but "
|
||||||
"enable_metrics is not True!"
|
"enable_metrics is not True!"
|
||||||
@ -156,7 +156,7 @@ class EventCreatorServer(HomeServer):
|
|||||||
else:
|
else:
|
||||||
_base.listen_metrics(listener["bind_addresses"], listener["port"])
|
_base.listen_metrics(listener["bind_addresses"], listener["port"])
|
||||||
else:
|
else:
|
||||||
logger.warn("Unrecognized listener type: %s", listener["type"])
|
logger.warning("Unrecognized listener type: %s", listener["type"])
|
||||||
|
|
||||||
self.get_tcp_replication().start_replication(self)
|
self.get_tcp_replication().start_replication(self)
|
||||||
|
|
||||||
|
@ -132,7 +132,7 @@ class FederationReaderServer(HomeServer):
|
|||||||
)
|
)
|
||||||
elif listener["type"] == "metrics":
|
elif listener["type"] == "metrics":
|
||||||
if not self.get_config().enable_metrics:
|
if not self.get_config().enable_metrics:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
(
|
(
|
||||||
"Metrics listener configured, but "
|
"Metrics listener configured, but "
|
||||||
"enable_metrics is not True!"
|
"enable_metrics is not True!"
|
||||||
@ -141,7 +141,7 @@ class FederationReaderServer(HomeServer):
|
|||||||
else:
|
else:
|
||||||
_base.listen_metrics(listener["bind_addresses"], listener["port"])
|
_base.listen_metrics(listener["bind_addresses"], listener["port"])
|
||||||
else:
|
else:
|
||||||
logger.warn("Unrecognized listener type: %s", listener["type"])
|
logger.warning("Unrecognized listener type: %s", listener["type"])
|
||||||
|
|
||||||
self.get_tcp_replication().start_replication(self)
|
self.get_tcp_replication().start_replication(self)
|
||||||
|
|
||||||
|
@ -123,7 +123,7 @@ class FederationSenderServer(HomeServer):
|
|||||||
)
|
)
|
||||||
elif listener["type"] == "metrics":
|
elif listener["type"] == "metrics":
|
||||||
if not self.get_config().enable_metrics:
|
if not self.get_config().enable_metrics:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
(
|
(
|
||||||
"Metrics listener configured, but "
|
"Metrics listener configured, but "
|
||||||
"enable_metrics is not True!"
|
"enable_metrics is not True!"
|
||||||
@ -132,7 +132,7 @@ class FederationSenderServer(HomeServer):
|
|||||||
else:
|
else:
|
||||||
_base.listen_metrics(listener["bind_addresses"], listener["port"])
|
_base.listen_metrics(listener["bind_addresses"], listener["port"])
|
||||||
else:
|
else:
|
||||||
logger.warn("Unrecognized listener type: %s", listener["type"])
|
logger.warning("Unrecognized listener type: %s", listener["type"])
|
||||||
|
|
||||||
self.get_tcp_replication().start_replication(self)
|
self.get_tcp_replication().start_replication(self)
|
||||||
|
|
||||||
|
@ -204,7 +204,7 @@ class FrontendProxyServer(HomeServer):
|
|||||||
)
|
)
|
||||||
elif listener["type"] == "metrics":
|
elif listener["type"] == "metrics":
|
||||||
if not self.get_config().enable_metrics:
|
if not self.get_config().enable_metrics:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
(
|
(
|
||||||
"Metrics listener configured, but "
|
"Metrics listener configured, but "
|
||||||
"enable_metrics is not True!"
|
"enable_metrics is not True!"
|
||||||
@ -213,7 +213,7 @@ class FrontendProxyServer(HomeServer):
|
|||||||
else:
|
else:
|
||||||
_base.listen_metrics(listener["bind_addresses"], listener["port"])
|
_base.listen_metrics(listener["bind_addresses"], listener["port"])
|
||||||
else:
|
else:
|
||||||
logger.warn("Unrecognized listener type: %s", listener["type"])
|
logger.warning("Unrecognized listener type: %s", listener["type"])
|
||||||
|
|
||||||
self.get_tcp_replication().start_replication(self)
|
self.get_tcp_replication().start_replication(self)
|
||||||
|
|
||||||
|
@ -282,7 +282,7 @@ class SynapseHomeServer(HomeServer):
|
|||||||
reactor.addSystemEventTrigger("before", "shutdown", s.stopListening)
|
reactor.addSystemEventTrigger("before", "shutdown", s.stopListening)
|
||||||
elif listener["type"] == "metrics":
|
elif listener["type"] == "metrics":
|
||||||
if not self.get_config().enable_metrics:
|
if not self.get_config().enable_metrics:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
(
|
(
|
||||||
"Metrics listener configured, but "
|
"Metrics listener configured, but "
|
||||||
"enable_metrics is not True!"
|
"enable_metrics is not True!"
|
||||||
@ -291,7 +291,7 @@ class SynapseHomeServer(HomeServer):
|
|||||||
else:
|
else:
|
||||||
_base.listen_metrics(listener["bind_addresses"], listener["port"])
|
_base.listen_metrics(listener["bind_addresses"], listener["port"])
|
||||||
else:
|
else:
|
||||||
logger.warn("Unrecognized listener type: %s", listener["type"])
|
logger.warning("Unrecognized listener type: %s", listener["type"])
|
||||||
|
|
||||||
def run_startup_checks(self, db_conn, database_engine):
|
def run_startup_checks(self, db_conn, database_engine):
|
||||||
all_users_native = are_all_users_on_domain(
|
all_users_native = are_all_users_on_domain(
|
||||||
@ -569,7 +569,7 @@ def run(hs):
|
|||||||
hs.config.report_stats_endpoint, stats
|
hs.config.report_stats_endpoint, stats
|
||||||
)
|
)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.warn("Error reporting stats: %s", e)
|
logger.warning("Error reporting stats: %s", e)
|
||||||
|
|
||||||
def performance_stats_init():
|
def performance_stats_init():
|
||||||
try:
|
try:
|
||||||
|
@ -120,7 +120,7 @@ class MediaRepositoryServer(HomeServer):
|
|||||||
)
|
)
|
||||||
elif listener["type"] == "metrics":
|
elif listener["type"] == "metrics":
|
||||||
if not self.get_config().enable_metrics:
|
if not self.get_config().enable_metrics:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
(
|
(
|
||||||
"Metrics listener configured, but "
|
"Metrics listener configured, but "
|
||||||
"enable_metrics is not True!"
|
"enable_metrics is not True!"
|
||||||
@ -129,7 +129,7 @@ class MediaRepositoryServer(HomeServer):
|
|||||||
else:
|
else:
|
||||||
_base.listen_metrics(listener["bind_addresses"], listener["port"])
|
_base.listen_metrics(listener["bind_addresses"], listener["port"])
|
||||||
else:
|
else:
|
||||||
logger.warn("Unrecognized listener type: %s", listener["type"])
|
logger.warning("Unrecognized listener type: %s", listener["type"])
|
||||||
|
|
||||||
self.get_tcp_replication().start_replication(self)
|
self.get_tcp_replication().start_replication(self)
|
||||||
|
|
||||||
|
@ -114,7 +114,7 @@ class PusherServer(HomeServer):
|
|||||||
)
|
)
|
||||||
elif listener["type"] == "metrics":
|
elif listener["type"] == "metrics":
|
||||||
if not self.get_config().enable_metrics:
|
if not self.get_config().enable_metrics:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
(
|
(
|
||||||
"Metrics listener configured, but "
|
"Metrics listener configured, but "
|
||||||
"enable_metrics is not True!"
|
"enable_metrics is not True!"
|
||||||
@ -123,7 +123,7 @@ class PusherServer(HomeServer):
|
|||||||
else:
|
else:
|
||||||
_base.listen_metrics(listener["bind_addresses"], listener["port"])
|
_base.listen_metrics(listener["bind_addresses"], listener["port"])
|
||||||
else:
|
else:
|
||||||
logger.warn("Unrecognized listener type: %s", listener["type"])
|
logger.warning("Unrecognized listener type: %s", listener["type"])
|
||||||
|
|
||||||
self.get_tcp_replication().start_replication(self)
|
self.get_tcp_replication().start_replication(self)
|
||||||
|
|
||||||
|
@ -326,7 +326,7 @@ class SynchrotronServer(HomeServer):
|
|||||||
)
|
)
|
||||||
elif listener["type"] == "metrics":
|
elif listener["type"] == "metrics":
|
||||||
if not self.get_config().enable_metrics:
|
if not self.get_config().enable_metrics:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
(
|
(
|
||||||
"Metrics listener configured, but "
|
"Metrics listener configured, but "
|
||||||
"enable_metrics is not True!"
|
"enable_metrics is not True!"
|
||||||
@ -335,7 +335,7 @@ class SynchrotronServer(HomeServer):
|
|||||||
else:
|
else:
|
||||||
_base.listen_metrics(listener["bind_addresses"], listener["port"])
|
_base.listen_metrics(listener["bind_addresses"], listener["port"])
|
||||||
else:
|
else:
|
||||||
logger.warn("Unrecognized listener type: %s", listener["type"])
|
logger.warning("Unrecognized listener type: %s", listener["type"])
|
||||||
|
|
||||||
self.get_tcp_replication().start_replication(self)
|
self.get_tcp_replication().start_replication(self)
|
||||||
|
|
||||||
|
@ -150,7 +150,7 @@ class UserDirectoryServer(HomeServer):
|
|||||||
)
|
)
|
||||||
elif listener["type"] == "metrics":
|
elif listener["type"] == "metrics":
|
||||||
if not self.get_config().enable_metrics:
|
if not self.get_config().enable_metrics:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
(
|
(
|
||||||
"Metrics listener configured, but "
|
"Metrics listener configured, but "
|
||||||
"enable_metrics is not True!"
|
"enable_metrics is not True!"
|
||||||
@ -159,7 +159,7 @@ class UserDirectoryServer(HomeServer):
|
|||||||
else:
|
else:
|
||||||
_base.listen_metrics(listener["bind_addresses"], listener["port"])
|
_base.listen_metrics(listener["bind_addresses"], listener["port"])
|
||||||
else:
|
else:
|
||||||
logger.warn("Unrecognized listener type: %s", listener["type"])
|
logger.warning("Unrecognized listener type: %s", listener["type"])
|
||||||
|
|
||||||
self.get_tcp_replication().start_replication(self)
|
self.get_tcp_replication().start_replication(self)
|
||||||
|
|
||||||
|
@ -125,7 +125,7 @@ class KeyConfig(Config):
|
|||||||
|
|
||||||
# if neither trusted_key_servers nor perspectives are given, use the default.
|
# if neither trusted_key_servers nor perspectives are given, use the default.
|
||||||
if "perspectives" not in config and "trusted_key_servers" not in config:
|
if "perspectives" not in config and "trusted_key_servers" not in config:
|
||||||
logger.warn(TRUSTED_KEY_SERVER_NOT_CONFIGURED_WARN)
|
logger.warning(TRUSTED_KEY_SERVER_NOT_CONFIGURED_WARN)
|
||||||
key_servers = [{"server_name": "matrix.org"}]
|
key_servers = [{"server_name": "matrix.org"}]
|
||||||
else:
|
else:
|
||||||
key_servers = config.get("trusted_key_servers", [])
|
key_servers = config.get("trusted_key_servers", [])
|
||||||
@ -156,7 +156,7 @@ class KeyConfig(Config):
|
|||||||
if not self.macaroon_secret_key:
|
if not self.macaroon_secret_key:
|
||||||
# Unfortunately, there are people out there that don't have this
|
# Unfortunately, there are people out there that don't have this
|
||||||
# set. Lets just be "nice" and derive one from their secret key.
|
# set. Lets just be "nice" and derive one from their secret key.
|
||||||
logger.warn("Config is missing macaroon_secret_key")
|
logger.warning("Config is missing macaroon_secret_key")
|
||||||
seed = bytes(self.signing_key[0])
|
seed = bytes(self.signing_key[0])
|
||||||
self.macaroon_secret_key = hashlib.sha256(seed).digest()
|
self.macaroon_secret_key = hashlib.sha256(seed).digest()
|
||||||
|
|
||||||
|
@ -182,7 +182,7 @@ def _reload_stdlib_logging(*args, log_config=None):
|
|||||||
logger = logging.getLogger("")
|
logger = logging.getLogger("")
|
||||||
|
|
||||||
if not log_config:
|
if not log_config:
|
||||||
logger.warn("Reloaded a blank config?")
|
logger.warning("Reloaded a blank config?")
|
||||||
|
|
||||||
logging.config.dictConfig(log_config)
|
logging.config.dictConfig(log_config)
|
||||||
|
|
||||||
|
@ -77,7 +77,7 @@ def check(room_version, event, auth_events, do_sig_check=True, do_size_check=Tru
|
|||||||
if auth_events is None:
|
if auth_events is None:
|
||||||
# Oh, we don't know what the state of the room was, so we
|
# Oh, we don't know what the state of the room was, so we
|
||||||
# are trusting that this is allowed (at least for now)
|
# are trusting that this is allowed (at least for now)
|
||||||
logger.warn("Trusting event: %s", event.event_id)
|
logger.warning("Trusting event: %s", event.event_id)
|
||||||
return
|
return
|
||||||
|
|
||||||
if event.type == EventTypes.Create:
|
if event.type == EventTypes.Create:
|
||||||
|
@ -102,7 +102,7 @@ class FederationBase(object):
|
|||||||
pass
|
pass
|
||||||
|
|
||||||
if not res:
|
if not res:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Failed to find copy of %s with valid signature", pdu.event_id
|
"Failed to find copy of %s with valid signature", pdu.event_id
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -173,7 +173,7 @@ class FederationBase(object):
|
|||||||
return redacted_event
|
return redacted_event
|
||||||
|
|
||||||
if self.spam_checker.check_event_for_spam(pdu):
|
if self.spam_checker.check_event_for_spam(pdu):
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Event contains spam, redacting %s: %s",
|
"Event contains spam, redacting %s: %s",
|
||||||
pdu.event_id,
|
pdu.event_id,
|
||||||
pdu.get_pdu_json(),
|
pdu.get_pdu_json(),
|
||||||
@ -185,7 +185,7 @@ class FederationBase(object):
|
|||||||
def errback(failure, pdu):
|
def errback(failure, pdu):
|
||||||
failure.trap(SynapseError)
|
failure.trap(SynapseError)
|
||||||
with PreserveLoggingContext(ctx):
|
with PreserveLoggingContext(ctx):
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Signature check failed for %s: %s",
|
"Signature check failed for %s: %s",
|
||||||
pdu.event_id,
|
pdu.event_id,
|
||||||
failure.getErrorMessage(),
|
failure.getErrorMessage(),
|
||||||
|
@ -522,12 +522,12 @@ class FederationClient(FederationBase):
|
|||||||
res = yield callback(destination)
|
res = yield callback(destination)
|
||||||
return res
|
return res
|
||||||
except InvalidResponseError as e:
|
except InvalidResponseError as e:
|
||||||
logger.warn("Failed to %s via %s: %s", description, destination, e)
|
logger.warning("Failed to %s via %s: %s", description, destination, e)
|
||||||
except HttpResponseException as e:
|
except HttpResponseException as e:
|
||||||
if not 500 <= e.code < 600:
|
if not 500 <= e.code < 600:
|
||||||
raise e.to_synapse_error()
|
raise e.to_synapse_error()
|
||||||
else:
|
else:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Failed to %s via %s: %i %s",
|
"Failed to %s via %s: %i %s",
|
||||||
description,
|
description,
|
||||||
destination,
|
destination,
|
||||||
@ -535,7 +535,9 @@ class FederationClient(FederationBase):
|
|||||||
e.args[0],
|
e.args[0],
|
||||||
)
|
)
|
||||||
except Exception:
|
except Exception:
|
||||||
logger.warn("Failed to %s via %s", description, destination, exc_info=1)
|
logger.warning(
|
||||||
|
"Failed to %s via %s", description, destination, exc_info=1
|
||||||
|
)
|
||||||
|
|
||||||
raise SynapseError(502, "Failed to %s via any server" % (description,))
|
raise SynapseError(502, "Failed to %s via any server" % (description,))
|
||||||
|
|
||||||
|
@ -220,7 +220,7 @@ class FederationServer(FederationBase):
|
|||||||
try:
|
try:
|
||||||
await self.check_server_matches_acl(origin_host, room_id)
|
await self.check_server_matches_acl(origin_host, room_id)
|
||||||
except AuthError as e:
|
except AuthError as e:
|
||||||
logger.warn("Ignoring PDUs for room %s from banned server", room_id)
|
logger.warning("Ignoring PDUs for room %s from banned server", room_id)
|
||||||
for pdu in pdus_by_room[room_id]:
|
for pdu in pdus_by_room[room_id]:
|
||||||
event_id = pdu.event_id
|
event_id = pdu.event_id
|
||||||
pdu_results[event_id] = e.error_dict()
|
pdu_results[event_id] = e.error_dict()
|
||||||
@ -233,7 +233,7 @@ class FederationServer(FederationBase):
|
|||||||
await self._handle_received_pdu(origin, pdu)
|
await self._handle_received_pdu(origin, pdu)
|
||||||
pdu_results[event_id] = {}
|
pdu_results[event_id] = {}
|
||||||
except FederationError as e:
|
except FederationError as e:
|
||||||
logger.warn("Error handling PDU %s: %s", event_id, e)
|
logger.warning("Error handling PDU %s: %s", event_id, e)
|
||||||
pdu_results[event_id] = {"error": str(e)}
|
pdu_results[event_id] = {"error": str(e)}
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
f = failure.Failure()
|
f = failure.Failure()
|
||||||
@ -333,7 +333,9 @@ class FederationServer(FederationBase):
|
|||||||
|
|
||||||
room_version = await self.store.get_room_version(room_id)
|
room_version = await self.store.get_room_version(room_id)
|
||||||
if room_version not in supported_versions:
|
if room_version not in supported_versions:
|
||||||
logger.warn("Room version %s not in %s", room_version, supported_versions)
|
logger.warning(
|
||||||
|
"Room version %s not in %s", room_version, supported_versions
|
||||||
|
)
|
||||||
raise IncompatibleRoomVersionError(room_version=room_version)
|
raise IncompatibleRoomVersionError(room_version=room_version)
|
||||||
|
|
||||||
pdu = await self.handler.on_make_join_request(origin, room_id, user_id)
|
pdu = await self.handler.on_make_join_request(origin, room_id, user_id)
|
||||||
@ -679,7 +681,7 @@ def server_matches_acl_event(server_name, acl_event):
|
|||||||
# server name is a literal IP
|
# server name is a literal IP
|
||||||
allow_ip_literals = acl_event.content.get("allow_ip_literals", True)
|
allow_ip_literals = acl_event.content.get("allow_ip_literals", True)
|
||||||
if not isinstance(allow_ip_literals, bool):
|
if not isinstance(allow_ip_literals, bool):
|
||||||
logger.warn("Ignorning non-bool allow_ip_literals flag")
|
logger.warning("Ignorning non-bool allow_ip_literals flag")
|
||||||
allow_ip_literals = True
|
allow_ip_literals = True
|
||||||
if not allow_ip_literals:
|
if not allow_ip_literals:
|
||||||
# check for ipv6 literals. These start with '['.
|
# check for ipv6 literals. These start with '['.
|
||||||
@ -693,7 +695,7 @@ def server_matches_acl_event(server_name, acl_event):
|
|||||||
# next, check the deny list
|
# next, check the deny list
|
||||||
deny = acl_event.content.get("deny", [])
|
deny = acl_event.content.get("deny", [])
|
||||||
if not isinstance(deny, (list, tuple)):
|
if not isinstance(deny, (list, tuple)):
|
||||||
logger.warn("Ignorning non-list deny ACL %s", deny)
|
logger.warning("Ignorning non-list deny ACL %s", deny)
|
||||||
deny = []
|
deny = []
|
||||||
for e in deny:
|
for e in deny:
|
||||||
if _acl_entry_matches(server_name, e):
|
if _acl_entry_matches(server_name, e):
|
||||||
@ -703,7 +705,7 @@ def server_matches_acl_event(server_name, acl_event):
|
|||||||
# then the allow list.
|
# then the allow list.
|
||||||
allow = acl_event.content.get("allow", [])
|
allow = acl_event.content.get("allow", [])
|
||||||
if not isinstance(allow, (list, tuple)):
|
if not isinstance(allow, (list, tuple)):
|
||||||
logger.warn("Ignorning non-list allow ACL %s", allow)
|
logger.warning("Ignorning non-list allow ACL %s", allow)
|
||||||
allow = []
|
allow = []
|
||||||
for e in allow:
|
for e in allow:
|
||||||
if _acl_entry_matches(server_name, e):
|
if _acl_entry_matches(server_name, e):
|
||||||
@ -717,7 +719,7 @@ def server_matches_acl_event(server_name, acl_event):
|
|||||||
|
|
||||||
def _acl_entry_matches(server_name, acl_entry):
|
def _acl_entry_matches(server_name, acl_entry):
|
||||||
if not isinstance(acl_entry, six.string_types):
|
if not isinstance(acl_entry, six.string_types):
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Ignoring non-str ACL entry '%s' (is %s)", acl_entry, type(acl_entry)
|
"Ignoring non-str ACL entry '%s' (is %s)", acl_entry, type(acl_entry)
|
||||||
)
|
)
|
||||||
return False
|
return False
|
||||||
@ -772,7 +774,7 @@ class FederationHandlerRegistry(object):
|
|||||||
async def on_edu(self, edu_type, origin, content):
|
async def on_edu(self, edu_type, origin, content):
|
||||||
handler = self.edu_handlers.get(edu_type)
|
handler = self.edu_handlers.get(edu_type)
|
||||||
if not handler:
|
if not handler:
|
||||||
logger.warn("No handler registered for EDU type %s", edu_type)
|
logger.warning("No handler registered for EDU type %s", edu_type)
|
||||||
|
|
||||||
with start_active_span_from_edu(content, "handle_edu"):
|
with start_active_span_from_edu(content, "handle_edu"):
|
||||||
try:
|
try:
|
||||||
@ -785,7 +787,7 @@ class FederationHandlerRegistry(object):
|
|||||||
def on_query(self, query_type, args):
|
def on_query(self, query_type, args):
|
||||||
handler = self.query_handlers.get(query_type)
|
handler = self.query_handlers.get(query_type)
|
||||||
if not handler:
|
if not handler:
|
||||||
logger.warn("No handler registered for query type %s", query_type)
|
logger.warning("No handler registered for query type %s", query_type)
|
||||||
raise NotFoundError("No handler for Query type '%s'" % (query_type,))
|
raise NotFoundError("No handler for Query type '%s'" % (query_type,))
|
||||||
|
|
||||||
return handler(args)
|
return handler(args)
|
||||||
|
@ -146,7 +146,7 @@ class TransactionManager(object):
|
|||||||
if code == 200:
|
if code == 200:
|
||||||
for e_id, r in response.get("pdus", {}).items():
|
for e_id, r in response.get("pdus", {}).items():
|
||||||
if "error" in r:
|
if "error" in r:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"TX [%s] {%s} Remote returned error for %s: %s",
|
"TX [%s] {%s} Remote returned error for %s: %s",
|
||||||
destination,
|
destination,
|
||||||
txn_id,
|
txn_id,
|
||||||
@ -155,7 +155,7 @@ class TransactionManager(object):
|
|||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
for p in pdus:
|
for p in pdus:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"TX [%s] {%s} Failed to send event %s",
|
"TX [%s] {%s} Failed to send event %s",
|
||||||
destination,
|
destination,
|
||||||
txn_id,
|
txn_id,
|
||||||
|
@ -202,7 +202,7 @@ def _parse_auth_header(header_bytes):
|
|||||||
sig = strip_quotes(param_dict["sig"])
|
sig = strip_quotes(param_dict["sig"])
|
||||||
return origin, key, sig
|
return origin, key, sig
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Error parsing auth header '%s': %s",
|
"Error parsing auth header '%s': %s",
|
||||||
header_bytes.decode("ascii", "replace"),
|
header_bytes.decode("ascii", "replace"),
|
||||||
e,
|
e,
|
||||||
@ -287,10 +287,12 @@ class BaseFederationServlet(object):
|
|||||||
except NoAuthenticationError:
|
except NoAuthenticationError:
|
||||||
origin = None
|
origin = None
|
||||||
if self.REQUIRE_AUTH:
|
if self.REQUIRE_AUTH:
|
||||||
logger.warn("authenticate_request failed: missing authentication")
|
logger.warning(
|
||||||
|
"authenticate_request failed: missing authentication"
|
||||||
|
)
|
||||||
raise
|
raise
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.warn("authenticate_request failed: %s", e)
|
logger.warning("authenticate_request failed: %s", e)
|
||||||
raise
|
raise
|
||||||
|
|
||||||
request_tags = {
|
request_tags = {
|
||||||
|
@ -181,7 +181,7 @@ class GroupAttestionRenewer(object):
|
|||||||
elif not self.is_mine_id(user_id):
|
elif not self.is_mine_id(user_id):
|
||||||
destination = get_domain_from_id(user_id)
|
destination = get_domain_from_id(user_id)
|
||||||
else:
|
else:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Incorrectly trying to do attestations for user: %r in %r",
|
"Incorrectly trying to do attestations for user: %r in %r",
|
||||||
user_id,
|
user_id,
|
||||||
group_id,
|
group_id,
|
||||||
|
@ -488,7 +488,7 @@ class GroupsServerHandler(object):
|
|||||||
profile = yield self.profile_handler.get_profile_from_cache(user_id)
|
profile = yield self.profile_handler.get_profile_from_cache(user_id)
|
||||||
user_profile.update(profile)
|
user_profile.update(profile)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.warn("Error getting profile for %s: %s", user_id, e)
|
logger.warning("Error getting profile for %s: %s", user_id, e)
|
||||||
user_profiles.append(user_profile)
|
user_profiles.append(user_profile)
|
||||||
|
|
||||||
return {"chunk": user_profiles, "total_user_count_estimate": len(invited_users)}
|
return {"chunk": user_profiles, "total_user_count_estimate": len(invited_users)}
|
||||||
|
@ -525,7 +525,7 @@ class AuthHandler(BaseHandler):
|
|||||||
|
|
||||||
result = None
|
result = None
|
||||||
if not user_infos:
|
if not user_infos:
|
||||||
logger.warn("Attempted to login as %s but they do not exist", user_id)
|
logger.warning("Attempted to login as %s but they do not exist", user_id)
|
||||||
elif len(user_infos) == 1:
|
elif len(user_infos) == 1:
|
||||||
# a single match (possibly not exact)
|
# a single match (possibly not exact)
|
||||||
result = user_infos.popitem()
|
result = user_infos.popitem()
|
||||||
@ -534,7 +534,7 @@ class AuthHandler(BaseHandler):
|
|||||||
result = (user_id, user_infos[user_id])
|
result = (user_id, user_infos[user_id])
|
||||||
else:
|
else:
|
||||||
# multiple matches, none of them exact
|
# multiple matches, none of them exact
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Attempted to login as %s but it matches more than one user "
|
"Attempted to login as %s but it matches more than one user "
|
||||||
"inexactly: %r",
|
"inexactly: %r",
|
||||||
user_id,
|
user_id,
|
||||||
@ -728,7 +728,7 @@ class AuthHandler(BaseHandler):
|
|||||||
|
|
||||||
result = yield self.validate_hash(password, password_hash)
|
result = yield self.validate_hash(password, password_hash)
|
||||||
if not result:
|
if not result:
|
||||||
logger.warn("Failed password login for user %s", user_id)
|
logger.warning("Failed password login for user %s", user_id)
|
||||||
return None
|
return None
|
||||||
return user_id
|
return user_id
|
||||||
|
|
||||||
|
@ -656,7 +656,7 @@ class DeviceListUpdater(object):
|
|||||||
except (NotRetryingDestination, RequestSendFailed, HttpResponseException):
|
except (NotRetryingDestination, RequestSendFailed, HttpResponseException):
|
||||||
# TODO: Remember that we are now out of sync and try again
|
# TODO: Remember that we are now out of sync and try again
|
||||||
# later
|
# later
|
||||||
logger.warn("Failed to handle device list update for %s", user_id)
|
logger.warning("Failed to handle device list update for %s", user_id)
|
||||||
# We abort on exceptions rather than accepting the update
|
# We abort on exceptions rather than accepting the update
|
||||||
# as otherwise synapse will 'forget' that its device list
|
# as otherwise synapse will 'forget' that its device list
|
||||||
# is out of date. If we bail then we will retry the resync
|
# is out of date. If we bail then we will retry the resync
|
||||||
@ -694,7 +694,7 @@ class DeviceListUpdater(object):
|
|||||||
# up on storing the total list of devices and only handle the
|
# up on storing the total list of devices and only handle the
|
||||||
# delta instead.
|
# delta instead.
|
||||||
if len(devices) > 1000:
|
if len(devices) > 1000:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Ignoring device list snapshot for %s as it has >1K devs (%d)",
|
"Ignoring device list snapshot for %s as it has >1K devs (%d)",
|
||||||
user_id,
|
user_id,
|
||||||
len(devices),
|
len(devices),
|
||||||
|
@ -52,7 +52,7 @@ class DeviceMessageHandler(object):
|
|||||||
local_messages = {}
|
local_messages = {}
|
||||||
sender_user_id = content["sender"]
|
sender_user_id = content["sender"]
|
||||||
if origin != get_domain_from_id(sender_user_id):
|
if origin != get_domain_from_id(sender_user_id):
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Dropping device message from %r with spoofed sender %r",
|
"Dropping device message from %r with spoofed sender %r",
|
||||||
origin,
|
origin,
|
||||||
sender_user_id,
|
sender_user_id,
|
||||||
|
@ -181,7 +181,7 @@ class FederationHandler(BaseHandler):
|
|||||||
try:
|
try:
|
||||||
self._sanity_check_event(pdu)
|
self._sanity_check_event(pdu)
|
||||||
except SynapseError as err:
|
except SynapseError as err:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"[%s %s] Received event failed sanity checks", room_id, event_id
|
"[%s %s] Received event failed sanity checks", room_id, event_id
|
||||||
)
|
)
|
||||||
raise FederationError("ERROR", err.code, err.msg, affected=pdu.event_id)
|
raise FederationError("ERROR", err.code, err.msg, affected=pdu.event_id)
|
||||||
@ -302,7 +302,7 @@ class FederationHandler(BaseHandler):
|
|||||||
# following.
|
# following.
|
||||||
|
|
||||||
if sent_to_us_directly:
|
if sent_to_us_directly:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"[%s %s] Rejecting: failed to fetch %d prev events: %s",
|
"[%s %s] Rejecting: failed to fetch %d prev events: %s",
|
||||||
room_id,
|
room_id,
|
||||||
event_id,
|
event_id,
|
||||||
@ -406,7 +406,7 @@ class FederationHandler(BaseHandler):
|
|||||||
state = [event_map[e] for e in six.itervalues(state_map)]
|
state = [event_map[e] for e in six.itervalues(state_map)]
|
||||||
auth_chain = list(auth_chains)
|
auth_chain = list(auth_chains)
|
||||||
except Exception:
|
except Exception:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"[%s %s] Error attempting to resolve state at missing "
|
"[%s %s] Error attempting to resolve state at missing "
|
||||||
"prev_events",
|
"prev_events",
|
||||||
room_id,
|
room_id,
|
||||||
@ -519,7 +519,9 @@ class FederationHandler(BaseHandler):
|
|||||||
# We failed to get the missing events, but since we need to handle
|
# We failed to get the missing events, but since we need to handle
|
||||||
# the case of `get_missing_events` not returning the necessary
|
# the case of `get_missing_events` not returning the necessary
|
||||||
# events anyway, it is safe to simply log the error and continue.
|
# events anyway, it is safe to simply log the error and continue.
|
||||||
logger.warn("[%s %s]: Failed to get prev_events: %s", room_id, event_id, e)
|
logger.warning(
|
||||||
|
"[%s %s]: Failed to get prev_events: %s", room_id, event_id, e
|
||||||
|
)
|
||||||
return
|
return
|
||||||
|
|
||||||
logger.info(
|
logger.info(
|
||||||
@ -546,7 +548,7 @@ class FederationHandler(BaseHandler):
|
|||||||
yield self.on_receive_pdu(origin, ev, sent_to_us_directly=False)
|
yield self.on_receive_pdu(origin, ev, sent_to_us_directly=False)
|
||||||
except FederationError as e:
|
except FederationError as e:
|
||||||
if e.code == 403:
|
if e.code == 403:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"[%s %s] Received prev_event %s failed history check.",
|
"[%s %s] Received prev_event %s failed history check.",
|
||||||
room_id,
|
room_id,
|
||||||
event_id,
|
event_id,
|
||||||
@ -1060,7 +1062,7 @@ class FederationHandler(BaseHandler):
|
|||||||
SynapseError if the event does not pass muster
|
SynapseError if the event does not pass muster
|
||||||
"""
|
"""
|
||||||
if len(ev.prev_event_ids()) > 20:
|
if len(ev.prev_event_ids()) > 20:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Rejecting event %s which has %i prev_events",
|
"Rejecting event %s which has %i prev_events",
|
||||||
ev.event_id,
|
ev.event_id,
|
||||||
len(ev.prev_event_ids()),
|
len(ev.prev_event_ids()),
|
||||||
@ -1068,7 +1070,7 @@ class FederationHandler(BaseHandler):
|
|||||||
raise SynapseError(http_client.BAD_REQUEST, "Too many prev_events")
|
raise SynapseError(http_client.BAD_REQUEST, "Too many prev_events")
|
||||||
|
|
||||||
if len(ev.auth_event_ids()) > 10:
|
if len(ev.auth_event_ids()) > 10:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Rejecting event %s which has %i auth_events",
|
"Rejecting event %s which has %i auth_events",
|
||||||
ev.event_id,
|
ev.event_id,
|
||||||
len(ev.auth_event_ids()),
|
len(ev.auth_event_ids()),
|
||||||
@ -1204,7 +1206,7 @@ class FederationHandler(BaseHandler):
|
|||||||
with nested_logging_context(p.event_id):
|
with nested_logging_context(p.event_id):
|
||||||
yield self.on_receive_pdu(origin, p, sent_to_us_directly=True)
|
yield self.on_receive_pdu(origin, p, sent_to_us_directly=True)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Error handling queued PDU %s from %s: %s", p.event_id, origin, e
|
"Error handling queued PDU %s from %s: %s", p.event_id, origin, e
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -1251,7 +1253,7 @@ class FederationHandler(BaseHandler):
|
|||||||
builder=builder
|
builder=builder
|
||||||
)
|
)
|
||||||
except AuthError as e:
|
except AuthError as e:
|
||||||
logger.warn("Failed to create join to %s because %s", room_id, e)
|
logger.warning("Failed to create join to %s because %s", room_id, e)
|
||||||
raise e
|
raise e
|
||||||
|
|
||||||
event_allowed = yield self.third_party_event_rules.check_event_allowed(
|
event_allowed = yield self.third_party_event_rules.check_event_allowed(
|
||||||
@ -1495,7 +1497,7 @@ class FederationHandler(BaseHandler):
|
|||||||
room_version, event, context, do_sig_check=False
|
room_version, event, context, do_sig_check=False
|
||||||
)
|
)
|
||||||
except AuthError as e:
|
except AuthError as e:
|
||||||
logger.warn("Failed to create new leave %r because %s", event, e)
|
logger.warning("Failed to create new leave %r because %s", event, e)
|
||||||
raise e
|
raise e
|
||||||
|
|
||||||
return event
|
return event
|
||||||
@ -1789,7 +1791,7 @@ class FederationHandler(BaseHandler):
|
|||||||
# cause SynapseErrors in auth.check. We don't want to give up
|
# cause SynapseErrors in auth.check. We don't want to give up
|
||||||
# the attempt to federate altogether in such cases.
|
# the attempt to federate altogether in such cases.
|
||||||
|
|
||||||
logger.warn("Rejecting %s because %s", e.event_id, err.msg)
|
logger.warning("Rejecting %s because %s", e.event_id, err.msg)
|
||||||
|
|
||||||
if e == event:
|
if e == event:
|
||||||
raise
|
raise
|
||||||
@ -1845,7 +1847,9 @@ class FederationHandler(BaseHandler):
|
|||||||
try:
|
try:
|
||||||
yield self.do_auth(origin, event, context, auth_events=auth_events)
|
yield self.do_auth(origin, event, context, auth_events=auth_events)
|
||||||
except AuthError as e:
|
except AuthError as e:
|
||||||
logger.warn("[%s %s] Rejecting: %s", event.room_id, event.event_id, e.msg)
|
logger.warning(
|
||||||
|
"[%s %s] Rejecting: %s", event.room_id, event.event_id, e.msg
|
||||||
|
)
|
||||||
|
|
||||||
context.rejected = RejectedReason.AUTH_ERROR
|
context.rejected = RejectedReason.AUTH_ERROR
|
||||||
|
|
||||||
@ -1939,7 +1943,7 @@ class FederationHandler(BaseHandler):
|
|||||||
try:
|
try:
|
||||||
event_auth.check(room_version, event, auth_events=current_auth_events)
|
event_auth.check(room_version, event, auth_events=current_auth_events)
|
||||||
except AuthError as e:
|
except AuthError as e:
|
||||||
logger.warn("Soft-failing %r because %s", event, e)
|
logger.warning("Soft-failing %r because %s", event, e)
|
||||||
event.internal_metadata.soft_failed = True
|
event.internal_metadata.soft_failed = True
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
@ -2038,7 +2042,7 @@ class FederationHandler(BaseHandler):
|
|||||||
try:
|
try:
|
||||||
event_auth.check(room_version, event, auth_events=auth_events)
|
event_auth.check(room_version, event, auth_events=auth_events)
|
||||||
except AuthError as e:
|
except AuthError as e:
|
||||||
logger.warn("Failed auth resolution for %r because %s", event, e)
|
logger.warning("Failed auth resolution for %r because %s", event, e)
|
||||||
raise e
|
raise e
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
@ -2432,7 +2436,7 @@ class FederationHandler(BaseHandler):
|
|||||||
try:
|
try:
|
||||||
yield self.auth.check_from_context(room_version, event, context)
|
yield self.auth.check_from_context(room_version, event, context)
|
||||||
except AuthError as e:
|
except AuthError as e:
|
||||||
logger.warn("Denying new third party invite %r because %s", event, e)
|
logger.warning("Denying new third party invite %r because %s", event, e)
|
||||||
raise e
|
raise e
|
||||||
|
|
||||||
yield self._check_signature(event, context)
|
yield self._check_signature(event, context)
|
||||||
@ -2488,7 +2492,7 @@ class FederationHandler(BaseHandler):
|
|||||||
try:
|
try:
|
||||||
yield self.auth.check_from_context(room_version, event, context)
|
yield self.auth.check_from_context(room_version, event, context)
|
||||||
except AuthError as e:
|
except AuthError as e:
|
||||||
logger.warn("Denying third party invite %r because %s", event, e)
|
logger.warning("Denying third party invite %r because %s", event, e)
|
||||||
raise e
|
raise e
|
||||||
yield self._check_signature(event, context)
|
yield self._check_signature(event, context)
|
||||||
|
|
||||||
|
@ -392,7 +392,7 @@ class GroupsLocalHandler(object):
|
|||||||
try:
|
try:
|
||||||
user_profile = yield self.profile_handler.get_profile(user_id)
|
user_profile = yield self.profile_handler.get_profile(user_id)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.warn("No profile for user %s: %s", user_id, e)
|
logger.warning("No profile for user %s: %s", user_id, e)
|
||||||
user_profile = {}
|
user_profile = {}
|
||||||
|
|
||||||
return {"state": "invite", "user_profile": user_profile}
|
return {"state": "invite", "user_profile": user_profile}
|
||||||
|
@ -272,7 +272,7 @@ class IdentityHandler(BaseHandler):
|
|||||||
changed = False
|
changed = False
|
||||||
if e.code in (400, 404, 501):
|
if e.code in (400, 404, 501):
|
||||||
# The remote server probably doesn't support unbinding (yet)
|
# The remote server probably doesn't support unbinding (yet)
|
||||||
logger.warn("Received %d response while unbinding threepid", e.code)
|
logger.warning("Received %d response while unbinding threepid", e.code)
|
||||||
else:
|
else:
|
||||||
logger.error("Failed to unbind threepid on identity server: %s", e)
|
logger.error("Failed to unbind threepid on identity server: %s", e)
|
||||||
raise SynapseError(500, "Failed to contact identity server")
|
raise SynapseError(500, "Failed to contact identity server")
|
||||||
@ -403,7 +403,7 @@ class IdentityHandler(BaseHandler):
|
|||||||
|
|
||||||
if self.hs.config.using_identity_server_from_trusted_list:
|
if self.hs.config.using_identity_server_from_trusted_list:
|
||||||
# Warn that a deprecated config option is in use
|
# Warn that a deprecated config option is in use
|
||||||
logger.warn(
|
logger.warning(
|
||||||
'The config option "trust_identity_server_for_password_resets" '
|
'The config option "trust_identity_server_for_password_resets" '
|
||||||
'has been replaced by "account_threepid_delegate". '
|
'has been replaced by "account_threepid_delegate". '
|
||||||
"Please consult the sample config at docs/sample_config.yaml for "
|
"Please consult the sample config at docs/sample_config.yaml for "
|
||||||
@ -457,7 +457,7 @@ class IdentityHandler(BaseHandler):
|
|||||||
|
|
||||||
if self.hs.config.using_identity_server_from_trusted_list:
|
if self.hs.config.using_identity_server_from_trusted_list:
|
||||||
# Warn that a deprecated config option is in use
|
# Warn that a deprecated config option is in use
|
||||||
logger.warn(
|
logger.warning(
|
||||||
'The config option "trust_identity_server_for_password_resets" '
|
'The config option "trust_identity_server_for_password_resets" '
|
||||||
'has been replaced by "account_threepid_delegate". '
|
'has been replaced by "account_threepid_delegate". '
|
||||||
"Please consult the sample config at docs/sample_config.yaml for "
|
"Please consult the sample config at docs/sample_config.yaml for "
|
||||||
|
@ -688,7 +688,7 @@ class EventCreationHandler(object):
|
|||||||
try:
|
try:
|
||||||
yield self.auth.check_from_context(room_version, event, context)
|
yield self.auth.check_from_context(room_version, event, context)
|
||||||
except AuthError as err:
|
except AuthError as err:
|
||||||
logger.warn("Denying new event %r because %s", event, err)
|
logger.warning("Denying new event %r because %s", event, err)
|
||||||
raise err
|
raise err
|
||||||
|
|
||||||
# Ensure that we can round trip before trying to persist in db
|
# Ensure that we can round trip before trying to persist in db
|
||||||
|
@ -275,7 +275,7 @@ class BaseProfileHandler(BaseHandler):
|
|||||||
ratelimit=False, # Try to hide that these events aren't atomic.
|
ratelimit=False, # Try to hide that these events aren't atomic.
|
||||||
)
|
)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Failed to update join event for room %s - %s", room_id, str(e)
|
"Failed to update join event for room %s - %s", room_id, str(e)
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -922,7 +922,7 @@ class RoomEventSource(object):
|
|||||||
|
|
||||||
from_token = RoomStreamToken.parse(from_key)
|
from_token = RoomStreamToken.parse(from_key)
|
||||||
if from_token.topological:
|
if from_token.topological:
|
||||||
logger.warn("Stream has topological part!!!! %r", from_key)
|
logger.warning("Stream has topological part!!!! %r", from_key)
|
||||||
from_key = "s%s" % (from_token.stream,)
|
from_key = "s%s" % (from_token.stream,)
|
||||||
|
|
||||||
app_service = self.store.get_app_service_by_user_id(user.to_string())
|
app_service = self.store.get_app_service_by_user_id(user.to_string())
|
||||||
|
@ -535,7 +535,7 @@ class SimpleHttpClient(object):
|
|||||||
b"Content-Length" in resp_headers
|
b"Content-Length" in resp_headers
|
||||||
and int(resp_headers[b"Content-Length"][0]) > max_size
|
and int(resp_headers[b"Content-Length"][0]) > max_size
|
||||||
):
|
):
|
||||||
logger.warn("Requested URL is too large > %r bytes" % (self.max_size,))
|
logger.warning("Requested URL is too large > %r bytes" % (self.max_size,))
|
||||||
raise SynapseError(
|
raise SynapseError(
|
||||||
502,
|
502,
|
||||||
"Requested file is too large > %r bytes" % (self.max_size,),
|
"Requested file is too large > %r bytes" % (self.max_size,),
|
||||||
@ -543,7 +543,7 @@ class SimpleHttpClient(object):
|
|||||||
)
|
)
|
||||||
|
|
||||||
if response.code > 299:
|
if response.code > 299:
|
||||||
logger.warn("Got %d when downloading %s" % (response.code, url))
|
logger.warning("Got %d when downloading %s" % (response.code, url))
|
||||||
raise SynapseError(502, "Got error %d" % (response.code,), Codes.UNKNOWN)
|
raise SynapseError(502, "Got error %d" % (response.code,), Codes.UNKNOWN)
|
||||||
|
|
||||||
# TODO: if our Content-Type is HTML or something, just read the first
|
# TODO: if our Content-Type is HTML or something, just read the first
|
||||||
|
@ -148,7 +148,7 @@ class SrvResolver(object):
|
|||||||
# Try something in the cache, else rereaise
|
# Try something in the cache, else rereaise
|
||||||
cache_entry = self._cache.get(service_name, None)
|
cache_entry = self._cache.get(service_name, None)
|
||||||
if cache_entry:
|
if cache_entry:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Failed to resolve %r, falling back to cache. %r", service_name, e
|
"Failed to resolve %r, falling back to cache. %r", service_name, e
|
||||||
)
|
)
|
||||||
return list(cache_entry)
|
return list(cache_entry)
|
||||||
|
@ -149,7 +149,7 @@ def _handle_json_response(reactor, timeout_sec, request, response):
|
|||||||
|
|
||||||
body = yield make_deferred_yieldable(d)
|
body = yield make_deferred_yieldable(d)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"{%s} [%s] Error reading response: %s",
|
"{%s} [%s] Error reading response: %s",
|
||||||
request.txn_id,
|
request.txn_id,
|
||||||
request.destination,
|
request.destination,
|
||||||
@ -457,7 +457,7 @@ class MatrixFederationHttpClient(object):
|
|||||||
except Exception as e:
|
except Exception as e:
|
||||||
# Eh, we're already going to raise an exception so lets
|
# Eh, we're already going to raise an exception so lets
|
||||||
# ignore if this fails.
|
# ignore if this fails.
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"{%s} [%s] Failed to get error response: %s %s: %s",
|
"{%s} [%s] Failed to get error response: %s %s: %s",
|
||||||
request.txn_id,
|
request.txn_id,
|
||||||
request.destination,
|
request.destination,
|
||||||
@ -478,7 +478,7 @@ class MatrixFederationHttpClient(object):
|
|||||||
|
|
||||||
break
|
break
|
||||||
except RequestSendFailed as e:
|
except RequestSendFailed as e:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"{%s} [%s] Request failed: %s %s: %s",
|
"{%s} [%s] Request failed: %s %s: %s",
|
||||||
request.txn_id,
|
request.txn_id,
|
||||||
request.destination,
|
request.destination,
|
||||||
@ -513,7 +513,7 @@ class MatrixFederationHttpClient(object):
|
|||||||
raise
|
raise
|
||||||
|
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"{%s} [%s] Request failed: %s %s: %s",
|
"{%s} [%s] Request failed: %s %s: %s",
|
||||||
request.txn_id,
|
request.txn_id,
|
||||||
request.destination,
|
request.destination,
|
||||||
@ -889,7 +889,7 @@ class MatrixFederationHttpClient(object):
|
|||||||
d.addTimeout(self.default_timeout, self.reactor)
|
d.addTimeout(self.default_timeout, self.reactor)
|
||||||
length = yield make_deferred_yieldable(d)
|
length = yield make_deferred_yieldable(d)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"{%s} [%s] Error reading response: %s",
|
"{%s} [%s] Error reading response: %s",
|
||||||
request.txn_id,
|
request.txn_id,
|
||||||
request.destination,
|
request.destination,
|
||||||
|
@ -170,7 +170,7 @@ class RequestMetrics(object):
|
|||||||
tag = context.tag
|
tag = context.tag
|
||||||
|
|
||||||
if context != self.start_context:
|
if context != self.start_context:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Context have unexpectedly changed %r, %r",
|
"Context have unexpectedly changed %r, %r",
|
||||||
context,
|
context,
|
||||||
self.start_context,
|
self.start_context,
|
||||||
|
@ -454,7 +454,7 @@ def respond_with_json(
|
|||||||
# the Deferred fires, but since the flag is RIGHT THERE it seems like
|
# the Deferred fires, but since the flag is RIGHT THERE it seems like
|
||||||
# a waste.
|
# a waste.
|
||||||
if request._disconnected:
|
if request._disconnected:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Not sending response to request %s, already disconnected.", request
|
"Not sending response to request %s, already disconnected.", request
|
||||||
)
|
)
|
||||||
return
|
return
|
||||||
|
@ -219,13 +219,13 @@ def parse_json_value_from_request(request, allow_empty_body=False):
|
|||||||
try:
|
try:
|
||||||
content_unicode = content_bytes.decode("utf8")
|
content_unicode = content_bytes.decode("utf8")
|
||||||
except UnicodeDecodeError:
|
except UnicodeDecodeError:
|
||||||
logger.warn("Unable to decode UTF-8")
|
logger.warning("Unable to decode UTF-8")
|
||||||
raise SynapseError(400, "Content not JSON.", errcode=Codes.NOT_JSON)
|
raise SynapseError(400, "Content not JSON.", errcode=Codes.NOT_JSON)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
content = json.loads(content_unicode)
|
content = json.loads(content_unicode)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.warn("Unable to parse JSON: %s", e)
|
logger.warning("Unable to parse JSON: %s", e)
|
||||||
raise SynapseError(400, "Content not JSON.", errcode=Codes.NOT_JSON)
|
raise SynapseError(400, "Content not JSON.", errcode=Codes.NOT_JSON)
|
||||||
|
|
||||||
return content
|
return content
|
||||||
|
@ -199,7 +199,7 @@ class SynapseRequest(Request):
|
|||||||
# It's useful to log it here so that we can get an idea of when
|
# It's useful to log it here so that we can get an idea of when
|
||||||
# the client disconnects.
|
# the client disconnects.
|
||||||
with PreserveLoggingContext(self.logcontext):
|
with PreserveLoggingContext(self.logcontext):
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Error processing request %r: %s %s", self, reason.type, reason.value
|
"Error processing request %r: %s %s", self, reason.type, reason.value
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -305,7 +305,7 @@ class SynapseRequest(Request):
|
|||||||
try:
|
try:
|
||||||
self.request_metrics.stop(self.finish_time, self.code, self.sentLength)
|
self.request_metrics.stop(self.finish_time, self.code, self.sentLength)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.warn("Failed to stop metrics: %r", e)
|
logger.warning("Failed to stop metrics: %r", e)
|
||||||
|
|
||||||
|
|
||||||
class XForwardedForRequest(SynapseRequest):
|
class XForwardedForRequest(SynapseRequest):
|
||||||
|
@ -294,7 +294,7 @@ class LoggingContext(object):
|
|||||||
"""Enters this logging context into thread local storage"""
|
"""Enters this logging context into thread local storage"""
|
||||||
old_context = self.set_current_context(self)
|
old_context = self.set_current_context(self)
|
||||||
if self.previous_context != old_context:
|
if self.previous_context != old_context:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Expected previous context %r, found %r",
|
"Expected previous context %r, found %r",
|
||||||
self.previous_context,
|
self.previous_context,
|
||||||
old_context,
|
old_context,
|
||||||
|
@ -246,7 +246,7 @@ class HttpPusher(object):
|
|||||||
# we really only give up so that if the URL gets
|
# we really only give up so that if the URL gets
|
||||||
# fixed, we don't suddenly deliver a load
|
# fixed, we don't suddenly deliver a load
|
||||||
# of old notifications.
|
# of old notifications.
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Giving up on a notification to user %s, " "pushkey %s",
|
"Giving up on a notification to user %s, " "pushkey %s",
|
||||||
self.user_id,
|
self.user_id,
|
||||||
self.pushkey,
|
self.pushkey,
|
||||||
@ -299,7 +299,7 @@ class HttpPusher(object):
|
|||||||
if pk != self.pushkey:
|
if pk != self.pushkey:
|
||||||
# for sanity, we only remove the pushkey if it
|
# for sanity, we only remove the pushkey if it
|
||||||
# was the one we actually sent...
|
# was the one we actually sent...
|
||||||
logger.warn(
|
logger.warning(
|
||||||
("Ignoring rejected pushkey %s because we" " didn't send it"),
|
("Ignoring rejected pushkey %s because we" " didn't send it"),
|
||||||
pk,
|
pk,
|
||||||
)
|
)
|
||||||
|
@ -117,7 +117,7 @@ class PushRuleEvaluatorForEvent(object):
|
|||||||
pattern = UserID.from_string(user_id).localpart
|
pattern = UserID.from_string(user_id).localpart
|
||||||
|
|
||||||
if not pattern:
|
if not pattern:
|
||||||
logger.warn("event_match condition with no pattern")
|
logger.warning("event_match condition with no pattern")
|
||||||
return False
|
return False
|
||||||
|
|
||||||
# XXX: optimisation: cache our pattern regexps
|
# XXX: optimisation: cache our pattern regexps
|
||||||
@ -173,7 +173,7 @@ def _glob_matches(glob, value, word_boundary=False):
|
|||||||
regex_cache[(glob, word_boundary)] = r
|
regex_cache[(glob, word_boundary)] = r
|
||||||
return r.search(value)
|
return r.search(value)
|
||||||
except re.error:
|
except re.error:
|
||||||
logger.warn("Failed to parse glob to regex: %r", glob)
|
logger.warning("Failed to parse glob to regex: %r", glob)
|
||||||
return False
|
return False
|
||||||
|
|
||||||
|
|
||||||
|
@ -180,7 +180,7 @@ class ReplicationEndpoint(object):
|
|||||||
if e.code != 504 or not cls.RETRY_ON_TIMEOUT:
|
if e.code != 504 or not cls.RETRY_ON_TIMEOUT:
|
||||||
raise
|
raise
|
||||||
|
|
||||||
logger.warn("%s request timed out", cls.NAME)
|
logger.warning("%s request timed out", cls.NAME)
|
||||||
|
|
||||||
# If we timed out we probably don't need to worry about backing
|
# If we timed out we probably don't need to worry about backing
|
||||||
# off too much, but lets just wait a little anyway.
|
# off too much, but lets just wait a little anyway.
|
||||||
|
@ -144,7 +144,7 @@ class ReplicationRemoteRejectInviteRestServlet(ReplicationEndpoint):
|
|||||||
# The 'except' clause is very broad, but we need to
|
# The 'except' clause is very broad, but we need to
|
||||||
# capture everything from DNS failures upwards
|
# capture everything from DNS failures upwards
|
||||||
#
|
#
|
||||||
logger.warn("Failed to reject invite: %s", e)
|
logger.warning("Failed to reject invite: %s", e)
|
||||||
|
|
||||||
await self.store.locally_reject_invite(user_id, room_id)
|
await self.store.locally_reject_invite(user_id, room_id)
|
||||||
ret = {}
|
ret = {}
|
||||||
|
@ -168,7 +168,7 @@ class ReplicationClientHandler(object):
|
|||||||
if self.connection:
|
if self.connection:
|
||||||
self.connection.send_command(cmd)
|
self.connection.send_command(cmd)
|
||||||
else:
|
else:
|
||||||
logger.warn("Queuing command as not connected: %r", cmd.NAME)
|
logger.warning("Queuing command as not connected: %r", cmd.NAME)
|
||||||
self.pending_commands.append(cmd)
|
self.pending_commands.append(cmd)
|
||||||
|
|
||||||
def send_federation_ack(self, token):
|
def send_federation_ack(self, token):
|
||||||
|
@ -249,7 +249,7 @@ class BaseReplicationStreamProtocol(LineOnlyReceiver):
|
|||||||
return handler(cmd)
|
return handler(cmd)
|
||||||
|
|
||||||
def close(self):
|
def close(self):
|
||||||
logger.warn("[%s] Closing connection", self.id())
|
logger.warning("[%s] Closing connection", self.id())
|
||||||
self.time_we_closed = self.clock.time_msec()
|
self.time_we_closed = self.clock.time_msec()
|
||||||
self.transport.loseConnection()
|
self.transport.loseConnection()
|
||||||
self.on_connection_closed()
|
self.on_connection_closed()
|
||||||
|
@ -286,7 +286,7 @@ class PurgeHistoryRestServlet(RestServlet):
|
|||||||
room_id, stream_ordering
|
room_id, stream_ordering
|
||||||
)
|
)
|
||||||
if not r:
|
if not r:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"[purge] purging events not possible: No event found "
|
"[purge] purging events not possible: No event found "
|
||||||
"(received_ts %i => stream_ordering %i)",
|
"(received_ts %i => stream_ordering %i)",
|
||||||
ts,
|
ts,
|
||||||
|
@ -221,7 +221,7 @@ class LoginRestServlet(RestServlet):
|
|||||||
medium, address
|
medium, address
|
||||||
)
|
)
|
||||||
if not user_id:
|
if not user_id:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"unknown 3pid identifier medium %s, address %r", medium, address
|
"unknown 3pid identifier medium %s, address %r", medium, address
|
||||||
)
|
)
|
||||||
raise LoginError(403, "", errcode=Codes.FORBIDDEN)
|
raise LoginError(403, "", errcode=Codes.FORBIDDEN)
|
||||||
|
@ -71,7 +71,7 @@ class EmailPasswordRequestTokenRestServlet(RestServlet):
|
|||||||
def on_POST(self, request):
|
def on_POST(self, request):
|
||||||
if self.config.threepid_behaviour_email == ThreepidBehaviour.OFF:
|
if self.config.threepid_behaviour_email == ThreepidBehaviour.OFF:
|
||||||
if self.config.local_threepid_handling_disabled_due_to_email_config:
|
if self.config.local_threepid_handling_disabled_due_to_email_config:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"User password resets have been disabled due to lack of email config"
|
"User password resets have been disabled due to lack of email config"
|
||||||
)
|
)
|
||||||
raise SynapseError(
|
raise SynapseError(
|
||||||
@ -162,7 +162,7 @@ class PasswordResetSubmitTokenServlet(RestServlet):
|
|||||||
)
|
)
|
||||||
if self.config.threepid_behaviour_email == ThreepidBehaviour.OFF:
|
if self.config.threepid_behaviour_email == ThreepidBehaviour.OFF:
|
||||||
if self.config.local_threepid_handling_disabled_due_to_email_config:
|
if self.config.local_threepid_handling_disabled_due_to_email_config:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Password reset emails have been disabled due to lack of an email config"
|
"Password reset emails have been disabled due to lack of an email config"
|
||||||
)
|
)
|
||||||
raise SynapseError(
|
raise SynapseError(
|
||||||
@ -183,7 +183,7 @@ class PasswordResetSubmitTokenServlet(RestServlet):
|
|||||||
# Perform a 302 redirect if next_link is set
|
# Perform a 302 redirect if next_link is set
|
||||||
if next_link:
|
if next_link:
|
||||||
if next_link.startswith("file:///"):
|
if next_link.startswith("file:///"):
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Not redirecting to next_link as it is a local file: address"
|
"Not redirecting to next_link as it is a local file: address"
|
||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
@ -350,7 +350,7 @@ class EmailThreepidRequestTokenRestServlet(RestServlet):
|
|||||||
def on_POST(self, request):
|
def on_POST(self, request):
|
||||||
if self.config.threepid_behaviour_email == ThreepidBehaviour.OFF:
|
if self.config.threepid_behaviour_email == ThreepidBehaviour.OFF:
|
||||||
if self.config.local_threepid_handling_disabled_due_to_email_config:
|
if self.config.local_threepid_handling_disabled_due_to_email_config:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Adding emails have been disabled due to lack of an email config"
|
"Adding emails have been disabled due to lack of an email config"
|
||||||
)
|
)
|
||||||
raise SynapseError(
|
raise SynapseError(
|
||||||
@ -441,7 +441,7 @@ class MsisdnThreepidRequestTokenRestServlet(RestServlet):
|
|||||||
raise SynapseError(400, "MSISDN is already in use", Codes.THREEPID_IN_USE)
|
raise SynapseError(400, "MSISDN is already in use", Codes.THREEPID_IN_USE)
|
||||||
|
|
||||||
if not self.hs.config.account_threepid_delegate_msisdn:
|
if not self.hs.config.account_threepid_delegate_msisdn:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"No upstream msisdn account_threepid_delegate configured on the server to "
|
"No upstream msisdn account_threepid_delegate configured on the server to "
|
||||||
"handle this request"
|
"handle this request"
|
||||||
)
|
)
|
||||||
@ -488,7 +488,7 @@ class AddThreepidEmailSubmitTokenServlet(RestServlet):
|
|||||||
def on_GET(self, request):
|
def on_GET(self, request):
|
||||||
if self.config.threepid_behaviour_email == ThreepidBehaviour.OFF:
|
if self.config.threepid_behaviour_email == ThreepidBehaviour.OFF:
|
||||||
if self.config.local_threepid_handling_disabled_due_to_email_config:
|
if self.config.local_threepid_handling_disabled_due_to_email_config:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Adding emails have been disabled due to lack of an email config"
|
"Adding emails have been disabled due to lack of an email config"
|
||||||
)
|
)
|
||||||
raise SynapseError(
|
raise SynapseError(
|
||||||
@ -515,7 +515,7 @@ class AddThreepidEmailSubmitTokenServlet(RestServlet):
|
|||||||
# Perform a 302 redirect if next_link is set
|
# Perform a 302 redirect if next_link is set
|
||||||
if next_link:
|
if next_link:
|
||||||
if next_link.startswith("file:///"):
|
if next_link.startswith("file:///"):
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Not redirecting to next_link as it is a local file: address"
|
"Not redirecting to next_link as it is a local file: address"
|
||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
|
@ -106,7 +106,7 @@ class EmailRegisterRequestTokenRestServlet(RestServlet):
|
|||||||
def on_POST(self, request):
|
def on_POST(self, request):
|
||||||
if self.hs.config.threepid_behaviour_email == ThreepidBehaviour.OFF:
|
if self.hs.config.threepid_behaviour_email == ThreepidBehaviour.OFF:
|
||||||
if self.hs.config.local_threepid_handling_disabled_due_to_email_config:
|
if self.hs.config.local_threepid_handling_disabled_due_to_email_config:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Email registration has been disabled due to lack of email config"
|
"Email registration has been disabled due to lack of email config"
|
||||||
)
|
)
|
||||||
raise SynapseError(
|
raise SynapseError(
|
||||||
@ -207,7 +207,7 @@ class MsisdnRegisterRequestTokenRestServlet(RestServlet):
|
|||||||
)
|
)
|
||||||
|
|
||||||
if not self.hs.config.account_threepid_delegate_msisdn:
|
if not self.hs.config.account_threepid_delegate_msisdn:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"No upstream msisdn account_threepid_delegate configured on the server to "
|
"No upstream msisdn account_threepid_delegate configured on the server to "
|
||||||
"handle this request"
|
"handle this request"
|
||||||
)
|
)
|
||||||
@ -266,7 +266,7 @@ class RegistrationSubmitTokenServlet(RestServlet):
|
|||||||
)
|
)
|
||||||
if self.config.threepid_behaviour_email == ThreepidBehaviour.OFF:
|
if self.config.threepid_behaviour_email == ThreepidBehaviour.OFF:
|
||||||
if self.config.local_threepid_handling_disabled_due_to_email_config:
|
if self.config.local_threepid_handling_disabled_due_to_email_config:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"User registration via email has been disabled due to lack of email config"
|
"User registration via email has been disabled due to lack of email config"
|
||||||
)
|
)
|
||||||
raise SynapseError(
|
raise SynapseError(
|
||||||
@ -287,7 +287,7 @@ class RegistrationSubmitTokenServlet(RestServlet):
|
|||||||
# Perform a 302 redirect if next_link is set
|
# Perform a 302 redirect if next_link is set
|
||||||
if next_link:
|
if next_link:
|
||||||
if next_link.startswith("file:///"):
|
if next_link.startswith("file:///"):
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Not redirecting to next_link as it is a local file: address"
|
"Not redirecting to next_link as it is a local file: address"
|
||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
@ -480,7 +480,7 @@ class RegisterRestServlet(RestServlet):
|
|||||||
# a password to work around a client bug where it sent
|
# a password to work around a client bug where it sent
|
||||||
# the 'initial_device_display_name' param alone, wiping out
|
# the 'initial_device_display_name' param alone, wiping out
|
||||||
# the original registration params
|
# the original registration params
|
||||||
logger.warn("Ignoring initial_device_display_name without password")
|
logger.warning("Ignoring initial_device_display_name without password")
|
||||||
del body["initial_device_display_name"]
|
del body["initial_device_display_name"]
|
||||||
|
|
||||||
session_id = self.auth_handler.get_session_id(body)
|
session_id = self.auth_handler.get_session_id(body)
|
||||||
|
@ -394,7 +394,7 @@ class SyncRestServlet(RestServlet):
|
|||||||
# We've had bug reports that events were coming down under the
|
# We've had bug reports that events were coming down under the
|
||||||
# wrong room.
|
# wrong room.
|
||||||
if event.room_id != room.room_id:
|
if event.room_id != room.room_id:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Event %r is under room %r instead of %r",
|
"Event %r is under room %r instead of %r",
|
||||||
event.event_id,
|
event.event_id,
|
||||||
room.room_id,
|
room.room_id,
|
||||||
|
@ -363,7 +363,7 @@ class MediaRepository(object):
|
|||||||
},
|
},
|
||||||
)
|
)
|
||||||
except RequestSendFailed as e:
|
except RequestSendFailed as e:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Request failed fetching remote media %s/%s: %r",
|
"Request failed fetching remote media %s/%s: %r",
|
||||||
server_name,
|
server_name,
|
||||||
media_id,
|
media_id,
|
||||||
@ -372,7 +372,7 @@ class MediaRepository(object):
|
|||||||
raise SynapseError(502, "Failed to fetch remote media")
|
raise SynapseError(502, "Failed to fetch remote media")
|
||||||
|
|
||||||
except HttpResponseException as e:
|
except HttpResponseException as e:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"HTTP error fetching remote media %s/%s: %s",
|
"HTTP error fetching remote media %s/%s: %s",
|
||||||
server_name,
|
server_name,
|
||||||
media_id,
|
media_id,
|
||||||
@ -383,10 +383,12 @@ class MediaRepository(object):
|
|||||||
raise SynapseError(502, "Failed to fetch remote media")
|
raise SynapseError(502, "Failed to fetch remote media")
|
||||||
|
|
||||||
except SynapseError:
|
except SynapseError:
|
||||||
logger.warn("Failed to fetch remote media %s/%s", server_name, media_id)
|
logger.warning(
|
||||||
|
"Failed to fetch remote media %s/%s", server_name, media_id
|
||||||
|
)
|
||||||
raise
|
raise
|
||||||
except NotRetryingDestination:
|
except NotRetryingDestination:
|
||||||
logger.warn("Not retrying destination %r", server_name)
|
logger.warning("Not retrying destination %r", server_name)
|
||||||
raise SynapseError(502, "Failed to fetch remote media")
|
raise SynapseError(502, "Failed to fetch remote media")
|
||||||
except Exception:
|
except Exception:
|
||||||
logger.exception(
|
logger.exception(
|
||||||
@ -691,7 +693,7 @@ class MediaRepository(object):
|
|||||||
try:
|
try:
|
||||||
os.remove(full_path)
|
os.remove(full_path)
|
||||||
except OSError as e:
|
except OSError as e:
|
||||||
logger.warn("Failed to remove file: %r", full_path)
|
logger.warning("Failed to remove file: %r", full_path)
|
||||||
if e.errno == errno.ENOENT:
|
if e.errno == errno.ENOENT:
|
||||||
pass
|
pass
|
||||||
else:
|
else:
|
||||||
|
@ -136,7 +136,7 @@ class PreviewUrlResource(DirectServeResource):
|
|||||||
match = False
|
match = False
|
||||||
continue
|
continue
|
||||||
if match:
|
if match:
|
||||||
logger.warn("URL %s blocked by url_blacklist entry %s", url, entry)
|
logger.warning("URL %s blocked by url_blacklist entry %s", url, entry)
|
||||||
raise SynapseError(
|
raise SynapseError(
|
||||||
403, "URL blocked by url pattern blacklist entry", Codes.UNKNOWN
|
403, "URL blocked by url pattern blacklist entry", Codes.UNKNOWN
|
||||||
)
|
)
|
||||||
@ -208,7 +208,7 @@ class PreviewUrlResource(DirectServeResource):
|
|||||||
og["og:image:width"] = dims["width"]
|
og["og:image:width"] = dims["width"]
|
||||||
og["og:image:height"] = dims["height"]
|
og["og:image:height"] = dims["height"]
|
||||||
else:
|
else:
|
||||||
logger.warn("Couldn't get dims for %s" % url)
|
logger.warning("Couldn't get dims for %s" % url)
|
||||||
|
|
||||||
# define our OG response for this media
|
# define our OG response for this media
|
||||||
elif _is_html(media_info["media_type"]):
|
elif _is_html(media_info["media_type"]):
|
||||||
@ -256,7 +256,7 @@ class PreviewUrlResource(DirectServeResource):
|
|||||||
og["og:image:width"] = dims["width"]
|
og["og:image:width"] = dims["width"]
|
||||||
og["og:image:height"] = dims["height"]
|
og["og:image:height"] = dims["height"]
|
||||||
else:
|
else:
|
||||||
logger.warn("Couldn't get dims for %s", og["og:image"])
|
logger.warning("Couldn't get dims for %s", og["og:image"])
|
||||||
|
|
||||||
og["og:image"] = "mxc://%s/%s" % (
|
og["og:image"] = "mxc://%s/%s" % (
|
||||||
self.server_name,
|
self.server_name,
|
||||||
@ -267,7 +267,7 @@ class PreviewUrlResource(DirectServeResource):
|
|||||||
else:
|
else:
|
||||||
del og["og:image"]
|
del og["og:image"]
|
||||||
else:
|
else:
|
||||||
logger.warn("Failed to find any OG data in %s", url)
|
logger.warning("Failed to find any OG data in %s", url)
|
||||||
og = {}
|
og = {}
|
||||||
|
|
||||||
logger.debug("Calculated OG for %s as %s", url, og)
|
logger.debug("Calculated OG for %s as %s", url, og)
|
||||||
@ -319,7 +319,7 @@ class PreviewUrlResource(DirectServeResource):
|
|||||||
)
|
)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
# FIXME: pass through 404s and other error messages nicely
|
# FIXME: pass through 404s and other error messages nicely
|
||||||
logger.warn("Error downloading %s: %r", url, e)
|
logger.warning("Error downloading %s: %r", url, e)
|
||||||
|
|
||||||
raise SynapseError(
|
raise SynapseError(
|
||||||
500,
|
500,
|
||||||
@ -400,7 +400,7 @@ class PreviewUrlResource(DirectServeResource):
|
|||||||
except OSError as e:
|
except OSError as e:
|
||||||
# If the path doesn't exist, meh
|
# If the path doesn't exist, meh
|
||||||
if e.errno != errno.ENOENT:
|
if e.errno != errno.ENOENT:
|
||||||
logger.warn("Failed to remove media: %r: %s", media_id, e)
|
logger.warning("Failed to remove media: %r: %s", media_id, e)
|
||||||
continue
|
continue
|
||||||
|
|
||||||
removed_media.append(media_id)
|
removed_media.append(media_id)
|
||||||
@ -432,7 +432,7 @@ class PreviewUrlResource(DirectServeResource):
|
|||||||
except OSError as e:
|
except OSError as e:
|
||||||
# If the path doesn't exist, meh
|
# If the path doesn't exist, meh
|
||||||
if e.errno != errno.ENOENT:
|
if e.errno != errno.ENOENT:
|
||||||
logger.warn("Failed to remove media: %r: %s", media_id, e)
|
logger.warning("Failed to remove media: %r: %s", media_id, e)
|
||||||
continue
|
continue
|
||||||
|
|
||||||
try:
|
try:
|
||||||
@ -448,7 +448,7 @@ class PreviewUrlResource(DirectServeResource):
|
|||||||
except OSError as e:
|
except OSError as e:
|
||||||
# If the path doesn't exist, meh
|
# If the path doesn't exist, meh
|
||||||
if e.errno != errno.ENOENT:
|
if e.errno != errno.ENOENT:
|
||||||
logger.warn("Failed to remove media: %r: %s", media_id, e)
|
logger.warning("Failed to remove media: %r: %s", media_id, e)
|
||||||
continue
|
continue
|
||||||
|
|
||||||
removed_media.append(media_id)
|
removed_media.append(media_id)
|
||||||
|
@ -182,7 +182,7 @@ class ThumbnailResource(DirectServeResource):
|
|||||||
if file_path:
|
if file_path:
|
||||||
yield respond_with_file(request, desired_type, file_path)
|
yield respond_with_file(request, desired_type, file_path)
|
||||||
else:
|
else:
|
||||||
logger.warn("Failed to generate thumbnail")
|
logger.warning("Failed to generate thumbnail")
|
||||||
respond_404(request)
|
respond_404(request)
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
@ -245,7 +245,7 @@ class ThumbnailResource(DirectServeResource):
|
|||||||
if file_path:
|
if file_path:
|
||||||
yield respond_with_file(request, desired_type, file_path)
|
yield respond_with_file(request, desired_type, file_path)
|
||||||
else:
|
else:
|
||||||
logger.warn("Failed to generate thumbnail")
|
logger.warning("Failed to generate thumbnail")
|
||||||
respond_404(request)
|
respond_404(request)
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
|
@ -83,7 +83,7 @@ class ResourceLimitsServerNotices(object):
|
|||||||
room_id = yield self._server_notices_manager.get_notice_room_for_user(user_id)
|
room_id = yield self._server_notices_manager.get_notice_room_for_user(user_id)
|
||||||
|
|
||||||
if not room_id:
|
if not room_id:
|
||||||
logger.warn("Failed to get server notices room")
|
logger.warning("Failed to get server notices room")
|
||||||
return
|
return
|
||||||
|
|
||||||
yield self._check_and_set_tags(user_id, room_id)
|
yield self._check_and_set_tags(user_id, room_id)
|
||||||
|
@ -494,7 +494,7 @@ class SQLBaseStore(object):
|
|||||||
exception_callbacks = []
|
exception_callbacks = []
|
||||||
|
|
||||||
if LoggingContext.current_context() == LoggingContext.sentinel:
|
if LoggingContext.current_context() == LoggingContext.sentinel:
|
||||||
logger.warn("Starting db txn '%s' from sentinel context", desc)
|
logger.warning("Starting db txn '%s' from sentinel context", desc)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
result = yield self.runWithConnection(
|
result = yield self.runWithConnection(
|
||||||
@ -532,7 +532,7 @@ class SQLBaseStore(object):
|
|||||||
"""
|
"""
|
||||||
parent_context = LoggingContext.current_context()
|
parent_context = LoggingContext.current_context()
|
||||||
if parent_context == LoggingContext.sentinel:
|
if parent_context == LoggingContext.sentinel:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Starting db connection from sentinel context: metrics will be lost"
|
"Starting db connection from sentinel context: metrics will be lost"
|
||||||
)
|
)
|
||||||
parent_context = None
|
parent_context = None
|
||||||
@ -719,7 +719,7 @@ class SQLBaseStore(object):
|
|||||||
raise
|
raise
|
||||||
|
|
||||||
# presumably we raced with another transaction: let's retry.
|
# presumably we raced with another transaction: let's retry.
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"IntegrityError when upserting into %s; retrying: %s", table, e
|
"IntegrityError when upserting into %s; retrying: %s", table, e
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -44,7 +44,7 @@ class PusherWorkerStore(SQLBaseStore):
|
|||||||
|
|
||||||
r["data"] = json.loads(dataJson)
|
r["data"] = json.loads(dataJson)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Invalid JSON in data for pusher %d: %s, %s",
|
"Invalid JSON in data for pusher %d: %s, %s",
|
||||||
r["id"],
|
r["id"],
|
||||||
dataJson,
|
dataJson,
|
||||||
|
@ -196,7 +196,7 @@ class SearchBackgroundUpdateStore(BackgroundUpdateStore):
|
|||||||
" ON event_search USING GIN (vector)"
|
" ON event_search USING GIN (vector)"
|
||||||
)
|
)
|
||||||
except psycopg2.ProgrammingError as e:
|
except psycopg2.ProgrammingError as e:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Ignoring error %r when trying to switch from GIST to GIN", e
|
"Ignoring error %r when trying to switch from GIST to GIN", e
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -309,7 +309,7 @@ class Linearizer(object):
|
|||||||
)
|
)
|
||||||
|
|
||||||
else:
|
else:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Unexpected exception waiting for linearizer lock %r for key %r",
|
"Unexpected exception waiting for linearizer lock %r for key %r",
|
||||||
self.name,
|
self.name,
|
||||||
key,
|
key,
|
||||||
|
@ -107,7 +107,7 @@ def register_cache(cache_type, cache_name, cache, collect_callback=None):
|
|||||||
if collect_callback:
|
if collect_callback:
|
||||||
collect_callback()
|
collect_callback()
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.warn("Error calculating metrics for %s: %s", cache_name, e)
|
logger.warning("Error calculating metrics for %s: %s", cache_name, e)
|
||||||
raise
|
raise
|
||||||
|
|
||||||
yield GaugeMetricFamily("__unused", "")
|
yield GaugeMetricFamily("__unused", "")
|
||||||
|
@ -119,7 +119,7 @@ class Measure(object):
|
|||||||
context = LoggingContext.current_context()
|
context = LoggingContext.current_context()
|
||||||
|
|
||||||
if context != self.start_context:
|
if context != self.start_context:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Context has unexpectedly changed from '%s' to '%s'. (%r)",
|
"Context has unexpectedly changed from '%s' to '%s'. (%r)",
|
||||||
self.start_context,
|
self.start_context,
|
||||||
context,
|
context,
|
||||||
@ -128,7 +128,7 @@ class Measure(object):
|
|||||||
return
|
return
|
||||||
|
|
||||||
if not context:
|
if not context:
|
||||||
logger.warn("Expected context. (%r)", self.name)
|
logger.warning("Expected context. (%r)", self.name)
|
||||||
return
|
return
|
||||||
|
|
||||||
current = context.get_resource_usage()
|
current = context.get_resource_usage()
|
||||||
@ -140,7 +140,7 @@ class Measure(object):
|
|||||||
block_db_txn_duration.labels(self.name).inc(usage.db_txn_duration_sec)
|
block_db_txn_duration.labels(self.name).inc(usage.db_txn_duration_sec)
|
||||||
block_db_sched_duration.labels(self.name).inc(usage.db_sched_duration_sec)
|
block_db_sched_duration.labels(self.name).inc(usage.db_sched_duration_sec)
|
||||||
except ValueError:
|
except ValueError:
|
||||||
logger.warn(
|
logger.warning(
|
||||||
"Failed to save metrics! OLD: %r, NEW: %r", self.start_usage, current
|
"Failed to save metrics! OLD: %r, NEW: %r", self.start_usage, current
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -33,4 +33,4 @@ def change_resource_limit(soft_file_no):
|
|||||||
resource.RLIMIT_CORE, (resource.RLIM_INFINITY, resource.RLIM_INFINITY)
|
resource.RLIMIT_CORE, (resource.RLIM_INFINITY, resource.RLIM_INFINITY)
|
||||||
)
|
)
|
||||||
except (ValueError, resource.error) as e:
|
except (ValueError, resource.error) as e:
|
||||||
logger.warn("Failed to set file or core limit: %s", e)
|
logger.warning("Failed to set file or core limit: %s", e)
|
||||||
|
Loading…
Reference in New Issue
Block a user