mirror of
https://mau.dev/maunium/synapse.git
synced 2024-10-01 01:36:05 -04:00
Merge branch 'develop' into http_client_refactor
This commit is contained in:
commit
f1c7f8e813
@ -272,7 +272,7 @@ class Auth(object):
|
|||||||
key = (RoomCreateEvent.TYPE, "", )
|
key = (RoomCreateEvent.TYPE, "", )
|
||||||
create_event = event.old_state_events.get(key)
|
create_event = event.old_state_events.get(key)
|
||||||
if (create_event is not None and
|
if (create_event is not None and
|
||||||
create_event.content["creator"] == user_id):
|
create_event.content["creator"] == user_id):
|
||||||
return 100
|
return 100
|
||||||
|
|
||||||
return level
|
return level
|
||||||
|
@ -17,6 +17,8 @@
|
|||||||
|
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
class Codes(object):
|
class Codes(object):
|
||||||
UNAUTHORIZED = "M_UNAUTHORIZED"
|
UNAUTHORIZED = "M_UNAUTHORIZED"
|
||||||
@ -38,7 +40,7 @@ class CodeMessageException(Exception):
|
|||||||
"""An exception with integer code and message string attributes."""
|
"""An exception with integer code and message string attributes."""
|
||||||
|
|
||||||
def __init__(self, code, msg):
|
def __init__(self, code, msg):
|
||||||
logging.error("%s: %s, %s", type(self).__name__, code, msg)
|
logger.info("%s: %s, %s", type(self).__name__, code, msg)
|
||||||
super(CodeMessageException, self).__init__("%d: %s" % (code, msg))
|
super(CodeMessageException, self).__init__("%d: %s" % (code, msg))
|
||||||
self.code = code
|
self.code = code
|
||||||
self.msg = msg
|
self.msg = msg
|
||||||
@ -140,7 +142,8 @@ def cs_exception(exception):
|
|||||||
if isinstance(exception, CodeMessageException):
|
if isinstance(exception, CodeMessageException):
|
||||||
return exception.error_dict()
|
return exception.error_dict()
|
||||||
else:
|
else:
|
||||||
logging.error("Unknown exception type: %s", type(exception))
|
logger.error("Unknown exception type: %s", type(exception))
|
||||||
|
return {}
|
||||||
|
|
||||||
|
|
||||||
def cs_error(msg, code=Codes.UNKNOWN, **kwargs):
|
def cs_error(msg, code=Codes.UNKNOWN, **kwargs):
|
||||||
|
@ -83,6 +83,8 @@ class SynapseEvent(JsonEncodedObject):
|
|||||||
"content",
|
"content",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
outlier = False
|
||||||
|
|
||||||
def __init__(self, raises=True, **kwargs):
|
def __init__(self, raises=True, **kwargs):
|
||||||
super(SynapseEvent, self).__init__(**kwargs)
|
super(SynapseEvent, self).__init__(**kwargs)
|
||||||
# if "content" in kwargs:
|
# if "content" in kwargs:
|
||||||
|
@ -116,7 +116,7 @@ class SynapseHomeServer(HomeServer):
|
|||||||
# extra resources to existing nodes. See self._resource_id for the key.
|
# extra resources to existing nodes. See self._resource_id for the key.
|
||||||
resource_mappings = {}
|
resource_mappings = {}
|
||||||
for (full_path, resource) in desired_tree:
|
for (full_path, resource) in desired_tree:
|
||||||
logging.info("Attaching %s to path %s", resource, full_path)
|
logger.info("Attaching %s to path %s", resource, full_path)
|
||||||
last_resource = self.root_resource
|
last_resource = self.root_resource
|
||||||
for path_seg in full_path.split('/')[1:-1]:
|
for path_seg in full_path.split('/')[1:-1]:
|
||||||
if not path_seg in last_resource.listNames():
|
if not path_seg in last_resource.listNames():
|
||||||
@ -221,12 +221,12 @@ def setup():
|
|||||||
|
|
||||||
db_name = hs.get_db_name()
|
db_name = hs.get_db_name()
|
||||||
|
|
||||||
logging.info("Preparing database: %s...", db_name)
|
logger.info("Preparing database: %s...", db_name)
|
||||||
|
|
||||||
with sqlite3.connect(db_name) as db_conn:
|
with sqlite3.connect(db_name) as db_conn:
|
||||||
prepare_database(db_conn)
|
prepare_database(db_conn)
|
||||||
|
|
||||||
logging.info("Database prepared in %s.", db_name)
|
logger.info("Database prepared in %s.", db_name)
|
||||||
|
|
||||||
hs.get_db_pool()
|
hs.get_db_pool()
|
||||||
|
|
||||||
@ -257,13 +257,16 @@ def setup():
|
|||||||
else:
|
else:
|
||||||
reactor.run()
|
reactor.run()
|
||||||
|
|
||||||
|
|
||||||
def run():
|
def run():
|
||||||
with LoggingContext("run"):
|
with LoggingContext("run"):
|
||||||
reactor.run()
|
reactor.run()
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
with LoggingContext("main"):
|
with LoggingContext("main"):
|
||||||
setup()
|
setup()
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
|
@ -21,11 +21,12 @@ import signal
|
|||||||
|
|
||||||
SYNAPSE = ["python", "-m", "synapse.app.homeserver"]
|
SYNAPSE = ["python", "-m", "synapse.app.homeserver"]
|
||||||
|
|
||||||
CONFIGFILE="homeserver.yaml"
|
CONFIGFILE = "homeserver.yaml"
|
||||||
PIDFILE="homeserver.pid"
|
PIDFILE = "homeserver.pid"
|
||||||
|
|
||||||
|
GREEN = "\x1b[1;32m"
|
||||||
|
NORMAL = "\x1b[m"
|
||||||
|
|
||||||
GREEN="\x1b[1;32m"
|
|
||||||
NORMAL="\x1b[m"
|
|
||||||
|
|
||||||
def start():
|
def start():
|
||||||
if not os.path.exists(CONFIGFILE):
|
if not os.path.exists(CONFIGFILE):
|
||||||
@ -43,12 +44,14 @@ def start():
|
|||||||
subprocess.check_call(args)
|
subprocess.check_call(args)
|
||||||
print GREEN + "started" + NORMAL
|
print GREEN + "started" + NORMAL
|
||||||
|
|
||||||
|
|
||||||
def stop():
|
def stop():
|
||||||
if os.path.exists(PIDFILE):
|
if os.path.exists(PIDFILE):
|
||||||
pid = int(open(PIDFILE).read())
|
pid = int(open(PIDFILE).read())
|
||||||
os.kill(pid, signal.SIGTERM)
|
os.kill(pid, signal.SIGTERM)
|
||||||
print GREEN + "stopped" + NORMAL
|
print GREEN + "stopped" + NORMAL
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
action = sys.argv[1] if sys.argv[1:] else "usage"
|
action = sys.argv[1] if sys.argv[1:] else "usage"
|
||||||
if action == "start":
|
if action == "start":
|
||||||
@ -62,5 +65,6 @@ def main():
|
|||||||
sys.stderr.write("Usage: %s [start|stop|restart]\n" % (sys.argv[0],))
|
sys.stderr.write("Usage: %s [start|stop|restart]\n" % (sys.argv[0],))
|
||||||
sys.exit(1)
|
sys.exit(1)
|
||||||
|
|
||||||
if __name__=='__main__':
|
|
||||||
|
if __name__ == "__main__":
|
||||||
main()
|
main()
|
||||||
|
@ -30,7 +30,7 @@ logger = logging.getLogger(__name__)
|
|||||||
def check_event_content_hash(event, hash_algorithm=hashlib.sha256):
|
def check_event_content_hash(event, hash_algorithm=hashlib.sha256):
|
||||||
"""Check whether the hash for this PDU matches the contents"""
|
"""Check whether the hash for this PDU matches the contents"""
|
||||||
computed_hash = _compute_content_hash(event, hash_algorithm)
|
computed_hash = _compute_content_hash(event, hash_algorithm)
|
||||||
logging.debug("Expecting hash: %s", encode_base64(computed_hash.digest()))
|
logger.debug("Expecting hash: %s", encode_base64(computed_hash.digest()))
|
||||||
if computed_hash.name not in event.hashes:
|
if computed_hash.name not in event.hashes:
|
||||||
raise SynapseError(
|
raise SynapseError(
|
||||||
400,
|
400,
|
||||||
|
@ -135,7 +135,7 @@ class Keyring(object):
|
|||||||
|
|
||||||
time_now_ms = self.clock.time_msec()
|
time_now_ms = self.clock.time_msec()
|
||||||
|
|
||||||
self.store.store_server_certificate(
|
yield self.store.store_server_certificate(
|
||||||
server_name,
|
server_name,
|
||||||
server_name,
|
server_name,
|
||||||
time_now_ms,
|
time_now_ms,
|
||||||
@ -143,7 +143,7 @@ class Keyring(object):
|
|||||||
)
|
)
|
||||||
|
|
||||||
for key_id, key in verify_keys.items():
|
for key_id, key in verify_keys.items():
|
||||||
self.store.store_server_verify_key(
|
yield self.store.store_server_verify_key(
|
||||||
server_name, server_name, time_now_ms, key
|
server_name, server_name, time_now_ms, key
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -24,6 +24,7 @@ from .units import Transaction, Edu
|
|||||||
from .persistence import TransactionActions
|
from .persistence import TransactionActions
|
||||||
|
|
||||||
from synapse.util.logutils import log_function
|
from synapse.util.logutils import log_function
|
||||||
|
from synapse.util.logcontext import PreserveLoggingContext
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
@ -319,19 +320,20 @@ class ReplicationLayer(object):
|
|||||||
|
|
||||||
logger.debug("[%s] Transacition is new", transaction.transaction_id)
|
logger.debug("[%s] Transacition is new", transaction.transaction_id)
|
||||||
|
|
||||||
dl = []
|
with PreserveLoggingContext():
|
||||||
for pdu in pdu_list:
|
dl = []
|
||||||
dl.append(self._handle_new_pdu(transaction.origin, pdu))
|
for pdu in pdu_list:
|
||||||
|
dl.append(self._handle_new_pdu(transaction.origin, pdu))
|
||||||
|
|
||||||
if hasattr(transaction, "edus"):
|
if hasattr(transaction, "edus"):
|
||||||
for edu in [Edu(**x) for x in transaction.edus]:
|
for edu in [Edu(**x) for x in transaction.edus]:
|
||||||
self.received_edu(
|
self.received_edu(
|
||||||
transaction.origin,
|
transaction.origin,
|
||||||
edu.edu_type,
|
edu.edu_type,
|
||||||
edu.content
|
edu.content
|
||||||
)
|
)
|
||||||
|
|
||||||
results = yield defer.DeferredList(dl)
|
results = yield defer.DeferredList(dl)
|
||||||
|
|
||||||
ret = []
|
ret = []
|
||||||
for r in results:
|
for r in results:
|
||||||
@ -425,7 +427,9 @@ class ReplicationLayer(object):
|
|||||||
time_now = self._clock.time_msec()
|
time_now = self._clock.time_msec()
|
||||||
defer.returnValue((200, {
|
defer.returnValue((200, {
|
||||||
"state": [p.get_pdu_json(time_now) for p in res_pdus["state"]],
|
"state": [p.get_pdu_json(time_now) for p in res_pdus["state"]],
|
||||||
"auth_chain": [p.get_pdu_json(time_now) for p in res_pdus["auth_chain"]],
|
"auth_chain": [
|
||||||
|
p.get_pdu_json(time_now) for p in res_pdus["auth_chain"]
|
||||||
|
],
|
||||||
}))
|
}))
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
@ -436,7 +440,9 @@ class ReplicationLayer(object):
|
|||||||
(
|
(
|
||||||
200,
|
200,
|
||||||
{
|
{
|
||||||
"auth_chain": [a.get_pdu_json(time_now) for a in auth_pdus],
|
"auth_chain": [
|
||||||
|
a.get_pdu_json(time_now) for a in auth_pdus
|
||||||
|
],
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
@ -457,7 +463,7 @@ class ReplicationLayer(object):
|
|||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def send_join(self, destination, pdu):
|
def send_join(self, destination, pdu):
|
||||||
time_now = self._clock.time_msec()
|
time_now = self._clock.time_msec()
|
||||||
_, content = yield self.transport_layer.send_join(
|
_, content = yield self.transport_layer.send_join(
|
||||||
destination,
|
destination,
|
||||||
pdu.room_id,
|
pdu.room_id,
|
||||||
@ -649,7 +655,8 @@ class _TransactionQueue(object):
|
|||||||
(pdu, deferred, order)
|
(pdu, deferred, order)
|
||||||
)
|
)
|
||||||
|
|
||||||
self._attempt_new_transaction(destination)
|
with PreserveLoggingContext():
|
||||||
|
self._attempt_new_transaction(destination)
|
||||||
|
|
||||||
deferreds.append(deferred)
|
deferreds.append(deferred)
|
||||||
|
|
||||||
@ -669,7 +676,9 @@ class _TransactionQueue(object):
|
|||||||
deferred.errback(failure)
|
deferred.errback(failure)
|
||||||
else:
|
else:
|
||||||
logger.exception("Failed to send edu", failure)
|
logger.exception("Failed to send edu", failure)
|
||||||
self._attempt_new_transaction(destination).addErrback(eb)
|
|
||||||
|
with PreserveLoggingContext():
|
||||||
|
self._attempt_new_transaction(destination).addErrback(eb)
|
||||||
|
|
||||||
return deferred
|
return deferred
|
||||||
|
|
||||||
|
@ -25,7 +25,6 @@ import logging
|
|||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
class Edu(JsonEncodedObject):
|
class Edu(JsonEncodedObject):
|
||||||
""" An Edu represents a piece of data sent from one homeserver to another.
|
""" An Edu represents a piece of data sent from one homeserver to another.
|
||||||
|
|
||||||
|
@ -112,7 +112,7 @@ class BaseHandler(object):
|
|||||||
|
|
||||||
event.destinations = list(destinations)
|
event.destinations = list(destinations)
|
||||||
|
|
||||||
self.notifier.on_new_room_event(event, extra_users=extra_users)
|
yield self.notifier.on_new_room_event(event, extra_users=extra_users)
|
||||||
|
|
||||||
federation_handler = self.hs.get_handlers().federation_handler
|
federation_handler = self.hs.get_handlers().federation_handler
|
||||||
yield federation_handler.handle_new_event(event, snapshot)
|
yield federation_handler.handle_new_event(event, snapshot)
|
||||||
|
@ -128,8 +128,9 @@ class DirectoryHandler(BaseHandler):
|
|||||||
"servers": result.servers,
|
"servers": result.servers,
|
||||||
})
|
})
|
||||||
else:
|
else:
|
||||||
raise SynapseError(404, "Room alias \"%s\" not found" % (room_alias,))
|
raise SynapseError(
|
||||||
|
404, "Room alias \"%s\" not found" % (room_alias,)
|
||||||
|
)
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def send_room_alias_update_event(self, user_id, room_id):
|
def send_room_alias_update_event(self, user_id, room_id):
|
||||||
|
@ -56,7 +56,7 @@ class EventStreamHandler(BaseHandler):
|
|||||||
self.clock.cancel_call_later(
|
self.clock.cancel_call_later(
|
||||||
self._stop_timer_per_user.pop(auth_user))
|
self._stop_timer_per_user.pop(auth_user))
|
||||||
else:
|
else:
|
||||||
self.distributor.fire(
|
yield self.distributor.fire(
|
||||||
"started_user_eventstream", auth_user
|
"started_user_eventstream", auth_user
|
||||||
)
|
)
|
||||||
self._streams_per_user[auth_user] += 1
|
self._streams_per_user[auth_user] += 1
|
||||||
@ -65,8 +65,10 @@ class EventStreamHandler(BaseHandler):
|
|||||||
pagin_config.from_token = None
|
pagin_config.from_token = None
|
||||||
|
|
||||||
rm_handler = self.hs.get_handlers().room_member_handler
|
rm_handler = self.hs.get_handlers().room_member_handler
|
||||||
|
logger.debug("BETA")
|
||||||
room_ids = yield rm_handler.get_rooms_for_user(auth_user)
|
room_ids = yield rm_handler.get_rooms_for_user(auth_user)
|
||||||
|
|
||||||
|
logger.debug("ALPHA")
|
||||||
with PreserveLoggingContext():
|
with PreserveLoggingContext():
|
||||||
events, tokens = yield self.notifier.get_events_for(
|
events, tokens = yield self.notifier.get_events_for(
|
||||||
auth_user, room_ids, pagin_config, timeout
|
auth_user, room_ids, pagin_config, timeout
|
||||||
@ -93,7 +95,7 @@ class EventStreamHandler(BaseHandler):
|
|||||||
logger.debug(
|
logger.debug(
|
||||||
"_later stopped_user_eventstream %s", auth_user
|
"_later stopped_user_eventstream %s", auth_user
|
||||||
)
|
)
|
||||||
self.distributor.fire(
|
yield self.distributor.fire(
|
||||||
"stopped_user_eventstream", auth_user
|
"stopped_user_eventstream", auth_user
|
||||||
)
|
)
|
||||||
del self._stop_timer_per_user[auth_user]
|
del self._stop_timer_per_user[auth_user]
|
||||||
|
@ -122,7 +122,8 @@ class FederationHandler(BaseHandler):
|
|||||||
event.origin, redacted_pdu_json
|
event.origin, redacted_pdu_json
|
||||||
)
|
)
|
||||||
except SynapseError as e:
|
except SynapseError as e:
|
||||||
logger.warn("Signature check failed for %s redacted to %s",
|
logger.warn(
|
||||||
|
"Signature check failed for %s redacted to %s",
|
||||||
encode_canonical_json(pdu.get_pdu_json()),
|
encode_canonical_json(pdu.get_pdu_json()),
|
||||||
encode_canonical_json(redacted_pdu_json),
|
encode_canonical_json(redacted_pdu_json),
|
||||||
)
|
)
|
||||||
@ -209,7 +210,7 @@ class FederationHandler(BaseHandler):
|
|||||||
if event.type == RoomMemberEvent.TYPE:
|
if event.type == RoomMemberEvent.TYPE:
|
||||||
if event.membership == Membership.JOIN:
|
if event.membership == Membership.JOIN:
|
||||||
user = self.hs.parse_userid(event.state_key)
|
user = self.hs.parse_userid(event.state_key)
|
||||||
self.distributor.fire(
|
yield self.distributor.fire(
|
||||||
"user_joined_room", user=user, room_id=event.room_id
|
"user_joined_room", user=user, room_id=event.room_id
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -390,7 +391,8 @@ class FederationHandler(BaseHandler):
|
|||||||
|
|
||||||
event.outlier = False
|
event.outlier = False
|
||||||
|
|
||||||
is_new_state = yield self.state_handler.annotate_event_with_state(event)
|
state_handler = self.state_handler
|
||||||
|
is_new_state = yield state_handler.annotate_event_with_state(event)
|
||||||
self.auth.check(event, raises=True)
|
self.auth.check(event, raises=True)
|
||||||
|
|
||||||
# FIXME (erikj): All this is duplicated above :(
|
# FIXME (erikj): All this is duplicated above :(
|
||||||
@ -414,7 +416,7 @@ class FederationHandler(BaseHandler):
|
|||||||
if event.type == RoomMemberEvent.TYPE:
|
if event.type == RoomMemberEvent.TYPE:
|
||||||
if event.membership == Membership.JOIN:
|
if event.membership == Membership.JOIN:
|
||||||
user = self.hs.parse_userid(event.state_key)
|
user = self.hs.parse_userid(event.state_key)
|
||||||
self.distributor.fire(
|
yield self.distributor.fire(
|
||||||
"user_joined_room", user=user, room_id=event.room_id
|
"user_joined_room", user=user, room_id=event.room_id
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -18,6 +18,7 @@ from twisted.internet import defer
|
|||||||
from synapse.api.constants import Membership
|
from synapse.api.constants import Membership
|
||||||
from synapse.api.errors import RoomError
|
from synapse.api.errors import RoomError
|
||||||
from synapse.streams.config import PaginationConfig
|
from synapse.streams.config import PaginationConfig
|
||||||
|
from synapse.util.logcontext import PreserveLoggingContext
|
||||||
from ._base import BaseHandler
|
from ._base import BaseHandler
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
@ -86,9 +87,10 @@ class MessageHandler(BaseHandler):
|
|||||||
event, snapshot, suppress_auth=suppress_auth
|
event, snapshot, suppress_auth=suppress_auth
|
||||||
)
|
)
|
||||||
|
|
||||||
self.hs.get_handlers().presence_handler.bump_presence_active_time(
|
with PreserveLoggingContext():
|
||||||
user
|
self.hs.get_handlers().presence_handler.bump_presence_active_time(
|
||||||
)
|
user
|
||||||
|
)
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def get_messages(self, user_id=None, room_id=None, pagin_config=None,
|
def get_messages(self, user_id=None, room_id=None, pagin_config=None,
|
||||||
@ -296,7 +298,7 @@ class MessageHandler(BaseHandler):
|
|||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def room_initial_sync(self, user_id, room_id, pagin_config=None,
|
def room_initial_sync(self, user_id, room_id, pagin_config=None,
|
||||||
feedback=False):
|
feedback=False):
|
||||||
yield self.auth.check_joined_room(room_id, user_id)
|
yield self.auth.check_joined_room(room_id, user_id)
|
||||||
|
|
||||||
# TODO(paul): I wish I was called with user objects not user_id
|
# TODO(paul): I wish I was called with user objects not user_id
|
||||||
@ -340,8 +342,8 @@ class MessageHandler(BaseHandler):
|
|||||||
)
|
)
|
||||||
presence.append(member_presence)
|
presence.append(member_presence)
|
||||||
except Exception:
|
except Exception:
|
||||||
logger.exception("Failed to get member presence of %r",
|
logger.exception(
|
||||||
m.user_id
|
"Failed to get member presence of %r", m.user_id
|
||||||
)
|
)
|
||||||
|
|
||||||
defer.returnValue({
|
defer.returnValue({
|
||||||
|
@ -19,6 +19,7 @@ from synapse.api.errors import SynapseError, AuthError
|
|||||||
from synapse.api.constants import PresenceState
|
from synapse.api.constants import PresenceState
|
||||||
|
|
||||||
from synapse.util.logutils import log_function
|
from synapse.util.logutils import log_function
|
||||||
|
from synapse.util.logcontext import PreserveLoggingContext
|
||||||
|
|
||||||
from ._base import BaseHandler
|
from ._base import BaseHandler
|
||||||
|
|
||||||
@ -142,7 +143,7 @@ class PresenceHandler(BaseHandler):
|
|||||||
return UserPresenceCache()
|
return UserPresenceCache()
|
||||||
|
|
||||||
def registered_user(self, user):
|
def registered_user(self, user):
|
||||||
self.store.create_presence(user.localpart)
|
return self.store.create_presence(user.localpart)
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def is_presence_visible(self, observer_user, observed_user):
|
def is_presence_visible(self, observer_user, observed_user):
|
||||||
@ -241,14 +242,12 @@ class PresenceHandler(BaseHandler):
|
|||||||
was_level = self.STATE_LEVELS[statuscache.get_state()["presence"]]
|
was_level = self.STATE_LEVELS[statuscache.get_state()["presence"]]
|
||||||
now_level = self.STATE_LEVELS[state["presence"]]
|
now_level = self.STATE_LEVELS[state["presence"]]
|
||||||
|
|
||||||
yield defer.DeferredList([
|
yield self.store.set_presence_state(
|
||||||
self.store.set_presence_state(
|
target_user.localpart, state_to_store
|
||||||
target_user.localpart, state_to_store
|
)
|
||||||
),
|
yield self.distributor.fire(
|
||||||
self.distributor.fire(
|
"collect_presencelike_data", target_user, state
|
||||||
"collect_presencelike_data", target_user, state
|
)
|
||||||
),
|
|
||||||
])
|
|
||||||
|
|
||||||
if now_level > was_level:
|
if now_level > was_level:
|
||||||
state["last_active"] = self.clock.time_msec()
|
state["last_active"] = self.clock.time_msec()
|
||||||
@ -256,14 +255,15 @@ class PresenceHandler(BaseHandler):
|
|||||||
now_online = state["presence"] != PresenceState.OFFLINE
|
now_online = state["presence"] != PresenceState.OFFLINE
|
||||||
was_polling = target_user in self._user_cachemap
|
was_polling = target_user in self._user_cachemap
|
||||||
|
|
||||||
if now_online and not was_polling:
|
with PreserveLoggingContext():
|
||||||
self.start_polling_presence(target_user, state=state)
|
if now_online and not was_polling:
|
||||||
elif not now_online and was_polling:
|
self.start_polling_presence(target_user, state=state)
|
||||||
self.stop_polling_presence(target_user)
|
elif not now_online and was_polling:
|
||||||
|
self.stop_polling_presence(target_user)
|
||||||
|
|
||||||
# TODO(paul): perform a presence push as part of start/stop poll so
|
# TODO(paul): perform a presence push as part of start/stop poll so
|
||||||
# we don't have to do this all the time
|
# we don't have to do this all the time
|
||||||
self.changed_presencelike_data(target_user, state)
|
self.changed_presencelike_data(target_user, state)
|
||||||
|
|
||||||
def bump_presence_active_time(self, user, now=None):
|
def bump_presence_active_time(self, user, now=None):
|
||||||
if now is None:
|
if now is None:
|
||||||
@ -277,7 +277,7 @@ class PresenceHandler(BaseHandler):
|
|||||||
self._user_cachemap_latest_serial += 1
|
self._user_cachemap_latest_serial += 1
|
||||||
statuscache.update(state, serial=self._user_cachemap_latest_serial)
|
statuscache.update(state, serial=self._user_cachemap_latest_serial)
|
||||||
|
|
||||||
self.push_presence(user, statuscache=statuscache)
|
return self.push_presence(user, statuscache=statuscache)
|
||||||
|
|
||||||
@log_function
|
@log_function
|
||||||
def started_user_eventstream(self, user):
|
def started_user_eventstream(self, user):
|
||||||
@ -381,8 +381,10 @@ class PresenceHandler(BaseHandler):
|
|||||||
yield self.store.set_presence_list_accepted(
|
yield self.store.set_presence_list_accepted(
|
||||||
observer_user.localpart, observed_user.to_string()
|
observer_user.localpart, observed_user.to_string()
|
||||||
)
|
)
|
||||||
|
with PreserveLoggingContext():
|
||||||
self.start_polling_presence(observer_user, target_user=observed_user)
|
self.start_polling_presence(
|
||||||
|
observer_user, target_user=observed_user
|
||||||
|
)
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def deny_presence(self, observed_user, observer_user):
|
def deny_presence(self, observed_user, observer_user):
|
||||||
@ -401,7 +403,10 @@ class PresenceHandler(BaseHandler):
|
|||||||
observer_user.localpart, observed_user.to_string()
|
observer_user.localpart, observed_user.to_string()
|
||||||
)
|
)
|
||||||
|
|
||||||
self.stop_polling_presence(observer_user, target_user=observed_user)
|
with PreserveLoggingContext():
|
||||||
|
self.stop_polling_presence(
|
||||||
|
observer_user, target_user=observed_user
|
||||||
|
)
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def get_presence_list(self, observer_user, accepted=None):
|
def get_presence_list(self, observer_user, accepted=None):
|
||||||
@ -710,7 +715,8 @@ class PresenceHandler(BaseHandler):
|
|||||||
if not self._remote_sendmap[user]:
|
if not self._remote_sendmap[user]:
|
||||||
del self._remote_sendmap[user]
|
del self._remote_sendmap[user]
|
||||||
|
|
||||||
yield defer.DeferredList(deferreds)
|
with PreserveLoggingContext():
|
||||||
|
yield defer.DeferredList(deferreds)
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def push_update_to_local_and_remote(self, observed_user, statuscache,
|
def push_update_to_local_and_remote(self, observed_user, statuscache,
|
||||||
|
@ -17,6 +17,7 @@ from twisted.internet import defer
|
|||||||
|
|
||||||
from synapse.api.errors import SynapseError, AuthError, CodeMessageException
|
from synapse.api.errors import SynapseError, AuthError, CodeMessageException
|
||||||
from synapse.api.constants import Membership
|
from synapse.api.constants import Membership
|
||||||
|
from synapse.util.logcontext import PreserveLoggingContext
|
||||||
|
|
||||||
from ._base import BaseHandler
|
from ._base import BaseHandler
|
||||||
|
|
||||||
@ -46,7 +47,7 @@ class ProfileHandler(BaseHandler):
|
|||||||
)
|
)
|
||||||
|
|
||||||
def registered_user(self, user):
|
def registered_user(self, user):
|
||||||
self.store.create_profile(user.localpart)
|
return self.store.create_profile(user.localpart)
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def get_displayname(self, target_user):
|
def get_displayname(self, target_user):
|
||||||
@ -152,13 +153,14 @@ class ProfileHandler(BaseHandler):
|
|||||||
if not user.is_mine:
|
if not user.is_mine:
|
||||||
defer.returnValue(None)
|
defer.returnValue(None)
|
||||||
|
|
||||||
(displayname, avatar_url) = yield defer.gatherResults(
|
with PreserveLoggingContext():
|
||||||
[
|
(displayname, avatar_url) = yield defer.gatherResults(
|
||||||
self.store.get_profile_displayname(user.localpart),
|
[
|
||||||
self.store.get_profile_avatar_url(user.localpart),
|
self.store.get_profile_displayname(user.localpart),
|
||||||
],
|
self.store.get_profile_avatar_url(user.localpart),
|
||||||
consumeErrors=True
|
],
|
||||||
)
|
consumeErrors=True
|
||||||
|
)
|
||||||
|
|
||||||
state["displayname"] = displayname
|
state["displayname"] = displayname
|
||||||
state["avatar_url"] = avatar_url
|
state["avatar_url"] = avatar_url
|
||||||
|
@ -69,7 +69,7 @@ class RegistrationHandler(BaseHandler):
|
|||||||
password_hash=password_hash
|
password_hash=password_hash
|
||||||
)
|
)
|
||||||
|
|
||||||
self.distributor.fire("registered_user", user)
|
yield self.distributor.fire("registered_user", user)
|
||||||
else:
|
else:
|
||||||
# autogen a random user ID
|
# autogen a random user ID
|
||||||
attempts = 0
|
attempts = 0
|
||||||
|
@ -178,7 +178,9 @@ class RoomCreationHandler(BaseHandler):
|
|||||||
|
|
||||||
if room_alias:
|
if room_alias:
|
||||||
result["room_alias"] = room_alias.to_string()
|
result["room_alias"] = room_alias.to_string()
|
||||||
directory_handler.send_room_alias_update_event(user_id, room_id)
|
yield directory_handler.send_room_alias_update_event(
|
||||||
|
user_id, room_id
|
||||||
|
)
|
||||||
|
|
||||||
defer.returnValue(result)
|
defer.returnValue(result)
|
||||||
|
|
||||||
@ -211,7 +213,6 @@ class RoomCreationHandler(BaseHandler):
|
|||||||
**event_keys
|
**event_keys
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
power_levels_event = self.event_factory.create_event(
|
power_levels_event = self.event_factory.create_event(
|
||||||
etype=RoomPowerLevelsEvent.TYPE,
|
etype=RoomPowerLevelsEvent.TYPE,
|
||||||
content={
|
content={
|
||||||
@ -480,7 +481,7 @@ class RoomMemberHandler(BaseHandler):
|
|||||||
)
|
)
|
||||||
|
|
||||||
user = self.hs.parse_userid(event.user_id)
|
user = self.hs.parse_userid(event.user_id)
|
||||||
self.distributor.fire(
|
yield self.distributor.fire(
|
||||||
"user_joined_room", user=user, room_id=room_id
|
"user_joined_room", user=user, room_id=room_id
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -131,11 +131,13 @@ class ContentRepoResource(resource.Resource):
|
|||||||
request.setHeader('Content-Type', content_type)
|
request.setHeader('Content-Type', content_type)
|
||||||
|
|
||||||
# cache for at least a day.
|
# cache for at least a day.
|
||||||
# XXX: we might want to turn this off for data we don't want to recommend
|
# XXX: we might want to turn this off for data we don't want to
|
||||||
# caching as it's sensitive or private - or at least select private.
|
# recommend caching as it's sensitive or private - or at least
|
||||||
# don't bother setting Expires as all our matrix clients are smart enough to
|
# select private. don't bother setting Expires as all our matrix
|
||||||
# be happy with Cache-Control (right?)
|
# clients are smart enough to be happy with Cache-Control (right?)
|
||||||
request.setHeader('Cache-Control', 'public,max-age=86400,s-maxage=86400')
|
request.setHeader(
|
||||||
|
"Cache-Control", "public,max-age=86400,s-maxage=86400"
|
||||||
|
)
|
||||||
|
|
||||||
d = FileSender().beginFileTransfer(f, request)
|
d = FileSender().beginFileTransfer(f, request)
|
||||||
|
|
||||||
|
@ -138,8 +138,7 @@ class JsonResource(HttpServer, resource.Resource):
|
|||||||
)
|
)
|
||||||
except CodeMessageException as e:
|
except CodeMessageException as e:
|
||||||
if isinstance(e, SynapseError):
|
if isinstance(e, SynapseError):
|
||||||
logger.error("%s SynapseError: %s - %s", request, e.code,
|
logger.info("%s SynapseError: %s - %s", request, e.code, e.msg)
|
||||||
e.msg)
|
|
||||||
else:
|
else:
|
||||||
logger.exception(e)
|
logger.exception(e)
|
||||||
self._send_response(
|
self._send_response(
|
||||||
|
@ -17,6 +17,7 @@ from twisted.internet import defer
|
|||||||
|
|
||||||
from synapse.util.logutils import log_function
|
from synapse.util.logutils import log_function
|
||||||
from synapse.util.logcontext import PreserveLoggingContext
|
from synapse.util.logcontext import PreserveLoggingContext
|
||||||
|
from synapse.util.async import run_on_reactor
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
@ -96,6 +97,7 @@ class Notifier(object):
|
|||||||
listening to the room, and any listeners for the users in the
|
listening to the room, and any listeners for the users in the
|
||||||
`extra_users` param.
|
`extra_users` param.
|
||||||
"""
|
"""
|
||||||
|
yield run_on_reactor()
|
||||||
room_id = event.room_id
|
room_id = event.room_id
|
||||||
|
|
||||||
room_source = self.event_sources.sources["room"]
|
room_source = self.event_sources.sources["room"]
|
||||||
@ -143,6 +145,7 @@ class Notifier(object):
|
|||||||
|
|
||||||
Will wake up all listeners for the given users and rooms.
|
Will wake up all listeners for the given users and rooms.
|
||||||
"""
|
"""
|
||||||
|
yield run_on_reactor()
|
||||||
presence_source = self.event_sources.sources["presence"]
|
presence_source = self.event_sources.sources["presence"]
|
||||||
|
|
||||||
listeners = set()
|
listeners = set()
|
||||||
@ -211,6 +214,7 @@ class Notifier(object):
|
|||||||
timeout,
|
timeout,
|
||||||
deferred,
|
deferred,
|
||||||
)
|
)
|
||||||
|
|
||||||
def _timeout_listener():
|
def _timeout_listener():
|
||||||
# TODO (erikj): We should probably set to_token to the current
|
# TODO (erikj): We should probably set to_token to the current
|
||||||
# max rather than reusing from_token.
|
# max rather than reusing from_token.
|
||||||
|
@ -26,7 +26,6 @@ import logging
|
|||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
class EventStreamRestServlet(RestServlet):
|
class EventStreamRestServlet(RestServlet):
|
||||||
PATTERN = client_path_pattern("/events$")
|
PATTERN = client_path_pattern("/events$")
|
||||||
|
|
||||||
|
@ -117,8 +117,6 @@ class PresenceListRestServlet(RestServlet):
|
|||||||
logger.exception("JSON parse error")
|
logger.exception("JSON parse error")
|
||||||
raise SynapseError(400, "Unable to parse content")
|
raise SynapseError(400, "Unable to parse content")
|
||||||
|
|
||||||
deferreds = []
|
|
||||||
|
|
||||||
if "invite" in content:
|
if "invite" in content:
|
||||||
for u in content["invite"]:
|
for u in content["invite"]:
|
||||||
if not isinstance(u, basestring):
|
if not isinstance(u, basestring):
|
||||||
@ -126,8 +124,9 @@ class PresenceListRestServlet(RestServlet):
|
|||||||
if len(u) == 0:
|
if len(u) == 0:
|
||||||
continue
|
continue
|
||||||
invited_user = self.hs.parse_userid(u)
|
invited_user = self.hs.parse_userid(u)
|
||||||
deferreds.append(self.handlers.presence_handler.send_invite(
|
yield self.handlers.presence_handler.send_invite(
|
||||||
observer_user=user, observed_user=invited_user))
|
observer_user=user, observed_user=invited_user
|
||||||
|
)
|
||||||
|
|
||||||
if "drop" in content:
|
if "drop" in content:
|
||||||
for u in content["drop"]:
|
for u in content["drop"]:
|
||||||
@ -136,10 +135,9 @@ class PresenceListRestServlet(RestServlet):
|
|||||||
if len(u) == 0:
|
if len(u) == 0:
|
||||||
continue
|
continue
|
||||||
dropped_user = self.hs.parse_userid(u)
|
dropped_user = self.hs.parse_userid(u)
|
||||||
deferreds.append(self.handlers.presence_handler.drop(
|
yield self.handlers.presence_handler.drop(
|
||||||
observer_user=user, observed_user=dropped_user))
|
observer_user=user, observed_user=dropped_user
|
||||||
|
)
|
||||||
yield defer.DeferredList(deferreds)
|
|
||||||
|
|
||||||
defer.returnValue((200, {}))
|
defer.returnValue((200, {}))
|
||||||
|
|
||||||
|
@ -508,7 +508,7 @@ def prepare_database(db_conn):
|
|||||||
"new for the server to understand"
|
"new for the server to understand"
|
||||||
)
|
)
|
||||||
elif user_version < SCHEMA_VERSION:
|
elif user_version < SCHEMA_VERSION:
|
||||||
logging.info(
|
logger.info(
|
||||||
"Upgrading database from version %d",
|
"Upgrading database from version %d",
|
||||||
user_version
|
user_version
|
||||||
)
|
)
|
||||||
|
@ -57,7 +57,7 @@ class LoggingTransaction(object):
|
|||||||
if args and args[0]:
|
if args and args[0]:
|
||||||
values = args[0]
|
values = args[0]
|
||||||
sql_logger.debug(
|
sql_logger.debug(
|
||||||
"[SQL values] {%s} " + ", ".join(("<%s>",) * len(values)),
|
"[SQL values] {%s} " + ", ".join(("<%r>",) * len(values)),
|
||||||
self.name,
|
self.name,
|
||||||
*values
|
*values
|
||||||
)
|
)
|
||||||
@ -91,6 +91,7 @@ class SQLBaseStore(object):
|
|||||||
def runInteraction(self, desc, func, *args, **kwargs):
|
def runInteraction(self, desc, func, *args, **kwargs):
|
||||||
"""Wraps the .runInteraction() method on the underlying db_pool."""
|
"""Wraps the .runInteraction() method on the underlying db_pool."""
|
||||||
current_context = LoggingContext.current_context()
|
current_context = LoggingContext.current_context()
|
||||||
|
|
||||||
def inner_func(txn, *args, **kwargs):
|
def inner_func(txn, *args, **kwargs):
|
||||||
with LoggingContext("runInteraction") as context:
|
with LoggingContext("runInteraction") as context:
|
||||||
current_context.copy_to(context)
|
current_context.copy_to(context)
|
||||||
@ -115,7 +116,6 @@ class SQLBaseStore(object):
|
|||||||
"[TXN END] {%s} %f",
|
"[TXN END] {%s} %f",
|
||||||
name, end - start
|
name, end - start
|
||||||
)
|
)
|
||||||
|
|
||||||
with PreserveLoggingContext():
|
with PreserveLoggingContext():
|
||||||
result = yield self._db_pool.runInteraction(
|
result = yield self._db_pool.runInteraction(
|
||||||
inner_func, *args, **kwargs
|
inner_func, *args, **kwargs
|
||||||
|
@ -75,7 +75,9 @@ class RegistrationStore(SQLBaseStore):
|
|||||||
"VALUES (?,?,?)",
|
"VALUES (?,?,?)",
|
||||||
[user_id, password_hash, now])
|
[user_id, password_hash, now])
|
||||||
except IntegrityError:
|
except IntegrityError:
|
||||||
raise StoreError(400, "User ID already taken.", errcode=Codes.USER_IN_USE)
|
raise StoreError(
|
||||||
|
400, "User ID already taken.", errcode=Codes.USER_IN_USE
|
||||||
|
)
|
||||||
|
|
||||||
# it's possible for this to get a conflict, but only for a single user
|
# it's possible for this to get a conflict, but only for a single user
|
||||||
# since tokens are namespaced based on their user ID
|
# since tokens are namespaced based on their user ID
|
||||||
@ -83,8 +85,8 @@ class RegistrationStore(SQLBaseStore):
|
|||||||
"VALUES (?,?)", [txn.lastrowid, token])
|
"VALUES (?,?)", [txn.lastrowid, token])
|
||||||
|
|
||||||
def get_user_by_id(self, user_id):
|
def get_user_by_id(self, user_id):
|
||||||
query = ("SELECT users.name, users.password_hash FROM users "
|
query = ("SELECT users.name, users.password_hash FROM users"
|
||||||
"WHERE users.name = ?")
|
" WHERE users.name = ?")
|
||||||
return self._execute(
|
return self._execute(
|
||||||
self.cursor_to_dict,
|
self.cursor_to_dict,
|
||||||
query, user_id
|
query, user_id
|
||||||
@ -120,10 +122,10 @@ class RegistrationStore(SQLBaseStore):
|
|||||||
|
|
||||||
def _query_for_auth(self, txn, token):
|
def _query_for_auth(self, txn, token):
|
||||||
sql = (
|
sql = (
|
||||||
"SELECT users.name, users.admin, access_tokens.device_id "
|
"SELECT users.name, users.admin, access_tokens.device_id"
|
||||||
"FROM users "
|
" FROM users"
|
||||||
"INNER JOIN access_tokens on users.id = access_tokens.user_id "
|
" INNER JOIN access_tokens on users.id = access_tokens.user_id"
|
||||||
"WHERE token = ?"
|
" WHERE token = ?"
|
||||||
)
|
)
|
||||||
|
|
||||||
cursor = txn.execute(sql, (token,))
|
cursor = txn.execute(sql, (token,))
|
||||||
|
@ -27,7 +27,9 @@ import logging
|
|||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
OpsLevel = collections.namedtuple("OpsLevel", ("ban_level", "kick_level", "redact_level"))
|
OpsLevel = collections.namedtuple("OpsLevel", (
|
||||||
|
"ban_level", "kick_level", "redact_level")
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
class RoomStore(SQLBaseStore):
|
class RoomStore(SQLBaseStore):
|
||||||
|
@ -177,8 +177,8 @@ class RoomMemberStore(SQLBaseStore):
|
|||||||
return self._get_members_query(clause, vals)
|
return self._get_members_query(clause, vals)
|
||||||
|
|
||||||
def _get_members_query(self, where_clause, where_values):
|
def _get_members_query(self, where_clause, where_values):
|
||||||
return self._db_pool.runInteraction(
|
return self.runInteraction(
|
||||||
self._get_members_query_txn,
|
"get_members_query", self._get_members_query_txn,
|
||||||
where_clause, where_values
|
where_clause, where_values
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -36,7 +36,7 @@ class SignatureStore(SQLBaseStore):
|
|||||||
return dict(txn.fetchall())
|
return dict(txn.fetchall())
|
||||||
|
|
||||||
def _store_event_content_hash_txn(self, txn, event_id, algorithm,
|
def _store_event_content_hash_txn(self, txn, event_id, algorithm,
|
||||||
hash_bytes):
|
hash_bytes):
|
||||||
"""Store a hash for a Event
|
"""Store a hash for a Event
|
||||||
Args:
|
Args:
|
||||||
txn (cursor):
|
txn (cursor):
|
||||||
@ -84,7 +84,7 @@ class SignatureStore(SQLBaseStore):
|
|||||||
return dict(txn.fetchall())
|
return dict(txn.fetchall())
|
||||||
|
|
||||||
def _store_event_reference_hash_txn(self, txn, event_id, algorithm,
|
def _store_event_reference_hash_txn(self, txn, event_id, algorithm,
|
||||||
hash_bytes):
|
hash_bytes):
|
||||||
"""Store a hash for a PDU
|
"""Store a hash for a PDU
|
||||||
Args:
|
Args:
|
||||||
txn (cursor):
|
txn (cursor):
|
||||||
@ -127,7 +127,7 @@ class SignatureStore(SQLBaseStore):
|
|||||||
return res
|
return res
|
||||||
|
|
||||||
def _store_event_signature_txn(self, txn, event_id, signature_name, key_id,
|
def _store_event_signature_txn(self, txn, event_id, signature_name, key_id,
|
||||||
signature_bytes):
|
signature_bytes):
|
||||||
"""Store a signature from the origin server for a PDU.
|
"""Store a signature from the origin server for a PDU.
|
||||||
Args:
|
Args:
|
||||||
txn (cursor):
|
txn (cursor):
|
||||||
@ -169,7 +169,7 @@ class SignatureStore(SQLBaseStore):
|
|||||||
return results
|
return results
|
||||||
|
|
||||||
def _store_prev_event_hash_txn(self, txn, event_id, prev_event_id,
|
def _store_prev_event_hash_txn(self, txn, event_id, prev_event_id,
|
||||||
algorithm, hash_bytes):
|
algorithm, hash_bytes):
|
||||||
self._simple_insert_txn(
|
self._simple_insert_txn(
|
||||||
txn,
|
txn,
|
||||||
"event_edge_hashes",
|
"event_edge_hashes",
|
||||||
|
@ -213,8 +213,8 @@ class StreamStore(SQLBaseStore):
|
|||||||
# Tokens really represent positions between elements, but we use
|
# Tokens really represent positions between elements, but we use
|
||||||
# the convention of pointing to the event before the gap. Hence
|
# the convention of pointing to the event before the gap. Hence
|
||||||
# we have a bit of asymmetry when it comes to equalities.
|
# we have a bit of asymmetry when it comes to equalities.
|
||||||
from_comp = '<=' if direction =='b' else '>'
|
from_comp = '<=' if direction == 'b' else '>'
|
||||||
to_comp = '>' if direction =='b' else '<='
|
to_comp = '>' if direction == 'b' else '<='
|
||||||
order = "DESC" if direction == 'b' else "ASC"
|
order = "DESC" if direction == 'b' else "ASC"
|
||||||
|
|
||||||
args = [room_id]
|
args = [room_id]
|
||||||
@ -235,9 +235,10 @@ class StreamStore(SQLBaseStore):
|
|||||||
)
|
)
|
||||||
|
|
||||||
sql = (
|
sql = (
|
||||||
"SELECT *, (%(redacted)s) AS redacted FROM events "
|
"SELECT *, (%(redacted)s) AS redacted FROM events"
|
||||||
"WHERE outlier = 0 AND room_id = ? AND %(bounds)s "
|
" WHERE outlier = 0 AND room_id = ? AND %(bounds)s"
|
||||||
"ORDER BY topological_ordering %(order)s, stream_ordering %(order)s %(limit)s "
|
" ORDER BY topological_ordering %(order)s,"
|
||||||
|
" stream_ordering %(order)s %(limit)s"
|
||||||
) % {
|
) % {
|
||||||
"redacted": del_sql,
|
"redacted": del_sql,
|
||||||
"bounds": bounds,
|
"bounds": bounds,
|
||||||
|
@ -37,6 +37,7 @@ class Clock(object):
|
|||||||
|
|
||||||
def call_later(self, delay, callback):
|
def call_later(self, delay, callback):
|
||||||
current_context = LoggingContext.current_context()
|
current_context = LoggingContext.current_context()
|
||||||
|
|
||||||
def wrapped_callback():
|
def wrapped_callback():
|
||||||
LoggingContext.thread_local.current_context = current_context
|
LoggingContext.thread_local.current_context = current_context
|
||||||
callback()
|
callback()
|
||||||
|
@ -18,6 +18,7 @@ from twisted.internet import defer, reactor
|
|||||||
|
|
||||||
from .logcontext import PreserveLoggingContext
|
from .logcontext import PreserveLoggingContext
|
||||||
|
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def sleep(seconds):
|
def sleep(seconds):
|
||||||
d = defer.Deferred()
|
d = defer.Deferred()
|
||||||
@ -25,6 +26,7 @@ def sleep(seconds):
|
|||||||
with PreserveLoggingContext():
|
with PreserveLoggingContext():
|
||||||
yield d
|
yield d
|
||||||
|
|
||||||
|
|
||||||
def run_on_reactor():
|
def run_on_reactor():
|
||||||
""" This will cause the rest of the function to be invoked upon the next
|
""" This will cause the rest of the function to be invoked upon the next
|
||||||
iteration of the main loop
|
iteration of the main loop
|
||||||
|
@ -13,6 +13,8 @@
|
|||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
|
||||||
|
from synapse.util.logcontext import PreserveLoggingContext
|
||||||
|
|
||||||
from twisted.internet import defer
|
from twisted.internet import defer
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
@ -91,6 +93,7 @@ class Signal(object):
|
|||||||
Each observer callable may return a Deferred."""
|
Each observer callable may return a Deferred."""
|
||||||
self.observers.append(observer)
|
self.observers.append(observer)
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
def fire(self, *args, **kwargs):
|
def fire(self, *args, **kwargs):
|
||||||
"""Invokes every callable in the observer list, passing in the args and
|
"""Invokes every callable in the observer list, passing in the args and
|
||||||
kwargs. Exceptions thrown by observers are logged but ignored. It is
|
kwargs. Exceptions thrown by observers are logged but ignored. It is
|
||||||
@ -98,22 +101,24 @@ class Signal(object):
|
|||||||
|
|
||||||
Returns a Deferred that will complete when all the observers have
|
Returns a Deferred that will complete when all the observers have
|
||||||
completed."""
|
completed."""
|
||||||
deferreds = []
|
with PreserveLoggingContext():
|
||||||
for observer in self.observers:
|
deferreds = []
|
||||||
d = defer.maybeDeferred(observer, *args, **kwargs)
|
for observer in self.observers:
|
||||||
|
d = defer.maybeDeferred(observer, *args, **kwargs)
|
||||||
|
|
||||||
def eb(failure):
|
def eb(failure):
|
||||||
logger.warning(
|
logger.warning(
|
||||||
"%s signal observer %s failed: %r",
|
"%s signal observer %s failed: %r",
|
||||||
self.name, observer, failure,
|
self.name, observer, failure,
|
||||||
exc_info=(
|
exc_info=(
|
||||||
failure.type,
|
failure.type,
|
||||||
failure.value,
|
failure.value,
|
||||||
failure.getTracebackObject()))
|
failure.getTracebackObject()))
|
||||||
if not self.suppress_failures:
|
if not self.suppress_failures:
|
||||||
raise failure
|
raise failure
|
||||||
deferreds.append(d.addErrback(eb))
|
deferreds.append(d.addErrback(eb))
|
||||||
|
|
||||||
return defer.DeferredList(
|
result = yield defer.DeferredList(
|
||||||
deferreds, fireOnOneErrback=not self.suppress_failures
|
deferreds, fireOnOneErrback=not self.suppress_failures
|
||||||
)
|
)
|
||||||
|
defer.returnValue(result)
|
||||||
|
@ -1,6 +1,8 @@
|
|||||||
import threading
|
import threading
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
class LoggingContext(object):
|
class LoggingContext(object):
|
||||||
"""Additional context for log formatting. Contexts are scoped within a
|
"""Additional context for log formatting. Contexts are scoped within a
|
||||||
@ -53,7 +55,7 @@ class LoggingContext(object):
|
|||||||
None to avoid suppressing any exeptions that were thrown.
|
None to avoid suppressing any exeptions that were thrown.
|
||||||
"""
|
"""
|
||||||
if self.thread_local.current_context is not self:
|
if self.thread_local.current_context is not self:
|
||||||
logging.error(
|
logger.error(
|
||||||
"Current logging context %s is not the expected context %s",
|
"Current logging context %s is not the expected context %s",
|
||||||
self.thread_local.current_context,
|
self.thread_local.current_context,
|
||||||
self
|
self
|
||||||
|
Loading…
Reference in New Issue
Block a user