mirror of
https://git.anonymousland.org/anonymousland/synapse.git
synced 2025-05-15 15:02:12 -04:00
Merge branch 'develop' of github.com:matrix-org/synapse into matrix-org-hotfixes
This commit is contained in:
commit
8a6196c6c8
5 changed files with 84 additions and 51 deletions
|
@ -332,9 +332,8 @@ https://obs.infoserver.lv/project/monitor/matrix-synapse
|
||||||
ArchLinux
|
ArchLinux
|
||||||
---------
|
---------
|
||||||
|
|
||||||
The quickest way to get up and running with ArchLinux is probably with Ivan
|
The quickest way to get up and running with ArchLinux is probably with the community package
|
||||||
Shapovalov's AUR package from
|
https://www.archlinux.org/packages/community/any/matrix-synapse/, which should pull in all
|
||||||
https://aur.archlinux.org/packages/matrix-synapse/, which should pull in all
|
|
||||||
the necessary dependencies.
|
the necessary dependencies.
|
||||||
|
|
||||||
Alternatively, to install using pip a few changes may be needed as ArchLinux
|
Alternatively, to install using pip a few changes may be needed as ArchLinux
|
||||||
|
|
|
@ -240,6 +240,9 @@ class MediaRepository(object):
|
||||||
if t_method == "crop":
|
if t_method == "crop":
|
||||||
t_len = thumbnailer.crop(t_path, t_width, t_height, t_type)
|
t_len = thumbnailer.crop(t_path, t_width, t_height, t_type)
|
||||||
elif t_method == "scale":
|
elif t_method == "scale":
|
||||||
|
t_width, t_height = thumbnailer.aspect(t_width, t_height)
|
||||||
|
t_width = min(m_width, t_width)
|
||||||
|
t_height = min(m_height, t_height)
|
||||||
t_len = thumbnailer.scale(t_path, t_width, t_height, t_type)
|
t_len = thumbnailer.scale(t_path, t_width, t_height, t_type)
|
||||||
else:
|
else:
|
||||||
t_len = None
|
t_len = None
|
||||||
|
|
|
@ -80,7 +80,13 @@ class LoggingTransaction(object):
|
||||||
def executemany(self, sql, *args):
|
def executemany(self, sql, *args):
|
||||||
self._do_execute(self.txn.executemany, sql, *args)
|
self._do_execute(self.txn.executemany, sql, *args)
|
||||||
|
|
||||||
|
def _make_sql_one_line(self, sql):
|
||||||
|
"Strip newlines out of SQL so that the loggers in the DB are on one line"
|
||||||
|
return " ".join(l.strip() for l in sql.splitlines() if l.strip())
|
||||||
|
|
||||||
def _do_execute(self, func, sql, *args):
|
def _do_execute(self, func, sql, *args):
|
||||||
|
sql = self._make_sql_one_line(sql)
|
||||||
|
|
||||||
# TODO(paul): Maybe use 'info' and 'debug' for values?
|
# TODO(paul): Maybe use 'info' and 'debug' for values?
|
||||||
sql_logger.debug("[SQL] {%s} %s", self.name, sql)
|
sql_logger.debug("[SQL] {%s} %s", self.name, sql)
|
||||||
|
|
||||||
|
|
|
@ -19,6 +19,8 @@ from twisted.internet import defer
|
||||||
|
|
||||||
from synapse.api.errors import StoreError
|
from synapse.api.errors import StoreError
|
||||||
from ._base import SQLBaseStore
|
from ._base import SQLBaseStore
|
||||||
|
from synapse.util.caches.descriptors import cached, cachedList, cachedInlineCallbacks
|
||||||
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
@ -144,6 +146,7 @@ class DeviceStore(SQLBaseStore):
|
||||||
|
|
||||||
defer.returnValue({d["device_id"]: d for d in devices})
|
defer.returnValue({d["device_id"]: d for d in devices})
|
||||||
|
|
||||||
|
@cached(max_entries=10000)
|
||||||
def get_device_list_last_stream_id_for_remote(self, user_id):
|
def get_device_list_last_stream_id_for_remote(self, user_id):
|
||||||
"""Get the last stream_id we got for a user. May be None if we haven't
|
"""Get the last stream_id we got for a user. May be None if we haven't
|
||||||
got any information for them.
|
got any information for them.
|
||||||
|
@ -156,16 +159,36 @@ class DeviceStore(SQLBaseStore):
|
||||||
allow_none=True,
|
allow_none=True,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@cachedList(cached_method_name="get_device_list_last_stream_id_for_remote",
|
||||||
|
list_name="user_ids", inlineCallbacks=True)
|
||||||
|
def get_device_list_last_stream_id_for_remotes(self, user_ids):
|
||||||
|
rows = yield self._simple_select_many_batch(
|
||||||
|
table="device_lists_remote_extremeties",
|
||||||
|
column="user_id",
|
||||||
|
iterable=user_ids,
|
||||||
|
retcols=("user_id", "stream_id",),
|
||||||
|
desc="get_user_devices_from_cache",
|
||||||
|
)
|
||||||
|
|
||||||
|
results = {user_id: None for user_id in user_ids}
|
||||||
|
results.update({
|
||||||
|
row["user_id"]: row["stream_id"] for row in rows
|
||||||
|
})
|
||||||
|
|
||||||
|
defer.returnValue(results)
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
def mark_remote_user_device_list_as_unsubscribed(self, user_id):
|
def mark_remote_user_device_list_as_unsubscribed(self, user_id):
|
||||||
"""Mark that we no longer track device lists for remote user.
|
"""Mark that we no longer track device lists for remote user.
|
||||||
"""
|
"""
|
||||||
return self._simple_delete(
|
yield self._simple_delete(
|
||||||
table="device_lists_remote_extremeties",
|
table="device_lists_remote_extremeties",
|
||||||
keyvalues={
|
keyvalues={
|
||||||
"user_id": user_id,
|
"user_id": user_id,
|
||||||
},
|
},
|
||||||
desc="mark_remote_user_device_list_as_unsubscribed",
|
desc="mark_remote_user_device_list_as_unsubscribed",
|
||||||
)
|
)
|
||||||
|
self.get_device_list_last_stream_id_for_remote.invalidate((user_id,))
|
||||||
|
|
||||||
def update_remote_device_list_cache_entry(self, user_id, device_id, content,
|
def update_remote_device_list_cache_entry(self, user_id, device_id, content,
|
||||||
stream_id):
|
stream_id):
|
||||||
|
@ -191,6 +214,12 @@ class DeviceStore(SQLBaseStore):
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
|
||||||
|
txn.call_after(self._get_cached_user_device.invalidate, (user_id, device_id,))
|
||||||
|
txn.call_after(self._get_cached_devices_for_user.invalidate, (user_id,))
|
||||||
|
txn.call_after(
|
||||||
|
self.get_device_list_last_stream_id_for_remote.invalidate, (user_id,)
|
||||||
|
)
|
||||||
|
|
||||||
self._simple_upsert_txn(
|
self._simple_upsert_txn(
|
||||||
txn,
|
txn,
|
||||||
table="device_lists_remote_extremeties",
|
table="device_lists_remote_extremeties",
|
||||||
|
@ -234,6 +263,12 @@ class DeviceStore(SQLBaseStore):
|
||||||
]
|
]
|
||||||
)
|
)
|
||||||
|
|
||||||
|
txn.call_after(self._get_cached_devices_for_user.invalidate, (user_id,))
|
||||||
|
txn.call_after(self._get_cached_user_device.invalidate_many, (user_id,))
|
||||||
|
txn.call_after(
|
||||||
|
self.get_device_list_last_stream_id_for_remote.invalidate, (user_id,)
|
||||||
|
)
|
||||||
|
|
||||||
self._simple_upsert_txn(
|
self._simple_upsert_txn(
|
||||||
txn,
|
txn,
|
||||||
table="device_lists_remote_extremeties",
|
table="device_lists_remote_extremeties",
|
||||||
|
@ -320,6 +355,7 @@ class DeviceStore(SQLBaseStore):
|
||||||
|
|
||||||
return (now_stream_id, results)
|
return (now_stream_id, results)
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
def get_user_devices_from_cache(self, query_list):
|
def get_user_devices_from_cache(self, query_list):
|
||||||
"""Get the devices (and keys if any) for remote users from the cache.
|
"""Get the devices (and keys if any) for remote users from the cache.
|
||||||
|
|
||||||
|
@ -332,27 +368,11 @@ class DeviceStore(SQLBaseStore):
|
||||||
a set of user_ids and results_map is a mapping of
|
a set of user_ids and results_map is a mapping of
|
||||||
user_id -> device_id -> device_info
|
user_id -> device_id -> device_info
|
||||||
"""
|
"""
|
||||||
return self.runInteraction(
|
user_ids = set(user_id for user_id, _ in query_list)
|
||||||
"get_user_devices_from_cache", self._get_user_devices_from_cache_txn,
|
user_map = yield self.get_device_list_last_stream_id_for_remotes(list(user_ids))
|
||||||
query_list,
|
user_ids_in_cache = set(
|
||||||
|
user_id for user_id, stream_id in user_map.items() if stream_id
|
||||||
)
|
)
|
||||||
|
|
||||||
def _get_user_devices_from_cache_txn(self, txn, query_list):
|
|
||||||
user_ids = {user_id for user_id, _ in query_list}
|
|
||||||
|
|
||||||
user_ids_in_cache = set()
|
|
||||||
for user_id in user_ids:
|
|
||||||
stream_ids = self._simple_select_onecol_txn(
|
|
||||||
txn,
|
|
||||||
table="device_lists_remote_extremeties",
|
|
||||||
keyvalues={
|
|
||||||
"user_id": user_id,
|
|
||||||
},
|
|
||||||
retcol="stream_id",
|
|
||||||
)
|
|
||||||
if stream_ids:
|
|
||||||
user_ids_in_cache.add(user_id)
|
|
||||||
|
|
||||||
user_ids_not_in_cache = user_ids - user_ids_in_cache
|
user_ids_not_in_cache = user_ids - user_ids_in_cache
|
||||||
|
|
||||||
results = {}
|
results = {}
|
||||||
|
@ -361,32 +381,40 @@ class DeviceStore(SQLBaseStore):
|
||||||
continue
|
continue
|
||||||
|
|
||||||
if device_id:
|
if device_id:
|
||||||
content = self._simple_select_one_onecol_txn(
|
device = yield self._get_cached_user_device(user_id, device_id)
|
||||||
txn,
|
results.setdefault(user_id, {})[device_id] = device
|
||||||
|
else:
|
||||||
|
results[user_id] = yield self._get_cached_devices_for_user(user_id)
|
||||||
|
|
||||||
|
defer.returnValue((user_ids_not_in_cache, results))
|
||||||
|
|
||||||
|
@cachedInlineCallbacks(num_args=2, tree=True)
|
||||||
|
def _get_cached_user_device(self, user_id, device_id):
|
||||||
|
content = yield self._simple_select_one_onecol(
|
||||||
table="device_lists_remote_cache",
|
table="device_lists_remote_cache",
|
||||||
keyvalues={
|
keyvalues={
|
||||||
"user_id": user_id,
|
"user_id": user_id,
|
||||||
"device_id": device_id,
|
"device_id": device_id,
|
||||||
},
|
},
|
||||||
retcol="content",
|
retcol="content",
|
||||||
|
desc="_get_cached_user_device",
|
||||||
)
|
)
|
||||||
results.setdefault(user_id, {})[device_id] = json.loads(content)
|
defer.returnValue(json.loads(content))
|
||||||
else:
|
|
||||||
devices = self._simple_select_list_txn(
|
@cachedInlineCallbacks()
|
||||||
txn,
|
def _get_cached_devices_for_user(self, user_id):
|
||||||
|
devices = yield self._simple_select_list(
|
||||||
table="device_lists_remote_cache",
|
table="device_lists_remote_cache",
|
||||||
keyvalues={
|
keyvalues={
|
||||||
"user_id": user_id,
|
"user_id": user_id,
|
||||||
},
|
},
|
||||||
retcols=("device_id", "content"),
|
retcols=("device_id", "content"),
|
||||||
|
desc="_get_cached_devices_for_user",
|
||||||
)
|
)
|
||||||
results[user_id] = {
|
defer.returnValue({
|
||||||
device["device_id"]: json.loads(device["content"])
|
device["device_id"]: json.loads(device["content"])
|
||||||
for device in devices
|
for device in devices
|
||||||
}
|
})
|
||||||
user_ids_in_cache.discard(user_id)
|
|
||||||
|
|
||||||
return user_ids_not_in_cache, results
|
|
||||||
|
|
||||||
def get_devices_with_keys_by_user(self, user_id):
|
def get_devices_with_keys_by_user(self, user_id):
|
||||||
"""Get all devices (with any device keys) for a user
|
"""Get all devices (with any device keys) for a user
|
||||||
|
|
|
@ -640,9 +640,6 @@ class EventPushActionsStore(SQLBaseStore):
|
||||||
the archiving process has caught up or not.
|
the archiving process has caught up or not.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
# We want to make sure that we only ever do this one at a time
|
|
||||||
# self.database_engine.lock_table(txn, "event_push_summary")
|
|
||||||
|
|
||||||
old_rotate_stream_ordering = self._simple_select_one_onecol_txn(
|
old_rotate_stream_ordering = self._simple_select_one_onecol_txn(
|
||||||
txn,
|
txn,
|
||||||
table="event_push_summary_stream_ordering",
|
table="event_push_summary_stream_ordering",
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue