mirror of
https://git.anonymousland.org/anonymousland/synapse.git
synced 2024-10-01 11:49:51 -04:00
Batch up various DB requests for event -> state
This commit is contained in:
parent
0211890134
commit
9eb5b23d3a
@ -13,11 +13,10 @@
|
|||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
|
||||||
from ._base import SQLBaseStore, cached, cachedInlineCallbacks
|
from ._base import SQLBaseStore, cached, cachedInlineCallbacks, cachedList
|
||||||
|
|
||||||
from twisted.internet import defer
|
from twisted.internet import defer
|
||||||
|
|
||||||
from synapse.util import unwrapFirstError
|
|
||||||
from synapse.util.stringutils import random_string
|
from synapse.util.stringutils import random_string
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
@ -50,32 +49,20 @@ class StateStore(SQLBaseStore):
|
|||||||
|
|
||||||
The return value is a dict mapping group names to lists of events.
|
The return value is a dict mapping group names to lists of events.
|
||||||
"""
|
"""
|
||||||
|
if not event_ids:
|
||||||
|
defer.returnValue({})
|
||||||
|
|
||||||
event_and_groups = yield defer.gatherResults(
|
event_to_groups = yield self._get_state_group_for_events(
|
||||||
[
|
room_id, event_ids,
|
||||||
self._get_state_group_for_event(
|
)
|
||||||
room_id, event_id,
|
|
||||||
).addCallback(lambda group, event_id: (event_id, group), event_id)
|
|
||||||
for event_id in event_ids
|
|
||||||
],
|
|
||||||
consumeErrors=True,
|
|
||||||
).addErrback(unwrapFirstError)
|
|
||||||
|
|
||||||
groups = set(group for _, group in event_and_groups if group)
|
groups = set(event_to_groups.values())
|
||||||
|
|
||||||
group_to_state = yield defer.gatherResults(
|
group_to_state = yield self._get_state_for_groups(groups)
|
||||||
[
|
|
||||||
self._get_state_for_group(
|
|
||||||
group,
|
|
||||||
).addCallback(lambda state_dict, group: (group, state_dict), group)
|
|
||||||
for group in groups
|
|
||||||
],
|
|
||||||
consumeErrors=True,
|
|
||||||
).addErrback(unwrapFirstError)
|
|
||||||
|
|
||||||
defer.returnValue({
|
defer.returnValue({
|
||||||
group: state_map.values()
|
group: state_map.values()
|
||||||
for group, state_map in group_to_state
|
for group, state_map in group_to_state.items()
|
||||||
})
|
})
|
||||||
|
|
||||||
@cached(num_args=1)
|
@cached(num_args=1)
|
||||||
@ -212,17 +199,48 @@ class StateStore(SQLBaseStore):
|
|||||||
|
|
||||||
txn.execute(sql, args)
|
txn.execute(sql, args)
|
||||||
|
|
||||||
return group, [
|
return [r[0] for r in txn.fetchall()]
|
||||||
r[0]
|
|
||||||
for r in txn.fetchall()
|
|
||||||
]
|
|
||||||
|
|
||||||
return self.runInteraction(
|
return self.runInteraction(
|
||||||
"_get_state_groups_from_group",
|
"_get_state_groups_from_group",
|
||||||
f,
|
f,
|
||||||
)
|
)
|
||||||
|
|
||||||
@cached(num_args=3, lru=True, max_entries=20000)
|
def _get_state_groups_from_groups(self, groups_and_types):
|
||||||
|
def f(txn):
|
||||||
|
results = {}
|
||||||
|
for group, types in groups_and_types:
|
||||||
|
if types is not None:
|
||||||
|
where_clause = "AND (%s)" % (
|
||||||
|
" OR ".join(["(type = ? AND state_key = ?)"] * len(types)),
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
where_clause = ""
|
||||||
|
|
||||||
|
sql = (
|
||||||
|
"SELECT event_id FROM state_groups_state WHERE"
|
||||||
|
" state_group = ? %s"
|
||||||
|
) % (where_clause,)
|
||||||
|
|
||||||
|
args = [group]
|
||||||
|
if types is not None:
|
||||||
|
args.extend([i for typ in types for i in typ])
|
||||||
|
|
||||||
|
txn.execute(sql, args)
|
||||||
|
|
||||||
|
results[group] = [
|
||||||
|
r[0]
|
||||||
|
for r in txn.fetchall()
|
||||||
|
]
|
||||||
|
|
||||||
|
return results
|
||||||
|
|
||||||
|
return self.runInteraction(
|
||||||
|
"_get_state_groups_from_groups",
|
||||||
|
f,
|
||||||
|
)
|
||||||
|
|
||||||
|
@cached(num_args=3, lru=True, max_entries=10000)
|
||||||
def _get_state_for_event_id(self, room_id, event_id, types):
|
def _get_state_for_event_id(self, room_id, event_id, types):
|
||||||
def f(txn):
|
def f(txn):
|
||||||
type_and_state_sql = " OR ".join([
|
type_and_state_sql = " OR ".join([
|
||||||
@ -274,33 +292,19 @@ class StateStore(SQLBaseStore):
|
|||||||
deferred: A list of dicts corresponding to the event_ids given.
|
deferred: A list of dicts corresponding to the event_ids given.
|
||||||
The dicts are mappings from (type, state_key) -> state_events
|
The dicts are mappings from (type, state_key) -> state_events
|
||||||
"""
|
"""
|
||||||
event_and_groups = yield defer.gatherResults(
|
event_to_groups = yield self._get_state_group_for_events(
|
||||||
[
|
room_id, event_ids,
|
||||||
self._get_state_group_for_event(
|
)
|
||||||
room_id, event_id,
|
|
||||||
).addCallback(lambda group, event_id: (event_id, group), event_id)
|
|
||||||
for event_id in event_ids
|
|
||||||
],
|
|
||||||
consumeErrors=True,
|
|
||||||
).addErrback(unwrapFirstError)
|
|
||||||
|
|
||||||
groups = set(group for _, group in event_and_groups)
|
groups = set(event_to_groups.values())
|
||||||
|
|
||||||
res = yield defer.gatherResults(
|
group_to_state = yield self._get_state_for_groups(
|
||||||
[
|
groups, types
|
||||||
self._get_state_for_group(
|
)
|
||||||
group, types
|
|
||||||
).addCallback(lambda state_dict, group: (group, state_dict), group)
|
|
||||||
for group in groups
|
|
||||||
],
|
|
||||||
consumeErrors=True,
|
|
||||||
).addErrback(unwrapFirstError)
|
|
||||||
|
|
||||||
group_to_state = dict(res)
|
|
||||||
|
|
||||||
event_to_state = {
|
event_to_state = {
|
||||||
event_id: group_to_state[group]
|
event_id: group_to_state[group]
|
||||||
for event_id, group in event_and_groups
|
for event_id, group in event_to_groups.items()
|
||||||
}
|
}
|
||||||
|
|
||||||
defer.returnValue([
|
defer.returnValue([
|
||||||
@ -320,8 +324,29 @@ class StateStore(SQLBaseStore):
|
|||||||
desc="_get_state_group_for_event",
|
desc="_get_state_group_for_event",
|
||||||
)
|
)
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@cachedList(cache=_get_state_group_for_event.cache, list_name="event_ids", num_args=2)
|
||||||
def _get_state_for_group(self, group, types=None):
|
def _get_state_group_for_events(self, room_id, event_ids):
|
||||||
|
def f(txn):
|
||||||
|
results = {}
|
||||||
|
for event_id in event_ids:
|
||||||
|
results[event_id] = self._simple_select_one_onecol_txn(
|
||||||
|
txn,
|
||||||
|
table="event_to_state_groups",
|
||||||
|
keyvalues={
|
||||||
|
"event_id": event_id,
|
||||||
|
},
|
||||||
|
retcol="state_group",
|
||||||
|
allow_none=True,
|
||||||
|
)
|
||||||
|
|
||||||
|
return results
|
||||||
|
|
||||||
|
return self.runInteraction(
|
||||||
|
"_get_state_group_for_events",
|
||||||
|
f,
|
||||||
|
)
|
||||||
|
|
||||||
|
def _get_state_for_group_from_cache(self, group, types=None):
|
||||||
is_all, state_dict = self._state_group_cache.get(group)
|
is_all, state_dict = self._state_group_cache.get(group)
|
||||||
|
|
||||||
type_to_key = {}
|
type_to_key = {}
|
||||||
@ -339,7 +364,7 @@ class StateStore(SQLBaseStore):
|
|||||||
missing_types.add((typ, state_key))
|
missing_types.add((typ, state_key))
|
||||||
|
|
||||||
if is_all and types is None:
|
if is_all and types is None:
|
||||||
defer.returnValue(state_dict)
|
return state_dict, missing_types
|
||||||
|
|
||||||
if is_all or (types is not None and not missing_types):
|
if is_all or (types is not None and not missing_types):
|
||||||
sentinel = object()
|
sentinel = object()
|
||||||
@ -354,41 +379,81 @@ class StateStore(SQLBaseStore):
|
|||||||
return True
|
return True
|
||||||
return False
|
return False
|
||||||
|
|
||||||
defer.returnValue({
|
return {
|
||||||
k: v
|
k: v
|
||||||
for k, v in state_dict.items()
|
for k, v in state_dict.items()
|
||||||
if v and include(k[0], k[1])
|
if v and include(k[0], k[1])
|
||||||
})
|
}, missing_types
|
||||||
|
|
||||||
|
return {}, missing_types
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def _get_state_for_groups(self, groups, types=None):
|
||||||
|
results = {}
|
||||||
|
missing_groups_and_types = []
|
||||||
|
for group in groups:
|
||||||
|
state_dict, missing_types = self._get_state_for_group_from_cache(
|
||||||
|
group, types
|
||||||
|
)
|
||||||
|
|
||||||
|
if types is not None and not missing_types:
|
||||||
|
results[group] = {
|
||||||
|
key: value
|
||||||
|
for key, value in state_dict.items()
|
||||||
|
if value
|
||||||
|
}
|
||||||
|
else:
|
||||||
|
missing_groups_and_types.append((
|
||||||
|
group,
|
||||||
|
missing_types if types else None
|
||||||
|
))
|
||||||
|
|
||||||
|
if not missing_groups_and_types:
|
||||||
|
defer.returnValue(results)
|
||||||
|
|
||||||
# Okay, so we have some missing_types, lets fetch them.
|
# Okay, so we have some missing_types, lets fetch them.
|
||||||
cache_seq_num = self._state_group_cache.sequence
|
cache_seq_num = self._state_group_cache.sequence
|
||||||
_, state_ids = yield self._get_state_groups_from_group(
|
|
||||||
group,
|
group_state_dict = yield self._get_state_groups_from_groups(
|
||||||
frozenset(missing_types) if types else None
|
missing_groups_and_types
|
||||||
)
|
)
|
||||||
state_events = yield self._get_events(state_ids, get_prev_content=False)
|
|
||||||
state_dict = {
|
state_events = yield self._get_events(
|
||||||
key: None
|
[e_id for l in group_state_dict.values() for e_id in l],
|
||||||
for key in missing_types
|
get_prev_content=False
|
||||||
}
|
)
|
||||||
state_dict.update({
|
|
||||||
(e.type, e.state_key): e
|
state_events = {
|
||||||
|
e.event_id: e
|
||||||
for e in state_events
|
for e in state_events
|
||||||
})
|
}
|
||||||
|
|
||||||
# Update the cache
|
for group, state_ids in group_state_dict.items():
|
||||||
self._state_group_cache.update(
|
state_dict = {
|
||||||
cache_seq_num,
|
key: None
|
||||||
key=group,
|
for key in missing_types
|
||||||
value=state_dict,
|
}
|
||||||
full=(types is None),
|
evs = [state_events[e_id] for e_id in state_ids]
|
||||||
)
|
state_dict.update({
|
||||||
|
(e.type, e.state_key): e
|
||||||
|
for e in evs
|
||||||
|
})
|
||||||
|
|
||||||
defer.returnValue({
|
# Update the cache
|
||||||
key: value
|
self._state_group_cache.update(
|
||||||
for key, value in state_dict.items()
|
cache_seq_num,
|
||||||
if value
|
key=group,
|
||||||
})
|
value=state_dict,
|
||||||
|
full=(types is None),
|
||||||
|
)
|
||||||
|
|
||||||
|
results[group] = {
|
||||||
|
key: value
|
||||||
|
for key, value in state_dict.items()
|
||||||
|
if value
|
||||||
|
}
|
||||||
|
|
||||||
|
defer.returnValue(results)
|
||||||
|
|
||||||
|
|
||||||
def _make_group_id(clock):
|
def _make_group_id(clock):
|
||||||
|
Loading…
Reference in New Issue
Block a user