2016-01-06 23:26:29 -05:00
|
|
|
# Copyright 2015, 2016 OpenMarket Ltd
|
2015-08-04 10:56:56 -04:00
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
2020-10-16 10:56:39 -04:00
|
|
|
import enum
|
2015-08-05 10:06:51 -04:00
|
|
|
import logging
|
2018-07-09 02:09:20 -04:00
|
|
|
import threading
|
2021-03-29 12:15:33 -04:00
|
|
|
from typing import Any, Dict, Generic, Iterable, Optional, Set, TypeVar
|
|
|
|
|
|
|
|
import attr
|
2018-07-09 02:09:20 -04:00
|
|
|
|
|
|
|
from synapse.util.caches.lrucache import LruCache
|
2015-08-05 10:06:51 -04:00
|
|
|
|
|
|
|
logger = logging.getLogger(__name__)
|
2015-08-04 10:56:56 -04:00
|
|
|
|
|
|
|
|
2021-03-29 12:15:33 -04:00
|
|
|
# The type of the cache keys.
|
|
|
|
KT = TypeVar("KT")
|
|
|
|
# The type of the dictionary keys.
|
|
|
|
DKT = TypeVar("DKT")
|
|
|
|
|
|
|
|
|
|
|
|
@attr.s(slots=True)
|
|
|
|
class DictionaryEntry:
|
2017-05-17 09:31:23 -04:00
|
|
|
"""Returned when getting an entry from the cache
|
|
|
|
|
|
|
|
Attributes:
|
2021-03-29 12:15:33 -04:00
|
|
|
full: Whether the cache has the full or dict or just some keys.
|
2017-05-17 09:31:23 -04:00
|
|
|
If not full then not all requested keys will necessarily be present
|
|
|
|
in `value`
|
2021-03-29 12:15:33 -04:00
|
|
|
known_absent: Keys that were looked up in the dict and were not
|
2017-05-17 09:31:23 -04:00
|
|
|
there.
|
2021-03-29 12:15:33 -04:00
|
|
|
value: The full or partial dict value
|
2017-05-17 09:31:23 -04:00
|
|
|
"""
|
2019-06-20 05:32:02 -04:00
|
|
|
|
2021-03-29 12:15:33 -04:00
|
|
|
full = attr.ib(type=bool)
|
|
|
|
known_absent = attr.ib()
|
|
|
|
value = attr.ib()
|
|
|
|
|
2017-01-17 06:18:13 -05:00
|
|
|
def __len__(self):
|
|
|
|
return len(self.value)
|
2015-08-04 10:56:56 -04:00
|
|
|
|
|
|
|
|
2020-10-16 10:56:39 -04:00
|
|
|
class _Sentinel(enum.Enum):
|
|
|
|
# defining a sentinel in this way allows mypy to correctly handle the
|
|
|
|
# type of a dictionary lookup.
|
|
|
|
sentinel = object()
|
|
|
|
|
|
|
|
|
2021-03-29 12:15:33 -04:00
|
|
|
class DictionaryCache(Generic[KT, DKT]):
|
2015-08-10 09:16:24 -04:00
|
|
|
"""Caches key -> dictionary lookups, supporting caching partial dicts, i.e.
|
|
|
|
fetching a subset of dictionary keys for a particular key.
|
|
|
|
"""
|
2015-08-04 10:56:56 -04:00
|
|
|
|
2021-03-29 12:15:33 -04:00
|
|
|
def __init__(self, name: str, max_entries: int = 1000):
|
2021-07-15 12:46:54 -04:00
|
|
|
self.cache: LruCache[KT, DictionaryEntry] = LruCache(
|
2020-10-16 10:56:39 -04:00
|
|
|
max_size=max_entries, cache_name=name, size_callback=len
|
2021-07-15 12:46:54 -04:00
|
|
|
)
|
2015-08-04 10:56:56 -04:00
|
|
|
|
|
|
|
self.name = name
|
|
|
|
self.sequence = 0
|
2021-07-15 12:46:54 -04:00
|
|
|
self.thread: Optional[threading.Thread] = None
|
2015-08-04 10:56:56 -04:00
|
|
|
|
2021-03-29 12:15:33 -04:00
|
|
|
def check_thread(self) -> None:
|
2015-08-04 10:56:56 -04:00
|
|
|
expected_thread = self.thread
|
|
|
|
if expected_thread is None:
|
|
|
|
self.thread = threading.current_thread()
|
|
|
|
else:
|
|
|
|
if expected_thread is not threading.current_thread():
|
|
|
|
raise ValueError(
|
|
|
|
"Cache objects can only be accessed from the main thread"
|
|
|
|
)
|
|
|
|
|
2021-03-29 12:15:33 -04:00
|
|
|
def get(
|
|
|
|
self, key: KT, dict_keys: Optional[Iterable[DKT]] = None
|
|
|
|
) -> DictionaryEntry:
|
2017-05-17 09:31:23 -04:00
|
|
|
"""Fetch an entry out of the cache
|
|
|
|
|
|
|
|
Args:
|
|
|
|
key
|
2021-03-29 12:15:33 -04:00
|
|
|
dict_key: If given a set of keys then return only those keys
|
2017-05-17 09:31:23 -04:00
|
|
|
that exist in the cache.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
DictionaryEntry
|
|
|
|
"""
|
2020-10-16 10:56:39 -04:00
|
|
|
entry = self.cache.get(key, _Sentinel.sentinel)
|
|
|
|
if entry is not _Sentinel.sentinel:
|
2015-08-12 05:13:35 -04:00
|
|
|
if dict_keys is None:
|
2019-06-20 05:32:02 -04:00
|
|
|
return DictionaryEntry(
|
|
|
|
entry.full, entry.known_absent, dict(entry.value)
|
|
|
|
)
|
2015-08-12 05:13:35 -04:00
|
|
|
else:
|
2019-06-20 05:32:02 -04:00
|
|
|
return DictionaryEntry(
|
|
|
|
entry.full,
|
|
|
|
entry.known_absent,
|
|
|
|
{k: entry.value[k] for k in dict_keys if k in entry.value},
|
|
|
|
)
|
2015-08-12 05:13:35 -04:00
|
|
|
|
2017-05-17 09:31:23 -04:00
|
|
|
return DictionaryEntry(False, set(), {})
|
2015-08-04 10:56:56 -04:00
|
|
|
|
2021-03-29 12:15:33 -04:00
|
|
|
def invalidate(self, key: KT) -> None:
|
2015-08-04 10:56:56 -04:00
|
|
|
self.check_thread()
|
|
|
|
|
|
|
|
# Increment the sequence number so that any SELECT statements that
|
|
|
|
# raced with the INSERT don't update the cache (SYN-369)
|
|
|
|
self.sequence += 1
|
|
|
|
self.cache.pop(key, None)
|
|
|
|
|
2021-03-29 12:15:33 -04:00
|
|
|
def invalidate_all(self) -> None:
|
2015-08-04 10:56:56 -04:00
|
|
|
self.check_thread()
|
|
|
|
self.sequence += 1
|
|
|
|
self.cache.clear()
|
|
|
|
|
2021-03-29 12:15:33 -04:00
|
|
|
def update(
|
|
|
|
self,
|
|
|
|
sequence: int,
|
|
|
|
key: KT,
|
|
|
|
value: Dict[DKT, Any],
|
|
|
|
fetched_keys: Optional[Set[DKT]] = None,
|
|
|
|
) -> None:
|
2017-05-17 09:31:23 -04:00
|
|
|
"""Updates the entry in the cache
|
|
|
|
|
|
|
|
Args:
|
|
|
|
sequence
|
2021-03-29 12:15:33 -04:00
|
|
|
key
|
|
|
|
value: The value to update the cache with.
|
|
|
|
fetched_keys: All of the dictionary keys which were
|
2018-06-11 18:13:06 -04:00
|
|
|
fetched from the database.
|
|
|
|
|
|
|
|
If None, this is the complete value for key K. Otherwise, it
|
|
|
|
is used to infer a list of keys which we know don't exist in
|
|
|
|
the full dict.
|
2017-05-17 09:31:23 -04:00
|
|
|
"""
|
2015-08-12 05:13:35 -04:00
|
|
|
self.check_thread()
|
|
|
|
if self.sequence == sequence:
|
|
|
|
# Only update the cache if the caches sequence number matches the
|
|
|
|
# number that the cache had before the SELECT was started (SYN-369)
|
2018-06-11 18:13:06 -04:00
|
|
|
if fetched_keys is None:
|
|
|
|
self._insert(key, value, set())
|
2015-08-12 05:13:35 -04:00
|
|
|
else:
|
2018-06-11 18:13:06 -04:00
|
|
|
self._update_or_insert(key, value, fetched_keys)
|
2015-08-04 10:56:56 -04:00
|
|
|
|
2021-03-29 12:15:33 -04:00
|
|
|
def _update_or_insert(
|
|
|
|
self, key: KT, value: Dict[DKT, Any], known_absent: Set[DKT]
|
|
|
|
) -> None:
|
2018-03-15 11:40:13 -04:00
|
|
|
# We pop and reinsert as we need to tell the cache the size may have
|
|
|
|
# changed
|
|
|
|
|
|
|
|
entry = self.cache.pop(key, DictionaryEntry(False, set(), {}))
|
2015-08-04 10:56:56 -04:00
|
|
|
entry.value.update(value)
|
2017-05-17 09:31:23 -04:00
|
|
|
entry.known_absent.update(known_absent)
|
2018-03-15 11:40:13 -04:00
|
|
|
self.cache[key] = entry
|
2015-08-04 10:56:56 -04:00
|
|
|
|
2021-03-29 12:15:33 -04:00
|
|
|
def _insert(self, key: KT, value: Dict[DKT, Any], known_absent: Set[DKT]) -> None:
|
2017-05-17 09:31:23 -04:00
|
|
|
self.cache[key] = DictionaryEntry(True, known_absent, value)
|