mirror of
https://mau.dev/maunium/synapse.git
synced 2024-10-01 01:36:05 -04:00
116 lines
3.6 KiB
Python
116 lines
3.6 KiB
Python
# Copyright 2019 New Vector Ltd
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
|
|
import logging
|
|
from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple
|
|
|
|
import attr
|
|
|
|
from synapse.api.errors import SynapseError
|
|
from synapse.types import JsonDict
|
|
|
|
if TYPE_CHECKING:
|
|
from synapse.storage.databases.main import DataStore
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
@attr.s(slots=True, auto_attribs=True)
|
|
class PaginationChunk:
|
|
"""Returned by relation pagination APIs.
|
|
|
|
Attributes:
|
|
chunk: The rows returned by pagination
|
|
next_batch: Token to fetch next set of results with, if
|
|
None then there are no more results.
|
|
prev_batch: Token to fetch previous set of results with, if
|
|
None then there are no previous results.
|
|
"""
|
|
|
|
chunk: List[JsonDict]
|
|
next_batch: Optional[Any] = None
|
|
prev_batch: Optional[Any] = None
|
|
|
|
async def to_dict(self, store: "DataStore") -> Dict[str, Any]:
|
|
d = {"chunk": self.chunk}
|
|
|
|
if self.next_batch:
|
|
d["next_batch"] = await self.next_batch.to_string(store)
|
|
|
|
if self.prev_batch:
|
|
d["prev_batch"] = await self.prev_batch.to_string(store)
|
|
|
|
return d
|
|
|
|
|
|
@attr.s(frozen=True, slots=True, auto_attribs=True)
|
|
class RelationPaginationToken:
|
|
"""Pagination token for relation pagination API.
|
|
|
|
As the results are in topological order, we can use the
|
|
`topological_ordering` and `stream_ordering` fields of the events at the
|
|
boundaries of the chunk as pagination tokens.
|
|
|
|
Attributes:
|
|
topological: The topological ordering of the boundary event
|
|
stream: The stream ordering of the boundary event.
|
|
"""
|
|
|
|
topological: int
|
|
stream: int
|
|
|
|
@staticmethod
|
|
def from_string(string: str) -> "RelationPaginationToken":
|
|
try:
|
|
t, s = string.split("-")
|
|
return RelationPaginationToken(int(t), int(s))
|
|
except ValueError:
|
|
raise SynapseError(400, "Invalid relation pagination token")
|
|
|
|
async def to_string(self, store: "DataStore") -> str:
|
|
return "%d-%d" % (self.topological, self.stream)
|
|
|
|
def as_tuple(self) -> Tuple[Any, ...]:
|
|
return attr.astuple(self)
|
|
|
|
|
|
@attr.s(frozen=True, slots=True, auto_attribs=True)
|
|
class AggregationPaginationToken:
|
|
"""Pagination token for relation aggregation pagination API.
|
|
|
|
As the results are order by count and then MAX(stream_ordering) of the
|
|
aggregation groups, we can just use them as our pagination token.
|
|
|
|
Attributes:
|
|
count: The count of relations in the boundary group.
|
|
stream: The MAX stream ordering in the boundary group.
|
|
"""
|
|
|
|
count: int
|
|
stream: int
|
|
|
|
@staticmethod
|
|
def from_string(string: str) -> "AggregationPaginationToken":
|
|
try:
|
|
c, s = string.split("-")
|
|
return AggregationPaginationToken(int(c), int(s))
|
|
except ValueError:
|
|
raise SynapseError(400, "Invalid aggregation pagination token")
|
|
|
|
async def to_string(self, store: "DataStore") -> str:
|
|
return "%d-%d" % (self.count, self.stream)
|
|
|
|
def as_tuple(self) -> Tuple[Any, ...]:
|
|
return attr.astuple(self)
|