2015-01-23 13:31:29 -05:00
|
|
|
#
|
2023-11-21 15:29:58 -05:00
|
|
|
# This file is licensed under the Affero General Public License (AGPL) version 3.
|
|
|
|
#
|
2024-01-23 06:26:48 -05:00
|
|
|
# Copyright 2015, 2016 OpenMarket Ltd
|
2023-11-21 15:29:58 -05:00
|
|
|
# Copyright (C) 2023 New Vector, Ltd
|
|
|
|
#
|
|
|
|
# This program is free software: you can redistribute it and/or modify
|
|
|
|
# it under the terms of the GNU Affero General Public License as
|
|
|
|
# published by the Free Software Foundation, either version 3 of the
|
|
|
|
# License, or (at your option) any later version.
|
|
|
|
#
|
|
|
|
# See the GNU Affero General Public License for more details:
|
|
|
|
# <https://www.gnu.org/licenses/agpl-3.0.html>.
|
|
|
|
#
|
|
|
|
# Originally licensed under the Apache License, Version 2.0:
|
|
|
|
# <http://www.apache.org/licenses/LICENSE-2.0>.
|
|
|
|
#
|
|
|
|
# [This file includes modifications made by New Vector Limited]
|
2015-01-23 13:31:29 -05:00
|
|
|
#
|
|
|
|
#
|
2018-07-09 02:09:20 -04:00
|
|
|
import itertools
|
|
|
|
import logging
|
2021-06-23 10:57:41 -04:00
|
|
|
from collections import defaultdict
|
2022-03-03 10:43:06 -05:00
|
|
|
from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union
|
2018-07-09 02:09:20 -04:00
|
|
|
|
2023-02-28 12:11:26 -05:00
|
|
|
from synapse.api.constants import AccountDataTypes, EduTypes, Membership, PresenceState
|
2019-10-10 07:59:55 -04:00
|
|
|
from synapse.api.errors import Codes, StoreError, SynapseError
|
2021-11-09 08:10:58 -05:00
|
|
|
from synapse.api.filtering import FilterCollection
|
2021-08-23 08:14:42 -04:00
|
|
|
from synapse.api.presence import UserPresenceState
|
2018-07-09 02:09:20 -04:00
|
|
|
from synapse.events.utils import (
|
2022-03-03 10:43:06 -05:00
|
|
|
SerializeEventConfig,
|
2018-07-09 02:09:20 -04:00
|
|
|
format_event_for_client_v2_without_room_id,
|
2018-09-04 10:18:25 -04:00
|
|
|
format_event_raw,
|
2015-04-21 11:35:53 -04:00
|
|
|
)
|
2017-03-15 10:27:34 -04:00
|
|
|
from synapse.handlers.presence import format_user_presence_state
|
Add Sliding Sync `/sync` endpoint (initial implementation) (#17187)
Based on [MSC3575](https://github.com/matrix-org/matrix-spec-proposals/pull/3575): Sliding Sync
This iteration only focuses on returning the list of room IDs in the sliding window API (without sorting/filtering).
Rooms appear in the Sliding sync response based on:
- `invite`, `join`, `knock`, `ban` membership events
- Kicks (`leave` membership events where `sender` is different from the `user_id`/`state_key`)
- `newly_left` (rooms that were left during the given token range, > `from_token` and <= `to_token`)
- In order for bans/kicks to not show up, you need to `/forget` those rooms. This doesn't modify the event itself though and only adds the `forgotten` flag to `room_memberships` in Synapse. There isn't a way to tell when a room was forgotten at the moment so we can't factor it into the from/to range.
### Example request
`POST http://localhost:8008/_matrix/client/unstable/org.matrix.msc3575/sync`
```json
{
"lists": {
"foo-list": {
"ranges": [ [0, 99] ],
"sort": [ "by_notification_level", "by_recency", "by_name" ],
"required_state": [
["m.room.join_rules", ""],
["m.room.history_visibility", ""],
["m.space.child", "*"]
],
"timeline_limit": 100
}
}
}
```
Response:
```json
{
"next_pos": "s58_224_0_13_10_1_1_16_0_1",
"lists": {
"foo-list": {
"count": 1,
"ops": [
{
"op": "SYNC",
"range": [0, 99],
"room_ids": [
"!MmgikIyFzsuvtnbvVG:my.synapse.linux.server"
]
}
]
}
},
"rooms": {},
"extensions": {}
}
```
2024-06-06 15:44:32 -04:00
|
|
|
from synapse.handlers.sliding_sync import SlidingSyncConfig, SlidingSyncResult
|
2021-08-23 08:14:42 -04:00
|
|
|
from synapse.handlers.sync import (
|
|
|
|
ArchivedSyncResult,
|
|
|
|
InvitedSyncResult,
|
|
|
|
JoinedSyncResult,
|
|
|
|
KnockedSyncResult,
|
|
|
|
SyncConfig,
|
|
|
|
SyncResult,
|
2024-05-16 06:36:54 -04:00
|
|
|
SyncVersion,
|
2021-08-23 08:14:42 -04:00
|
|
|
)
|
|
|
|
from synapse.http.server import HttpServer
|
Add Sliding Sync `/sync` endpoint (initial implementation) (#17187)
Based on [MSC3575](https://github.com/matrix-org/matrix-spec-proposals/pull/3575): Sliding Sync
This iteration only focuses on returning the list of room IDs in the sliding window API (without sorting/filtering).
Rooms appear in the Sliding sync response based on:
- `invite`, `join`, `knock`, `ban` membership events
- Kicks (`leave` membership events where `sender` is different from the `user_id`/`state_key`)
- `newly_left` (rooms that were left during the given token range, > `from_token` and <= `to_token`)
- In order for bans/kicks to not show up, you need to `/forget` those rooms. This doesn't modify the event itself though and only adds the `forgotten` flag to `room_memberships` in Synapse. There isn't a way to tell when a room was forgotten at the moment so we can't factor it into the from/to range.
### Example request
`POST http://localhost:8008/_matrix/client/unstable/org.matrix.msc3575/sync`
```json
{
"lists": {
"foo-list": {
"ranges": [ [0, 99] ],
"sort": [ "by_notification_level", "by_recency", "by_name" ],
"required_state": [
["m.room.join_rules", ""],
["m.room.history_visibility", ""],
["m.space.child", "*"]
],
"timeline_limit": 100
}
}
}
```
Response:
```json
{
"next_pos": "s58_224_0_13_10_1_1_16_0_1",
"lists": {
"foo-list": {
"count": 1,
"ops": [
{
"op": "SYNC",
"range": [0, 99],
"room_ids": [
"!MmgikIyFzsuvtnbvVG:my.synapse.linux.server"
]
}
]
}
},
"rooms": {},
"extensions": {}
}
```
2024-06-06 15:44:32 -04:00
|
|
|
from synapse.http.servlet import (
|
|
|
|
RestServlet,
|
|
|
|
parse_and_validate_json_object_from_request,
|
|
|
|
parse_boolean,
|
|
|
|
parse_integer,
|
|
|
|
parse_string,
|
|
|
|
)
|
2021-03-24 06:48:46 -04:00
|
|
|
from synapse.http.site import SynapseRequest
|
2022-07-19 14:14:30 -04:00
|
|
|
from synapse.logging.opentracing import trace_with_opname
|
2023-03-06 11:08:39 -05:00
|
|
|
from synapse.types import JsonDict, Requester, StreamToken
|
2024-06-10 16:03:50 -04:00
|
|
|
from synapse.types.rest.client import SlidingSyncBody
|
2020-08-19 07:26:03 -04:00
|
|
|
from synapse.util import json_decoder
|
2024-05-14 10:08:46 -04:00
|
|
|
from synapse.util.caches.lrucache import LruCache
|
2015-01-23 13:31:29 -05:00
|
|
|
|
2019-06-03 07:28:59 -04:00
|
|
|
from ._base import client_patterns, set_timeline_upper_limit
|
2015-12-09 07:56:50 -05:00
|
|
|
|
2021-03-24 06:48:46 -04:00
|
|
|
if TYPE_CHECKING:
|
|
|
|
from synapse.server import HomeServer
|
|
|
|
|
2015-01-23 13:31:29 -05:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
|
|
|
class SyncRestServlet(RestServlet):
|
|
|
|
"""
|
|
|
|
|
|
|
|
GET parameters::
|
|
|
|
timeout(int): How long to wait for new events in milliseconds.
|
|
|
|
since(batch_token): Batch token when asking for incremental deltas.
|
|
|
|
set_presence(str): What state the device presence should be set to.
|
|
|
|
default is "online".
|
|
|
|
filter(filter_id): A filter to apply to the events returned.
|
|
|
|
|
|
|
|
Response JSON::
|
|
|
|
{
|
2015-10-01 12:53:07 -04:00
|
|
|
"next_batch": // batch token for the next /sync
|
|
|
|
"presence": // presence data for the user.
|
2015-10-07 10:55:20 -04:00
|
|
|
"rooms": {
|
2015-12-09 07:56:50 -05:00
|
|
|
"join": { // Joined rooms being updated.
|
2015-10-07 10:55:20 -04:00
|
|
|
"${room_id}": { // Id of the room being updated
|
|
|
|
"event_map": // Map of EventID -> event JSON.
|
|
|
|
"timeline": { // The recent events in the room if gap is "true"
|
2015-10-01 12:53:07 -04:00
|
|
|
"limited": // Was the per-room event limit exceeded?
|
2015-10-07 10:55:20 -04:00
|
|
|
// otherwise the next events in the room.
|
|
|
|
"events": [] // list of EventIDs in the "event_map".
|
2015-10-01 12:53:07 -04:00
|
|
|
"prev_batch": // back token for getting previous events.
|
2015-10-07 10:55:20 -04:00
|
|
|
}
|
|
|
|
"state": {"events": []} // list of EventIDs updating the
|
|
|
|
// current state to be what it should
|
|
|
|
// be at the end of the batch.
|
|
|
|
"ephemeral": {"events": []} // list of event objects
|
2015-10-01 12:53:07 -04:00
|
|
|
}
|
2015-10-07 10:55:20 -04:00
|
|
|
},
|
2015-12-09 07:56:50 -05:00
|
|
|
"invite": {}, // Invited rooms being updated.
|
|
|
|
"leave": {} // Archived rooms being updated.
|
2015-10-01 12:53:07 -04:00
|
|
|
}
|
2015-01-23 13:31:29 -05:00
|
|
|
}
|
|
|
|
"""
|
|
|
|
|
2019-06-03 07:28:59 -04:00
|
|
|
PATTERNS = client_patterns("/sync$")
|
2020-02-21 07:15:07 -05:00
|
|
|
ALLOWED_PRESENCE = {"online", "offline", "unavailable"}
|
2023-03-23 08:11:14 -04:00
|
|
|
CATEGORY = "Sync requests"
|
2015-01-23 13:31:29 -05:00
|
|
|
|
2021-03-24 06:48:46 -04:00
|
|
|
def __init__(self, hs: "HomeServer"):
|
2020-09-18 09:56:44 -04:00
|
|
|
super().__init__()
|
2017-05-15 08:51:43 -04:00
|
|
|
self.hs = hs
|
2015-01-23 13:31:29 -05:00
|
|
|
self.auth = hs.get_auth()
|
2022-02-23 06:04:02 -05:00
|
|
|
self.store = hs.get_datastores().main
|
2016-05-16 15:19:26 -04:00
|
|
|
self.sync_handler = hs.get_sync_handler()
|
2015-01-26 13:53:31 -05:00
|
|
|
self.clock = hs.get_clock()
|
2015-01-29 13:11:28 -05:00
|
|
|
self.filtering = hs.get_filtering()
|
2016-05-16 13:56:37 -04:00
|
|
|
self.presence_handler = hs.get_presence_handler()
|
2018-05-22 05:57:56 -04:00
|
|
|
self._server_notices_sender = hs.get_server_notices_sender()
|
2019-05-09 08:21:57 -04:00
|
|
|
self._event_serializer = hs.get_event_client_serializer()
|
2022-03-31 15:05:13 -04:00
|
|
|
self._msc2654_enabled = hs.config.experimental.msc2654_enabled
|
2022-10-07 09:26:40 -04:00
|
|
|
self._msc3773_enabled = hs.config.experimental.msc3773_enabled
|
2015-01-23 13:31:29 -05:00
|
|
|
|
2024-05-14 10:08:46 -04:00
|
|
|
self._json_filter_cache: LruCache[str, bool] = LruCache(
|
|
|
|
max_size=1000,
|
|
|
|
cache_name="sync_valid_filter",
|
|
|
|
)
|
|
|
|
|
2021-03-24 06:48:46 -04:00
|
|
|
async def on_GET(self, request: SynapseRequest) -> Tuple[int, JsonDict]:
|
2021-03-26 12:49:46 -04:00
|
|
|
# This will always be set by the time Twisted calls us.
|
|
|
|
assert request.args is not None
|
|
|
|
|
2018-09-12 06:41:31 -04:00
|
|
|
if b"from" in request.args:
|
2016-01-20 10:42:57 -05:00
|
|
|
# /events used to use 'from', but /sync uses 'since'.
|
|
|
|
# Lets be helpful and whine if we see a 'from'.
|
|
|
|
raise SynapseError(
|
|
|
|
400, "'from' is not a valid query parameter. Did you mean 'since'?"
|
|
|
|
)
|
|
|
|
|
2019-12-05 11:46:37 -05:00
|
|
|
requester = await self.auth.get_user_by_req(request, allow_guest=True)
|
2016-01-11 10:29:57 -05:00
|
|
|
user = requester.user
|
2016-08-25 12:35:37 -04:00
|
|
|
device_id = requester.device_id
|
2015-01-23 13:31:29 -05:00
|
|
|
|
2015-04-21 11:35:53 -04:00
|
|
|
timeout = parse_integer(request, "timeout", default=0)
|
|
|
|
since = parse_string(request, "since")
|
|
|
|
set_presence = parse_string(
|
2019-06-20 05:32:02 -04:00
|
|
|
request,
|
|
|
|
"set_presence",
|
|
|
|
default="online",
|
|
|
|
allowed_values=self.ALLOWED_PRESENCE,
|
2015-01-23 13:31:29 -05:00
|
|
|
)
|
2021-07-21 09:47:56 -04:00
|
|
|
filter_id = parse_string(request, "filter")
|
2015-10-26 14:47:18 -04:00
|
|
|
full_state = parse_boolean(request, "full_state", default=False)
|
2015-01-23 13:31:29 -05:00
|
|
|
|
2017-08-02 12:29:51 -04:00
|
|
|
logger.debug(
|
2019-10-24 13:31:53 -04:00
|
|
|
"/sync: user=%r, timeout=%r, since=%r, "
|
|
|
|
"set_presence=%r, filter_id=%r, device_id=%r",
|
2019-10-24 13:43:13 -04:00
|
|
|
user,
|
|
|
|
timeout,
|
|
|
|
since,
|
|
|
|
set_presence,
|
|
|
|
filter_id,
|
|
|
|
device_id,
|
2015-01-23 13:31:29 -05:00
|
|
|
)
|
|
|
|
|
2023-02-28 12:11:26 -05:00
|
|
|
# Stream position of the last ignored users account data event for this user,
|
|
|
|
# if we're initial syncing.
|
|
|
|
# We include this in the request key to invalidate an initial sync
|
|
|
|
# in the response cache once the set of ignored users has changed.
|
|
|
|
# (We filter out ignored users from timeline events, so our sync response
|
|
|
|
# is invalid once the set of ignored users changes.)
|
|
|
|
last_ignore_accdata_streampos: Optional[int] = None
|
|
|
|
if not since:
|
|
|
|
# No `since`, so this is an initial sync.
|
|
|
|
last_ignore_accdata_streampos = await self.store.get_latest_stream_id_for_global_account_data_by_type_for_user(
|
|
|
|
user.to_string(), AccountDataTypes.IGNORED_USER_LIST
|
|
|
|
)
|
|
|
|
|
|
|
|
request_key = (
|
|
|
|
user,
|
|
|
|
timeout,
|
|
|
|
since,
|
|
|
|
filter_id,
|
|
|
|
full_state,
|
|
|
|
device_id,
|
|
|
|
last_ignore_accdata_streampos,
|
|
|
|
)
|
2016-03-24 13:47:31 -04:00
|
|
|
|
2019-10-10 07:59:55 -04:00
|
|
|
if filter_id is None:
|
2021-11-09 08:10:58 -05:00
|
|
|
filter_collection = self.filtering.DEFAULT_FILTER_COLLECTION
|
2019-10-10 07:59:55 -04:00
|
|
|
elif filter_id.startswith("{"):
|
|
|
|
try:
|
2020-08-19 07:26:03 -04:00
|
|
|
filter_object = json_decoder.decode(filter_id)
|
2019-10-10 07:59:55 -04:00
|
|
|
except Exception:
|
2022-10-24 11:55:06 -04:00
|
|
|
raise SynapseError(400, "Invalid filter JSON", errcode=Codes.NOT_JSON)
|
2024-05-14 10:08:46 -04:00
|
|
|
|
|
|
|
# We cache the validation, as this can get quite expensive if people use
|
|
|
|
# a literal json blob as a query param.
|
|
|
|
if not self._json_filter_cache.get(filter_id):
|
|
|
|
self.filtering.check_valid_filter(filter_object)
|
|
|
|
self._json_filter_cache[filter_id] = True
|
|
|
|
|
2022-10-24 11:55:06 -04:00
|
|
|
set_timeline_upper_limit(
|
|
|
|
filter_object, self.hs.config.server.filter_timeline_limit
|
|
|
|
)
|
2021-11-09 08:10:58 -05:00
|
|
|
filter_collection = FilterCollection(self.hs, filter_object)
|
2016-01-22 05:41:30 -05:00
|
|
|
else:
|
2019-10-10 07:59:55 -04:00
|
|
|
try:
|
2019-12-05 11:46:37 -05:00
|
|
|
filter_collection = await self.filtering.get_user_filter(
|
2023-06-02 20:24:13 -04:00
|
|
|
user, filter_id
|
2019-10-10 07:59:55 -04:00
|
|
|
)
|
|
|
|
except StoreError as err:
|
|
|
|
if err.code != 404:
|
|
|
|
raise
|
|
|
|
# fix up the description and errcode to be more useful
|
|
|
|
raise SynapseError(400, "No such filter", errcode=Codes.INVALID_PARAM)
|
2015-01-23 13:31:29 -05:00
|
|
|
|
2015-01-26 13:53:31 -05:00
|
|
|
sync_config = SyncConfig(
|
|
|
|
user=user,
|
2019-10-10 07:59:55 -04:00
|
|
|
filter_collection=filter_collection,
|
2016-01-11 10:29:57 -05:00
|
|
|
is_guest=requester.is_guest,
|
2016-08-25 12:35:37 -04:00
|
|
|
device_id=device_id,
|
2015-01-26 13:53:31 -05:00
|
|
|
)
|
|
|
|
|
2020-09-30 15:29:19 -04:00
|
|
|
since_token = None
|
2015-01-26 13:53:31 -05:00
|
|
|
if since is not None:
|
2020-09-30 15:29:19 -04:00
|
|
|
since_token = await StreamToken.from_string(self.store, since)
|
2015-01-23 13:31:29 -05:00
|
|
|
|
2018-05-22 05:57:56 -04:00
|
|
|
# send any outstanding server notices to the user.
|
2019-12-05 11:46:37 -05:00
|
|
|
await self._server_notices_sender.on_user_syncing(user.to_string())
|
2018-05-22 05:57:56 -04:00
|
|
|
|
2016-02-15 12:10:40 -05:00
|
|
|
affect_presence = set_presence != PresenceState.OFFLINE
|
2015-10-09 14:57:50 -04:00
|
|
|
|
2019-12-05 11:46:37 -05:00
|
|
|
context = await self.presence_handler.user_syncing(
|
2022-04-13 11:21:07 -04:00
|
|
|
user.to_string(),
|
2023-08-28 13:08:49 -04:00
|
|
|
requester.device_id,
|
2022-04-13 11:21:07 -04:00
|
|
|
affect_presence=affect_presence,
|
|
|
|
presence_state=set_presence,
|
2016-02-15 12:10:40 -05:00
|
|
|
)
|
|
|
|
with context:
|
2019-12-05 11:46:37 -05:00
|
|
|
sync_result = await self.sync_handler.wait_for_sync_for_user(
|
2020-11-17 05:51:25 -05:00
|
|
|
requester,
|
2019-06-20 05:32:02 -04:00
|
|
|
sync_config,
|
2024-05-16 06:36:54 -04:00
|
|
|
SyncVersion.SYNC_V2,
|
2024-05-16 12:54:46 -04:00
|
|
|
request_key,
|
2019-06-20 05:32:02 -04:00
|
|
|
since_token=since_token,
|
|
|
|
timeout=timeout,
|
|
|
|
full_state=full_state,
|
2015-10-09 14:57:50 -04:00
|
|
|
)
|
2015-01-23 13:31:29 -05:00
|
|
|
|
2020-07-22 08:44:16 -04:00
|
|
|
# the client may have disconnected by now; don't bother to serialize the
|
|
|
|
# response if so.
|
|
|
|
if request._disconnected:
|
|
|
|
logger.info("Client has disconnected; not serializing response.")
|
|
|
|
return 200, {}
|
|
|
|
|
2015-01-26 13:53:31 -05:00
|
|
|
time_now = self.clock.time_msec()
|
2021-08-23 08:14:42 -04:00
|
|
|
# We know that the the requester has an access token since appservices
|
|
|
|
# cannot use sync.
|
2019-12-05 11:46:37 -05:00
|
|
|
response_content = await self.encode_response(
|
2023-03-06 11:08:39 -05:00
|
|
|
time_now, sync_result, requester, filter_collection
|
2017-07-10 11:34:58 -04:00
|
|
|
)
|
2015-01-26 13:53:31 -05:00
|
|
|
|
2020-07-22 08:43:10 -04:00
|
|
|
logger.debug("Event formatting complete")
|
2019-08-30 11:28:26 -04:00
|
|
|
return 200, response_content
|
2015-10-19 12:26:18 -04:00
|
|
|
|
2022-07-19 14:14:30 -04:00
|
|
|
@trace_with_opname("sync.encode_response")
|
2021-08-23 08:14:42 -04:00
|
|
|
async def encode_response(
|
|
|
|
self,
|
|
|
|
time_now: int,
|
|
|
|
sync_result: SyncResult,
|
2023-03-06 11:08:39 -05:00
|
|
|
requester: Requester,
|
2021-08-23 08:14:42 -04:00
|
|
|
filter: FilterCollection,
|
|
|
|
) -> JsonDict:
|
2020-07-22 08:43:10 -04:00
|
|
|
logger.debug("Formatting events in sync response")
|
2019-06-20 05:32:02 -04:00
|
|
|
if filter.event_format == "client":
|
2018-09-04 10:18:25 -04:00
|
|
|
event_formatter = format_event_for_client_v2_without_room_id
|
2019-06-20 05:32:02 -04:00
|
|
|
elif filter.event_format == "federation":
|
2018-09-04 10:18:25 -04:00
|
|
|
event_formatter = format_event_raw
|
|
|
|
else:
|
2019-06-20 05:32:02 -04:00
|
|
|
raise Exception("Unknown event format %s" % (filter.event_format,))
|
2018-09-04 10:18:25 -04:00
|
|
|
|
2022-03-03 10:43:06 -05:00
|
|
|
serialize_options = SerializeEventConfig(
|
|
|
|
event_format=event_formatter,
|
2023-03-06 11:08:39 -05:00
|
|
|
requester=requester,
|
2022-03-03 10:43:06 -05:00
|
|
|
only_event_fields=filter.event_fields,
|
|
|
|
)
|
|
|
|
stripped_serialize_options = SerializeEventConfig(
|
|
|
|
event_format=event_formatter,
|
2023-03-06 11:08:39 -05:00
|
|
|
requester=requester,
|
2022-03-03 10:43:06 -05:00
|
|
|
include_stripped_room_state=True,
|
|
|
|
)
|
|
|
|
|
2019-12-05 11:46:37 -05:00
|
|
|
joined = await self.encode_joined(
|
2022-03-03 10:43:06 -05:00
|
|
|
sync_result.joined, time_now, serialize_options
|
2017-07-11 07:14:35 -04:00
|
|
|
)
|
2017-07-10 10:42:17 -04:00
|
|
|
|
2019-12-05 11:46:37 -05:00
|
|
|
invited = await self.encode_invited(
|
2022-03-03 10:43:06 -05:00
|
|
|
sync_result.invited, time_now, stripped_serialize_options
|
2017-07-11 07:14:35 -04:00
|
|
|
)
|
2017-07-10 10:42:17 -04:00
|
|
|
|
2021-06-09 14:39:51 -04:00
|
|
|
knocked = await self.encode_knocked(
|
2022-03-03 10:43:06 -05:00
|
|
|
sync_result.knocked, time_now, stripped_serialize_options
|
2021-06-09 14:39:51 -04:00
|
|
|
)
|
|
|
|
|
2019-12-05 11:46:37 -05:00
|
|
|
archived = await self.encode_archived(
|
2022-03-03 10:43:06 -05:00
|
|
|
sync_result.archived, time_now, serialize_options
|
2017-07-11 07:14:35 -04:00
|
|
|
)
|
2017-07-10 10:42:17 -04:00
|
|
|
|
2020-07-22 08:43:10 -04:00
|
|
|
logger.debug("building sync response dict")
|
2021-06-23 10:57:41 -04:00
|
|
|
|
2021-08-23 08:14:42 -04:00
|
|
|
response: JsonDict = defaultdict(dict)
|
2021-06-23 10:57:41 -04:00
|
|
|
response["next_batch"] = await sync_result.next_batch.to_string(self.store)
|
|
|
|
|
|
|
|
if sync_result.account_data:
|
|
|
|
response["account_data"] = {"events": sync_result.account_data}
|
|
|
|
if sync_result.presence:
|
|
|
|
response["presence"] = SyncRestServlet.encode_presence(
|
|
|
|
sync_result.presence, time_now
|
|
|
|
)
|
|
|
|
|
|
|
|
if sync_result.to_device:
|
|
|
|
response["to_device"] = {"events": sync_result.to_device}
|
|
|
|
|
|
|
|
if sync_result.device_lists.changed:
|
|
|
|
response["device_lists"]["changed"] = list(sync_result.device_lists.changed)
|
|
|
|
if sync_result.device_lists.left:
|
|
|
|
response["device_lists"]["left"] = list(sync_result.device_lists.left)
|
|
|
|
|
2021-07-22 10:29:27 -04:00
|
|
|
# We always include this because https://github.com/vector-im/element-android/issues/3725
|
|
|
|
# The spec isn't terribly clear on when this can be omitted and how a client would tell
|
|
|
|
# the difference between "no keys present" and "nothing changed" in terms of whole field
|
|
|
|
# absent / individual key type entry absent
|
|
|
|
# Corresponding synapse issue: https://github.com/matrix-org/synapse/issues/10456
|
|
|
|
response["device_one_time_keys_count"] = sync_result.device_one_time_keys_count
|
|
|
|
|
2021-08-17 07:32:25 -04:00
|
|
|
# https://github.com/matrix-org/matrix-doc/blob/54255851f642f84a4f1aaf7bc063eebe3d76752b/proposals/2732-olm-fallback-keys.md
|
|
|
|
# states that this field should always be included, as long as the server supports the feature.
|
2024-03-13 12:46:44 -04:00
|
|
|
response["org.matrix.msc2732.device_unused_fallback_key_types"] = (
|
|
|
|
sync_result.device_unused_fallback_key_types
|
|
|
|
)
|
|
|
|
response["device_unused_fallback_key_types"] = (
|
|
|
|
sync_result.device_unused_fallback_key_types
|
|
|
|
)
|
2021-06-23 10:57:41 -04:00
|
|
|
|
|
|
|
if joined:
|
|
|
|
response["rooms"][Membership.JOIN] = joined
|
|
|
|
if invited:
|
|
|
|
response["rooms"][Membership.INVITE] = invited
|
|
|
|
if knocked:
|
|
|
|
response["rooms"][Membership.KNOCK] = knocked
|
|
|
|
if archived:
|
|
|
|
response["rooms"][Membership.LEAVE] = archived
|
|
|
|
|
|
|
|
return response
|
2017-07-10 10:42:17 -04:00
|
|
|
|
|
|
|
@staticmethod
|
2021-08-23 08:14:42 -04:00
|
|
|
def encode_presence(events: List[UserPresenceState], time_now: int) -> JsonDict:
|
2017-03-15 10:27:34 -04:00
|
|
|
return {
|
|
|
|
"events": [
|
|
|
|
{
|
2022-05-27 07:14:36 -04:00
|
|
|
"type": EduTypes.PRESENCE,
|
2017-03-15 10:27:34 -04:00
|
|
|
"sender": event.user_id,
|
|
|
|
"content": format_user_presence_state(
|
|
|
|
event, time_now, include_user_id=False
|
|
|
|
),
|
|
|
|
}
|
|
|
|
for event in events
|
|
|
|
]
|
|
|
|
}
|
2015-12-01 13:41:32 -05:00
|
|
|
|
2022-07-19 14:14:30 -04:00
|
|
|
@trace_with_opname("sync.encode_joined")
|
2019-12-05 11:46:37 -05:00
|
|
|
async def encode_joined(
|
2021-08-23 08:14:42 -04:00
|
|
|
self,
|
|
|
|
rooms: List[JoinedSyncResult],
|
|
|
|
time_now: int,
|
2022-03-03 10:43:06 -05:00
|
|
|
serialize_options: SerializeEventConfig,
|
2021-08-23 08:14:42 -04:00
|
|
|
) -> JsonDict:
|
2015-11-13 05:31:15 -05:00
|
|
|
"""
|
|
|
|
Encode the joined rooms in a sync result
|
|
|
|
|
2016-04-01 11:08:59 -04:00
|
|
|
Args:
|
2021-08-23 08:14:42 -04:00
|
|
|
rooms: list of sync results for rooms this user is joined to
|
|
|
|
time_now: current time - used as a baseline for age calculations
|
2022-03-03 10:43:06 -05:00
|
|
|
serialize_options: Event serializer options
|
2016-04-01 11:08:59 -04:00
|
|
|
Returns:
|
2021-08-23 08:14:42 -04:00
|
|
|
The joined rooms list, in our response format
|
2015-11-13 05:31:15 -05:00
|
|
|
"""
|
2015-10-07 10:55:20 -04:00
|
|
|
joined = {}
|
2015-10-05 11:39:22 -04:00
|
|
|
for room in rooms:
|
2019-12-05 11:46:37 -05:00
|
|
|
joined[room.room_id] = await self.encode_room(
|
2022-03-03 10:43:06 -05:00
|
|
|
room, time_now, joined=True, serialize_options=serialize_options
|
2015-10-05 11:39:22 -04:00
|
|
|
)
|
|
|
|
|
2019-07-23 09:00:55 -04:00
|
|
|
return joined
|
2015-01-26 13:53:31 -05:00
|
|
|
|
2022-07-19 14:14:30 -04:00
|
|
|
@trace_with_opname("sync.encode_invited")
|
2021-08-23 08:14:42 -04:00
|
|
|
async def encode_invited(
|
|
|
|
self,
|
|
|
|
rooms: List[InvitedSyncResult],
|
|
|
|
time_now: int,
|
2022-03-03 10:43:06 -05:00
|
|
|
serialize_options: SerializeEventConfig,
|
2021-08-23 08:14:42 -04:00
|
|
|
) -> JsonDict:
|
2015-11-13 05:31:15 -05:00
|
|
|
"""
|
|
|
|
Encode the invited rooms in a sync result
|
|
|
|
|
2016-04-01 11:08:59 -04:00
|
|
|
Args:
|
2021-08-23 08:14:42 -04:00
|
|
|
rooms: list of sync results for rooms this user is invited to
|
|
|
|
time_now: current time - used as a baseline for age calculations
|
2022-03-03 10:43:06 -05:00
|
|
|
serialize_options: Event serializer options
|
2015-11-13 05:31:15 -05:00
|
|
|
|
2016-04-01 11:08:59 -04:00
|
|
|
Returns:
|
2021-08-23 08:14:42 -04:00
|
|
|
The invited rooms list, in our response format
|
2015-11-13 05:31:15 -05:00
|
|
|
"""
|
2015-10-13 06:03:48 -04:00
|
|
|
invited = {}
|
|
|
|
for room in rooms:
|
2023-10-27 05:04:08 -04:00
|
|
|
invite = await self._event_serializer.serialize_event(
|
2022-03-03 10:43:06 -05:00
|
|
|
room.invite, time_now, config=serialize_options
|
2015-10-13 06:03:48 -04:00
|
|
|
)
|
2016-01-25 05:10:44 -05:00
|
|
|
unsigned = dict(invite.get("unsigned", {}))
|
|
|
|
invite["unsigned"] = unsigned
|
|
|
|
invited_state = list(unsigned.pop("invite_room_state", []))
|
2015-10-13 06:03:48 -04:00
|
|
|
invited_state.append(invite)
|
2019-06-20 05:32:02 -04:00
|
|
|
invited[room.room_id] = {"invite_state": {"events": invited_state}}
|
2015-10-13 06:03:48 -04:00
|
|
|
|
2019-07-23 09:00:55 -04:00
|
|
|
return invited
|
2015-10-13 06:03:48 -04:00
|
|
|
|
2022-07-19 14:14:30 -04:00
|
|
|
@trace_with_opname("sync.encode_knocked")
|
2021-06-09 14:39:51 -04:00
|
|
|
async def encode_knocked(
|
|
|
|
self,
|
|
|
|
rooms: List[KnockedSyncResult],
|
|
|
|
time_now: int,
|
2022-03-03 10:43:06 -05:00
|
|
|
serialize_options: SerializeEventConfig,
|
2021-06-09 14:39:51 -04:00
|
|
|
) -> Dict[str, Dict[str, Any]]:
|
|
|
|
"""
|
|
|
|
Encode the rooms we've knocked on in a sync result.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
rooms: list of sync results for rooms this user is knocking on
|
|
|
|
time_now: current time - used as a baseline for age calculations
|
2022-03-03 10:43:06 -05:00
|
|
|
serialize_options: Event serializer options
|
2021-06-09 14:39:51 -04:00
|
|
|
|
|
|
|
Returns:
|
|
|
|
The list of rooms the user has knocked on, in our response format.
|
|
|
|
"""
|
|
|
|
knocked = {}
|
|
|
|
for room in rooms:
|
2023-10-27 05:04:08 -04:00
|
|
|
knock = await self._event_serializer.serialize_event(
|
2022-03-03 10:43:06 -05:00
|
|
|
room.knock, time_now, config=serialize_options
|
2021-06-09 14:39:51 -04:00
|
|
|
)
|
|
|
|
|
|
|
|
# Extract the `unsigned` key from the knock event.
|
|
|
|
# This is where we (cheekily) store the knock state events
|
|
|
|
unsigned = knock.setdefault("unsigned", {})
|
|
|
|
|
|
|
|
# Duplicate the dictionary in order to avoid modifying the original
|
|
|
|
unsigned = dict(unsigned)
|
|
|
|
|
|
|
|
# Extract the stripped room state from the unsigned dict
|
|
|
|
# This is for clients to get a little bit of information about
|
|
|
|
# the room they've knocked on, without revealing any sensitive information
|
|
|
|
knocked_state = list(unsigned.pop("knock_room_state", []))
|
|
|
|
|
|
|
|
# Append the actual knock membership event itself as well. This provides
|
|
|
|
# the client with:
|
|
|
|
#
|
|
|
|
# * A knock state event that they can use for easier internal tracking
|
|
|
|
# * The rough timestamp of when the knock occurred contained within the event
|
|
|
|
knocked_state.append(knock)
|
|
|
|
|
|
|
|
# Build the `knock_state` dictionary, which will contain the state of the
|
|
|
|
# room that the client has knocked on
|
|
|
|
knocked[room.room_id] = {"knock_state": {"events": knocked_state}}
|
|
|
|
|
|
|
|
return knocked
|
|
|
|
|
2022-07-19 14:14:30 -04:00
|
|
|
@trace_with_opname("sync.encode_archived")
|
2019-12-05 11:46:37 -05:00
|
|
|
async def encode_archived(
|
2021-08-23 08:14:42 -04:00
|
|
|
self,
|
|
|
|
rooms: List[ArchivedSyncResult],
|
|
|
|
time_now: int,
|
2022-03-03 10:43:06 -05:00
|
|
|
serialize_options: SerializeEventConfig,
|
2021-08-23 08:14:42 -04:00
|
|
|
) -> JsonDict:
|
2015-11-13 05:31:15 -05:00
|
|
|
"""
|
|
|
|
Encode the archived rooms in a sync result
|
|
|
|
|
2016-04-01 11:08:59 -04:00
|
|
|
Args:
|
2021-08-23 08:14:42 -04:00
|
|
|
rooms: list of sync results for rooms this user is joined to
|
|
|
|
time_now: current time - used as a baseline for age calculations
|
2022-03-03 10:43:06 -05:00
|
|
|
serialize_options: Event serializer options
|
2016-04-01 11:08:59 -04:00
|
|
|
Returns:
|
2021-08-23 08:14:42 -04:00
|
|
|
The archived rooms list, in our response format
|
2015-11-13 05:31:15 -05:00
|
|
|
"""
|
2015-10-19 12:26:18 -04:00
|
|
|
joined = {}
|
|
|
|
for room in rooms:
|
2019-12-05 11:46:37 -05:00
|
|
|
joined[room.room_id] = await self.encode_room(
|
2022-03-03 10:43:06 -05:00
|
|
|
room, time_now, joined=False, serialize_options=serialize_options
|
2015-10-19 12:26:18 -04:00
|
|
|
)
|
|
|
|
|
2019-07-23 09:00:55 -04:00
|
|
|
return joined
|
2015-10-19 12:26:18 -04:00
|
|
|
|
2019-12-05 11:46:37 -05:00
|
|
|
async def encode_room(
|
2021-08-23 08:14:42 -04:00
|
|
|
self,
|
|
|
|
room: Union[JoinedSyncResult, ArchivedSyncResult],
|
|
|
|
time_now: int,
|
|
|
|
joined: bool,
|
2022-03-03 10:43:06 -05:00
|
|
|
serialize_options: SerializeEventConfig,
|
2021-08-23 08:14:42 -04:00
|
|
|
) -> JsonDict:
|
2015-11-13 05:31:15 -05:00
|
|
|
"""
|
2016-04-01 11:08:59 -04:00
|
|
|
Args:
|
2021-08-23 08:14:42 -04:00
|
|
|
room: sync result for a single room
|
|
|
|
time_now: current time - used as a baseline for age calculations
|
|
|
|
token_id: ID of the user's auth token - used for namespacing
|
2016-04-01 11:08:59 -04:00
|
|
|
of transaction IDs
|
2021-08-23 08:14:42 -04:00
|
|
|
joined: True if the user is joined to this room - will mean
|
2016-04-01 11:08:59 -04:00
|
|
|
we handle ephemeral events
|
2021-08-23 08:14:42 -04:00
|
|
|
only_fields: Optional. The list of event fields to include.
|
|
|
|
event_formatter: function to convert from federation format
|
2018-09-04 10:18:25 -04:00
|
|
|
to client format
|
2016-04-01 11:08:59 -04:00
|
|
|
Returns:
|
2021-08-23 08:14:42 -04:00
|
|
|
The room, encoded in our response format
|
2015-11-13 05:31:15 -05:00
|
|
|
"""
|
2015-11-10 13:29:25 -05:00
|
|
|
state_dict = room.state
|
2016-01-25 05:10:44 -05:00
|
|
|
timeline_events = room.timeline.events
|
2015-11-10 13:29:25 -05:00
|
|
|
|
2016-01-25 05:10:44 -05:00
|
|
|
state_events = state_dict.values()
|
2015-01-26 13:53:31 -05:00
|
|
|
|
2016-02-11 04:22:37 -05:00
|
|
|
for event in itertools.chain(state_events, timeline_events):
|
|
|
|
# We've had bug reports that events were coming down under the
|
|
|
|
# wrong room.
|
|
|
|
if event.room_id != room.room_id:
|
2019-10-31 06:23:24 -04:00
|
|
|
logger.warning(
|
2016-02-11 04:22:37 -05:00
|
|
|
"Event %r is under room %r instead of %r",
|
2019-06-20 05:32:02 -04:00
|
|
|
event.event_id,
|
|
|
|
room.room_id,
|
|
|
|
event.room_id,
|
2016-02-11 04:22:37 -05:00
|
|
|
)
|
|
|
|
|
2023-10-27 05:04:08 -04:00
|
|
|
serialized_state = await self._event_serializer.serialize_events(
|
2022-03-03 10:43:06 -05:00
|
|
|
state_events, time_now, config=serialize_options
|
|
|
|
)
|
2023-10-27 05:04:08 -04:00
|
|
|
serialized_timeline = await self._event_serializer.serialize_events(
|
2022-03-03 10:43:06 -05:00
|
|
|
timeline_events,
|
|
|
|
time_now,
|
|
|
|
config=serialize_options,
|
|
|
|
bundle_aggregations=room.timeline.bundled_aggregations,
|
2022-01-13 10:45:28 -05:00
|
|
|
)
|
2015-10-19 12:26:18 -04:00
|
|
|
|
2016-01-25 05:10:44 -05:00
|
|
|
account_data = room.account_data
|
2015-11-02 11:23:15 -05:00
|
|
|
|
2021-08-23 08:14:42 -04:00
|
|
|
result: JsonDict = {
|
2015-10-08 10:17:43 -04:00
|
|
|
"timeline": {
|
2015-11-12 05:33:19 -05:00
|
|
|
"events": serialized_timeline,
|
2020-09-30 15:29:19 -04:00
|
|
|
"prev_batch": await room.timeline.prev_batch.to_string(self.store),
|
2015-10-08 10:17:43 -04:00
|
|
|
"limited": room.timeline.limited,
|
2015-01-26 13:53:31 -05:00
|
|
|
},
|
2015-11-19 07:23:42 -05:00
|
|
|
"state": {"events": serialized_state},
|
2015-11-18 10:31:04 -05:00
|
|
|
"account_data": {"events": account_data},
|
2015-01-26 13:53:31 -05:00
|
|
|
}
|
2015-10-19 12:26:18 -04:00
|
|
|
|
|
|
|
if joined:
|
2021-08-23 08:14:42 -04:00
|
|
|
assert isinstance(room, JoinedSyncResult)
|
2016-01-25 05:10:44 -05:00
|
|
|
ephemeral_events = room.ephemeral
|
2015-10-19 12:26:18 -04:00
|
|
|
result["ephemeral"] = {"events": ephemeral_events}
|
2016-01-19 12:19:53 -05:00
|
|
|
result["unread_notifications"] = room.unread_notifications
|
2022-10-04 09:47:04 -04:00
|
|
|
if room.unread_thread_notifications:
|
2022-10-07 09:26:40 -04:00
|
|
|
result["unread_thread_notifications"] = room.unread_thread_notifications
|
|
|
|
if self._msc3773_enabled:
|
2024-03-13 12:46:44 -04:00
|
|
|
result["org.matrix.msc3773.unread_thread_notifications"] = (
|
|
|
|
room.unread_thread_notifications
|
|
|
|
)
|
2018-08-16 04:46:50 -04:00
|
|
|
result["summary"] = room.summary
|
2022-03-31 15:05:13 -04:00
|
|
|
if self._msc2654_enabled:
|
|
|
|
result["org.matrix.msc2654.unread_count"] = room.unread_count
|
2015-10-19 12:26:18 -04:00
|
|
|
|
2019-07-23 09:00:55 -04:00
|
|
|
return result
|
2015-01-26 13:53:31 -05:00
|
|
|
|
2015-01-23 13:31:29 -05:00
|
|
|
|
2024-05-23 13:06:16 -04:00
|
|
|
class SlidingSyncE2eeRestServlet(RestServlet):
|
|
|
|
"""
|
|
|
|
API endpoint for MSC3575 Sliding Sync `/sync/e2ee`. This is being introduced as part
|
|
|
|
of Sliding Sync but doesn't have any sliding window component. It's just a way to
|
|
|
|
get E2EE events without having to sit through a big initial sync (`/sync` v2). And
|
|
|
|
we can avoid encryption events being backed up by the main sync response.
|
|
|
|
|
|
|
|
Having To-Device messages split out to this sync endpoint also helps when clients
|
|
|
|
need to have 2 or more sync streams open at a time, e.g a push notification process
|
|
|
|
and a main process. This can cause the two processes to race to fetch the To-Device
|
|
|
|
events, resulting in the need for complex synchronisation rules to ensure the token
|
|
|
|
is correctly and atomically exchanged between processes.
|
|
|
|
|
|
|
|
GET parameters::
|
|
|
|
timeout(int): How long to wait for new events in milliseconds.
|
|
|
|
since(batch_token): Batch token when asking for incremental deltas.
|
|
|
|
|
|
|
|
Response JSON::
|
|
|
|
{
|
|
|
|
"next_batch": // batch token for the next /sync
|
|
|
|
"to_device": {
|
|
|
|
// list of to-device events
|
|
|
|
"events": [
|
|
|
|
{
|
|
|
|
"content: { "algorithm": "m.olm.v1.curve25519-aes-sha2", "ciphertext": { ... }, "org.matrix.msgid": "abcd", "session_id": "abcd" },
|
|
|
|
"type": "m.room.encrypted",
|
|
|
|
"sender": "@alice:example.com",
|
|
|
|
}
|
|
|
|
// ...
|
|
|
|
]
|
|
|
|
},
|
|
|
|
"device_lists": {
|
|
|
|
"changed": ["@alice:example.com"],
|
|
|
|
"left": ["@bob:example.com"]
|
|
|
|
},
|
|
|
|
"device_one_time_keys_count": {
|
|
|
|
"signed_curve25519": 50
|
|
|
|
},
|
|
|
|
"device_unused_fallback_key_types": [
|
|
|
|
"signed_curve25519"
|
|
|
|
]
|
|
|
|
}
|
|
|
|
"""
|
|
|
|
|
|
|
|
PATTERNS = client_patterns(
|
|
|
|
"/org.matrix.msc3575/sync/e2ee$", releases=[], v1=False, unstable=True
|
|
|
|
)
|
|
|
|
|
|
|
|
def __init__(self, hs: "HomeServer"):
|
|
|
|
super().__init__()
|
|
|
|
self.hs = hs
|
|
|
|
self.auth = hs.get_auth()
|
|
|
|
self.store = hs.get_datastores().main
|
|
|
|
self.sync_handler = hs.get_sync_handler()
|
|
|
|
|
|
|
|
# Filtering only matters for the `device_lists` because it requires a bunch of
|
|
|
|
# derived information from rooms (see how `_generate_sync_entry_for_rooms()`
|
|
|
|
# prepares a bunch of data for `_generate_sync_entry_for_device_list()`).
|
|
|
|
self.only_member_events_filter_collection = FilterCollection(
|
|
|
|
self.hs,
|
|
|
|
{
|
|
|
|
"room": {
|
|
|
|
# We only care about membership events for the `device_lists`.
|
|
|
|
# Membership will tell us whether a user has joined/left a room and
|
|
|
|
# if there are new devices to encrypt for.
|
|
|
|
"timeline": {
|
|
|
|
"types": ["m.room.member"],
|
|
|
|
},
|
|
|
|
"state": {
|
|
|
|
"types": ["m.room.member"],
|
|
|
|
},
|
|
|
|
# We don't want any extra account_data generated because it's not
|
|
|
|
# returned by this endpoint. This helps us avoid work in
|
|
|
|
# `_generate_sync_entry_for_rooms()`
|
|
|
|
"account_data": {
|
|
|
|
"not_types": ["*"],
|
|
|
|
},
|
|
|
|
# We don't want any extra ephemeral data generated because it's not
|
|
|
|
# returned by this endpoint. This helps us avoid work in
|
|
|
|
# `_generate_sync_entry_for_rooms()`
|
|
|
|
"ephemeral": {
|
|
|
|
"not_types": ["*"],
|
|
|
|
},
|
|
|
|
},
|
|
|
|
# We don't want any extra account_data generated because it's not
|
|
|
|
# returned by this endpoint. (This is just here for good measure)
|
|
|
|
"account_data": {
|
|
|
|
"not_types": ["*"],
|
|
|
|
},
|
|
|
|
# We don't want any extra presence data generated because it's not
|
|
|
|
# returned by this endpoint. (This is just here for good measure)
|
|
|
|
"presence": {
|
|
|
|
"not_types": ["*"],
|
|
|
|
},
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
async def on_GET(self, request: SynapseRequest) -> Tuple[int, JsonDict]:
|
|
|
|
requester = await self.auth.get_user_by_req(request, allow_guest=True)
|
|
|
|
user = requester.user
|
|
|
|
device_id = requester.device_id
|
|
|
|
|
|
|
|
timeout = parse_integer(request, "timeout", default=0)
|
|
|
|
since = parse_string(request, "since")
|
|
|
|
|
|
|
|
sync_config = SyncConfig(
|
|
|
|
user=user,
|
|
|
|
filter_collection=self.only_member_events_filter_collection,
|
|
|
|
is_guest=requester.is_guest,
|
|
|
|
device_id=device_id,
|
|
|
|
)
|
|
|
|
|
|
|
|
since_token = None
|
|
|
|
if since is not None:
|
|
|
|
since_token = await StreamToken.from_string(self.store, since)
|
|
|
|
|
|
|
|
# Request cache key
|
|
|
|
request_key = (
|
|
|
|
SyncVersion.E2EE_SYNC,
|
|
|
|
user,
|
|
|
|
timeout,
|
|
|
|
since,
|
|
|
|
)
|
|
|
|
|
|
|
|
# Gather data for the response
|
|
|
|
sync_result = await self.sync_handler.wait_for_sync_for_user(
|
|
|
|
requester,
|
|
|
|
sync_config,
|
|
|
|
SyncVersion.E2EE_SYNC,
|
|
|
|
request_key,
|
|
|
|
since_token=since_token,
|
|
|
|
timeout=timeout,
|
|
|
|
full_state=False,
|
|
|
|
)
|
|
|
|
|
|
|
|
# The client may have disconnected by now; don't bother to serialize the
|
|
|
|
# response if so.
|
|
|
|
if request._disconnected:
|
|
|
|
logger.info("Client has disconnected; not serializing response.")
|
|
|
|
return 200, {}
|
|
|
|
|
|
|
|
response: JsonDict = defaultdict(dict)
|
|
|
|
response["next_batch"] = await sync_result.next_batch.to_string(self.store)
|
|
|
|
|
|
|
|
if sync_result.to_device:
|
|
|
|
response["to_device"] = {"events": sync_result.to_device}
|
|
|
|
|
|
|
|
if sync_result.device_lists.changed:
|
|
|
|
response["device_lists"]["changed"] = list(sync_result.device_lists.changed)
|
|
|
|
if sync_result.device_lists.left:
|
|
|
|
response["device_lists"]["left"] = list(sync_result.device_lists.left)
|
|
|
|
|
|
|
|
# We always include this because https://github.com/vector-im/element-android/issues/3725
|
|
|
|
# The spec isn't terribly clear on when this can be omitted and how a client would tell
|
|
|
|
# the difference between "no keys present" and "nothing changed" in terms of whole field
|
|
|
|
# absent / individual key type entry absent
|
|
|
|
# Corresponding synapse issue: https://github.com/matrix-org/synapse/issues/10456
|
|
|
|
response["device_one_time_keys_count"] = sync_result.device_one_time_keys_count
|
|
|
|
|
|
|
|
# https://github.com/matrix-org/matrix-doc/blob/54255851f642f84a4f1aaf7bc063eebe3d76752b/proposals/2732-olm-fallback-keys.md
|
|
|
|
# states that this field should always be included, as long as the server supports the feature.
|
|
|
|
response["device_unused_fallback_key_types"] = (
|
|
|
|
sync_result.device_unused_fallback_key_types
|
|
|
|
)
|
|
|
|
|
|
|
|
return 200, response
|
|
|
|
|
|
|
|
|
Add Sliding Sync `/sync` endpoint (initial implementation) (#17187)
Based on [MSC3575](https://github.com/matrix-org/matrix-spec-proposals/pull/3575): Sliding Sync
This iteration only focuses on returning the list of room IDs in the sliding window API (without sorting/filtering).
Rooms appear in the Sliding sync response based on:
- `invite`, `join`, `knock`, `ban` membership events
- Kicks (`leave` membership events where `sender` is different from the `user_id`/`state_key`)
- `newly_left` (rooms that were left during the given token range, > `from_token` and <= `to_token`)
- In order for bans/kicks to not show up, you need to `/forget` those rooms. This doesn't modify the event itself though and only adds the `forgotten` flag to `room_memberships` in Synapse. There isn't a way to tell when a room was forgotten at the moment so we can't factor it into the from/to range.
### Example request
`POST http://localhost:8008/_matrix/client/unstable/org.matrix.msc3575/sync`
```json
{
"lists": {
"foo-list": {
"ranges": [ [0, 99] ],
"sort": [ "by_notification_level", "by_recency", "by_name" ],
"required_state": [
["m.room.join_rules", ""],
["m.room.history_visibility", ""],
["m.space.child", "*"]
],
"timeline_limit": 100
}
}
}
```
Response:
```json
{
"next_pos": "s58_224_0_13_10_1_1_16_0_1",
"lists": {
"foo-list": {
"count": 1,
"ops": [
{
"op": "SYNC",
"range": [0, 99],
"room_ids": [
"!MmgikIyFzsuvtnbvVG:my.synapse.linux.server"
]
}
]
}
},
"rooms": {},
"extensions": {}
}
```
2024-06-06 15:44:32 -04:00
|
|
|
class SlidingSyncRestServlet(RestServlet):
|
|
|
|
"""
|
|
|
|
API endpoint for MSC3575 Sliding Sync `/sync`. Allows for clients to request a
|
|
|
|
subset (sliding window) of rooms, state, and timeline events (just what they need)
|
|
|
|
in order to bootstrap quickly and subscribe to only what the client cares about.
|
|
|
|
Because the client can specify what it cares about, we can respond quickly and skip
|
|
|
|
all of the work we would normally have to do with a sync v2 response.
|
|
|
|
|
|
|
|
Request query parameters:
|
|
|
|
timeout: How long to wait for new events in milliseconds.
|
|
|
|
pos: Stream position token when asking for incremental deltas.
|
|
|
|
|
|
|
|
Request body::
|
|
|
|
{
|
|
|
|
// Sliding Window API
|
|
|
|
"lists": {
|
|
|
|
"foo-list": {
|
|
|
|
"ranges": [ [0, 99] ],
|
|
|
|
"sort": [ "by_notification_level", "by_recency", "by_name" ],
|
|
|
|
"required_state": [
|
|
|
|
["m.room.join_rules", ""],
|
|
|
|
["m.room.history_visibility", ""],
|
|
|
|
["m.space.child", "*"]
|
|
|
|
],
|
|
|
|
"timeline_limit": 10,
|
|
|
|
"filters": {
|
|
|
|
"is_dm": true
|
|
|
|
},
|
|
|
|
"bump_event_types": [ "m.room.message", "m.room.encrypted" ],
|
|
|
|
}
|
|
|
|
},
|
|
|
|
// Room Subscriptions API
|
|
|
|
"room_subscriptions": {
|
|
|
|
"!sub1:bar": {
|
|
|
|
"required_state": [ ["*","*"] ],
|
|
|
|
"timeline_limit": 10,
|
|
|
|
"include_old_rooms": {
|
|
|
|
"timeline_limit": 1,
|
|
|
|
"required_state": [ ["m.room.tombstone", ""], ["m.room.create", ""] ],
|
|
|
|
}
|
|
|
|
}
|
|
|
|
},
|
|
|
|
// Extensions API
|
|
|
|
"extensions": {}
|
|
|
|
}
|
|
|
|
|
|
|
|
Response JSON::
|
|
|
|
{
|
|
|
|
"next_pos": "s58_224_0_13_10_1_1_16_0_1",
|
|
|
|
"lists": {
|
|
|
|
"foo-list": {
|
|
|
|
"count": 1337,
|
|
|
|
"ops": [{
|
|
|
|
"op": "SYNC",
|
|
|
|
"range": [0, 99],
|
|
|
|
"room_ids": [
|
|
|
|
"!foo:bar",
|
|
|
|
// ... 99 more room IDs
|
|
|
|
]
|
|
|
|
}]
|
|
|
|
}
|
|
|
|
},
|
|
|
|
// Aggregated rooms from lists and room subscriptions
|
|
|
|
"rooms": {
|
|
|
|
// Room from room subscription
|
|
|
|
"!sub1:bar": {
|
|
|
|
"name": "Alice and Bob",
|
|
|
|
"avatar": "mxc://...",
|
|
|
|
"initial": true,
|
|
|
|
"required_state": [
|
|
|
|
{"sender":"@alice:example.com","type":"m.room.create", "state_key":"", "content":{"creator":"@alice:example.com"}},
|
|
|
|
{"sender":"@alice:example.com","type":"m.room.join_rules", "state_key":"", "content":{"join_rule":"invite"}},
|
|
|
|
{"sender":"@alice:example.com","type":"m.room.history_visibility", "state_key":"", "content":{"history_visibility":"joined"}},
|
|
|
|
{"sender":"@alice:example.com","type":"m.room.member", "state_key":"@alice:example.com", "content":{"membership":"join"}}
|
|
|
|
],
|
|
|
|
"timeline": [
|
|
|
|
{"sender":"@alice:example.com","type":"m.room.create", "state_key":"", "content":{"creator":"@alice:example.com"}},
|
|
|
|
{"sender":"@alice:example.com","type":"m.room.join_rules", "state_key":"", "content":{"join_rule":"invite"}},
|
|
|
|
{"sender":"@alice:example.com","type":"m.room.history_visibility", "state_key":"", "content":{"history_visibility":"joined"}},
|
|
|
|
{"sender":"@alice:example.com","type":"m.room.member", "state_key":"@alice:example.com", "content":{"membership":"join"}},
|
|
|
|
{"sender":"@alice:example.com","type":"m.room.message", "content":{"body":"A"}},
|
|
|
|
{"sender":"@alice:example.com","type":"m.room.message", "content":{"body":"B"}},
|
|
|
|
],
|
|
|
|
"prev_batch": "t111_222_333",
|
|
|
|
"joined_count": 41,
|
|
|
|
"invited_count": 1,
|
|
|
|
"notification_count": 1,
|
|
|
|
"highlight_count": 0
|
|
|
|
},
|
|
|
|
// rooms from list
|
|
|
|
"!foo:bar": {
|
|
|
|
"name": "The calculated room name",
|
|
|
|
"avatar": "mxc://...",
|
|
|
|
"initial": true,
|
|
|
|
"required_state": [
|
|
|
|
{"sender":"@alice:example.com","type":"m.room.join_rules", "state_key":"", "content":{"join_rule":"invite"}},
|
|
|
|
{"sender":"@alice:example.com","type":"m.room.history_visibility", "state_key":"", "content":{"history_visibility":"joined"}},
|
|
|
|
{"sender":"@alice:example.com","type":"m.space.child", "state_key":"!foo:example.com", "content":{"via":["example.com"]}},
|
|
|
|
{"sender":"@alice:example.com","type":"m.space.child", "state_key":"!bar:example.com", "content":{"via":["example.com"]}},
|
|
|
|
{"sender":"@alice:example.com","type":"m.space.child", "state_key":"!baz:example.com", "content":{"via":["example.com"]}}
|
|
|
|
],
|
|
|
|
"timeline": [
|
|
|
|
{"sender":"@alice:example.com","type":"m.room.join_rules", "state_key":"", "content":{"join_rule":"invite"}},
|
|
|
|
{"sender":"@alice:example.com","type":"m.room.message", "content":{"body":"A"}},
|
|
|
|
{"sender":"@alice:example.com","type":"m.room.message", "content":{"body":"B"}},
|
|
|
|
{"sender":"@alice:example.com","type":"m.room.message", "content":{"body":"C"}},
|
|
|
|
{"sender":"@alice:example.com","type":"m.room.message", "content":{"body":"D"}},
|
|
|
|
],
|
|
|
|
"prev_batch": "t111_222_333",
|
|
|
|
"joined_count": 4,
|
|
|
|
"invited_count": 0,
|
|
|
|
"notification_count": 54,
|
|
|
|
"highlight_count": 3
|
|
|
|
},
|
|
|
|
// ... 99 more items
|
|
|
|
},
|
|
|
|
"extensions": {}
|
|
|
|
}
|
|
|
|
"""
|
|
|
|
|
|
|
|
PATTERNS = client_patterns(
|
|
|
|
"/org.matrix.msc3575/sync$", releases=[], v1=False, unstable=True
|
|
|
|
)
|
|
|
|
|
|
|
|
def __init__(self, hs: "HomeServer"):
|
|
|
|
super().__init__()
|
|
|
|
self.auth = hs.get_auth()
|
|
|
|
self.store = hs.get_datastores().main
|
|
|
|
self.filtering = hs.get_filtering()
|
|
|
|
self.sliding_sync_handler = hs.get_sliding_sync_handler()
|
|
|
|
|
|
|
|
# TODO: Update this to `on_GET` once we figure out how we want to handle params
|
|
|
|
async def on_POST(self, request: SynapseRequest) -> Tuple[int, JsonDict]:
|
|
|
|
requester = await self.auth.get_user_by_req(request, allow_guest=True)
|
|
|
|
user = requester.user
|
|
|
|
device_id = requester.device_id
|
|
|
|
|
|
|
|
timeout = parse_integer(request, "timeout", default=0)
|
|
|
|
# Position in the stream
|
|
|
|
from_token_string = parse_string(request, "pos")
|
|
|
|
|
|
|
|
from_token = None
|
|
|
|
if from_token_string is not None:
|
|
|
|
from_token = await StreamToken.from_string(self.store, from_token_string)
|
|
|
|
|
|
|
|
# TODO: We currently don't know whether we're going to use sticky params or
|
|
|
|
# maybe some filters like sync v2 where they are built up once and referenced
|
|
|
|
# by filter ID. For now, we will just prototype with always passing everything
|
|
|
|
# in.
|
|
|
|
body = parse_and_validate_json_object_from_request(request, SlidingSyncBody)
|
|
|
|
logger.info("Sliding sync request: %r", body)
|
|
|
|
|
|
|
|
sync_config = SlidingSyncConfig(
|
|
|
|
user=user,
|
|
|
|
device_id=device_id,
|
|
|
|
# FIXME: Currently, we're just manually copying the fields from the
|
|
|
|
# `SlidingSyncBody` into the config. How can we gurantee into the future
|
|
|
|
# that we don't forget any? I would like something more structured like
|
|
|
|
# `copy_attributes(from=body, to=config)`
|
|
|
|
lists=body.lists,
|
|
|
|
room_subscriptions=body.room_subscriptions,
|
|
|
|
extensions=body.extensions,
|
|
|
|
)
|
|
|
|
|
|
|
|
sliding_sync_results = await self.sliding_sync_handler.wait_for_sync_for_user(
|
|
|
|
requester,
|
|
|
|
sync_config,
|
|
|
|
from_token,
|
|
|
|
timeout,
|
|
|
|
)
|
|
|
|
|
|
|
|
# The client may have disconnected by now; don't bother to serialize the
|
|
|
|
# response if so.
|
|
|
|
if request._disconnected:
|
|
|
|
logger.info("Client has disconnected; not serializing response.")
|
|
|
|
return 200, {}
|
|
|
|
|
|
|
|
response_content = await self.encode_response(sliding_sync_results)
|
|
|
|
|
|
|
|
return 200, response_content
|
|
|
|
|
|
|
|
# TODO: Is there a better way to encode things?
|
|
|
|
async def encode_response(
|
|
|
|
self,
|
|
|
|
sliding_sync_result: SlidingSyncResult,
|
|
|
|
) -> JsonDict:
|
|
|
|
response: JsonDict = defaultdict(dict)
|
|
|
|
|
|
|
|
response["next_pos"] = await sliding_sync_result.next_pos.to_string(self.store)
|
|
|
|
serialized_lists = self.encode_lists(sliding_sync_result.lists)
|
|
|
|
if serialized_lists:
|
|
|
|
response["lists"] = serialized_lists
|
|
|
|
response["rooms"] = {} # TODO: sliding_sync_result.rooms
|
|
|
|
response["extensions"] = {} # TODO: sliding_sync_result.extensions
|
|
|
|
|
|
|
|
return response
|
|
|
|
|
|
|
|
def encode_lists(
|
|
|
|
self, lists: Dict[str, SlidingSyncResult.SlidingWindowList]
|
|
|
|
) -> JsonDict:
|
|
|
|
def encode_operation(
|
|
|
|
operation: SlidingSyncResult.SlidingWindowList.Operation,
|
|
|
|
) -> JsonDict:
|
|
|
|
return {
|
|
|
|
"op": operation.op.value,
|
|
|
|
"range": operation.range,
|
|
|
|
"room_ids": operation.room_ids,
|
|
|
|
}
|
|
|
|
|
|
|
|
serialized_lists = {}
|
|
|
|
for list_key, list_result in lists.items():
|
|
|
|
serialized_lists[list_key] = {
|
|
|
|
"count": list_result.count,
|
|
|
|
"ops": [encode_operation(op) for op in list_result.ops],
|
|
|
|
}
|
|
|
|
|
|
|
|
return serialized_lists
|
|
|
|
|
|
|
|
|
2021-08-23 08:14:42 -04:00
|
|
|
def register_servlets(hs: "HomeServer", http_server: HttpServer) -> None:
|
2015-01-23 13:31:29 -05:00
|
|
|
SyncRestServlet(hs).register(http_server)
|
2024-05-23 13:06:16 -04:00
|
|
|
|
|
|
|
if hs.config.experimental.msc3575_enabled:
|
Add Sliding Sync `/sync` endpoint (initial implementation) (#17187)
Based on [MSC3575](https://github.com/matrix-org/matrix-spec-proposals/pull/3575): Sliding Sync
This iteration only focuses on returning the list of room IDs in the sliding window API (without sorting/filtering).
Rooms appear in the Sliding sync response based on:
- `invite`, `join`, `knock`, `ban` membership events
- Kicks (`leave` membership events where `sender` is different from the `user_id`/`state_key`)
- `newly_left` (rooms that were left during the given token range, > `from_token` and <= `to_token`)
- In order for bans/kicks to not show up, you need to `/forget` those rooms. This doesn't modify the event itself though and only adds the `forgotten` flag to `room_memberships` in Synapse. There isn't a way to tell when a room was forgotten at the moment so we can't factor it into the from/to range.
### Example request
`POST http://localhost:8008/_matrix/client/unstable/org.matrix.msc3575/sync`
```json
{
"lists": {
"foo-list": {
"ranges": [ [0, 99] ],
"sort": [ "by_notification_level", "by_recency", "by_name" ],
"required_state": [
["m.room.join_rules", ""],
["m.room.history_visibility", ""],
["m.space.child", "*"]
],
"timeline_limit": 100
}
}
}
```
Response:
```json
{
"next_pos": "s58_224_0_13_10_1_1_16_0_1",
"lists": {
"foo-list": {
"count": 1,
"ops": [
{
"op": "SYNC",
"range": [0, 99],
"room_ids": [
"!MmgikIyFzsuvtnbvVG:my.synapse.linux.server"
]
}
]
}
},
"rooms": {},
"extensions": {}
}
```
2024-06-06 15:44:32 -04:00
|
|
|
SlidingSyncRestServlet(hs).register(http_server)
|
2024-05-23 13:06:16 -04:00
|
|
|
SlidingSyncE2eeRestServlet(hs).register(http_server)
|