mirror of
https://git.anonymousland.org/anonymousland/synapse-product.git
synced 2024-12-23 16:09:26 -05:00
54dd5dc12b
Implement part [MSC2228](https://github.com/matrix-org/matrix-doc/pull/2228). The parts that differ are: * the feature is hidden behind a configuration flag (`enable_ephemeral_messages`) * self-destruction doesn't happen for state events * only implement support for the `m.self_destruct_after` field (not the `m.self_destruct` one) * doesn't send synthetic redactions to clients because for this specific case we consider the clients to be able to destroy an event themselves, instead we just censor it (by pruning its JSON) in the database
1107 lines
43 KiB
Python
1107 lines
43 KiB
Python
# -*- coding: utf-8 -*-
|
|
# Copyright 2014-2016 OpenMarket Ltd
|
|
# Copyright 2017-2018 New Vector Ltd
|
|
# Copyright 2019 The Matrix.org Foundation C.I.C.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
import logging
|
|
from typing import Optional
|
|
|
|
from six import iteritems, itervalues, string_types
|
|
|
|
from canonicaljson import encode_canonical_json, json
|
|
|
|
from twisted.internet import defer
|
|
from twisted.internet.defer import succeed
|
|
from twisted.internet.interfaces import IDelayedCall
|
|
|
|
from synapse import event_auth
|
|
from synapse.api.constants import (
|
|
EventContentFields,
|
|
EventTypes,
|
|
Membership,
|
|
RelationTypes,
|
|
UserTypes,
|
|
)
|
|
from synapse.api.errors import (
|
|
AuthError,
|
|
Codes,
|
|
ConsentNotGivenError,
|
|
NotFoundError,
|
|
SynapseError,
|
|
)
|
|
from synapse.api.room_versions import RoomVersions
|
|
from synapse.api.urls import ConsentURIBuilder
|
|
from synapse.events.validator import EventValidator
|
|
from synapse.logging.context import run_in_background
|
|
from synapse.metrics.background_process_metrics import run_as_background_process
|
|
from synapse.replication.http.send_event import ReplicationSendEventRestServlet
|
|
from synapse.storage.state import StateFilter
|
|
from synapse.types import RoomAlias, UserID, create_requester
|
|
from synapse.util.async_helpers import Linearizer
|
|
from synapse.util.frozenutils import frozendict_json_encoder
|
|
from synapse.util.metrics import measure_func
|
|
from synapse.visibility import filter_events_for_client
|
|
|
|
from ._base import BaseHandler
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
class MessageHandler(object):
|
|
"""Contains some read only APIs to get state about a room
|
|
"""
|
|
|
|
def __init__(self, hs):
|
|
self.auth = hs.get_auth()
|
|
self.clock = hs.get_clock()
|
|
self.state = hs.get_state_handler()
|
|
self.store = hs.get_datastore()
|
|
self.storage = hs.get_storage()
|
|
self.state_store = self.storage.state
|
|
self._event_serializer = hs.get_event_client_serializer()
|
|
self._ephemeral_events_enabled = hs.config.enable_ephemeral_messages
|
|
self._is_worker_app = bool(hs.config.worker_app)
|
|
|
|
# The scheduled call to self._expire_event. None if no call is currently
|
|
# scheduled.
|
|
self._scheduled_expiry = None # type: Optional[IDelayedCall]
|
|
|
|
if not hs.config.worker_app:
|
|
run_as_background_process(
|
|
"_schedule_next_expiry", self._schedule_next_expiry
|
|
)
|
|
|
|
@defer.inlineCallbacks
|
|
def get_room_data(
|
|
self, user_id=None, room_id=None, event_type=None, state_key="", is_guest=False
|
|
):
|
|
""" Get data from a room.
|
|
|
|
Args:
|
|
event : The room path event
|
|
Returns:
|
|
The path data content.
|
|
Raises:
|
|
SynapseError if something went wrong.
|
|
"""
|
|
(
|
|
membership,
|
|
membership_event_id,
|
|
) = yield self.auth.check_in_room_or_world_readable(room_id, user_id)
|
|
|
|
if membership == Membership.JOIN:
|
|
data = yield self.state.get_current_state(room_id, event_type, state_key)
|
|
elif membership == Membership.LEAVE:
|
|
key = (event_type, state_key)
|
|
room_state = yield self.state_store.get_state_for_events(
|
|
[membership_event_id], StateFilter.from_types([key])
|
|
)
|
|
data = room_state[membership_event_id].get(key)
|
|
|
|
return data
|
|
|
|
@defer.inlineCallbacks
|
|
def get_state_events(
|
|
self,
|
|
user_id,
|
|
room_id,
|
|
state_filter=StateFilter.all(),
|
|
at_token=None,
|
|
is_guest=False,
|
|
):
|
|
"""Retrieve all state events for a given room. If the user is
|
|
joined to the room then return the current state. If the user has
|
|
left the room return the state events from when they left. If an explicit
|
|
'at' parameter is passed, return the state events as of that event, if
|
|
visible.
|
|
|
|
Args:
|
|
user_id(str): The user requesting state events.
|
|
room_id(str): The room ID to get all state events from.
|
|
state_filter (StateFilter): The state filter used to fetch state
|
|
from the database.
|
|
at_token(StreamToken|None): the stream token of the at which we are requesting
|
|
the stats. If the user is not allowed to view the state as of that
|
|
stream token, we raise a 403 SynapseError. If None, returns the current
|
|
state based on the current_state_events table.
|
|
is_guest(bool): whether this user is a guest
|
|
Returns:
|
|
A list of dicts representing state events. [{}, {}, {}]
|
|
Raises:
|
|
NotFoundError (404) if the at token does not yield an event
|
|
|
|
AuthError (403) if the user doesn't have permission to view
|
|
members of this room.
|
|
"""
|
|
if at_token:
|
|
# FIXME this claims to get the state at a stream position, but
|
|
# get_recent_events_for_room operates by topo ordering. This therefore
|
|
# does not reliably give you the state at the given stream position.
|
|
# (https://github.com/matrix-org/synapse/issues/3305)
|
|
last_events, _ = yield self.store.get_recent_events_for_room(
|
|
room_id, end_token=at_token.room_key, limit=1
|
|
)
|
|
|
|
if not last_events:
|
|
raise NotFoundError("Can't find event for token %s" % (at_token,))
|
|
|
|
visible_events = yield filter_events_for_client(
|
|
self.storage, user_id, last_events, apply_retention_policies=False
|
|
)
|
|
|
|
event = last_events[0]
|
|
if visible_events:
|
|
room_state = yield self.state_store.get_state_for_events(
|
|
[event.event_id], state_filter=state_filter
|
|
)
|
|
room_state = room_state[event.event_id]
|
|
else:
|
|
raise AuthError(
|
|
403,
|
|
"User %s not allowed to view events in room %s at token %s"
|
|
% (user_id, room_id, at_token),
|
|
)
|
|
else:
|
|
(
|
|
membership,
|
|
membership_event_id,
|
|
) = yield self.auth.check_in_room_or_world_readable(room_id, user_id)
|
|
|
|
if membership == Membership.JOIN:
|
|
state_ids = yield self.store.get_filtered_current_state_ids(
|
|
room_id, state_filter=state_filter
|
|
)
|
|
room_state = yield self.store.get_events(state_ids.values())
|
|
elif membership == Membership.LEAVE:
|
|
room_state = yield self.state_store.get_state_for_events(
|
|
[membership_event_id], state_filter=state_filter
|
|
)
|
|
room_state = room_state[membership_event_id]
|
|
|
|
now = self.clock.time_msec()
|
|
events = yield self._event_serializer.serialize_events(
|
|
room_state.values(),
|
|
now,
|
|
# We don't bother bundling aggregations in when asked for state
|
|
# events, as clients won't use them.
|
|
bundle_aggregations=False,
|
|
)
|
|
return events
|
|
|
|
@defer.inlineCallbacks
|
|
def get_joined_members(self, requester, room_id):
|
|
"""Get all the joined members in the room and their profile information.
|
|
|
|
If the user has left the room return the state events from when they left.
|
|
|
|
Args:
|
|
requester(Requester): The user requesting state events.
|
|
room_id(str): The room ID to get all state events from.
|
|
Returns:
|
|
A dict of user_id to profile info
|
|
"""
|
|
user_id = requester.user.to_string()
|
|
if not requester.app_service:
|
|
# We check AS auth after fetching the room membership, as it
|
|
# requires us to pull out all joined members anyway.
|
|
membership, _ = yield self.auth.check_in_room_or_world_readable(
|
|
room_id, user_id
|
|
)
|
|
if membership != Membership.JOIN:
|
|
raise NotImplementedError(
|
|
"Getting joined members after leaving is not implemented"
|
|
)
|
|
|
|
users_with_profile = yield self.state.get_current_users_in_room(room_id)
|
|
|
|
# If this is an AS, double check that they are allowed to see the members.
|
|
# This can either be because the AS user is in the room or because there
|
|
# is a user in the room that the AS is "interested in"
|
|
if requester.app_service and user_id not in users_with_profile:
|
|
for uid in users_with_profile:
|
|
if requester.app_service.is_interested_in_user(uid):
|
|
break
|
|
else:
|
|
# Loop fell through, AS has no interested users in room
|
|
raise AuthError(403, "Appservice not in room")
|
|
|
|
return {
|
|
user_id: {
|
|
"avatar_url": profile.avatar_url,
|
|
"display_name": profile.display_name,
|
|
}
|
|
for user_id, profile in iteritems(users_with_profile)
|
|
}
|
|
|
|
def maybe_schedule_expiry(self, event):
|
|
"""Schedule the expiry of an event if there's not already one scheduled,
|
|
or if the one running is for an event that will expire after the provided
|
|
timestamp.
|
|
|
|
This function needs to invalidate the event cache, which is only possible on
|
|
the master process, and therefore needs to be run on there.
|
|
|
|
Args:
|
|
event (EventBase): The event to schedule the expiry of.
|
|
"""
|
|
assert not self._is_worker_app
|
|
|
|
expiry_ts = event.content.get(EventContentFields.SELF_DESTRUCT_AFTER)
|
|
if not isinstance(expiry_ts, int) or event.is_state():
|
|
return
|
|
|
|
# _schedule_expiry_for_event won't actually schedule anything if there's already
|
|
# a task scheduled for a timestamp that's sooner than the provided one.
|
|
self._schedule_expiry_for_event(event.event_id, expiry_ts)
|
|
|
|
@defer.inlineCallbacks
|
|
def _schedule_next_expiry(self):
|
|
"""Retrieve the ID and the expiry timestamp of the next event to be expired,
|
|
and schedule an expiry task for it.
|
|
|
|
If there's no event left to expire, set _expiry_scheduled to None so that a
|
|
future call to save_expiry_ts can schedule a new expiry task.
|
|
"""
|
|
# Try to get the expiry timestamp of the next event to expire.
|
|
res = yield self.store.get_next_event_to_expire()
|
|
if res:
|
|
event_id, expiry_ts = res
|
|
self._schedule_expiry_for_event(event_id, expiry_ts)
|
|
|
|
def _schedule_expiry_for_event(self, event_id, expiry_ts):
|
|
"""Schedule an expiry task for the provided event if there's not already one
|
|
scheduled at a timestamp that's sooner than the provided one.
|
|
|
|
Args:
|
|
event_id (str): The ID of the event to expire.
|
|
expiry_ts (int): The timestamp at which to expire the event.
|
|
"""
|
|
if self._scheduled_expiry:
|
|
# If the provided timestamp refers to a time before the scheduled time of the
|
|
# next expiry task, cancel that task and reschedule it for this timestamp.
|
|
next_scheduled_expiry_ts = self._scheduled_expiry.getTime() * 1000
|
|
if expiry_ts < next_scheduled_expiry_ts:
|
|
self._scheduled_expiry.cancel()
|
|
else:
|
|
return
|
|
|
|
# Figure out how many seconds we need to wait before expiring the event.
|
|
now_ms = self.clock.time_msec()
|
|
delay = (expiry_ts - now_ms) / 1000
|
|
|
|
# callLater doesn't support negative delays, so trim the delay to 0 if we're
|
|
# in that case.
|
|
if delay < 0:
|
|
delay = 0
|
|
|
|
logger.info("Scheduling expiry for event %s in %.3fs", event_id, delay)
|
|
|
|
self._scheduled_expiry = self.clock.call_later(
|
|
delay,
|
|
run_as_background_process,
|
|
"_expire_event",
|
|
self._expire_event,
|
|
event_id,
|
|
)
|
|
|
|
@defer.inlineCallbacks
|
|
def _expire_event(self, event_id):
|
|
"""Retrieve and expire an event that needs to be expired from the database.
|
|
|
|
If the event doesn't exist in the database, log it and delete the expiry date
|
|
from the database (so that we don't try to expire it again).
|
|
"""
|
|
assert self._ephemeral_events_enabled
|
|
|
|
self._scheduled_expiry = None
|
|
|
|
logger.info("Expiring event %s", event_id)
|
|
|
|
try:
|
|
# Expire the event if we know about it. This function also deletes the expiry
|
|
# date from the database in the same database transaction.
|
|
yield self.store.expire_event(event_id)
|
|
except Exception as e:
|
|
logger.error("Could not expire event %s: %r", event_id, e)
|
|
|
|
# Schedule the expiry of the next event to expire.
|
|
yield self._schedule_next_expiry()
|
|
|
|
|
|
# The duration (in ms) after which rooms should be removed
|
|
# `_rooms_to_exclude_from_dummy_event_insertion` (with the effect that we will try
|
|
# to generate a dummy event for them once more)
|
|
#
|
|
_DUMMY_EVENT_ROOM_EXCLUSION_EXPIRY = 7 * 24 * 60 * 60 * 1000
|
|
|
|
|
|
class EventCreationHandler(object):
|
|
def __init__(self, hs):
|
|
self.hs = hs
|
|
self.auth = hs.get_auth()
|
|
self.store = hs.get_datastore()
|
|
self.storage = hs.get_storage()
|
|
self.state = hs.get_state_handler()
|
|
self.clock = hs.get_clock()
|
|
self.validator = EventValidator()
|
|
self.profile_handler = hs.get_profile_handler()
|
|
self.event_builder_factory = hs.get_event_builder_factory()
|
|
self.server_name = hs.hostname
|
|
self.ratelimiter = hs.get_ratelimiter()
|
|
self.notifier = hs.get_notifier()
|
|
self.config = hs.config
|
|
self.require_membership_for_aliases = hs.config.require_membership_for_aliases
|
|
|
|
self.send_event_to_master = ReplicationSendEventRestServlet.make_client(hs)
|
|
|
|
# This is only used to get at ratelimit function, and maybe_kick_guest_users
|
|
self.base_handler = BaseHandler(hs)
|
|
|
|
self.pusher_pool = hs.get_pusherpool()
|
|
|
|
# We arbitrarily limit concurrent event creation for a room to 5.
|
|
# This is to stop us from diverging history *too* much.
|
|
self.limiter = Linearizer(max_count=5, name="room_event_creation_limit")
|
|
|
|
self.action_generator = hs.get_action_generator()
|
|
|
|
self.spam_checker = hs.get_spam_checker()
|
|
self.third_party_event_rules = hs.get_third_party_event_rules()
|
|
|
|
self._block_events_without_consent_error = (
|
|
self.config.block_events_without_consent_error
|
|
)
|
|
|
|
# Rooms which should be excluded from dummy insertion. (For instance,
|
|
# those without local users who can send events into the room).
|
|
#
|
|
# map from room id to time-of-last-attempt.
|
|
#
|
|
self._rooms_to_exclude_from_dummy_event_insertion = {} # type: dict[str, int]
|
|
|
|
# we need to construct a ConsentURIBuilder here, as it checks that the necessary
|
|
# config options, but *only* if we have a configuration for which we are
|
|
# going to need it.
|
|
if self._block_events_without_consent_error:
|
|
self._consent_uri_builder = ConsentURIBuilder(self.config)
|
|
|
|
if (
|
|
not self.config.worker_app
|
|
and self.config.cleanup_extremities_with_dummy_events
|
|
):
|
|
self.clock.looping_call(
|
|
lambda: run_as_background_process(
|
|
"send_dummy_events_to_fill_extremities",
|
|
self._send_dummy_events_to_fill_extremities,
|
|
),
|
|
5 * 60 * 1000,
|
|
)
|
|
|
|
self._message_handler = hs.get_message_handler()
|
|
|
|
self._ephemeral_events_enabled = hs.config.enable_ephemeral_messages
|
|
|
|
@defer.inlineCallbacks
|
|
def create_event(
|
|
self,
|
|
requester,
|
|
event_dict,
|
|
token_id=None,
|
|
txn_id=None,
|
|
prev_events_and_hashes=None,
|
|
require_consent=True,
|
|
):
|
|
"""
|
|
Given a dict from a client, create a new event.
|
|
|
|
Creates an FrozenEvent object, filling out auth_events, prev_events,
|
|
etc.
|
|
|
|
Adds display names to Join membership events.
|
|
|
|
Args:
|
|
requester
|
|
event_dict (dict): An entire event
|
|
token_id (str)
|
|
txn_id (str)
|
|
|
|
prev_events_and_hashes (list[(str, dict[str, str], int)]|None):
|
|
the forward extremities to use as the prev_events for the
|
|
new event. For each event, a tuple of (event_id, hashes, depth)
|
|
where *hashes* is a map from algorithm to hash.
|
|
|
|
If None, they will be requested from the database.
|
|
|
|
require_consent (bool): Whether to check if the requester has
|
|
consented to privacy policy.
|
|
Raises:
|
|
ResourceLimitError if server is blocked to some resource being
|
|
exceeded
|
|
Returns:
|
|
Tuple of created event (FrozenEvent), Context
|
|
"""
|
|
yield self.auth.check_auth_blocking(requester.user.to_string())
|
|
|
|
if event_dict["type"] == EventTypes.Create and event_dict["state_key"] == "":
|
|
room_version = event_dict["content"]["room_version"]
|
|
else:
|
|
try:
|
|
room_version = yield self.store.get_room_version(event_dict["room_id"])
|
|
except NotFoundError:
|
|
raise AuthError(403, "Unknown room")
|
|
|
|
builder = self.event_builder_factory.new(room_version, event_dict)
|
|
|
|
self.validator.validate_builder(builder)
|
|
|
|
if builder.type == EventTypes.Member:
|
|
membership = builder.content.get("membership", None)
|
|
target = UserID.from_string(builder.state_key)
|
|
|
|
if membership in {Membership.JOIN, Membership.INVITE}:
|
|
# If event doesn't include a display name, add one.
|
|
profile = self.profile_handler
|
|
content = builder.content
|
|
|
|
try:
|
|
if "displayname" not in content:
|
|
content["displayname"] = yield profile.get_displayname(target)
|
|
if "avatar_url" not in content:
|
|
content["avatar_url"] = yield profile.get_avatar_url(target)
|
|
except Exception as e:
|
|
logger.info(
|
|
"Failed to get profile information for %r: %s", target, e
|
|
)
|
|
|
|
is_exempt = yield self._is_exempt_from_privacy_policy(builder, requester)
|
|
if require_consent and not is_exempt:
|
|
yield self.assert_accepted_privacy_policy(requester)
|
|
|
|
if token_id is not None:
|
|
builder.internal_metadata.token_id = token_id
|
|
|
|
if txn_id is not None:
|
|
builder.internal_metadata.txn_id = txn_id
|
|
|
|
event, context = yield self.create_new_client_event(
|
|
builder=builder,
|
|
requester=requester,
|
|
prev_events_and_hashes=prev_events_and_hashes,
|
|
)
|
|
|
|
# In an ideal world we wouldn't need the second part of this condition. However,
|
|
# this behaviour isn't spec'd yet, meaning we should be able to deactivate this
|
|
# behaviour. Another reason is that this code is also evaluated each time a new
|
|
# m.room.aliases event is created, which includes hitting a /directory route.
|
|
# Therefore not including this condition here would render the similar one in
|
|
# synapse.handlers.directory pointless.
|
|
if builder.type == EventTypes.Aliases and self.require_membership_for_aliases:
|
|
# Ideally we'd do the membership check in event_auth.check(), which
|
|
# describes a spec'd algorithm for authenticating events received over
|
|
# federation as well as those created locally. As of room v3, aliases events
|
|
# can be created by users that are not in the room, therefore we have to
|
|
# tolerate them in event_auth.check().
|
|
prev_state_ids = yield context.get_prev_state_ids(self.store)
|
|
prev_event_id = prev_state_ids.get((EventTypes.Member, event.sender))
|
|
prev_event = (
|
|
yield self.store.get_event(prev_event_id, allow_none=True)
|
|
if prev_event_id
|
|
else None
|
|
)
|
|
if not prev_event or prev_event.membership != Membership.JOIN:
|
|
logger.warning(
|
|
(
|
|
"Attempt to send `m.room.aliases` in room %s by user %s but"
|
|
" membership is %s"
|
|
),
|
|
event.room_id,
|
|
event.sender,
|
|
prev_event.membership if prev_event else None,
|
|
)
|
|
|
|
raise AuthError(
|
|
403, "You must be in the room to create an alias for it"
|
|
)
|
|
|
|
self.validator.validate_new(event, self.config)
|
|
|
|
return (event, context)
|
|
|
|
def _is_exempt_from_privacy_policy(self, builder, requester):
|
|
""""Determine if an event to be sent is exempt from having to consent
|
|
to the privacy policy
|
|
|
|
Args:
|
|
builder (synapse.events.builder.EventBuilder): event being created
|
|
requester (Requster): user requesting this event
|
|
|
|
Returns:
|
|
Deferred[bool]: true if the event can be sent without the user
|
|
consenting
|
|
"""
|
|
# the only thing the user can do is join the server notices room.
|
|
if builder.type == EventTypes.Member:
|
|
membership = builder.content.get("membership", None)
|
|
if membership == Membership.JOIN:
|
|
return self._is_server_notices_room(builder.room_id)
|
|
elif membership == Membership.LEAVE:
|
|
# the user is always allowed to leave (but not kick people)
|
|
return builder.state_key == requester.user.to_string()
|
|
return succeed(False)
|
|
|
|
@defer.inlineCallbacks
|
|
def _is_server_notices_room(self, room_id):
|
|
if self.config.server_notices_mxid is None:
|
|
return False
|
|
user_ids = yield self.store.get_users_in_room(room_id)
|
|
return self.config.server_notices_mxid in user_ids
|
|
|
|
@defer.inlineCallbacks
|
|
def assert_accepted_privacy_policy(self, requester):
|
|
"""Check if a user has accepted the privacy policy
|
|
|
|
Called when the given user is about to do something that requires
|
|
privacy consent. We see if the user is exempt and otherwise check that
|
|
they have given consent. If they have not, a ConsentNotGiven error is
|
|
raised.
|
|
|
|
Args:
|
|
requester (synapse.types.Requester):
|
|
The user making the request
|
|
|
|
Returns:
|
|
Deferred[None]: returns normally if the user has consented or is
|
|
exempt
|
|
|
|
Raises:
|
|
ConsentNotGivenError: if the user has not given consent yet
|
|
"""
|
|
if self._block_events_without_consent_error is None:
|
|
return
|
|
|
|
# exempt AS users from needing consent
|
|
if requester.app_service is not None:
|
|
return
|
|
|
|
user_id = requester.user.to_string()
|
|
|
|
# exempt the system notices user
|
|
if (
|
|
self.config.server_notices_mxid is not None
|
|
and user_id == self.config.server_notices_mxid
|
|
):
|
|
return
|
|
|
|
u = yield self.store.get_user_by_id(user_id)
|
|
assert u is not None
|
|
if u["user_type"] in (UserTypes.SUPPORT, UserTypes.BOT):
|
|
# support and bot users are not required to consent
|
|
return
|
|
if u["appservice_id"] is not None:
|
|
# users registered by an appservice are exempt
|
|
return
|
|
if u["consent_version"] == self.config.user_consent_version:
|
|
return
|
|
|
|
consent_uri = self._consent_uri_builder.build_user_consent_uri(
|
|
requester.user.localpart
|
|
)
|
|
msg = self._block_events_without_consent_error % {"consent_uri": consent_uri}
|
|
raise ConsentNotGivenError(msg=msg, consent_uri=consent_uri)
|
|
|
|
@defer.inlineCallbacks
|
|
def send_nonmember_event(self, requester, event, context, ratelimit=True):
|
|
"""
|
|
Persists and notifies local clients and federation of an event.
|
|
|
|
Args:
|
|
event (FrozenEvent) the event to send.
|
|
context (Context) the context of the event.
|
|
ratelimit (bool): Whether to rate limit this send.
|
|
is_guest (bool): Whether the sender is a guest.
|
|
"""
|
|
if event.type == EventTypes.Member:
|
|
raise SynapseError(
|
|
500, "Tried to send member event through non-member codepath"
|
|
)
|
|
|
|
user = UserID.from_string(event.sender)
|
|
|
|
assert self.hs.is_mine(user), "User must be our own: %s" % (user,)
|
|
|
|
if event.is_state():
|
|
prev_state = yield self.deduplicate_state_event(event, context)
|
|
if prev_state is not None:
|
|
logger.info(
|
|
"Not bothering to persist state event %s duplicated by %s",
|
|
event.event_id,
|
|
prev_state.event_id,
|
|
)
|
|
return prev_state
|
|
|
|
yield self.handle_new_client_event(
|
|
requester=requester, event=event, context=context, ratelimit=ratelimit
|
|
)
|
|
|
|
@defer.inlineCallbacks
|
|
def deduplicate_state_event(self, event, context):
|
|
"""
|
|
Checks whether event is in the latest resolved state in context.
|
|
|
|
If so, returns the version of the event in context.
|
|
Otherwise, returns None.
|
|
"""
|
|
prev_state_ids = yield context.get_prev_state_ids(self.store)
|
|
prev_event_id = prev_state_ids.get((event.type, event.state_key))
|
|
if not prev_event_id:
|
|
return
|
|
prev_event = yield self.store.get_event(prev_event_id, allow_none=True)
|
|
if not prev_event:
|
|
return
|
|
|
|
if prev_event and event.user_id == prev_event.user_id:
|
|
prev_content = encode_canonical_json(prev_event.content)
|
|
next_content = encode_canonical_json(event.content)
|
|
if prev_content == next_content:
|
|
return prev_event
|
|
return
|
|
|
|
@defer.inlineCallbacks
|
|
def create_and_send_nonmember_event(
|
|
self, requester, event_dict, ratelimit=True, txn_id=None
|
|
):
|
|
"""
|
|
Creates an event, then sends it.
|
|
|
|
See self.create_event and self.send_nonmember_event.
|
|
"""
|
|
|
|
# We limit the number of concurrent event sends in a room so that we
|
|
# don't fork the DAG too much. If we don't limit then we can end up in
|
|
# a situation where event persistence can't keep up, causing
|
|
# extremities to pile up, which in turn leads to state resolution
|
|
# taking longer.
|
|
with (yield self.limiter.queue(event_dict["room_id"])):
|
|
event, context = yield self.create_event(
|
|
requester, event_dict, token_id=requester.access_token_id, txn_id=txn_id
|
|
)
|
|
|
|
spam_error = self.spam_checker.check_event_for_spam(event)
|
|
if spam_error:
|
|
if not isinstance(spam_error, string_types):
|
|
spam_error = "Spam is not permitted here"
|
|
raise SynapseError(403, spam_error, Codes.FORBIDDEN)
|
|
|
|
yield self.send_nonmember_event(
|
|
requester, event, context, ratelimit=ratelimit
|
|
)
|
|
return event
|
|
|
|
@measure_func("create_new_client_event")
|
|
@defer.inlineCallbacks
|
|
def create_new_client_event(
|
|
self, builder, requester=None, prev_events_and_hashes=None
|
|
):
|
|
"""Create a new event for a local client
|
|
|
|
Args:
|
|
builder (EventBuilder):
|
|
|
|
requester (synapse.types.Requester|None):
|
|
|
|
prev_events_and_hashes (list[(str, dict[str, str], int)]|None):
|
|
the forward extremities to use as the prev_events for the
|
|
new event. For each event, a tuple of (event_id, hashes, depth)
|
|
where *hashes* is a map from algorithm to hash.
|
|
|
|
If None, they will be requested from the database.
|
|
|
|
Returns:
|
|
Deferred[(synapse.events.EventBase, synapse.events.snapshot.EventContext)]
|
|
"""
|
|
|
|
if prev_events_and_hashes is not None:
|
|
assert len(prev_events_and_hashes) <= 10, (
|
|
"Attempting to create an event with %i prev_events"
|
|
% (len(prev_events_and_hashes),)
|
|
)
|
|
else:
|
|
prev_events_and_hashes = yield self.store.get_prev_events_for_room(
|
|
builder.room_id
|
|
)
|
|
|
|
prev_events = [
|
|
(event_id, prev_hashes)
|
|
for event_id, prev_hashes, _ in prev_events_and_hashes
|
|
]
|
|
|
|
event = yield builder.build(prev_event_ids=[p for p, _ in prev_events])
|
|
context = yield self.state.compute_event_context(event)
|
|
if requester:
|
|
context.app_service = requester.app_service
|
|
|
|
self.validator.validate_new(event, self.config)
|
|
|
|
# If this event is an annotation then we check that that the sender
|
|
# can't annotate the same way twice (e.g. stops users from liking an
|
|
# event multiple times).
|
|
relation = event.content.get("m.relates_to", {})
|
|
if relation.get("rel_type") == RelationTypes.ANNOTATION:
|
|
relates_to = relation["event_id"]
|
|
aggregation_key = relation["key"]
|
|
|
|
already_exists = yield self.store.has_user_annotated_event(
|
|
relates_to, event.type, aggregation_key, event.sender
|
|
)
|
|
if already_exists:
|
|
raise SynapseError(400, "Can't send same reaction twice")
|
|
|
|
logger.debug("Created event %s", event.event_id)
|
|
|
|
return (event, context)
|
|
|
|
@measure_func("handle_new_client_event")
|
|
@defer.inlineCallbacks
|
|
def handle_new_client_event(
|
|
self, requester, event, context, ratelimit=True, extra_users=[]
|
|
):
|
|
"""Processes a new event. This includes checking auth, persisting it,
|
|
notifying users, sending to remote servers, etc.
|
|
|
|
If called from a worker will hit out to the master process for final
|
|
processing.
|
|
|
|
Args:
|
|
requester (Requester)
|
|
event (FrozenEvent)
|
|
context (EventContext)
|
|
ratelimit (bool)
|
|
extra_users (list(UserID)): Any extra users to notify about event
|
|
"""
|
|
|
|
if event.is_state() and (event.type, event.state_key) == (
|
|
EventTypes.Create,
|
|
"",
|
|
):
|
|
room_version = event.content.get("room_version", RoomVersions.V1.identifier)
|
|
else:
|
|
room_version = yield self.store.get_room_version(event.room_id)
|
|
|
|
event_allowed = yield self.third_party_event_rules.check_event_allowed(
|
|
event, context
|
|
)
|
|
if not event_allowed:
|
|
raise SynapseError(
|
|
403, "This event is not allowed in this context", Codes.FORBIDDEN
|
|
)
|
|
|
|
try:
|
|
yield self.auth.check_from_context(room_version, event, context)
|
|
except AuthError as err:
|
|
logger.warning("Denying new event %r because %s", event, err)
|
|
raise err
|
|
|
|
# Ensure that we can round trip before trying to persist in db
|
|
try:
|
|
dump = frozendict_json_encoder.encode(event.content)
|
|
json.loads(dump)
|
|
except Exception:
|
|
logger.exception("Failed to encode content: %r", event.content)
|
|
raise
|
|
|
|
yield self.action_generator.handle_push_actions_for_event(event, context)
|
|
|
|
# reraise does not allow inlineCallbacks to preserve the stacktrace, so we
|
|
# hack around with a try/finally instead.
|
|
success = False
|
|
try:
|
|
# If we're a worker we need to hit out to the master.
|
|
if self.config.worker_app:
|
|
yield self.send_event_to_master(
|
|
event_id=event.event_id,
|
|
store=self.store,
|
|
requester=requester,
|
|
event=event,
|
|
context=context,
|
|
ratelimit=ratelimit,
|
|
extra_users=extra_users,
|
|
)
|
|
success = True
|
|
return
|
|
|
|
yield self.persist_and_notify_client_event(
|
|
requester, event, context, ratelimit=ratelimit, extra_users=extra_users
|
|
)
|
|
|
|
success = True
|
|
finally:
|
|
if not success:
|
|
# Ensure that we actually remove the entries in the push actions
|
|
# staging area, if we calculated them.
|
|
run_in_background(
|
|
self.store.remove_push_actions_from_staging, event.event_id
|
|
)
|
|
|
|
@defer.inlineCallbacks
|
|
def persist_and_notify_client_event(
|
|
self, requester, event, context, ratelimit=True, extra_users=[]
|
|
):
|
|
"""Called when we have fully built the event, have already
|
|
calculated the push actions for the event, and checked auth.
|
|
|
|
This should only be run on master.
|
|
"""
|
|
assert not self.config.worker_app
|
|
|
|
if ratelimit:
|
|
# We check if this is a room admin redacting an event so that we
|
|
# can apply different ratelimiting. We do this by simply checking
|
|
# it's not a self-redaction (to avoid having to look up whether the
|
|
# user is actually admin or not).
|
|
is_admin_redaction = False
|
|
if event.type == EventTypes.Redaction:
|
|
original_event = yield self.store.get_event(
|
|
event.redacts,
|
|
check_redacted=False,
|
|
get_prev_content=False,
|
|
allow_rejected=False,
|
|
allow_none=True,
|
|
)
|
|
|
|
is_admin_redaction = (
|
|
original_event and event.sender != original_event.sender
|
|
)
|
|
|
|
yield self.base_handler.ratelimit(
|
|
requester, is_admin_redaction=is_admin_redaction
|
|
)
|
|
|
|
yield self.base_handler.maybe_kick_guest_users(event, context)
|
|
|
|
if event.type == EventTypes.CanonicalAlias:
|
|
# Check the alias is acually valid (at this time at least)
|
|
room_alias_str = event.content.get("alias", None)
|
|
if room_alias_str:
|
|
room_alias = RoomAlias.from_string(room_alias_str)
|
|
directory_handler = self.hs.get_handlers().directory_handler
|
|
mapping = yield directory_handler.get_association(room_alias)
|
|
|
|
if mapping["room_id"] != event.room_id:
|
|
raise SynapseError(
|
|
400,
|
|
"Room alias %s does not point to the room" % (room_alias_str,),
|
|
)
|
|
|
|
federation_handler = self.hs.get_handlers().federation_handler
|
|
|
|
if event.type == EventTypes.Member:
|
|
if event.content["membership"] == Membership.INVITE:
|
|
|
|
def is_inviter_member_event(e):
|
|
return e.type == EventTypes.Member and e.sender == event.sender
|
|
|
|
current_state_ids = yield context.get_current_state_ids(self.store)
|
|
|
|
state_to_include_ids = [
|
|
e_id
|
|
for k, e_id in iteritems(current_state_ids)
|
|
if k[0] in self.hs.config.room_invite_state_types
|
|
or k == (EventTypes.Member, event.sender)
|
|
]
|
|
|
|
state_to_include = yield self.store.get_events(state_to_include_ids)
|
|
|
|
event.unsigned["invite_room_state"] = [
|
|
{
|
|
"type": e.type,
|
|
"state_key": e.state_key,
|
|
"content": e.content,
|
|
"sender": e.sender,
|
|
}
|
|
for e in itervalues(state_to_include)
|
|
]
|
|
|
|
invitee = UserID.from_string(event.state_key)
|
|
if not self.hs.is_mine(invitee):
|
|
# TODO: Can we add signature from remote server in a nicer
|
|
# way? If we have been invited by a remote server, we need
|
|
# to get them to sign the event.
|
|
|
|
returned_invite = yield federation_handler.send_invite(
|
|
invitee.domain, event
|
|
)
|
|
|
|
event.unsigned.pop("room_state", None)
|
|
|
|
# TODO: Make sure the signatures actually are correct.
|
|
event.signatures.update(returned_invite.signatures)
|
|
|
|
if event.type == EventTypes.Redaction:
|
|
original_event = yield self.store.get_event(
|
|
event.redacts,
|
|
check_redacted=False,
|
|
get_prev_content=False,
|
|
allow_rejected=False,
|
|
allow_none=True,
|
|
)
|
|
|
|
# we can make some additional checks now if we have the original event.
|
|
if original_event:
|
|
if original_event.type == EventTypes.Create:
|
|
raise AuthError(403, "Redacting create events is not permitted")
|
|
|
|
if original_event.room_id != event.room_id:
|
|
raise SynapseError(400, "Cannot redact event from a different room")
|
|
|
|
prev_state_ids = yield context.get_prev_state_ids(self.store)
|
|
auth_events_ids = yield self.auth.compute_auth_events(
|
|
event, prev_state_ids, for_verification=True
|
|
)
|
|
auth_events = yield self.store.get_events(auth_events_ids)
|
|
auth_events = {(e.type, e.state_key): e for e in auth_events.values()}
|
|
room_version = yield self.store.get_room_version(event.room_id)
|
|
|
|
if event_auth.check_redaction(room_version, event, auth_events=auth_events):
|
|
# this user doesn't have 'redact' rights, so we need to do some more
|
|
# checks on the original event. Let's start by checking the original
|
|
# event exists.
|
|
if not original_event:
|
|
raise NotFoundError("Could not find event %s" % (event.redacts,))
|
|
|
|
if event.user_id != original_event.user_id:
|
|
raise AuthError(403, "You don't have permission to redact events")
|
|
|
|
# all the checks are done.
|
|
event.internal_metadata.recheck_redaction = False
|
|
|
|
if event.type == EventTypes.Create:
|
|
prev_state_ids = yield context.get_prev_state_ids(self.store)
|
|
if prev_state_ids:
|
|
raise AuthError(403, "Changing the room create event is forbidden")
|
|
|
|
event_stream_id, max_stream_id = yield self.storage.persistence.persist_event(
|
|
event, context=context
|
|
)
|
|
|
|
if self._ephemeral_events_enabled:
|
|
# If there's an expiry timestamp on the event, schedule its expiry.
|
|
self._message_handler.maybe_schedule_expiry(event)
|
|
|
|
yield self.pusher_pool.on_new_notifications(event_stream_id, max_stream_id)
|
|
|
|
def _notify():
|
|
try:
|
|
self.notifier.on_new_room_event(
|
|
event, event_stream_id, max_stream_id, extra_users=extra_users
|
|
)
|
|
except Exception:
|
|
logger.exception("Error notifying about new room event")
|
|
|
|
run_in_background(_notify)
|
|
|
|
if event.type == EventTypes.Message:
|
|
# We don't want to block sending messages on any presence code. This
|
|
# matters as sometimes presence code can take a while.
|
|
run_in_background(self._bump_active_time, requester.user)
|
|
|
|
@defer.inlineCallbacks
|
|
def _bump_active_time(self, user):
|
|
try:
|
|
presence = self.hs.get_presence_handler()
|
|
yield presence.bump_presence_active_time(user)
|
|
except Exception:
|
|
logger.exception("Error bumping presence active time")
|
|
|
|
@defer.inlineCallbacks
|
|
def _send_dummy_events_to_fill_extremities(self):
|
|
"""Background task to send dummy events into rooms that have a large
|
|
number of extremities
|
|
"""
|
|
self._expire_rooms_to_exclude_from_dummy_event_insertion()
|
|
room_ids = yield self.store.get_rooms_with_many_extremities(
|
|
min_count=10,
|
|
limit=5,
|
|
room_id_filter=self._rooms_to_exclude_from_dummy_event_insertion.keys(),
|
|
)
|
|
|
|
for room_id in room_ids:
|
|
# For each room we need to find a joined member we can use to send
|
|
# the dummy event with.
|
|
|
|
prev_events_and_hashes = yield self.store.get_prev_events_for_room(room_id)
|
|
|
|
latest_event_ids = (event_id for (event_id, _, _) in prev_events_and_hashes)
|
|
|
|
members = yield self.state.get_current_users_in_room(
|
|
room_id, latest_event_ids=latest_event_ids
|
|
)
|
|
dummy_event_sent = False
|
|
for user_id in members:
|
|
if not self.hs.is_mine_id(user_id):
|
|
continue
|
|
requester = create_requester(user_id)
|
|
try:
|
|
event, context = yield self.create_event(
|
|
requester,
|
|
{
|
|
"type": "org.matrix.dummy_event",
|
|
"content": {},
|
|
"room_id": room_id,
|
|
"sender": user_id,
|
|
},
|
|
prev_events_and_hashes=prev_events_and_hashes,
|
|
)
|
|
|
|
event.internal_metadata.proactively_send = False
|
|
|
|
yield self.send_nonmember_event(
|
|
requester, event, context, ratelimit=False
|
|
)
|
|
dummy_event_sent = True
|
|
break
|
|
except ConsentNotGivenError:
|
|
logger.info(
|
|
"Failed to send dummy event into room %s for user %s due to "
|
|
"lack of consent. Will try another user" % (room_id, user_id)
|
|
)
|
|
except AuthError:
|
|
logger.info(
|
|
"Failed to send dummy event into room %s for user %s due to "
|
|
"lack of power. Will try another user" % (room_id, user_id)
|
|
)
|
|
|
|
if not dummy_event_sent:
|
|
# Did not find a valid user in the room, so remove from future attempts
|
|
# Exclusion is time limited, so the room will be rechecked in the future
|
|
# dependent on _DUMMY_EVENT_ROOM_EXCLUSION_EXPIRY
|
|
logger.info(
|
|
"Failed to send dummy event into room %s. Will exclude it from "
|
|
"future attempts until cache expires" % (room_id,)
|
|
)
|
|
now = self.clock.time_msec()
|
|
self._rooms_to_exclude_from_dummy_event_insertion[room_id] = now
|
|
|
|
def _expire_rooms_to_exclude_from_dummy_event_insertion(self):
|
|
expire_before = self.clock.time_msec() - _DUMMY_EVENT_ROOM_EXCLUSION_EXPIRY
|
|
to_expire = set()
|
|
for room_id, time in self._rooms_to_exclude_from_dummy_event_insertion.items():
|
|
if time < expire_before:
|
|
to_expire.add(room_id)
|
|
for room_id in to_expire:
|
|
logger.debug(
|
|
"Expiring room id %s from dummy event insertion exclusion cache",
|
|
room_id,
|
|
)
|
|
del self._rooms_to_exclude_from_dummy_event_insertion[room_id]
|