2017-05-31 06:51:01 -04:00
|
|
|
# -*- coding: utf-8 -*-
|
|
|
|
# Copyright 2017 Vector Creations Ltd
|
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
|
|
|
import logging
|
2018-07-09 02:09:20 -04:00
|
|
|
|
2019-03-07 04:22:53 -05:00
|
|
|
from six import iteritems, iterkeys
|
2018-07-09 02:09:20 -04:00
|
|
|
|
2017-05-31 06:51:01 -04:00
|
|
|
from twisted.internet import defer
|
|
|
|
|
2019-01-23 10:01:09 -05:00
|
|
|
import synapse.metrics
|
2017-05-31 06:51:01 -04:00
|
|
|
from synapse.api.constants import EventTypes, JoinRules, Membership
|
2018-10-23 09:29:17 -04:00
|
|
|
from synapse.metrics.background_process_metrics import run_as_background_process
|
2017-05-31 06:51:01 -04:00
|
|
|
from synapse.storage.roommember import ProfileInfo
|
2017-11-29 20:17:15 -05:00
|
|
|
from synapse.types import get_localpart_from_id
|
2018-07-09 02:09:20 -04:00
|
|
|
from synapse.util.metrics import Measure
|
2017-05-31 06:51:01 -04:00
|
|
|
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
2017-11-29 11:46:45 -05:00
|
|
|
class UserDirectoryHandler(object):
|
2017-05-31 10:00:29 -04:00
|
|
|
"""Handles querying of and keeping updated the user_directory.
|
|
|
|
|
|
|
|
N.B.: ASSUMES IT IS THE ONLY THING THAT MODIFIES THE USER DIRECTORY
|
2017-05-31 12:11:18 -04:00
|
|
|
|
|
|
|
The user directory is filled with users who this server can see are joined to a
|
|
|
|
world_readable or publically joinable room. We keep a database table up to date
|
|
|
|
by streaming changes of the current state and recalculating whether users should
|
|
|
|
be in the directory or not when necessary.
|
|
|
|
|
|
|
|
For each user in the directory we also store a room_id which is public and that the
|
|
|
|
user is joined to. This allows us to ignore history_visibility and join_rules changes
|
|
|
|
for that user in all other public rooms, as we know they'll still be in at least
|
|
|
|
one public room.
|
2017-05-31 10:00:29 -04:00
|
|
|
"""
|
|
|
|
|
2017-12-04 13:05:28 -05:00
|
|
|
INITIAL_ROOM_SLEEP_MS = 50
|
|
|
|
INITIAL_ROOM_SLEEP_COUNT = 100
|
|
|
|
INITIAL_ROOM_BATCH_SIZE = 100
|
|
|
|
INITIAL_USER_SLEEP_MS = 10
|
2017-06-15 04:59:04 -04:00
|
|
|
|
2017-05-31 06:51:01 -04:00
|
|
|
def __init__(self, hs):
|
|
|
|
self.store = hs.get_datastore()
|
|
|
|
self.state = hs.get_state_handler()
|
|
|
|
self.server_name = hs.hostname
|
|
|
|
self.clock = hs.get_clock()
|
2017-06-07 07:02:53 -04:00
|
|
|
self.notifier = hs.get_notifier()
|
2017-06-15 04:59:04 -04:00
|
|
|
self.is_mine_id = hs.is_mine_id
|
2017-06-15 07:47:05 -04:00
|
|
|
self.update_user_directory = hs.config.update_user_directory
|
2017-12-04 09:58:39 -05:00
|
|
|
self.search_all_users = hs.config.user_directory_search_all_users
|
2017-05-31 06:51:01 -04:00
|
|
|
|
2019-03-11 09:35:31 -04:00
|
|
|
# If we're a worker, don't sleep when doing the initial room work, as it
|
|
|
|
# won't monopolise the master's CPU.
|
|
|
|
if hs.config.worker_app:
|
|
|
|
self.INITIAL_ROOM_SLEEP_MS = 0
|
|
|
|
self.INITIAL_USER_SLEEP_MS = 0
|
|
|
|
|
2017-05-31 10:00:29 -04:00
|
|
|
# When start up for the first time we need to populate the user_directory.
|
|
|
|
# This is a set of user_id's we've inserted already
|
2017-05-31 06:51:01 -04:00
|
|
|
self.initially_handled_users = set()
|
2017-06-15 04:59:04 -04:00
|
|
|
|
2017-05-31 10:00:29 -04:00
|
|
|
# The current position in the current_state_delta stream
|
2017-05-31 06:51:01 -04:00
|
|
|
self.pos = None
|
|
|
|
|
2017-05-31 10:00:29 -04:00
|
|
|
# Guard to ensure we only process deltas one at a time
|
2017-05-31 06:51:01 -04:00
|
|
|
self._is_processing = False
|
|
|
|
|
2017-06-15 07:47:05 -04:00
|
|
|
if self.update_user_directory:
|
|
|
|
self.notifier.add_replication_callback(self.notify_new_event)
|
|
|
|
|
|
|
|
# We kick this off so that we don't have to wait for a change before
|
|
|
|
# we start populating the user directory
|
|
|
|
self.clock.call_later(0, self.notify_new_event)
|
2017-05-31 06:59:36 -04:00
|
|
|
|
2017-06-15 05:00:28 -04:00
|
|
|
def search_users(self, user_id, search_term, limit):
|
2017-05-31 10:00:29 -04:00
|
|
|
"""Searches for users in directory
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
dict of the form::
|
|
|
|
|
|
|
|
{
|
|
|
|
"limited": <bool>, # whether there were more results or not
|
|
|
|
"results": [ # Ordered by best match first
|
|
|
|
{
|
|
|
|
"user_id": <user_id>,
|
|
|
|
"display_name": <display_name>,
|
|
|
|
"avatar_url": <avatar_url>
|
|
|
|
}
|
|
|
|
]
|
|
|
|
}
|
|
|
|
"""
|
2017-06-15 05:00:28 -04:00
|
|
|
return self.store.search_user_dir(user_id, search_term, limit)
|
2017-05-31 09:00:01 -04:00
|
|
|
|
2017-05-31 06:51:01 -04:00
|
|
|
def notify_new_event(self):
|
2017-05-31 10:00:29 -04:00
|
|
|
"""Called when there may be more deltas to process
|
|
|
|
"""
|
2017-06-15 07:47:05 -04:00
|
|
|
if not self.update_user_directory:
|
|
|
|
return
|
|
|
|
|
2017-05-31 06:51:01 -04:00
|
|
|
if self._is_processing:
|
|
|
|
return
|
|
|
|
|
2018-10-23 09:29:17 -04:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def process():
|
|
|
|
try:
|
|
|
|
yield self._unsafe_process()
|
|
|
|
finally:
|
|
|
|
self._is_processing = False
|
|
|
|
|
2017-05-31 06:51:01 -04:00
|
|
|
self._is_processing = True
|
2018-10-23 09:29:17 -04:00
|
|
|
run_as_background_process("user_directory.notify_new_event", process)
|
2017-05-31 06:51:01 -04:00
|
|
|
|
2017-11-29 13:27:05 -05:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def handle_local_profile_change(self, user_id, profile):
|
|
|
|
"""Called to update index of our local user profiles when they change
|
|
|
|
irrespective of any rooms the user may be in.
|
|
|
|
"""
|
2018-08-17 10:30:31 -04:00
|
|
|
# FIXME(#3714): We should probably do this in the same worker as all
|
|
|
|
# the other changes.
|
2018-12-14 13:20:59 -05:00
|
|
|
is_support = yield self.store.is_support_user(user_id)
|
|
|
|
# Support users are for diagnostics and should not appear in the user directory.
|
|
|
|
if not is_support:
|
|
|
|
yield self.store.update_profile_in_user_dir(
|
2019-02-13 07:05:32 -05:00
|
|
|
user_id, profile.display_name, profile.avatar_url, None
|
2018-12-14 13:20:59 -05:00
|
|
|
)
|
2017-11-29 13:27:05 -05:00
|
|
|
|
2018-05-24 10:59:58 -04:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def handle_user_deactivated(self, user_id):
|
|
|
|
"""Called when a user ID is deactivated
|
|
|
|
"""
|
2018-08-17 10:30:31 -04:00
|
|
|
# FIXME(#3714): We should probably do this in the same worker as all
|
|
|
|
# the other changes.
|
2018-05-24 10:59:58 -04:00
|
|
|
yield self.store.remove_from_user_dir(user_id)
|
|
|
|
|
2017-05-31 06:51:01 -04:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def _unsafe_process(self):
|
2017-05-31 10:00:29 -04:00
|
|
|
# If self.pos is None then means we haven't fetched it from DB
|
2017-05-31 06:51:01 -04:00
|
|
|
if self.pos is None:
|
|
|
|
self.pos = yield self.store.get_user_directory_stream_pos()
|
|
|
|
|
2017-05-31 10:00:29 -04:00
|
|
|
# If still None then we need to do the initial fill of directory
|
2017-05-31 06:51:01 -04:00
|
|
|
if self.pos is None:
|
|
|
|
yield self._do_initial_spam()
|
|
|
|
self.pos = yield self.store.get_user_directory_stream_pos()
|
|
|
|
|
2017-05-31 10:00:29 -04:00
|
|
|
# Loop round handling deltas until we're up to date
|
2017-05-31 06:51:01 -04:00
|
|
|
while True:
|
|
|
|
with Measure(self.clock, "user_dir_delta"):
|
|
|
|
deltas = yield self.store.get_current_state_deltas(self.pos)
|
|
|
|
if not deltas:
|
|
|
|
return
|
|
|
|
|
2017-06-22 09:59:52 -04:00
|
|
|
logger.info("Handling %d state deltas", len(deltas))
|
2017-05-31 06:51:01 -04:00
|
|
|
yield self._handle_deltas(deltas)
|
|
|
|
|
2017-05-31 06:55:13 -04:00
|
|
|
self.pos = deltas[-1]["stream_id"]
|
2019-01-23 10:01:09 -05:00
|
|
|
|
|
|
|
# Expose current event processing position to prometheus
|
2019-02-13 07:05:32 -05:00
|
|
|
synapse.metrics.event_processing_positions.labels("user_dir").set(
|
|
|
|
self.pos
|
|
|
|
)
|
2019-01-23 10:01:09 -05:00
|
|
|
|
2017-05-31 06:55:13 -04:00
|
|
|
yield self.store.update_user_directory_stream_pos(self.pos)
|
2017-05-31 06:51:01 -04:00
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def _do_initial_spam(self):
|
2017-05-31 10:00:29 -04:00
|
|
|
"""Populates the user_directory from the current state of the DB, used
|
|
|
|
when synapse first starts with user_directory support
|
|
|
|
"""
|
2017-05-31 10:13:49 -04:00
|
|
|
new_pos = yield self.store.get_max_stream_id_in_current_state_deltas()
|
2017-05-31 06:56:27 -04:00
|
|
|
|
2017-05-31 10:00:29 -04:00
|
|
|
# Delete any existing entries just in case there are any
|
2017-05-31 06:51:01 -04:00
|
|
|
yield self.store.delete_all_from_user_dir()
|
|
|
|
|
2017-05-31 10:00:29 -04:00
|
|
|
# We process by going through each existing room at a time.
|
2017-05-31 06:51:01 -04:00
|
|
|
room_ids = yield self.store.get_all_rooms()
|
|
|
|
|
2017-06-07 06:50:44 -04:00
|
|
|
logger.info("Doing initial update of user directory. %d rooms", len(room_ids))
|
2017-11-29 13:27:05 -05:00
|
|
|
num_processed_rooms = 0
|
2017-06-07 06:50:44 -04:00
|
|
|
|
2017-05-31 06:51:01 -04:00
|
|
|
for room_id in room_ids:
|
2017-12-04 10:11:38 -05:00
|
|
|
logger.info("Handling room %d/%d", num_processed_rooms + 1, len(room_ids))
|
2017-11-04 15:35:14 -04:00
|
|
|
yield self._handle_initial_room(room_id)
|
2017-06-07 06:50:44 -04:00
|
|
|
num_processed_rooms += 1
|
2019-02-13 07:05:32 -05:00
|
|
|
yield self.clock.sleep(self.INITIAL_ROOM_SLEEP_MS / 1000.0)
|
2017-06-07 06:50:44 -04:00
|
|
|
|
|
|
|
logger.info("Processed all rooms.")
|
2017-05-31 06:51:01 -04:00
|
|
|
|
2017-12-04 09:58:39 -05:00
|
|
|
if self.search_all_users:
|
2017-11-29 13:27:05 -05:00
|
|
|
num_processed_users = 0
|
|
|
|
user_ids = yield self.store.get_all_local_users()
|
2019-02-13 07:05:32 -05:00
|
|
|
logger.info(
|
|
|
|
"Doing initial update of user directory. %d users", len(user_ids)
|
|
|
|
)
|
2017-11-29 13:27:05 -05:00
|
|
|
for user_id in user_ids:
|
|
|
|
# We add profiles for all users even if they don't match the
|
|
|
|
# include pattern, just in case we want to change it in future
|
2019-02-13 07:05:32 -05:00
|
|
|
logger.info(
|
|
|
|
"Handling user %d/%d", num_processed_users + 1, len(user_ids)
|
|
|
|
)
|
2017-11-29 13:27:05 -05:00
|
|
|
yield self._handle_local_user(user_id)
|
|
|
|
num_processed_users += 1
|
2019-02-13 07:05:32 -05:00
|
|
|
yield self.clock.sleep(self.INITIAL_USER_SLEEP_MS / 1000.0)
|
2017-11-29 13:27:05 -05:00
|
|
|
|
|
|
|
logger.info("Processed all users")
|
|
|
|
|
2017-05-31 06:51:01 -04:00
|
|
|
self.initially_handled_users = None
|
|
|
|
|
2017-05-31 06:56:27 -04:00
|
|
|
yield self.store.update_user_directory_stream_pos(new_pos)
|
2017-05-31 06:51:01 -04:00
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
2017-11-04 15:35:14 -04:00
|
|
|
def _handle_initial_room(self, room_id):
|
2019-03-07 04:22:53 -05:00
|
|
|
"""
|
|
|
|
Called when we initially fill out user_directory one room at a time
|
2017-05-31 10:00:29 -04:00
|
|
|
"""
|
2017-06-09 05:52:26 -04:00
|
|
|
is_in_room = yield self.store.is_host_joined(room_id, self.server_name)
|
2017-05-31 10:11:36 -04:00
|
|
|
if not is_in_room:
|
|
|
|
return
|
2017-05-31 06:51:01 -04:00
|
|
|
|
2019-02-13 07:05:32 -05:00
|
|
|
is_public = yield self.store.is_room_world_readable_or_publicly_joinable(
|
|
|
|
room_id
|
|
|
|
)
|
2017-05-31 06:51:01 -04:00
|
|
|
|
2017-05-31 10:00:29 -04:00
|
|
|
users_with_profile = yield self.state.get_current_user_in_room(room_id)
|
2017-06-15 04:59:04 -04:00
|
|
|
user_ids = set(users_with_profile)
|
|
|
|
unhandled_users = user_ids - self.initially_handled_users
|
2017-05-31 06:51:01 -04:00
|
|
|
|
2017-05-31 10:00:29 -04:00
|
|
|
yield self.store.add_profiles_to_user_dir(
|
2019-03-11 06:11:36 -04:00
|
|
|
{user_id: users_with_profile[user_id] for user_id in unhandled_users}
|
2017-05-31 10:00:29 -04:00
|
|
|
)
|
2017-05-31 06:51:01 -04:00
|
|
|
|
2017-05-31 10:00:29 -04:00
|
|
|
self.initially_handled_users |= unhandled_users
|
2017-05-31 06:51:01 -04:00
|
|
|
|
2017-06-15 04:59:04 -04:00
|
|
|
# We now go and figure out the new users who share rooms with user entries
|
|
|
|
# We sleep aggressively here as otherwise it can starve resources.
|
|
|
|
# We also batch up inserts/updates, but try to avoid too many at once.
|
|
|
|
to_insert = set()
|
|
|
|
count = 0
|
|
|
|
|
2019-03-12 06:47:14 -04:00
|
|
|
if is_public:
|
|
|
|
for user_id in user_ids:
|
|
|
|
if count % self.INITIAL_ROOM_SLEEP_COUNT == 0:
|
|
|
|
yield self.clock.sleep(self.INITIAL_ROOM_SLEEP_MS / 1000.0)
|
2017-06-21 07:00:41 -04:00
|
|
|
|
2019-03-12 06:47:14 -04:00
|
|
|
if self.store.get_if_app_services_interested_in_user(user_id):
|
|
|
|
count += 1
|
2017-06-15 04:59:04 -04:00
|
|
|
continue
|
|
|
|
|
2019-03-12 06:47:14 -04:00
|
|
|
to_insert.add(user_id)
|
|
|
|
if len(to_insert) > self.INITIAL_ROOM_BATCH_SIZE:
|
|
|
|
yield self.store.add_users_in_public_rooms(room_id, to_insert)
|
|
|
|
to_insert.clear()
|
|
|
|
|
|
|
|
if to_insert:
|
|
|
|
yield self.store.add_users_in_public_rooms(room_id, to_insert)
|
|
|
|
to_insert.clear()
|
|
|
|
else:
|
|
|
|
|
|
|
|
for user_id in user_ids:
|
2017-12-04 13:05:28 -05:00
|
|
|
if count % self.INITIAL_ROOM_SLEEP_COUNT == 0:
|
2019-02-13 07:05:32 -05:00
|
|
|
yield self.clock.sleep(self.INITIAL_ROOM_SLEEP_MS / 1000.0)
|
2017-06-15 04:59:04 -04:00
|
|
|
|
2019-03-12 06:47:14 -04:00
|
|
|
if not self.is_mine_id(user_id):
|
|
|
|
count += 1
|
|
|
|
continue
|
2017-06-15 04:59:04 -04:00
|
|
|
|
2019-03-12 06:47:14 -04:00
|
|
|
if self.store.get_if_app_services_interested_in_user(user_id):
|
|
|
|
count += 1
|
|
|
|
continue
|
2017-06-15 04:59:04 -04:00
|
|
|
|
2019-03-12 06:47:14 -04:00
|
|
|
for other_user_id in user_ids:
|
|
|
|
if user_id == other_user_id:
|
|
|
|
continue
|
|
|
|
|
|
|
|
if count % self.INITIAL_ROOM_SLEEP_COUNT == 0:
|
|
|
|
yield self.clock.sleep(self.INITIAL_ROOM_SLEEP_MS / 1000.0)
|
|
|
|
count += 1
|
|
|
|
|
|
|
|
user_set = (user_id, other_user_id)
|
|
|
|
to_insert.add(user_set)
|
|
|
|
|
|
|
|
if len(to_insert) > self.INITIAL_ROOM_BATCH_SIZE:
|
|
|
|
yield self.store.add_users_who_share_private_room(
|
|
|
|
room_id, not is_public, to_insert
|
|
|
|
)
|
|
|
|
to_insert.clear()
|
|
|
|
|
|
|
|
if to_insert:
|
|
|
|
yield self.store.add_users_who_share_private_room(room_id, to_insert)
|
|
|
|
to_insert.clear()
|
2017-06-15 04:59:04 -04:00
|
|
|
|
2017-05-31 06:51:01 -04:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def _handle_deltas(self, deltas):
|
2017-05-31 10:00:29 -04:00
|
|
|
"""Called with the state deltas to process
|
|
|
|
"""
|
2017-05-31 06:51:01 -04:00
|
|
|
for delta in deltas:
|
|
|
|
typ = delta["type"]
|
|
|
|
state_key = delta["state_key"]
|
|
|
|
room_id = delta["room_id"]
|
|
|
|
event_id = delta["event_id"]
|
|
|
|
prev_event_id = delta["prev_event_id"]
|
|
|
|
|
2017-06-01 06:09:49 -04:00
|
|
|
logger.debug("Handling: %r %r, %s", typ, state_key, event_id)
|
|
|
|
|
2017-05-31 10:00:29 -04:00
|
|
|
# For join rule and visibility changes we need to check if the room
|
|
|
|
# may have become public or not and add/remove the users in said room
|
2017-05-31 12:00:24 -04:00
|
|
|
if typ in (EventTypes.RoomHistoryVisibility, EventTypes.JoinRules):
|
|
|
|
yield self._handle_room_publicity_change(
|
2019-02-13 07:05:32 -05:00
|
|
|
room_id, prev_event_id, event_id, typ
|
2017-05-31 08:30:12 -04:00
|
|
|
)
|
2017-05-31 06:51:01 -04:00
|
|
|
elif typ == EventTypes.Member:
|
|
|
|
change = yield self._get_key_change(
|
2019-02-13 07:05:32 -05:00
|
|
|
prev_event_id,
|
|
|
|
event_id,
|
2017-05-31 06:51:01 -04:00
|
|
|
key_name="membership",
|
|
|
|
public_value=Membership.JOIN,
|
|
|
|
)
|
|
|
|
|
2018-12-14 13:20:59 -05:00
|
|
|
if change is False:
|
2017-05-31 10:11:36 -04:00
|
|
|
# Need to check if the server left the room entirely, if so
|
|
|
|
# we might need to remove all the users in that room
|
2017-06-09 05:52:26 -04:00
|
|
|
is_in_room = yield self.store.is_host_joined(
|
2019-02-13 07:05:32 -05:00
|
|
|
room_id, self.server_name
|
2017-05-31 10:11:36 -04:00
|
|
|
)
|
|
|
|
if not is_in_room:
|
2017-06-22 09:59:52 -04:00
|
|
|
logger.info("Server left room: %r", room_id)
|
2017-05-31 10:11:36 -04:00
|
|
|
# Fetch all the users that we marked as being in user
|
|
|
|
# directory due to being in the room and then check if
|
|
|
|
# need to remove those users or not
|
2019-02-13 07:05:32 -05:00
|
|
|
user_ids = yield self.store.get_users_in_dir_due_to_room(
|
|
|
|
room_id
|
|
|
|
)
|
2019-03-07 04:22:53 -05:00
|
|
|
|
2017-05-31 10:11:36 -04:00
|
|
|
for user_id in user_ids:
|
|
|
|
yield self._handle_remove_user(room_id, user_id)
|
|
|
|
return
|
2017-06-01 08:05:39 -04:00
|
|
|
else:
|
|
|
|
logger.debug("Server is still in room: %r", room_id)
|
2017-05-31 10:11:36 -04:00
|
|
|
|
2018-12-14 13:20:59 -05:00
|
|
|
is_support = yield self.store.is_support_user(state_key)
|
|
|
|
if not is_support:
|
|
|
|
if change is None:
|
|
|
|
# Handle any profile changes
|
|
|
|
yield self._handle_profile_change(
|
2019-02-13 07:05:32 -05:00
|
|
|
state_key, room_id, prev_event_id, event_id
|
2018-12-14 13:20:59 -05:00
|
|
|
)
|
|
|
|
continue
|
|
|
|
|
|
|
|
if change: # The user joined
|
|
|
|
event = yield self.store.get_event(event_id, allow_none=True)
|
|
|
|
profile = ProfileInfo(
|
|
|
|
avatar_url=event.content.get("avatar_url"),
|
|
|
|
display_name=event.content.get("displayname"),
|
|
|
|
)
|
2017-05-31 06:51:01 -04:00
|
|
|
|
2018-12-14 13:20:59 -05:00
|
|
|
yield self._handle_new_user(room_id, state_key, profile)
|
|
|
|
else: # The user left
|
|
|
|
yield self._handle_remove_user(room_id, state_key)
|
2017-06-01 06:09:49 -04:00
|
|
|
else:
|
|
|
|
logger.debug("Ignoring irrelevant type: %r", typ)
|
2017-05-31 06:51:01 -04:00
|
|
|
|
2017-06-01 06:09:49 -04:00
|
|
|
@defer.inlineCallbacks
|
2017-05-31 12:00:24 -04:00
|
|
|
def _handle_room_publicity_change(self, room_id, prev_event_id, event_id, typ):
|
|
|
|
"""Handle a room having potentially changed from/to world_readable/publically
|
|
|
|
joinable.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
room_id (str)
|
|
|
|
prev_event_id (str|None): The previous event before the state change
|
|
|
|
event_id (str|None): The new event after the state change
|
|
|
|
typ (str): Type of the event
|
|
|
|
"""
|
2017-06-15 07:47:05 -04:00
|
|
|
logger.debug("Handling change for %s: %s", typ, room_id)
|
2017-06-01 06:09:49 -04:00
|
|
|
|
2017-05-31 12:00:24 -04:00
|
|
|
if typ == EventTypes.RoomHistoryVisibility:
|
|
|
|
change = yield self._get_key_change(
|
2019-02-13 07:05:32 -05:00
|
|
|
prev_event_id,
|
|
|
|
event_id,
|
2017-05-31 12:00:24 -04:00
|
|
|
key_name="history_visibility",
|
|
|
|
public_value="world_readable",
|
|
|
|
)
|
|
|
|
elif typ == EventTypes.JoinRules:
|
|
|
|
change = yield self._get_key_change(
|
2019-02-13 07:05:32 -05:00
|
|
|
prev_event_id,
|
|
|
|
event_id,
|
2017-06-01 06:09:49 -04:00
|
|
|
key_name="join_rule",
|
2017-05-31 12:00:24 -04:00
|
|
|
public_value=JoinRules.PUBLIC,
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
raise Exception("Invalid event type")
|
|
|
|
# If change is None, no change. True => become world_readable/public,
|
|
|
|
# False => was world_readable/public
|
|
|
|
if change is None:
|
2017-06-01 06:09:49 -04:00
|
|
|
logger.debug("No change")
|
2017-05-31 12:00:24 -04:00
|
|
|
return
|
|
|
|
|
|
|
|
# There's been a change to or from being world readable.
|
|
|
|
|
|
|
|
is_public = yield self.store.is_room_world_readable_or_publicly_joinable(
|
|
|
|
room_id
|
|
|
|
)
|
|
|
|
|
2017-06-01 06:09:49 -04:00
|
|
|
logger.debug("Change: %r, is_public: %r", change, is_public)
|
|
|
|
|
2017-05-31 12:00:24 -04:00
|
|
|
if change and not is_public:
|
|
|
|
# If we became world readable but room isn't currently public then
|
|
|
|
# we ignore the change
|
|
|
|
return
|
|
|
|
elif not change and is_public:
|
|
|
|
# If we stopped being world readable but are still public,
|
|
|
|
# ignore the change
|
|
|
|
return
|
|
|
|
|
2019-03-07 04:22:53 -05:00
|
|
|
users_with_profile = yield self.state.get_current_user_in_room(room_id)
|
|
|
|
|
|
|
|
# Remove every user from the sharing tables for that room.
|
|
|
|
for user_id in iterkeys(users_with_profile):
|
|
|
|
yield self.store.remove_user_who_share_room(user_id, room_id)
|
|
|
|
|
|
|
|
# Then, re-add them to the tables.
|
|
|
|
# NOTE: this is not the most efficient method, as handle_new_user sets
|
|
|
|
# up local_user -> other_user and other_user_whos_local -> local_user,
|
|
|
|
# which when ran over an entire room, will result in the same values
|
|
|
|
# being added multiple times. The batching upserts shouldn't make this
|
|
|
|
# too bad, though.
|
|
|
|
for user_id, profile in iteritems(users_with_profile):
|
|
|
|
yield self._handle_new_user(room_id, user_id, profile)
|
2017-05-31 12:00:24 -04:00
|
|
|
|
2017-11-29 13:27:05 -05:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def _handle_local_user(self, user_id):
|
|
|
|
"""Adds a new local roomless user into the user_directory_search table.
|
|
|
|
Used to populate up the user index when we have an
|
2017-12-04 09:58:39 -05:00
|
|
|
user_directory_search_all_users specified.
|
2017-11-29 13:27:05 -05:00
|
|
|
"""
|
|
|
|
logger.debug("Adding new local user to dir, %r", user_id)
|
|
|
|
|
2017-11-29 20:17:15 -05:00
|
|
|
profile = yield self.store.get_profileinfo(get_localpart_from_id(user_id))
|
2017-11-29 13:27:05 -05:00
|
|
|
|
|
|
|
row = yield self.store.get_user_in_directory(user_id)
|
|
|
|
if not row:
|
2019-03-07 04:22:53 -05:00
|
|
|
yield self.store.add_profiles_to_user_dir({user_id: profile})
|
2017-11-29 13:27:05 -05:00
|
|
|
|
2017-05-31 10:00:29 -04:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def _handle_new_user(self, room_id, user_id, profile):
|
|
|
|
"""Called when we might need to add user to directory
|
|
|
|
|
|
|
|
Args:
|
2017-11-29 11:46:45 -05:00
|
|
|
room_id (str): room_id that user joined or started being public
|
2017-05-31 10:00:29 -04:00
|
|
|
user_id (str)
|
|
|
|
"""
|
2017-11-29 13:27:05 -05:00
|
|
|
logger.debug("Adding new user to dir, %r", user_id)
|
2017-06-01 09:50:46 -04:00
|
|
|
|
2017-05-31 10:00:29 -04:00
|
|
|
row = yield self.store.get_user_in_directory(user_id)
|
2017-06-01 09:50:46 -04:00
|
|
|
if not row:
|
2019-03-07 04:22:53 -05:00
|
|
|
yield self.store.add_profiles_to_user_dir({user_id: profile})
|
2017-06-01 09:50:46 -04:00
|
|
|
|
|
|
|
is_public = yield self.store.is_room_world_readable_or_publicly_joinable(
|
|
|
|
room_id
|
|
|
|
)
|
2019-03-07 04:22:53 -05:00
|
|
|
# Now we update users who share rooms with users.
|
2017-06-15 05:15:00 -04:00
|
|
|
users_with_profile = yield self.state.get_current_user_in_room(room_id)
|
|
|
|
|
2019-03-12 06:47:14 -04:00
|
|
|
if is_public:
|
|
|
|
yield self.store.add_users_in_public_rooms(room_id, (user_id,))
|
|
|
|
else:
|
|
|
|
to_insert = set()
|
2017-06-21 07:00:41 -04:00
|
|
|
|
2019-03-12 06:47:14 -04:00
|
|
|
# First, if they're our user then we need to update for every user
|
|
|
|
if self.is_mine_id(user_id):
|
2017-06-15 05:15:00 -04:00
|
|
|
|
2019-03-12 06:47:14 -04:00
|
|
|
is_appservice = self.store.get_if_app_services_interested_in_user(user_id)
|
2019-03-07 04:22:53 -05:00
|
|
|
|
2019-03-12 06:47:14 -04:00
|
|
|
# We don't care about appservice users.
|
|
|
|
if not is_appservice:
|
|
|
|
for other_user_id in users_with_profile:
|
|
|
|
if user_id == other_user_id:
|
|
|
|
continue
|
2017-06-15 05:15:00 -04:00
|
|
|
|
2019-03-12 06:47:14 -04:00
|
|
|
to_insert.add((user_id, other_user_id))
|
2017-06-15 05:15:00 -04:00
|
|
|
|
2019-03-12 06:47:14 -04:00
|
|
|
# Next we need to update for every local user in the room
|
|
|
|
for other_user_id in users_with_profile:
|
|
|
|
if user_id == other_user_id:
|
|
|
|
continue
|
2017-06-15 05:15:00 -04:00
|
|
|
|
2019-03-12 06:47:14 -04:00
|
|
|
is_appservice = self.store.get_if_app_services_interested_in_user(
|
|
|
|
other_user_id
|
|
|
|
)
|
|
|
|
if self.is_mine_id(other_user_id) and not is_appservice:
|
|
|
|
to_insert.add((other_user_id, user_id))
|
2017-06-15 05:15:00 -04:00
|
|
|
|
2019-03-12 06:47:14 -04:00
|
|
|
if to_insert:
|
|
|
|
yield self.store.add_users_who_share_private_room(room_id, to_insert)
|
2017-06-15 05:15:00 -04:00
|
|
|
|
2017-06-01 06:36:50 -04:00
|
|
|
@defer.inlineCallbacks
|
2017-05-31 10:00:29 -04:00
|
|
|
def _handle_remove_user(self, room_id, user_id):
|
|
|
|
"""Called when we might need to remove user to directory
|
|
|
|
|
|
|
|
Args:
|
|
|
|
room_id (str): room_id that user left or stopped being public that
|
|
|
|
user_id (str)
|
|
|
|
"""
|
2019-03-07 04:22:53 -05:00
|
|
|
logger.debug("Removing user %r", user_id)
|
2017-05-31 10:00:29 -04:00
|
|
|
|
2019-03-07 04:22:53 -05:00
|
|
|
# Remove user from sharing tables
|
|
|
|
yield self.store.remove_user_who_share_room(user_id, room_id)
|
2017-06-01 09:50:46 -04:00
|
|
|
|
2019-03-12 06:47:14 -04:00
|
|
|
# Are they still in any rooms? If not, remove them entirely.
|
|
|
|
rooms_user_is_in = yield self.store.get_rooms_user_is_in(user_id)
|
2017-06-01 09:50:46 -04:00
|
|
|
|
2019-03-12 06:47:14 -04:00
|
|
|
if len(rooms_user_is_in) == 0:
|
2017-06-01 09:50:46 -04:00
|
|
|
yield self.store.remove_from_user_dir(user_id)
|
2017-06-15 05:15:00 -04:00
|
|
|
|
2017-06-01 10:39:51 -04:00
|
|
|
@defer.inlineCallbacks
|
2017-06-13 06:49:39 -04:00
|
|
|
def _handle_profile_change(self, user_id, room_id, prev_event_id, event_id):
|
2017-06-01 10:39:51 -04:00
|
|
|
"""Check member event changes for any profile changes and update the
|
|
|
|
database if there are.
|
|
|
|
"""
|
|
|
|
if not prev_event_id or not event_id:
|
|
|
|
return
|
|
|
|
|
2017-06-13 05:19:26 -04:00
|
|
|
prev_event = yield self.store.get_event(prev_event_id, allow_none=True)
|
|
|
|
event = yield self.store.get_event(event_id, allow_none=True)
|
|
|
|
|
|
|
|
if not prev_event or not event:
|
|
|
|
return
|
2017-06-01 10:39:51 -04:00
|
|
|
|
|
|
|
if event.membership != Membership.JOIN:
|
|
|
|
return
|
|
|
|
|
|
|
|
prev_name = prev_event.content.get("displayname")
|
|
|
|
new_name = event.content.get("displayname")
|
|
|
|
|
|
|
|
prev_avatar = prev_event.content.get("avatar_url")
|
|
|
|
new_avatar = event.content.get("avatar_url")
|
|
|
|
|
|
|
|
if prev_name != new_name or prev_avatar != new_avatar:
|
2017-06-13 06:49:39 -04:00
|
|
|
yield self.store.update_profile_in_user_dir(
|
2019-02-13 07:05:32 -05:00
|
|
|
user_id, new_name, new_avatar, room_id
|
2017-06-13 06:49:39 -04:00
|
|
|
)
|
2017-06-01 10:39:51 -04:00
|
|
|
|
2017-05-31 06:51:01 -04:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def _get_key_change(self, prev_event_id, event_id, key_name, public_value):
|
2017-05-31 10:00:29 -04:00
|
|
|
"""Given two events check if the `key_name` field in content changed
|
|
|
|
from not matching `public_value` to doing so.
|
|
|
|
|
|
|
|
For example, check if `history_visibility` (`key_name`) changed from
|
|
|
|
`shared` to `world_readable` (`public_value`).
|
|
|
|
|
|
|
|
Returns:
|
2017-05-31 11:50:57 -04:00
|
|
|
None if the field in the events either both match `public_value`
|
2017-06-01 08:11:38 -04:00
|
|
|
or if neither do, i.e. there has been no change.
|
2017-05-31 10:00:29 -04:00
|
|
|
True if it didnt match `public_value` but now does
|
2017-05-31 11:50:57 -04:00
|
|
|
False if it did match `public_value` but now doesn't
|
2017-05-31 10:00:29 -04:00
|
|
|
"""
|
2017-05-31 06:51:01 -04:00
|
|
|
prev_event = None
|
|
|
|
event = None
|
|
|
|
if prev_event_id:
|
|
|
|
prev_event = yield self.store.get_event(prev_event_id, allow_none=True)
|
|
|
|
|
|
|
|
if event_id:
|
|
|
|
event = yield self.store.get_event(event_id, allow_none=True)
|
|
|
|
|
|
|
|
if not event and not prev_event:
|
2017-06-01 06:09:49 -04:00
|
|
|
logger.debug("Neither event exists: %r %r", prev_event_id, event_id)
|
2017-05-31 06:51:01 -04:00
|
|
|
defer.returnValue(None)
|
|
|
|
|
2017-05-31 11:50:57 -04:00
|
|
|
prev_value = None
|
|
|
|
value = None
|
2017-05-31 06:51:01 -04:00
|
|
|
|
|
|
|
if prev_event:
|
2017-06-01 08:11:38 -04:00
|
|
|
prev_value = prev_event.content.get(key_name)
|
2017-05-31 06:51:01 -04:00
|
|
|
|
|
|
|
if event:
|
2017-06-01 08:11:38 -04:00
|
|
|
value = event.content.get(key_name)
|
2017-05-31 06:51:01 -04:00
|
|
|
|
2017-06-01 06:09:49 -04:00
|
|
|
logger.debug("prev_value: %r -> value: %r", prev_value, value)
|
|
|
|
|
2017-05-31 11:50:57 -04:00
|
|
|
if value == public_value and prev_value != public_value:
|
2017-05-31 06:51:01 -04:00
|
|
|
defer.returnValue(True)
|
2017-05-31 11:50:57 -04:00
|
|
|
elif value != public_value and prev_value == public_value:
|
2017-05-31 06:51:01 -04:00
|
|
|
defer.returnValue(False)
|
|
|
|
else:
|
|
|
|
defer.returnValue(None)
|