2017-05-31 06:51:01 -04:00
|
|
|
# -*- coding: utf-8 -*-
|
|
|
|
# Copyright 2017 Vector Creations Ltd
|
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
2018-07-09 02:09:20 -04:00
|
|
|
import logging
|
|
|
|
import re
|
2017-05-31 06:51:01 -04:00
|
|
|
|
2018-07-09 02:09:20 -04:00
|
|
|
from six import iteritems
|
|
|
|
|
|
|
|
from twisted.internet import defer
|
2017-06-15 04:59:04 -04:00
|
|
|
|
2017-05-31 06:51:01 -04:00
|
|
|
from synapse.api.constants import EventTypes, JoinRules
|
2019-03-11 09:35:31 -04:00
|
|
|
from synapse.storage.background_updates import BackgroundUpdateStore
|
2019-03-11 09:39:12 -04:00
|
|
|
from synapse.storage.engines import PostgresEngine, Sqlite3Engine
|
2019-02-05 07:16:28 -05:00
|
|
|
from synapse.storage.state import StateFilter
|
2017-05-31 09:29:32 -04:00
|
|
|
from synapse.types import get_domain_from_id, get_localpart_from_id
|
2018-07-09 02:09:20 -04:00
|
|
|
from synapse.util.caches.descriptors import cached, cachedInlineCallbacks
|
2017-05-31 06:51:01 -04:00
|
|
|
|
2017-06-15 04:59:04 -04:00
|
|
|
logger = logging.getLogger(__name__)
|
2017-05-31 06:51:01 -04:00
|
|
|
|
|
|
|
|
2019-03-11 09:35:31 -04:00
|
|
|
class UserDirectoryStore(BackgroundUpdateStore):
|
|
|
|
def __init__(self, dbconn, hs):
|
|
|
|
super(UserDirectoryStore, self).__init__(dbconn, hs)
|
|
|
|
|
|
|
|
self.register_background_update_handler(
|
|
|
|
"users_in_public_rooms_initial", self._populate_users_in_public_rooms
|
|
|
|
)
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def _populate_users_in_public_rooms(self, progress, batch_size):
|
|
|
|
"""
|
|
|
|
Populate the users_in_public_rooms table with the contents of the
|
|
|
|
users_who_share_public_rooms table.
|
|
|
|
"""
|
|
|
|
|
|
|
|
def _fetch(txn):
|
|
|
|
sql = "SELECT DISTINCT other_user_id FROM users_who_share_public_rooms"
|
|
|
|
txn.execute(sql)
|
|
|
|
return txn.fetchall()
|
|
|
|
|
|
|
|
users = yield self.runInteraction(
|
|
|
|
"populate_users_in_public_rooms_fetch", _fetch
|
|
|
|
)
|
|
|
|
|
|
|
|
if users:
|
|
|
|
def _fill(txn):
|
|
|
|
self._simple_upsert_many_txn(
|
|
|
|
txn,
|
|
|
|
table="users_in_public_rooms",
|
|
|
|
key_names=["user_id"],
|
|
|
|
key_values=users,
|
|
|
|
value_names=(),
|
|
|
|
value_values=None,
|
|
|
|
)
|
|
|
|
|
|
|
|
users = yield self.runInteraction(
|
|
|
|
"populate_users_in_public_rooms_fill", _fill
|
|
|
|
)
|
|
|
|
|
|
|
|
yield self._end_background_update("users_in_public_rooms_initial")
|
|
|
|
defer.returnValue(1)
|
|
|
|
|
2019-02-05 07:16:28 -05:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def is_room_world_readable_or_publicly_joinable(self, room_id):
|
2017-05-31 10:00:29 -04:00
|
|
|
"""Check if the room is either world_readable or publically joinable
|
|
|
|
"""
|
2019-02-05 07:16:28 -05:00
|
|
|
|
|
|
|
# Create a state filter that only queries join and history state event
|
|
|
|
types_to_filter = (
|
|
|
|
(EventTypes.JoinRules, ""),
|
|
|
|
(EventTypes.RoomHistoryVisibility, ""),
|
|
|
|
)
|
|
|
|
|
|
|
|
current_state_ids = yield self.get_filtered_current_state_ids(
|
2019-02-13 07:05:32 -05:00
|
|
|
room_id, StateFilter.from_types(types_to_filter)
|
2017-05-31 06:51:01 -04:00
|
|
|
)
|
|
|
|
|
|
|
|
join_rules_id = current_state_ids.get((EventTypes.JoinRules, ""))
|
|
|
|
if join_rules_id:
|
|
|
|
join_rule_ev = yield self.get_event(join_rules_id, allow_none=True)
|
|
|
|
if join_rule_ev:
|
2017-06-01 06:09:49 -04:00
|
|
|
if join_rule_ev.content.get("join_rule") == JoinRules.PUBLIC:
|
2017-05-31 06:51:01 -04:00
|
|
|
defer.returnValue(True)
|
|
|
|
|
|
|
|
hist_vis_id = current_state_ids.get((EventTypes.RoomHistoryVisibility, ""))
|
|
|
|
if hist_vis_id:
|
|
|
|
hist_vis_ev = yield self.get_event(hist_vis_id, allow_none=True)
|
|
|
|
if hist_vis_ev:
|
|
|
|
if hist_vis_ev.content.get("history_visibility") == "world_readable":
|
|
|
|
defer.returnValue(True)
|
|
|
|
|
|
|
|
defer.returnValue(False)
|
|
|
|
|
2019-03-07 04:22:53 -05:00
|
|
|
def add_profiles_to_user_dir(self, users_with_profile):
|
2017-06-01 09:50:46 -04:00
|
|
|
"""Add profiles to the user directory
|
|
|
|
|
|
|
|
Args:
|
2017-05-31 10:00:29 -04:00
|
|
|
users_with_profile (dict): Users to add to directory in the form of
|
|
|
|
mapping of user_id -> ProfileInfo
|
|
|
|
"""
|
2019-03-07 04:22:53 -05:00
|
|
|
|
2017-05-31 06:51:01 -04:00
|
|
|
if isinstance(self.database_engine, PostgresEngine):
|
2017-05-31 10:00:29 -04:00
|
|
|
# We weight the loclpart most highly, then display name and finally
|
|
|
|
# server name
|
2017-05-31 06:51:01 -04:00
|
|
|
sql = """
|
2017-05-31 10:23:49 -04:00
|
|
|
INSERT INTO user_directory_search(user_id, vector)
|
|
|
|
VALUES (?,
|
2017-05-31 09:29:32 -04:00
|
|
|
setweight(to_tsvector('english', ?), 'A')
|
2017-05-31 10:00:29 -04:00
|
|
|
|| setweight(to_tsvector('english', ?), 'D')
|
|
|
|
|| setweight(to_tsvector('english', COALESCE(?, '')), 'B')
|
2017-05-31 09:29:32 -04:00
|
|
|
)
|
2017-05-31 06:51:01 -04:00
|
|
|
"""
|
2017-05-31 09:29:32 -04:00
|
|
|
args = (
|
|
|
|
(
|
2019-02-13 07:05:32 -05:00
|
|
|
user_id,
|
|
|
|
get_localpart_from_id(user_id),
|
|
|
|
get_domain_from_id(user_id),
|
2017-05-31 10:23:49 -04:00
|
|
|
profile.display_name,
|
2017-05-31 09:29:32 -04:00
|
|
|
)
|
2018-05-31 05:03:47 -04:00
|
|
|
for user_id, profile in iteritems(users_with_profile)
|
2017-05-31 09:29:32 -04:00
|
|
|
)
|
2017-05-31 06:51:01 -04:00
|
|
|
elif isinstance(self.database_engine, Sqlite3Engine):
|
|
|
|
sql = """
|
2017-05-31 10:23:49 -04:00
|
|
|
INSERT INTO user_directory_search(user_id, value)
|
|
|
|
VALUES (?,?)
|
2017-05-31 06:51:01 -04:00
|
|
|
"""
|
2019-03-07 04:22:53 -05:00
|
|
|
args = tuple(
|
2017-05-31 06:51:01 -04:00
|
|
|
(
|
2017-05-31 10:23:49 -04:00
|
|
|
user_id,
|
2019-02-13 07:05:32 -05:00
|
|
|
"%s %s" % (user_id, p.display_name) if p.display_name else user_id,
|
2017-05-31 06:51:01 -04:00
|
|
|
)
|
2018-05-31 05:03:47 -04:00
|
|
|
for user_id, p in iteritems(users_with_profile)
|
2017-05-31 09:29:32 -04:00
|
|
|
)
|
|
|
|
else:
|
|
|
|
# This should be unreachable.
|
|
|
|
raise Exception("Unrecognized database engine")
|
|
|
|
|
|
|
|
def _add_profiles_to_user_dir_txn(txn):
|
|
|
|
txn.executemany(sql, args)
|
2017-05-31 10:23:49 -04:00
|
|
|
self._simple_insert_many_txn(
|
|
|
|
txn,
|
|
|
|
table="user_directory",
|
|
|
|
values=[
|
|
|
|
{
|
|
|
|
"user_id": user_id,
|
2019-03-07 04:22:53 -05:00
|
|
|
"room_id": None,
|
2017-05-31 10:23:49 -04:00
|
|
|
"display_name": profile.display_name,
|
|
|
|
"avatar_url": profile.avatar_url,
|
|
|
|
}
|
2018-05-31 05:03:47 -04:00
|
|
|
for user_id, profile in iteritems(users_with_profile)
|
2019-02-13 07:05:32 -05:00
|
|
|
],
|
2017-05-31 10:23:49 -04:00
|
|
|
)
|
2017-05-31 06:51:01 -04:00
|
|
|
for user_id in users_with_profile:
|
2019-02-13 07:05:32 -05:00
|
|
|
txn.call_after(self.get_user_in_directory.invalidate, (user_id,))
|
2017-05-31 06:51:01 -04:00
|
|
|
|
|
|
|
return self.runInteraction(
|
|
|
|
"add_profiles_to_user_dir", _add_profiles_to_user_dir_txn
|
|
|
|
)
|
|
|
|
|
2017-05-31 08:30:12 -04:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def update_user_in_user_dir(self, user_id, room_id):
|
|
|
|
yield self._simple_update_one(
|
|
|
|
table="user_directory",
|
|
|
|
keyvalues={"user_id": user_id},
|
|
|
|
updatevalues={"room_id": room_id},
|
|
|
|
desc="update_user_in_user_dir",
|
|
|
|
)
|
2017-06-01 10:39:51 -04:00
|
|
|
self.get_user_in_directory.invalidate((user_id,))
|
|
|
|
|
2017-06-13 06:49:39 -04:00
|
|
|
def update_profile_in_user_dir(self, user_id, display_name, avatar_url, room_id):
|
2017-06-01 10:39:51 -04:00
|
|
|
def _update_profile_in_user_dir_txn(txn):
|
2017-06-13 06:11:26 -04:00
|
|
|
new_entry = self._simple_upsert_txn(
|
2017-06-01 10:39:51 -04:00
|
|
|
txn,
|
|
|
|
table="user_directory",
|
|
|
|
keyvalues={"user_id": user_id},
|
2017-06-13 06:49:39 -04:00
|
|
|
insertion_values={"room_id": room_id},
|
2017-06-13 06:11:26 -04:00
|
|
|
values={"display_name": display_name, "avatar_url": avatar_url},
|
|
|
|
lock=False, # We're only inserter
|
2017-06-01 10:39:51 -04:00
|
|
|
)
|
|
|
|
|
|
|
|
if isinstance(self.database_engine, PostgresEngine):
|
2017-11-29 13:27:05 -05:00
|
|
|
# We weight the localpart most highly, then display name and finally
|
2017-06-01 10:39:51 -04:00
|
|
|
# server name
|
2019-01-24 05:31:54 -05:00
|
|
|
if self.database_engine.can_native_upsert:
|
2017-06-13 06:11:26 -04:00
|
|
|
sql = """
|
|
|
|
INSERT INTO user_directory_search(user_id, vector)
|
|
|
|
VALUES (?,
|
|
|
|
setweight(to_tsvector('english', ?), 'A')
|
|
|
|
|| setweight(to_tsvector('english', ?), 'D')
|
|
|
|
|| setweight(to_tsvector('english', COALESCE(?, '')), 'B')
|
2019-01-24 05:31:54 -05:00
|
|
|
) ON CONFLICT (user_id) DO UPDATE SET vector=EXCLUDED.vector
|
2017-06-13 06:11:26 -04:00
|
|
|
"""
|
2017-06-13 06:19:18 -04:00
|
|
|
txn.execute(
|
|
|
|
sql,
|
|
|
|
(
|
2019-02-13 07:05:32 -05:00
|
|
|
user_id,
|
|
|
|
get_localpart_from_id(user_id),
|
|
|
|
get_domain_from_id(user_id),
|
|
|
|
display_name,
|
|
|
|
),
|
2017-06-13 06:11:26 -04:00
|
|
|
)
|
|
|
|
else:
|
2019-01-24 05:31:54 -05:00
|
|
|
# TODO: Remove this code after we've bumped the minimum version
|
|
|
|
# of postgres to always support upserts, so we can get rid of
|
|
|
|
# `new_entry` usage
|
|
|
|
if new_entry is True:
|
|
|
|
sql = """
|
|
|
|
INSERT INTO user_directory_search(user_id, vector)
|
|
|
|
VALUES (?,
|
|
|
|
setweight(to_tsvector('english', ?), 'A')
|
|
|
|
|| setweight(to_tsvector('english', ?), 'D')
|
|
|
|
|| setweight(to_tsvector('english', COALESCE(?, '')), 'B')
|
|
|
|
)
|
|
|
|
"""
|
|
|
|
txn.execute(
|
|
|
|
sql,
|
|
|
|
(
|
2019-02-13 07:05:32 -05:00
|
|
|
user_id,
|
|
|
|
get_localpart_from_id(user_id),
|
|
|
|
get_domain_from_id(user_id),
|
|
|
|
display_name,
|
|
|
|
),
|
2019-01-24 05:31:54 -05:00
|
|
|
)
|
|
|
|
elif new_entry is False:
|
|
|
|
sql = """
|
|
|
|
UPDATE user_directory_search
|
|
|
|
SET vector = setweight(to_tsvector('english', ?), 'A')
|
|
|
|
|| setweight(to_tsvector('english', ?), 'D')
|
|
|
|
|| setweight(to_tsvector('english', COALESCE(?, '')), 'B')
|
|
|
|
WHERE user_id = ?
|
|
|
|
"""
|
|
|
|
txn.execute(
|
|
|
|
sql,
|
|
|
|
(
|
|
|
|
get_localpart_from_id(user_id),
|
|
|
|
get_domain_from_id(user_id),
|
2019-02-13 07:05:32 -05:00
|
|
|
display_name,
|
|
|
|
user_id,
|
|
|
|
),
|
2019-01-24 05:31:54 -05:00
|
|
|
)
|
|
|
|
else:
|
|
|
|
raise RuntimeError(
|
|
|
|
"upsert returned None when 'can_native_upsert' is False"
|
2017-06-13 06:19:18 -04:00
|
|
|
)
|
2017-06-01 10:39:51 -04:00
|
|
|
elif isinstance(self.database_engine, Sqlite3Engine):
|
2019-02-13 07:05:32 -05:00
|
|
|
value = "%s %s" % (user_id, display_name) if display_name else user_id
|
2017-06-13 06:11:26 -04:00
|
|
|
self._simple_upsert_txn(
|
|
|
|
txn,
|
|
|
|
table="user_directory_search",
|
|
|
|
keyvalues={"user_id": user_id},
|
|
|
|
values={"value": value},
|
|
|
|
lock=False, # We're only inserter
|
2017-06-01 10:39:51 -04:00
|
|
|
)
|
|
|
|
else:
|
|
|
|
# This should be unreachable.
|
|
|
|
raise Exception("Unrecognized database engine")
|
|
|
|
|
|
|
|
txn.call_after(self.get_user_in_directory.invalidate, (user_id,))
|
|
|
|
|
|
|
|
return self.runInteraction(
|
|
|
|
"update_profile_in_user_dir", _update_profile_in_user_dir_txn
|
|
|
|
)
|
2017-06-01 09:50:46 -04:00
|
|
|
|
2017-05-31 06:51:01 -04:00
|
|
|
def remove_from_user_dir(self, user_id):
|
2017-05-31 10:23:49 -04:00
|
|
|
def _remove_from_user_dir_txn(txn):
|
|
|
|
self._simple_delete_txn(
|
2019-02-13 07:05:32 -05:00
|
|
|
txn, table="user_directory", keyvalues={"user_id": user_id}
|
2017-05-31 10:23:49 -04:00
|
|
|
)
|
|
|
|
self._simple_delete_txn(
|
2019-02-13 07:05:32 -05:00
|
|
|
txn, table="user_directory_search", keyvalues={"user_id": user_id}
|
2017-05-31 10:23:49 -04:00
|
|
|
)
|
2019-03-11 06:11:36 -04:00
|
|
|
self._simple_delete_txn(
|
|
|
|
txn, table="users_in_public_rooms", keyvalues={"user_id": user_id}
|
|
|
|
)
|
2017-06-01 09:50:46 -04:00
|
|
|
self._simple_delete_txn(
|
2019-03-07 04:22:53 -05:00
|
|
|
txn,
|
|
|
|
table="users_who_share_public_rooms",
|
|
|
|
keyvalues={"user_id": user_id},
|
|
|
|
)
|
|
|
|
self._simple_delete_txn(
|
|
|
|
txn,
|
|
|
|
table="users_who_share_public_rooms",
|
|
|
|
keyvalues={"other_user_id": user_id},
|
|
|
|
)
|
|
|
|
self._simple_delete_txn(
|
|
|
|
txn,
|
|
|
|
table="users_who_share_private_rooms",
|
|
|
|
keyvalues={"user_id": user_id},
|
|
|
|
)
|
|
|
|
self._simple_delete_txn(
|
|
|
|
txn,
|
|
|
|
table="users_who_share_private_rooms",
|
|
|
|
keyvalues={"other_user_id": user_id},
|
2017-06-01 09:50:46 -04:00
|
|
|
)
|
2019-02-13 07:05:32 -05:00
|
|
|
txn.call_after(self.get_user_in_directory.invalidate, (user_id,))
|
|
|
|
|
|
|
|
return self.runInteraction("remove_from_user_dir", _remove_from_user_dir_txn)
|
2017-05-31 06:51:01 -04:00
|
|
|
|
2017-06-15 05:15:00 -04:00
|
|
|
@defer.inlineCallbacks
|
2017-05-31 10:11:36 -04:00
|
|
|
def get_users_in_dir_due_to_room(self, room_id):
|
2018-07-10 12:58:09 -04:00
|
|
|
"""Get all user_ids that are in the room directory because they're
|
2017-05-31 10:11:36 -04:00
|
|
|
in the given room_id
|
|
|
|
"""
|
2019-03-07 04:22:53 -05:00
|
|
|
user_ids_share_pub = yield self._simple_select_onecol(
|
|
|
|
table="users_who_share_public_rooms",
|
2017-06-15 05:15:00 -04:00
|
|
|
keyvalues={"room_id": room_id},
|
2019-03-07 04:22:53 -05:00
|
|
|
retcol="other_user_id",
|
2017-06-15 05:15:00 -04:00
|
|
|
desc="get_users_in_dir_due_to_room",
|
|
|
|
)
|
|
|
|
|
2019-03-07 04:22:53 -05:00
|
|
|
user_ids_share_priv = yield self._simple_select_onecol(
|
|
|
|
table="users_who_share_private_rooms",
|
2017-06-15 05:15:00 -04:00
|
|
|
keyvalues={"room_id": room_id},
|
2019-03-07 04:22:53 -05:00
|
|
|
retcol="other_user_id",
|
2017-06-15 05:15:00 -04:00
|
|
|
desc="get_users_in_dir_due_to_room",
|
|
|
|
)
|
|
|
|
|
2019-03-07 04:22:53 -05:00
|
|
|
user_ids = set(user_ids_share_pub)
|
|
|
|
user_ids.update(user_ids_share_priv)
|
2017-06-15 05:15:00 -04:00
|
|
|
|
|
|
|
defer.returnValue(user_ids)
|
|
|
|
|
2017-06-15 04:59:04 -04:00
|
|
|
@defer.inlineCallbacks
|
2017-05-31 06:51:01 -04:00
|
|
|
def get_all_rooms(self):
|
2017-06-15 04:59:04 -04:00
|
|
|
"""Get all room_ids we've ever known about, in ascending order of "size"
|
2017-05-31 10:00:29 -04:00
|
|
|
"""
|
2017-06-15 04:59:04 -04:00
|
|
|
sql = """
|
|
|
|
SELECT room_id FROM current_state_events
|
|
|
|
GROUP BY room_id
|
|
|
|
ORDER BY count(*) ASC
|
|
|
|
"""
|
|
|
|
rows = yield self._execute("get_all_rooms", None, sql)
|
|
|
|
defer.returnValue([room_id for room_id, in rows])
|
|
|
|
|
2017-11-29 13:27:05 -05:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def get_all_local_users(self):
|
|
|
|
"""Get all local users
|
|
|
|
"""
|
|
|
|
sql = """
|
|
|
|
SELECT name FROM users
|
|
|
|
"""
|
|
|
|
rows = yield self._execute("get_all_local_users", None, sql)
|
|
|
|
defer.returnValue([name for name, in rows])
|
|
|
|
|
2017-06-15 04:59:04 -04:00
|
|
|
def add_users_who_share_room(self, room_id, share_private, user_id_tuples):
|
2019-03-07 04:22:53 -05:00
|
|
|
"""Insert entries into the users_who_share_*_rooms table. The first
|
2017-06-15 04:59:04 -04:00
|
|
|
user should be a local user.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
room_id (str)
|
|
|
|
share_private (bool): Is the room private
|
|
|
|
user_id_tuples([(str, str)]): iterable of 2-tuple of user IDs.
|
|
|
|
"""
|
2019-02-13 07:05:32 -05:00
|
|
|
|
2017-06-15 04:59:04 -04:00
|
|
|
def _add_users_who_share_room_txn(txn):
|
2019-03-07 04:22:53 -05:00
|
|
|
|
|
|
|
if share_private:
|
|
|
|
tbl = "users_who_share_private_rooms"
|
|
|
|
else:
|
|
|
|
tbl = "users_who_share_public_rooms"
|
|
|
|
|
|
|
|
self._simple_upsert_many_txn(
|
2017-06-15 04:59:04 -04:00
|
|
|
txn,
|
2019-03-07 04:22:53 -05:00
|
|
|
table=tbl,
|
|
|
|
key_names=["user_id", "other_user_id", "room_id"],
|
|
|
|
key_values=[
|
|
|
|
(user_id, other_user_id, room_id)
|
2017-06-15 04:59:04 -04:00
|
|
|
for user_id, other_user_id in user_id_tuples
|
|
|
|
],
|
2019-03-07 04:22:53 -05:00
|
|
|
value_names=(),
|
|
|
|
value_values=None,
|
2017-06-15 04:59:04 -04:00
|
|
|
)
|
2019-03-11 06:11:36 -04:00
|
|
|
|
|
|
|
# If it's a public room, also update them in users_in_public_rooms.
|
|
|
|
# We don't look before they're in the table before we do it, as it's
|
|
|
|
# more efficient to simply have Postgres do that (one UPSERT vs one
|
|
|
|
# SELECT and maybe one INSERT).
|
|
|
|
if not share_private:
|
|
|
|
for user_id in set([x[1] for x in user_id_tuples]):
|
|
|
|
self._simple_upsert_txn(
|
|
|
|
txn,
|
|
|
|
"users_in_public_rooms",
|
|
|
|
keyvalues={"user_id": user_id},
|
2019-03-11 09:35:31 -04:00
|
|
|
values=None,
|
2019-03-11 06:11:36 -04:00
|
|
|
)
|
|
|
|
|
2017-06-15 04:59:04 -04:00
|
|
|
for user_id, other_user_id in user_id_tuples:
|
|
|
|
txn.call_after(
|
2019-02-13 07:05:32 -05:00
|
|
|
self.get_users_who_share_room_from_dir.invalidate, (user_id,)
|
2017-06-15 04:59:04 -04:00
|
|
|
)
|
2019-02-13 07:05:32 -05:00
|
|
|
|
2017-06-15 04:59:04 -04:00
|
|
|
return self.runInteraction(
|
|
|
|
"add_users_who_share_room", _add_users_who_share_room_txn
|
|
|
|
)
|
|
|
|
|
2019-03-07 04:22:53 -05:00
|
|
|
def remove_user_who_share_room(self, user_id, room_id):
|
2017-06-15 04:59:04 -04:00
|
|
|
"""
|
2019-03-07 04:22:53 -05:00
|
|
|
Deletes entries in the users_who_share_*_rooms table. The first
|
2017-06-15 04:59:04 -04:00
|
|
|
user should be a local user.
|
|
|
|
|
|
|
|
Args:
|
2019-03-07 04:22:53 -05:00
|
|
|
user_id (str)
|
2017-06-15 04:59:04 -04:00
|
|
|
room_id (str)
|
|
|
|
"""
|
2019-02-13 07:05:32 -05:00
|
|
|
|
2017-06-15 04:59:04 -04:00
|
|
|
def _remove_user_who_share_room_txn(txn):
|
|
|
|
self._simple_delete_txn(
|
|
|
|
txn,
|
2019-03-07 04:22:53 -05:00
|
|
|
table="users_who_share_private_rooms",
|
|
|
|
keyvalues={"user_id": user_id, "room_id": room_id},
|
2017-06-15 04:59:04 -04:00
|
|
|
)
|
2019-03-07 04:22:53 -05:00
|
|
|
self._simple_delete_txn(
|
|
|
|
txn,
|
|
|
|
table="users_who_share_private_rooms",
|
|
|
|
keyvalues={"other_user_id": user_id, "room_id": room_id},
|
|
|
|
)
|
|
|
|
self._simple_delete_txn(
|
|
|
|
txn,
|
|
|
|
table="users_who_share_public_rooms",
|
|
|
|
keyvalues={"user_id": user_id, "room_id": room_id},
|
|
|
|
)
|
|
|
|
self._simple_delete_txn(
|
|
|
|
txn,
|
|
|
|
table="users_who_share_public_rooms",
|
|
|
|
keyvalues={"other_user_id": user_id, "room_id": room_id},
|
2017-06-15 04:59:04 -04:00
|
|
|
)
|
2019-03-11 06:11:36 -04:00
|
|
|
|
|
|
|
# Are the users still in a public room after we deleted them from this one?
|
|
|
|
still_in_public = self._simple_select_one_onecol_txn(
|
|
|
|
txn,
|
|
|
|
"users_who_share_public_rooms",
|
|
|
|
keyvalues={"other_user_id": user_id},
|
|
|
|
retcol="other_user_id",
|
|
|
|
allow_none=True,
|
|
|
|
)
|
|
|
|
|
|
|
|
if still_in_public is None:
|
|
|
|
self._simple_delete_txn(
|
|
|
|
txn, table="users_in_public_rooms", keyvalues={"user_id": user_id}
|
|
|
|
)
|
|
|
|
|
2017-06-15 04:59:04 -04:00
|
|
|
txn.call_after(
|
2019-03-07 04:22:53 -05:00
|
|
|
self.get_users_who_share_room_from_dir.invalidate, (user_id,)
|
2017-06-15 04:59:04 -04:00
|
|
|
)
|
|
|
|
|
|
|
|
return self.runInteraction(
|
|
|
|
"remove_user_who_share_room", _remove_user_who_share_room_txn
|
2017-05-31 06:51:01 -04:00
|
|
|
)
|
|
|
|
|
2017-06-15 05:15:00 -04:00
|
|
|
@cachedInlineCallbacks(max_entries=500000, iterable=True)
|
|
|
|
def get_users_who_share_room_from_dir(self, user_id):
|
|
|
|
"""Returns the set of users who share a room with `user_id`
|
|
|
|
|
|
|
|
Args:
|
|
|
|
user_id(str): Must be a local user
|
|
|
|
|
|
|
|
Returns:
|
2019-03-07 04:22:53 -05:00
|
|
|
list: user_id
|
2017-06-15 05:15:00 -04:00
|
|
|
"""
|
2019-03-07 04:22:53 -05:00
|
|
|
rows = yield self._simple_select_onecol(
|
|
|
|
table="users_who_share_private_rooms",
|
|
|
|
keyvalues={"user_id": user_id},
|
|
|
|
retcol="other_user_id",
|
|
|
|
desc="get_users_who_share_room_with_user",
|
|
|
|
)
|
|
|
|
|
|
|
|
pub_rows = yield self._simple_select_onecol(
|
|
|
|
table="users_who_share_public_rooms",
|
2019-02-13 07:05:32 -05:00
|
|
|
keyvalues={"user_id": user_id},
|
2019-03-07 04:22:53 -05:00
|
|
|
retcol="other_user_id",
|
2017-06-15 05:15:00 -04:00
|
|
|
desc="get_users_who_share_room_with_user",
|
|
|
|
)
|
|
|
|
|
2019-03-07 04:22:53 -05:00
|
|
|
users = set(pub_rows)
|
|
|
|
users.update(rows)
|
2017-06-15 05:15:00 -04:00
|
|
|
|
2019-03-07 04:22:53 -05:00
|
|
|
# Remove the user themselves from this list.
|
|
|
|
users.discard(user_id)
|
2017-06-15 05:15:00 -04:00
|
|
|
|
2019-03-07 04:22:53 -05:00
|
|
|
defer.returnValue(list(users))
|
2017-06-15 05:15:00 -04:00
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def get_rooms_in_common_for_users(self, user_id, other_user_id):
|
|
|
|
"""Given two user_ids find out the list of rooms they share.
|
|
|
|
"""
|
|
|
|
sql = """
|
|
|
|
SELECT room_id FROM (
|
|
|
|
SELECT c.room_id FROM current_state_events AS c
|
|
|
|
INNER JOIN room_memberships USING (event_id)
|
|
|
|
WHERE type = 'm.room.member'
|
|
|
|
AND membership = 'join'
|
|
|
|
AND state_key = ?
|
|
|
|
) AS f1 INNER JOIN (
|
|
|
|
SELECT c.room_id FROM current_state_events AS c
|
|
|
|
INNER JOIN room_memberships USING (event_id)
|
|
|
|
WHERE type = 'm.room.member'
|
|
|
|
AND membership = 'join'
|
|
|
|
AND state_key = ?
|
|
|
|
) f2 USING (room_id)
|
|
|
|
"""
|
|
|
|
|
|
|
|
rows = yield self._execute(
|
|
|
|
"get_rooms_in_common_for_users", None, sql, user_id, other_user_id
|
|
|
|
)
|
|
|
|
|
|
|
|
defer.returnValue([room_id for room_id, in rows])
|
|
|
|
|
2017-05-31 06:51:01 -04:00
|
|
|
def delete_all_from_user_dir(self):
|
2017-05-31 10:00:29 -04:00
|
|
|
"""Delete the entire user directory
|
|
|
|
"""
|
2019-02-13 07:05:32 -05:00
|
|
|
|
2017-05-31 06:51:01 -04:00
|
|
|
def _delete_all_from_user_dir_txn(txn):
|
|
|
|
txn.execute("DELETE FROM user_directory")
|
2017-05-31 10:23:49 -04:00
|
|
|
txn.execute("DELETE FROM user_directory_search")
|
2019-03-11 06:11:36 -04:00
|
|
|
txn.execute("DELETE FROM users_in_public_rooms")
|
2019-03-07 04:22:53 -05:00
|
|
|
txn.execute("DELETE FROM users_who_share_public_rooms")
|
|
|
|
txn.execute("DELETE FROM users_who_share_private_rooms")
|
2017-05-31 06:51:01 -04:00
|
|
|
txn.call_after(self.get_user_in_directory.invalidate_all)
|
2017-06-15 04:59:04 -04:00
|
|
|
txn.call_after(self.get_users_who_share_room_from_dir.invalidate_all)
|
2019-02-13 07:05:32 -05:00
|
|
|
|
2017-05-31 06:51:01 -04:00
|
|
|
return self.runInteraction(
|
|
|
|
"delete_all_from_user_dir", _delete_all_from_user_dir_txn
|
|
|
|
)
|
|
|
|
|
|
|
|
@cached()
|
|
|
|
def get_user_in_directory(self, user_id):
|
|
|
|
return self._simple_select_one(
|
|
|
|
table="user_directory",
|
|
|
|
keyvalues={"user_id": user_id},
|
2019-03-07 04:22:53 -05:00
|
|
|
retcols=("display_name", "avatar_url"),
|
2017-05-31 06:51:01 -04:00
|
|
|
allow_none=True,
|
|
|
|
desc="get_user_in_directory",
|
|
|
|
)
|
|
|
|
|
|
|
|
def get_user_directory_stream_pos(self):
|
|
|
|
return self._simple_select_one_onecol(
|
|
|
|
table="user_directory_stream_pos",
|
|
|
|
keyvalues={},
|
|
|
|
retcol="stream_id",
|
|
|
|
desc="get_user_directory_stream_pos",
|
|
|
|
)
|
|
|
|
|
|
|
|
def update_user_directory_stream_pos(self, stream_id):
|
|
|
|
return self._simple_update_one(
|
|
|
|
table="user_directory_stream_pos",
|
|
|
|
keyvalues={},
|
|
|
|
updatevalues={"stream_id": stream_id},
|
|
|
|
desc="update_user_directory_stream_pos",
|
|
|
|
)
|
|
|
|
|
|
|
|
def get_current_state_deltas(self, prev_stream_id):
|
2017-05-31 12:03:08 -04:00
|
|
|
prev_stream_id = int(prev_stream_id)
|
2019-02-13 07:05:32 -05:00
|
|
|
if not self._curr_state_delta_stream_cache.has_any_entity_changed(
|
|
|
|
prev_stream_id
|
|
|
|
):
|
2017-05-31 10:46:36 -04:00
|
|
|
return []
|
|
|
|
|
2017-05-31 11:17:58 -04:00
|
|
|
def get_current_state_deltas_txn(txn):
|
|
|
|
# First we calculate the max stream id that will give us less than
|
2017-05-31 12:30:26 -04:00
|
|
|
# N results.
|
|
|
|
# We arbitarily limit to 100 stream_id entries to ensure we don't
|
|
|
|
# select toooo many.
|
2017-05-31 11:17:58 -04:00
|
|
|
sql = """
|
|
|
|
SELECT stream_id, count(*)
|
|
|
|
FROM current_state_delta_stream
|
|
|
|
WHERE stream_id > ?
|
|
|
|
GROUP BY stream_id
|
|
|
|
ORDER BY stream_id ASC
|
|
|
|
LIMIT 100
|
|
|
|
"""
|
|
|
|
txn.execute(sql, (prev_stream_id,))
|
|
|
|
|
|
|
|
total = 0
|
2017-05-31 11:34:40 -04:00
|
|
|
max_stream_id = prev_stream_id
|
2017-05-31 11:17:58 -04:00
|
|
|
for max_stream_id, count in txn:
|
|
|
|
total += count
|
2017-05-31 12:30:26 -04:00
|
|
|
if total > 100:
|
|
|
|
# We arbitarily limit to 100 entries to ensure we don't
|
|
|
|
# select toooo many.
|
2017-05-31 11:17:58 -04:00
|
|
|
break
|
2017-05-31 06:51:01 -04:00
|
|
|
|
2017-05-31 11:17:58 -04:00
|
|
|
# Now actually get the deltas
|
|
|
|
sql = """
|
|
|
|
SELECT stream_id, room_id, type, state_key, event_id, prev_event_id
|
|
|
|
FROM current_state_delta_stream
|
|
|
|
WHERE ? < stream_id AND stream_id <= ?
|
|
|
|
ORDER BY stream_id ASC
|
|
|
|
"""
|
2019-02-13 07:05:32 -05:00
|
|
|
txn.execute(sql, (prev_stream_id, max_stream_id))
|
2017-05-31 11:17:58 -04:00
|
|
|
return self.cursor_to_dict(txn)
|
|
|
|
|
|
|
|
return self.runInteraction(
|
|
|
|
"get_current_state_deltas", get_current_state_deltas_txn
|
2017-05-31 06:51:01 -04:00
|
|
|
)
|
2017-05-31 09:00:01 -04:00
|
|
|
|
2017-05-31 10:13:49 -04:00
|
|
|
def get_max_stream_id_in_current_state_deltas(self):
|
|
|
|
return self._simple_select_one_onecol(
|
|
|
|
table="current_state_delta_stream",
|
|
|
|
keyvalues={},
|
|
|
|
retcol="COALESCE(MAX(stream_id), -1)",
|
|
|
|
desc="get_max_stream_id_in_current_state_deltas",
|
|
|
|
)
|
|
|
|
|
2017-05-31 09:00:01 -04:00
|
|
|
@defer.inlineCallbacks
|
2017-06-15 05:00:28 -04:00
|
|
|
def search_user_dir(self, user_id, search_term, limit):
|
2017-05-31 10:00:29 -04:00
|
|
|
"""Searches for users in directory
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
dict of the form::
|
|
|
|
|
|
|
|
{
|
|
|
|
"limited": <bool>, # whether there were more results or not
|
|
|
|
"results": [ # Ordered by best match first
|
|
|
|
{
|
|
|
|
"user_id": <user_id>,
|
|
|
|
"display_name": <display_name>,
|
|
|
|
"avatar_url": <avatar_url>
|
|
|
|
}
|
|
|
|
]
|
|
|
|
}
|
|
|
|
"""
|
2017-11-29 11:46:45 -05:00
|
|
|
|
2017-12-04 09:58:39 -05:00
|
|
|
if self.hs.config.user_directory_search_all_users:
|
2018-01-25 16:20:28 -05:00
|
|
|
# make s.user_id null to keep the ordering algorithm happy
|
2018-01-17 10:58:52 -05:00
|
|
|
join_clause = """
|
2018-01-25 16:20:28 -05:00
|
|
|
CROSS JOIN (SELECT NULL as user_id) AS s
|
2018-01-17 10:58:52 -05:00
|
|
|
"""
|
2018-01-25 16:20:28 -05:00
|
|
|
join_args = ()
|
|
|
|
where_clause = "1=1"
|
2017-12-04 09:58:39 -05:00
|
|
|
else:
|
|
|
|
join_clause = """
|
|
|
|
LEFT JOIN (
|
2019-03-11 09:35:31 -04:00
|
|
|
SELECT user_id FROM users_in_public_rooms
|
2019-03-07 04:22:53 -05:00
|
|
|
UNION
|
|
|
|
SELECT other_user_id AS user_id FROM users_who_share_private_rooms
|
|
|
|
WHERE user_id = ?
|
|
|
|
) AS p USING (user_id)
|
2017-12-04 09:58:39 -05:00
|
|
|
"""
|
2018-01-25 16:20:28 -05:00
|
|
|
join_args = (user_id,)
|
2019-03-07 04:22:53 -05:00
|
|
|
where_clause = "p.user_id IS NOT NULL"
|
2017-11-29 11:46:45 -05:00
|
|
|
|
2017-05-31 09:00:01 -04:00
|
|
|
if isinstance(self.database_engine, PostgresEngine):
|
2017-06-13 05:16:31 -04:00
|
|
|
full_query, exact_query, prefix_query = _parse_query_postgres(search_term)
|
|
|
|
|
2017-05-31 13:17:47 -04:00
|
|
|
# We order by rank and then if they have profile info
|
2017-06-13 05:16:31 -04:00
|
|
|
# The ranking algorithm is hand tweaked for "best" results. Broadly
|
|
|
|
# the idea is we give a higher weight to exact matches.
|
|
|
|
# The array of numbers are the weights for the various part of the
|
|
|
|
# search: (domain, _, display name, localpart)
|
2017-05-31 09:00:01 -04:00
|
|
|
sql = """
|
2018-03-28 06:19:45 -04:00
|
|
|
SELECT d.user_id AS user_id, display_name, avatar_url
|
2017-05-31 10:23:49 -04:00
|
|
|
FROM user_directory_search
|
2017-06-15 05:00:28 -04:00
|
|
|
INNER JOIN user_directory AS d USING (user_id)
|
2017-11-29 20:17:15 -05:00
|
|
|
%s
|
2017-06-15 05:00:28 -04:00
|
|
|
WHERE
|
2017-12-04 09:58:39 -05:00
|
|
|
%s
|
2017-06-15 05:00:28 -04:00
|
|
|
AND vector @@ to_tsquery('english', ?)
|
2017-05-31 13:17:47 -04:00
|
|
|
ORDER BY
|
2019-03-07 04:22:53 -05:00
|
|
|
(CASE WHEN d.user_id IS NOT NULL THEN 4.0 ELSE 1.0 END)
|
2017-06-15 05:00:28 -04:00
|
|
|
* (CASE WHEN display_name IS NOT NULL THEN 1.2 ELSE 1.0 END)
|
|
|
|
* (CASE WHEN avatar_url IS NOT NULL THEN 1.2 ELSE 1.0 END)
|
|
|
|
* (
|
|
|
|
3 * ts_rank_cd(
|
|
|
|
'{0.1, 0.1, 0.9, 1.0}',
|
|
|
|
vector,
|
|
|
|
to_tsquery('english', ?),
|
|
|
|
8
|
|
|
|
)
|
|
|
|
+ ts_rank_cd(
|
|
|
|
'{0.1, 0.1, 0.9, 1.0}',
|
|
|
|
vector,
|
|
|
|
to_tsquery('english', ?),
|
|
|
|
8
|
|
|
|
)
|
2017-06-13 05:16:31 -04:00
|
|
|
)
|
|
|
|
DESC,
|
2017-05-31 13:17:47 -04:00
|
|
|
display_name IS NULL,
|
|
|
|
avatar_url IS NULL
|
2017-05-31 09:00:01 -04:00
|
|
|
LIMIT ?
|
2019-02-13 07:05:32 -05:00
|
|
|
""" % (
|
|
|
|
join_clause,
|
|
|
|
where_clause,
|
|
|
|
)
|
|
|
|
args = join_args + (full_query, exact_query, prefix_query, limit + 1)
|
2017-05-31 09:00:01 -04:00
|
|
|
elif isinstance(self.database_engine, Sqlite3Engine):
|
2017-06-13 05:16:31 -04:00
|
|
|
search_query = _parse_query_sqlite(search_term)
|
|
|
|
|
2017-05-31 09:00:01 -04:00
|
|
|
sql = """
|
2018-03-28 06:19:45 -04:00
|
|
|
SELECT d.user_id AS user_id, display_name, avatar_url
|
2017-05-31 10:23:49 -04:00
|
|
|
FROM user_directory_search
|
2017-06-15 05:00:28 -04:00
|
|
|
INNER JOIN user_directory AS d USING (user_id)
|
2017-11-29 20:17:15 -05:00
|
|
|
%s
|
2017-06-15 05:00:28 -04:00
|
|
|
WHERE
|
2017-12-04 09:58:39 -05:00
|
|
|
%s
|
2017-06-15 05:00:28 -04:00
|
|
|
AND value MATCH ?
|
2017-05-31 13:17:47 -04:00
|
|
|
ORDER BY
|
2017-06-01 09:58:48 -04:00
|
|
|
rank(matchinfo(user_directory_search)) DESC,
|
2017-05-31 13:17:47 -04:00
|
|
|
display_name IS NULL,
|
|
|
|
avatar_url IS NULL
|
2017-05-31 09:00:01 -04:00
|
|
|
LIMIT ?
|
2019-02-13 07:05:32 -05:00
|
|
|
""" % (
|
|
|
|
join_clause,
|
|
|
|
where_clause,
|
|
|
|
)
|
2018-01-25 16:20:28 -05:00
|
|
|
args = join_args + (search_query, limit + 1)
|
2017-05-31 09:00:01 -04:00
|
|
|
else:
|
|
|
|
# This should be unreachable.
|
|
|
|
raise Exception("Unrecognized database engine")
|
|
|
|
|
|
|
|
results = yield self._execute(
|
|
|
|
"search_user_dir", self.cursor_to_dict, sql, *args
|
|
|
|
)
|
|
|
|
|
|
|
|
limited = len(results) > limit
|
|
|
|
|
2019-02-13 07:05:32 -05:00
|
|
|
defer.returnValue({"limited": limited, "results": results})
|
2017-05-31 13:07:12 -04:00
|
|
|
|
|
|
|
|
2017-06-13 05:16:31 -04:00
|
|
|
def _parse_query_sqlite(search_term):
|
2017-05-31 13:07:12 -04:00
|
|
|
"""Takes a plain unicode string from the user and converts it into a form
|
|
|
|
that can be passed to database.
|
|
|
|
We use this so that we can add prefix matching, which isn't something
|
|
|
|
that is supported by default.
|
|
|
|
|
|
|
|
We specifically add both a prefix and non prefix matching term so that
|
|
|
|
exact matches get ranked higher.
|
|
|
|
"""
|
|
|
|
|
|
|
|
# Pull out the individual words, discarding any non-word characters.
|
|
|
|
results = re.findall(r"([\w\-]+)", search_term, re.UNICODE)
|
2019-02-13 07:05:32 -05:00
|
|
|
return " & ".join("(%s* OR %s)" % (result, result) for result in results)
|
2017-06-13 05:16:31 -04:00
|
|
|
|
|
|
|
|
|
|
|
def _parse_query_postgres(search_term):
|
|
|
|
"""Takes a plain unicode string from the user and converts it into a form
|
|
|
|
that can be passed to database.
|
|
|
|
We use this so that we can add prefix matching, which isn't something
|
|
|
|
that is supported by default.
|
|
|
|
"""
|
|
|
|
|
|
|
|
# Pull out the individual words, discarding any non-word characters.
|
|
|
|
results = re.findall(r"([\w\-]+)", search_term, re.UNICODE)
|
|
|
|
|
2019-02-13 07:05:32 -05:00
|
|
|
both = " & ".join("(%s:* | %s)" % (result, result) for result in results)
|
2017-06-13 05:16:31 -04:00
|
|
|
exact = " & ".join("%s" % (result,) for result in results)
|
|
|
|
prefix = " & ".join("%s:*" % (result,) for result in results)
|
2017-05-31 13:07:12 -04:00
|
|
|
|
2017-06-13 05:16:31 -04:00
|
|
|
return both, exact, prefix
|