2014-08-12 10:10:52 -04:00
|
|
|
# -*- coding: utf-8 -*-
|
2016-01-05 13:01:18 -05:00
|
|
|
# Copyright 2014 - 2016 OpenMarket Ltd
|
2014-08-12 10:10:52 -04:00
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
2014-08-12 22:14:34 -04:00
|
|
|
|
2016-02-05 06:22:30 -05:00
|
|
|
import re
|
|
|
|
|
2018-07-09 02:09:20 -04:00
|
|
|
from six.moves import range
|
|
|
|
|
2014-08-12 10:10:52 -04:00
|
|
|
from twisted.internet import defer
|
|
|
|
|
2018-07-09 02:09:20 -04:00
|
|
|
from synapse.api.errors import Codes, StoreError
|
2016-07-22 09:52:53 -04:00
|
|
|
from synapse.storage import background_updates
|
2018-03-01 13:19:34 -05:00
|
|
|
from synapse.storage._base import SQLBaseStore
|
2016-04-06 11:50:47 -04:00
|
|
|
from synapse.util.caches.descriptors import cached, cachedInlineCallbacks
|
2014-08-12 10:10:52 -04:00
|
|
|
|
|
|
|
|
2018-03-01 13:19:34 -05:00
|
|
|
class RegistrationWorkerStore(SQLBaseStore):
|
2018-08-23 14:17:08 -04:00
|
|
|
def __init__(self, db_conn, hs):
|
|
|
|
super(RegistrationWorkerStore, self).__init__(db_conn, hs)
|
|
|
|
|
|
|
|
self.config = hs.config
|
|
|
|
|
2018-03-01 13:19:34 -05:00
|
|
|
@cached()
|
|
|
|
def get_user_by_id(self, user_id):
|
|
|
|
return self._simple_select_one(
|
|
|
|
table="users",
|
|
|
|
keyvalues={
|
|
|
|
"name": user_id,
|
|
|
|
},
|
2018-05-17 12:35:31 -04:00
|
|
|
retcols=[
|
|
|
|
"name", "password_hash", "is_guest",
|
|
|
|
"consent_version", "consent_server_notice_sent",
|
2018-08-23 14:17:08 -04:00
|
|
|
"appservice_id", "creation_ts",
|
2018-05-17 12:35:31 -04:00
|
|
|
],
|
2018-03-01 13:19:34 -05:00
|
|
|
allow_none=True,
|
|
|
|
desc="get_user_by_id",
|
|
|
|
)
|
|
|
|
|
2018-08-23 14:17:08 -04:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def is_trial_user(self, user_id):
|
|
|
|
"""Checks if user is in the "trial" period, i.e. within the first
|
|
|
|
N days of registration defined by `mau_trial_days` config
|
|
|
|
|
|
|
|
Args:
|
|
|
|
user_id (str)
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
Deferred[bool]
|
|
|
|
"""
|
|
|
|
|
|
|
|
info = yield self.get_user_by_id(user_id)
|
|
|
|
if not info:
|
|
|
|
defer.returnValue(False)
|
|
|
|
|
|
|
|
now = self.clock.time_msec()
|
|
|
|
trial_duration_ms = self.config.mau_trial_days * 24 * 60 * 60 * 1000
|
|
|
|
is_trial = (now - info["creation_ts"] * 1000) < trial_duration_ms
|
|
|
|
defer.returnValue(is_trial)
|
|
|
|
|
2018-03-01 13:19:34 -05:00
|
|
|
@cached()
|
|
|
|
def get_user_by_access_token(self, token):
|
|
|
|
"""Get a user from the given access token.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
token (str): The access token of a user.
|
|
|
|
Returns:
|
|
|
|
defer.Deferred: None, if the token did not match, otherwise dict
|
|
|
|
including the keys `name`, `is_guest`, `device_id`, `token_id`.
|
|
|
|
"""
|
|
|
|
return self.runInteraction(
|
|
|
|
"get_user_by_access_token",
|
|
|
|
self._query_for_auth,
|
|
|
|
token
|
|
|
|
)
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def is_server_admin(self, user):
|
|
|
|
res = yield self._simple_select_one_onecol(
|
|
|
|
table="users",
|
|
|
|
keyvalues={"name": user.to_string()},
|
|
|
|
retcol="admin",
|
|
|
|
allow_none=True,
|
|
|
|
desc="is_server_admin",
|
|
|
|
)
|
|
|
|
|
|
|
|
defer.returnValue(res if res else False)
|
|
|
|
|
|
|
|
def _query_for_auth(self, txn, token):
|
|
|
|
sql = (
|
|
|
|
"SELECT users.name, users.is_guest, access_tokens.id as token_id,"
|
|
|
|
" access_tokens.device_id"
|
|
|
|
" FROM users"
|
|
|
|
" INNER JOIN access_tokens on users.name = access_tokens.user_id"
|
|
|
|
" WHERE token = ?"
|
|
|
|
)
|
|
|
|
|
|
|
|
txn.execute(sql, (token,))
|
|
|
|
rows = self.cursor_to_dict(txn)
|
|
|
|
if rows:
|
|
|
|
return rows[0]
|
|
|
|
|
|
|
|
return None
|
|
|
|
|
|
|
|
|
|
|
|
class RegistrationStore(RegistrationWorkerStore,
|
|
|
|
background_updates.BackgroundUpdateStore):
|
2014-08-12 10:10:52 -04:00
|
|
|
|
2017-11-09 13:51:27 -05:00
|
|
|
def __init__(self, db_conn, hs):
|
|
|
|
super(RegistrationStore, self).__init__(db_conn, hs)
|
2014-08-12 10:10:52 -04:00
|
|
|
|
|
|
|
self.clock = hs.get_clock()
|
|
|
|
|
2016-07-22 09:52:53 -04:00
|
|
|
self.register_background_index_update(
|
|
|
|
"access_tokens_device_index",
|
|
|
|
index_name="access_tokens_device_id",
|
|
|
|
table="access_tokens",
|
|
|
|
columns=["user_id", "device_id"],
|
|
|
|
)
|
|
|
|
|
2018-05-29 12:47:55 -04:00
|
|
|
self.register_background_index_update(
|
|
|
|
"users_creation_ts",
|
|
|
|
index_name="users_creation_ts",
|
|
|
|
table="users",
|
|
|
|
columns=["creation_ts"],
|
|
|
|
)
|
|
|
|
|
2017-10-31 16:35:58 -04:00
|
|
|
# we no longer use refresh tokens, but it's possible that some people
|
|
|
|
# might have a background update queued to build this index. Just
|
|
|
|
# clear the background update.
|
2018-02-02 09:32:51 -05:00
|
|
|
self.register_noop_background_update("refresh_tokens_device_index")
|
2016-07-22 09:52:53 -04:00
|
|
|
|
2014-08-12 10:10:52 -04:00
|
|
|
@defer.inlineCallbacks
|
2016-07-15 08:19:07 -04:00
|
|
|
def add_access_token_to_user(self, user_id, token, device_id=None):
|
2014-08-12 10:10:52 -04:00
|
|
|
"""Adds an access token for the given user.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
user_id (str): The user ID.
|
|
|
|
token (str): The new access token to add.
|
2016-07-15 08:19:07 -04:00
|
|
|
device_id (str): ID of the device to associate with the access
|
|
|
|
token
|
2014-08-12 10:10:52 -04:00
|
|
|
Raises:
|
|
|
|
StoreError if there was a problem adding this.
|
|
|
|
"""
|
2016-03-01 09:32:56 -05:00
|
|
|
next_id = self._access_tokens_id_gen.get_next()
|
2015-04-07 07:05:36 -04:00
|
|
|
|
2015-04-15 05:24:07 -04:00
|
|
|
yield self._simple_insert(
|
2014-08-12 10:10:52 -04:00
|
|
|
"access_tokens",
|
|
|
|
{
|
2015-04-07 07:05:36 -04:00
|
|
|
"id": next_id,
|
2015-03-19 11:59:48 -04:00
|
|
|
"user_id": user_id,
|
2016-07-15 08:19:07 -04:00
|
|
|
"token": token,
|
|
|
|
"device_id": device_id,
|
2015-03-20 11:59:18 -04:00
|
|
|
},
|
|
|
|
desc="add_access_token_to_user",
|
2014-08-12 10:10:52 -04:00
|
|
|
)
|
|
|
|
|
2016-07-19 13:46:19 -04:00
|
|
|
def register(self, user_id, token=None, password_hash=None,
|
2016-06-17 14:14:16 -04:00
|
|
|
was_guest=False, make_guest=False, appservice_id=None,
|
2016-07-05 12:30:22 -04:00
|
|
|
create_profile_with_localpart=None, admin=False):
|
2014-08-12 10:10:52 -04:00
|
|
|
"""Attempts to register an account.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
user_id (str): The desired user ID to register.
|
2016-07-19 13:46:19 -04:00
|
|
|
token (str): The desired access token to use for this user. If this
|
|
|
|
is not None, the given access token is associated with the user
|
|
|
|
id.
|
2014-08-12 10:10:52 -04:00
|
|
|
password_hash (str): Optional. The password hash for this user.
|
2016-01-05 13:01:18 -05:00
|
|
|
was_guest (bool): Optional. Whether this is a guest account being
|
|
|
|
upgraded to a non-guest account.
|
2016-01-06 06:38:09 -05:00
|
|
|
make_guest (boolean): True if the the new user should be guest,
|
|
|
|
false to add a regular user account.
|
2016-03-10 10:58:22 -05:00
|
|
|
appservice_id (str): The ID of the appservice registering the user.
|
2016-06-17 14:14:16 -04:00
|
|
|
create_profile_with_localpart (str): Optionally create a profile for
|
|
|
|
the given localpart.
|
2014-08-12 10:10:52 -04:00
|
|
|
Raises:
|
|
|
|
StoreError if the user_id could not be registered.
|
|
|
|
"""
|
2016-08-18 09:59:55 -04:00
|
|
|
return self.runInteraction(
|
2014-10-28 07:18:04 -04:00
|
|
|
"register",
|
2016-03-10 10:58:22 -05:00
|
|
|
self._register,
|
|
|
|
user_id,
|
|
|
|
token,
|
|
|
|
password_hash,
|
|
|
|
was_guest,
|
|
|
|
make_guest,
|
2016-06-17 14:14:16 -04:00
|
|
|
appservice_id,
|
|
|
|
create_profile_with_localpart,
|
2016-07-05 12:30:22 -04:00
|
|
|
admin
|
2014-10-28 07:18:04 -04:00
|
|
|
)
|
2014-08-12 10:10:52 -04:00
|
|
|
|
2016-03-10 10:58:22 -05:00
|
|
|
def _register(
|
|
|
|
self,
|
|
|
|
txn,
|
|
|
|
user_id,
|
|
|
|
token,
|
|
|
|
password_hash,
|
|
|
|
was_guest,
|
|
|
|
make_guest,
|
2016-06-17 14:14:16 -04:00
|
|
|
appservice_id,
|
|
|
|
create_profile_with_localpart,
|
2016-07-05 12:30:22 -04:00
|
|
|
admin,
|
2016-03-10 10:58:22 -05:00
|
|
|
):
|
2014-08-12 10:10:52 -04:00
|
|
|
now = int(self.clock.time())
|
|
|
|
|
2016-03-01 09:32:56 -05:00
|
|
|
next_id = self._access_tokens_id_gen.get_next()
|
2015-04-07 07:05:36 -04:00
|
|
|
|
2014-08-12 10:10:52 -04:00
|
|
|
try:
|
2016-01-05 13:01:18 -05:00
|
|
|
if was_guest:
|
2016-07-08 10:15:55 -04:00
|
|
|
# Ensure that the guest user actually exists
|
2016-07-08 10:57:06 -04:00
|
|
|
# ``allow_none=False`` makes this raise an exception
|
|
|
|
# if the row isn't in the database.
|
2016-07-08 10:15:55 -04:00
|
|
|
self._simple_select_one_txn(
|
|
|
|
txn,
|
|
|
|
"users",
|
|
|
|
keyvalues={
|
|
|
|
"name": user_id,
|
|
|
|
"is_guest": 1,
|
|
|
|
},
|
|
|
|
retcols=("name",),
|
|
|
|
allow_none=False,
|
|
|
|
)
|
|
|
|
|
2016-07-05 12:30:22 -04:00
|
|
|
self._simple_update_one_txn(
|
|
|
|
txn,
|
|
|
|
"users",
|
|
|
|
keyvalues={
|
|
|
|
"name": user_id,
|
2016-07-08 10:15:55 -04:00
|
|
|
"is_guest": 1,
|
2016-07-05 12:30:22 -04:00
|
|
|
},
|
|
|
|
updatevalues={
|
|
|
|
"password_hash": password_hash,
|
|
|
|
"upgrade_ts": now,
|
|
|
|
"is_guest": 1 if make_guest else 0,
|
|
|
|
"appservice_id": appservice_id,
|
2016-07-05 12:34:25 -04:00
|
|
|
"admin": 1 if admin else 0,
|
2016-07-05 12:30:22 -04:00
|
|
|
}
|
|
|
|
)
|
2016-01-05 13:01:18 -05:00
|
|
|
else:
|
2016-07-05 12:30:22 -04:00
|
|
|
self._simple_insert_txn(
|
|
|
|
txn,
|
|
|
|
"users",
|
|
|
|
values={
|
|
|
|
"name": user_id,
|
|
|
|
"password_hash": password_hash,
|
|
|
|
"creation_ts": now,
|
|
|
|
"is_guest": 1 if make_guest else 0,
|
|
|
|
"appservice_id": appservice_id,
|
2016-07-05 12:34:25 -04:00
|
|
|
"admin": 1 if admin else 0,
|
2016-07-05 12:30:22 -04:00
|
|
|
}
|
|
|
|
)
|
2015-04-08 11:53:48 -04:00
|
|
|
except self.database_engine.module.IntegrityError:
|
2014-11-20 12:26:36 -05:00
|
|
|
raise StoreError(
|
|
|
|
400, "User ID already taken.", errcode=Codes.USER_IN_USE
|
|
|
|
)
|
2014-08-12 10:10:52 -04:00
|
|
|
|
2015-11-04 12:29:07 -05:00
|
|
|
if token:
|
|
|
|
# it's possible for this to get a conflict, but only for a single user
|
|
|
|
# since tokens are namespaced based on their user ID
|
|
|
|
txn.execute(
|
|
|
|
"INSERT INTO access_tokens(id, user_id, token)"
|
|
|
|
" VALUES (?,?,?)",
|
|
|
|
(next_id, user_id, token,)
|
|
|
|
)
|
2014-08-12 10:10:52 -04:00
|
|
|
|
2016-06-17 14:14:16 -04:00
|
|
|
if create_profile_with_localpart:
|
2017-11-01 11:51:25 -04:00
|
|
|
# set a default displayname serverside to avoid ugly race
|
|
|
|
# between auto-joins and clients trying to set displaynames
|
2016-06-17 14:14:16 -04:00
|
|
|
txn.execute(
|
2017-11-01 11:51:25 -04:00
|
|
|
"INSERT INTO profiles(user_id, displayname) VALUES (?,?)",
|
|
|
|
(create_profile_with_localpart, create_profile_with_localpart)
|
2016-06-17 14:14:16 -04:00
|
|
|
)
|
|
|
|
|
2016-08-18 09:59:55 -04:00
|
|
|
self._invalidate_cache_and_stream(
|
|
|
|
txn, self.get_user_by_id, (user_id,)
|
|
|
|
)
|
|
|
|
txn.call_after(self.is_guest.invalidate, (user_id,))
|
|
|
|
|
2015-08-26 08:42:45 -04:00
|
|
|
def get_users_by_id_case_insensitive(self, user_id):
|
|
|
|
"""Gets users that match user_id case insensitively.
|
|
|
|
Returns a mapping of user_id -> password_hash.
|
|
|
|
"""
|
|
|
|
def f(txn):
|
|
|
|
sql = (
|
2016-01-06 12:16:02 -05:00
|
|
|
"SELECT name, password_hash FROM users"
|
2015-08-26 08:42:45 -04:00
|
|
|
" WHERE lower(name) = lower(?)"
|
|
|
|
)
|
|
|
|
txn.execute(sql, (user_id,))
|
2017-03-23 13:53:49 -04:00
|
|
|
return dict(txn)
|
2015-08-26 08:42:45 -04:00
|
|
|
|
|
|
|
return self.runInteraction("get_users_by_id_case_insensitive", f)
|
|
|
|
|
2015-03-23 10:20:28 -04:00
|
|
|
def user_set_password_hash(self, user_id, password_hash):
|
|
|
|
"""
|
|
|
|
NB. This does *not* evict any cache because the one use for this
|
|
|
|
removes most of the entries subsequently anyway so it would be
|
|
|
|
pointless. Use flush_user separately.
|
|
|
|
"""
|
2016-08-18 09:59:55 -04:00
|
|
|
def user_set_password_hash_txn(txn):
|
|
|
|
self._simple_update_one_txn(
|
|
|
|
txn,
|
|
|
|
'users', {
|
|
|
|
'name': user_id
|
|
|
|
},
|
|
|
|
{
|
|
|
|
'password_hash': password_hash
|
|
|
|
}
|
|
|
|
)
|
|
|
|
self._invalidate_cache_and_stream(
|
|
|
|
txn, self.get_user_by_id, (user_id,)
|
|
|
|
)
|
|
|
|
return self.runInteraction(
|
|
|
|
"user_set_password_hash", user_set_password_hash_txn
|
|
|
|
)
|
2015-03-23 10:20:28 -04:00
|
|
|
|
2018-05-10 19:17:11 -04:00
|
|
|
def user_set_consent_version(self, user_id, consent_version):
|
|
|
|
"""Updates the user table to record privacy policy consent
|
|
|
|
|
|
|
|
Args:
|
|
|
|
user_id (str): full mxid of the user to update
|
|
|
|
consent_version (str): version of the policy the user has consented
|
|
|
|
to
|
|
|
|
|
|
|
|
Raises:
|
|
|
|
StoreError(404) if user not found
|
|
|
|
"""
|
2018-05-17 12:35:31 -04:00
|
|
|
def f(txn):
|
|
|
|
self._simple_update_one_txn(
|
|
|
|
txn,
|
|
|
|
table='users',
|
|
|
|
keyvalues={'name': user_id, },
|
|
|
|
updatevalues={'consent_version': consent_version, },
|
|
|
|
)
|
|
|
|
self._invalidate_cache_and_stream(
|
|
|
|
txn, self.get_user_by_id, (user_id,)
|
|
|
|
)
|
|
|
|
return self.runInteraction("user_set_consent_version", f)
|
|
|
|
|
|
|
|
def user_set_consent_server_notice_sent(self, user_id, consent_version):
|
|
|
|
"""Updates the user table to record that we have sent the user a server
|
|
|
|
notice about privacy policy consent
|
|
|
|
|
|
|
|
Args:
|
|
|
|
user_id (str): full mxid of the user to update
|
|
|
|
consent_version (str): version of the policy we have notified the
|
|
|
|
user about
|
|
|
|
|
|
|
|
Raises:
|
|
|
|
StoreError(404) if user not found
|
|
|
|
"""
|
|
|
|
def f(txn):
|
|
|
|
self._simple_update_one_txn(
|
|
|
|
txn,
|
|
|
|
table='users',
|
|
|
|
keyvalues={'name': user_id, },
|
|
|
|
updatevalues={'consent_server_notice_sent': consent_version, },
|
|
|
|
)
|
|
|
|
self._invalidate_cache_and_stream(
|
|
|
|
txn, self.get_user_by_id, (user_id,)
|
|
|
|
)
|
|
|
|
return self.runInteraction("user_set_consent_server_notice_sent", f)
|
2018-05-10 19:17:11 -04:00
|
|
|
|
2016-08-15 12:04:39 -04:00
|
|
|
def user_delete_access_tokens(self, user_id, except_token_id=None,
|
2017-10-31 16:35:58 -04:00
|
|
|
device_id=None):
|
2016-07-26 06:09:47 -04:00
|
|
|
"""
|
2017-10-31 16:35:58 -04:00
|
|
|
Invalidate access tokens belonging to a user
|
2016-07-26 06:09:47 -04:00
|
|
|
|
|
|
|
Args:
|
|
|
|
user_id (str): ID of user the tokens belong to
|
2016-08-15 12:04:39 -04:00
|
|
|
except_token_id (str): list of access_tokens IDs which should
|
2016-07-26 06:09:47 -04:00
|
|
|
*not* be deleted
|
|
|
|
device_id (str|None): ID of device the tokens are associated with.
|
|
|
|
If None, tokens associated with any device (or no device) will
|
|
|
|
be deleted
|
|
|
|
Returns:
|
2017-11-29 09:33:05 -05:00
|
|
|
defer.Deferred[list[str, int, str|None, int]]: a list of
|
|
|
|
(token, token id, device id) for each of the deleted tokens
|
2016-07-26 06:09:47 -04:00
|
|
|
"""
|
2016-08-15 12:04:39 -04:00
|
|
|
def f(txn):
|
|
|
|
keyvalues = {
|
|
|
|
"user_id": user_id,
|
|
|
|
}
|
2016-07-22 09:52:53 -04:00
|
|
|
if device_id is not None:
|
2016-08-15 12:04:39 -04:00
|
|
|
keyvalues["device_id"] = device_id
|
2016-07-22 09:52:53 -04:00
|
|
|
|
2016-08-15 12:04:39 -04:00
|
|
|
items = keyvalues.items()
|
|
|
|
where_clause = " AND ".join(k + " = ?" for k, _ in items)
|
|
|
|
values = [v for _, v in items]
|
|
|
|
if except_token_id:
|
|
|
|
where_clause += " AND id != ?"
|
|
|
|
values.append(except_token_id)
|
2016-03-11 11:27:50 -05:00
|
|
|
|
2016-08-15 12:04:39 -04:00
|
|
|
txn.execute(
|
2017-11-29 09:33:05 -05:00
|
|
|
"SELECT token, id, device_id FROM access_tokens WHERE %s" % where_clause,
|
2016-08-15 12:04:39 -04:00
|
|
|
values
|
|
|
|
)
|
2017-11-29 09:33:05 -05:00
|
|
|
tokens_and_devices = [(r[0], r[1], r[2]) for r in txn]
|
2016-03-11 11:27:50 -05:00
|
|
|
|
2017-11-29 09:33:05 -05:00
|
|
|
for token, _, _ in tokens_and_devices:
|
2016-08-15 12:04:39 -04:00
|
|
|
self._invalidate_cache_and_stream(
|
2017-11-01 11:42:38 -04:00
|
|
|
txn, self.get_user_by_access_token, (token,)
|
2016-03-11 11:27:50 -05:00
|
|
|
)
|
|
|
|
|
2016-08-15 12:04:39 -04:00
|
|
|
txn.execute(
|
|
|
|
"DELETE FROM access_tokens WHERE %s" % where_clause,
|
|
|
|
values
|
2016-07-26 06:09:47 -04:00
|
|
|
)
|
|
|
|
|
2017-11-01 11:42:38 -04:00
|
|
|
return tokens_and_devices
|
|
|
|
|
2017-11-09 10:15:33 -05:00
|
|
|
return self.runInteraction(
|
2016-07-26 06:09:47 -04:00
|
|
|
"user_delete_access_tokens", f,
|
|
|
|
)
|
2016-03-11 11:27:50 -05:00
|
|
|
|
|
|
|
def delete_access_token(self, access_token):
|
|
|
|
def f(txn):
|
|
|
|
self._simple_delete_one_txn(
|
|
|
|
txn,
|
|
|
|
table="access_tokens",
|
|
|
|
keyvalues={
|
|
|
|
"token": access_token
|
|
|
|
},
|
2016-03-11 08:14:18 -05:00
|
|
|
)
|
2016-03-11 11:27:50 -05:00
|
|
|
|
2016-08-15 12:04:39 -04:00
|
|
|
self._invalidate_cache_and_stream(
|
|
|
|
txn, self.get_user_by_access_token, (access_token,)
|
|
|
|
)
|
2016-03-11 11:27:50 -05:00
|
|
|
|
|
|
|
return self.runInteraction("delete_access_token", f)
|
2015-03-25 13:15:20 -04:00
|
|
|
|
2016-01-06 06:38:09 -05:00
|
|
|
@cachedInlineCallbacks()
|
2016-01-18 09:09:47 -05:00
|
|
|
def is_guest(self, user_id):
|
2016-01-06 06:38:09 -05:00
|
|
|
res = yield self._simple_select_one_onecol(
|
|
|
|
table="users",
|
2016-01-18 09:09:47 -05:00
|
|
|
keyvalues={"name": user_id},
|
2016-01-06 06:38:09 -05:00
|
|
|
retcol="is_guest",
|
|
|
|
allow_none=True,
|
|
|
|
desc="is_guest",
|
|
|
|
)
|
|
|
|
|
|
|
|
defer.returnValue(res if res else False)
|
|
|
|
|
2015-04-17 11:44:49 -04:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def user_add_threepid(self, user_id, medium, address, validated_at, added_at):
|
|
|
|
yield self._simple_upsert("user_threepids", {
|
|
|
|
"medium": medium,
|
|
|
|
"address": address,
|
|
|
|
}, {
|
2015-12-15 12:02:21 -05:00
|
|
|
"user_id": user_id,
|
2015-04-17 11:44:49 -04:00
|
|
|
"validated_at": validated_at,
|
|
|
|
"added_at": added_at,
|
2015-04-17 11:46:45 -04:00
|
|
|
})
|
2015-04-17 12:20:18 -04:00
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def user_get_threepids(self, user_id):
|
|
|
|
ret = yield self._simple_select_list(
|
|
|
|
"user_threepids", {
|
2015-04-29 11:57:14 -04:00
|
|
|
"user_id": user_id
|
2015-04-17 12:20:18 -04:00
|
|
|
},
|
|
|
|
['medium', 'address', 'validated_at', 'added_at'],
|
|
|
|
'user_get_threepids'
|
|
|
|
)
|
2015-04-17 14:53:47 -04:00
|
|
|
defer.returnValue(ret)
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
2015-05-01 10:04:20 -04:00
|
|
|
def get_user_id_by_threepid(self, medium, address):
|
2015-04-17 14:53:47 -04:00
|
|
|
ret = yield self._simple_select_one(
|
|
|
|
"user_threepids",
|
|
|
|
{
|
|
|
|
"medium": medium,
|
|
|
|
"address": address
|
|
|
|
},
|
2015-05-01 10:04:20 -04:00
|
|
|
['user_id'], True, 'get_user_id_by_threepid'
|
2015-04-17 14:53:47 -04:00
|
|
|
)
|
|
|
|
if ret:
|
2015-05-01 10:04:20 -04:00
|
|
|
defer.returnValue(ret['user_id'])
|
2015-04-17 14:56:04 -04:00
|
|
|
defer.returnValue(None)
|
2015-09-22 07:57:40 -04:00
|
|
|
|
2016-12-20 13:27:30 -05:00
|
|
|
def user_delete_threepid(self, user_id, medium, address):
|
|
|
|
return self._simple_delete(
|
|
|
|
"user_threepids",
|
|
|
|
keyvalues={
|
|
|
|
"user_id": user_id,
|
|
|
|
"medium": medium,
|
|
|
|
"address": address,
|
|
|
|
},
|
|
|
|
desc="user_delete_threepids",
|
|
|
|
)
|
|
|
|
|
2015-09-22 07:57:40 -04:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def count_all_users(self):
|
2015-09-22 08:47:40 -04:00
|
|
|
"""Counts all users registered on the homeserver."""
|
2015-09-22 07:57:40 -04:00
|
|
|
def _count_users(txn):
|
|
|
|
txn.execute("SELECT COUNT(*) AS users FROM users")
|
|
|
|
rows = self.cursor_to_dict(txn)
|
|
|
|
if rows:
|
|
|
|
return rows[0]["users"]
|
|
|
|
return 0
|
|
|
|
|
|
|
|
ret = yield self.runInteraction("count_users", _count_users)
|
2017-06-15 04:39:39 -04:00
|
|
|
defer.returnValue(ret)
|
|
|
|
|
2018-05-22 13:09:09 -04:00
|
|
|
def count_daily_user_type(self):
|
|
|
|
"""
|
|
|
|
Counts 1) native non guest users
|
|
|
|
2) native guests users
|
|
|
|
3) bridged users
|
|
|
|
who registered on the homeserver in the past 24 hours
|
|
|
|
"""
|
|
|
|
def _count_daily_user_type(txn):
|
|
|
|
yesterday = int(self._clock.time()) - (60 * 60 * 24)
|
|
|
|
|
|
|
|
sql = """
|
|
|
|
SELECT user_type, COALESCE(count(*), 0) AS count FROM (
|
|
|
|
SELECT
|
|
|
|
CASE
|
|
|
|
WHEN is_guest=0 AND appservice_id IS NULL THEN 'native'
|
|
|
|
WHEN is_guest=1 AND appservice_id IS NULL THEN 'guest'
|
|
|
|
WHEN is_guest=0 AND appservice_id IS NOT NULL THEN 'bridged'
|
|
|
|
END AS user_type
|
|
|
|
FROM users
|
|
|
|
WHERE creation_ts > ?
|
|
|
|
) AS t GROUP BY user_type
|
|
|
|
"""
|
|
|
|
results = {'native': 0, 'guest': 0, 'bridged': 0}
|
|
|
|
txn.execute(sql, (yesterday,))
|
|
|
|
for row in txn:
|
|
|
|
results[row[0]] = row[1]
|
|
|
|
return results
|
|
|
|
return self.runInteraction("count_daily_user_type", _count_daily_user_type)
|
|
|
|
|
2017-06-15 04:39:39 -04:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def count_nonbridged_users(self):
|
|
|
|
def _count_users(txn):
|
|
|
|
txn.execute("""
|
|
|
|
SELECT COALESCE(COUNT(*), 0) FROM users
|
|
|
|
WHERE appservice_id IS NULL
|
|
|
|
""")
|
|
|
|
count, = txn.fetchone()
|
|
|
|
return count
|
|
|
|
|
|
|
|
ret = yield self.runInteraction("count_users", _count_users)
|
2015-09-22 07:57:40 -04:00
|
|
|
defer.returnValue(ret)
|
2016-02-05 06:22:30 -05:00
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def find_next_generated_user_id_localpart(self):
|
|
|
|
"""
|
|
|
|
Gets the localpart of the next generated user ID.
|
|
|
|
|
|
|
|
Generated user IDs are integers, and we aim for them to be as small as
|
|
|
|
we can. Unfortunately, it's possible some of them are already taken by
|
|
|
|
existing users, and there may be gaps in the already taken range. This
|
|
|
|
function returns the start of the first allocatable gap. This is to
|
|
|
|
avoid the case of ID 10000000 being pre-allocated, so us wasting the
|
|
|
|
first (and shortest) many generated user IDs.
|
|
|
|
"""
|
|
|
|
def _find_next_generated_user_id(txn):
|
|
|
|
txn.execute("SELECT name FROM users")
|
|
|
|
|
|
|
|
regex = re.compile("^@(\d+):")
|
|
|
|
|
|
|
|
found = set()
|
|
|
|
|
2018-03-26 07:02:44 -04:00
|
|
|
for user_id, in txn:
|
2016-02-05 06:22:30 -05:00
|
|
|
match = regex.search(user_id)
|
|
|
|
if match:
|
|
|
|
found.add(int(match.group(1)))
|
2018-04-28 07:57:00 -04:00
|
|
|
for i in range(len(found) + 1):
|
2016-02-05 06:22:30 -05:00
|
|
|
if i not in found:
|
|
|
|
return i
|
|
|
|
|
|
|
|
defer.returnValue((yield self.runInteraction(
|
|
|
|
"find_next_generated_user_id",
|
|
|
|
_find_next_generated_user_id
|
|
|
|
)))
|
2016-02-24 09:41:25 -05:00
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def get_3pid_guest_access_token(self, medium, address):
|
|
|
|
ret = yield self._simple_select_one(
|
|
|
|
"threepid_guest_access_tokens",
|
|
|
|
{
|
|
|
|
"medium": medium,
|
|
|
|
"address": address
|
|
|
|
},
|
|
|
|
["guest_access_token"], True, 'get_3pid_guest_access_token'
|
|
|
|
)
|
|
|
|
if ret:
|
|
|
|
defer.returnValue(ret["guest_access_token"])
|
|
|
|
defer.returnValue(None)
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def save_or_get_3pid_guest_access_token(
|
|
|
|
self, medium, address, access_token, inviter_user_id
|
|
|
|
):
|
|
|
|
"""
|
|
|
|
Gets the 3pid's guest access token if exists, else saves access_token.
|
|
|
|
|
2016-04-01 11:08:59 -04:00
|
|
|
Args:
|
|
|
|
medium (str): Medium of the 3pid. Must be "email".
|
|
|
|
address (str): 3pid address.
|
|
|
|
access_token (str): The access token to persist if none is
|
|
|
|
already persisted.
|
|
|
|
inviter_user_id (str): User ID of the inviter.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
deferred str: Whichever access token is persisted at the end
|
2016-02-24 09:41:25 -05:00
|
|
|
of this function call.
|
|
|
|
"""
|
|
|
|
def insert(txn):
|
|
|
|
txn.execute(
|
|
|
|
"INSERT INTO threepid_guest_access_tokens "
|
|
|
|
"(medium, address, guest_access_token, first_inviter) "
|
|
|
|
"VALUES (?, ?, ?, ?)",
|
|
|
|
(medium, address, access_token, inviter_user_id)
|
|
|
|
)
|
|
|
|
|
|
|
|
try:
|
|
|
|
yield self.runInteraction("save_3pid_guest_access_token", insert)
|
|
|
|
defer.returnValue(access_token)
|
|
|
|
except self.database_engine.module.IntegrityError:
|
|
|
|
ret = yield self.get_3pid_guest_access_token(medium, address)
|
|
|
|
defer.returnValue(ret)
|
2018-05-09 09:54:28 -04:00
|
|
|
|
|
|
|
def add_user_pending_deactivation(self, user_id):
|
2018-05-10 07:20:40 -04:00
|
|
|
"""
|
|
|
|
Adds a user to the table of users who need to be parted from all the rooms they're
|
|
|
|
in
|
|
|
|
"""
|
2018-05-09 09:54:28 -04:00
|
|
|
return self._simple_insert(
|
|
|
|
"users_pending_deactivation",
|
|
|
|
values={
|
|
|
|
"user_id": user_id,
|
|
|
|
},
|
|
|
|
desc="add_user_pending_deactivation",
|
|
|
|
)
|
|
|
|
|
|
|
|
def del_user_pending_deactivation(self, user_id):
|
2018-05-10 07:20:40 -04:00
|
|
|
"""
|
|
|
|
Removes the given user to the table of users who need to be parted from all the
|
|
|
|
rooms they're in, effectively marking that user as fully deactivated.
|
|
|
|
"""
|
2018-06-26 05:52:52 -04:00
|
|
|
# XXX: This should be simple_delete_one but we failed to put a unique index on
|
|
|
|
# the table, so somehow duplicate entries have ended up in it.
|
|
|
|
return self._simple_delete(
|
2018-05-09 09:54:28 -04:00
|
|
|
"users_pending_deactivation",
|
|
|
|
keyvalues={
|
|
|
|
"user_id": user_id,
|
|
|
|
},
|
|
|
|
desc="del_user_pending_deactivation",
|
|
|
|
)
|
|
|
|
|
|
|
|
def get_user_pending_deactivation(self):
|
2018-05-10 07:20:40 -04:00
|
|
|
"""
|
|
|
|
Gets one user from the table of users waiting to be parted from all the rooms
|
|
|
|
they're in.
|
|
|
|
"""
|
2018-05-09 09:54:28 -04:00
|
|
|
return self._simple_select_one_onecol(
|
|
|
|
"users_pending_deactivation",
|
|
|
|
keyvalues={},
|
|
|
|
retcol="user_id",
|
|
|
|
allow_none=True,
|
|
|
|
desc="get_users_pending_deactivation",
|
|
|
|
)
|