mirror of
https://git.anonymousland.org/anonymousland/synapse-product.git
synced 2024-10-01 08:25:44 -04:00
Merge branch 'develop' of github.com:matrix-org/synapse into erikj/ratelimit_3pid_invite
This commit is contained in:
commit
d6118c5be6
32
INSTALL.md
32
INSTALL.md
@ -257,18 +257,40 @@ https://github.com/spantaleev/matrix-docker-ansible-deploy
|
|||||||
#### Matrix.org packages
|
#### Matrix.org packages
|
||||||
|
|
||||||
Matrix.org provides Debian/Ubuntu packages of the latest stable version of
|
Matrix.org provides Debian/Ubuntu packages of the latest stable version of
|
||||||
Synapse via https://matrix.org/packages/debian/. To use them:
|
Synapse via https://packages.matrix.org/debian/. To use them:
|
||||||
|
|
||||||
|
For Debian 9 (Stretch), Ubuntu 16.04 (Xenial), and later:
|
||||||
|
|
||||||
```
|
```
|
||||||
sudo apt install -y lsb-release curl apt-transport-https
|
sudo apt install -y lsb-release wget apt-transport-https
|
||||||
echo "deb https://matrix.org/packages/debian `lsb_release -cs` main" |
|
sudo wget -O /usr/share/keyrings/matrix-org-archive-keyring.gpg https://packages.matrix.org/debian/matrix-org-archive-keyring.gpg
|
||||||
|
echo "deb [signed-by=/usr/share/keyrings/matrix-org-archive-keyring.gpg] https://packages.matrix.org/debian/ $(lsb_release -cs) main" |
|
||||||
sudo tee /etc/apt/sources.list.d/matrix-org.list
|
sudo tee /etc/apt/sources.list.d/matrix-org.list
|
||||||
curl "https://matrix.org/packages/debian/repo-key.asc" |
|
|
||||||
sudo apt-key add -
|
|
||||||
sudo apt update
|
sudo apt update
|
||||||
sudo apt install matrix-synapse-py3
|
sudo apt install matrix-synapse-py3
|
||||||
```
|
```
|
||||||
|
|
||||||
|
For Debian 8 (Jessie):
|
||||||
|
|
||||||
|
```
|
||||||
|
sudo apt install -y lsb-release wget apt-transport-https
|
||||||
|
sudo wget -O /etc/apt/trusted.gpg.d/matrix-org-archive-keyring.gpg https://packages.matrix.org/debian/matrix-org-archive-keyring.gpg
|
||||||
|
echo "deb [signed-by=5586CCC0CBBBEFC7A25811ADF473DD4473365DE1] https://packages.matrix.org/debian/ $(lsb_release -cs) main" |
|
||||||
|
sudo tee /etc/apt/sources.list.d/matrix-org.list
|
||||||
|
sudo apt update
|
||||||
|
sudo apt install matrix-synapse-py3
|
||||||
|
```
|
||||||
|
|
||||||
|
The fingerprint of the repository signing key is AAF9AE843A7584B5A3E4CD2BCF45A512DE2DA058.
|
||||||
|
|
||||||
|
**Note**: if you followed a previous version of these instructions which
|
||||||
|
recommended using `apt-key add` to add an old key from
|
||||||
|
`https://matrix.org/packages/debian/`, you should note that this key has been
|
||||||
|
revoked. You should remove the old key with `sudo apt-key remove
|
||||||
|
C35EB17E1EAE708E6603A9B3AD0592FE47F0DF61`, and follow the above instructions to
|
||||||
|
update your configuration.
|
||||||
|
|
||||||
|
|
||||||
#### Downstream Debian/Ubuntu packages
|
#### Downstream Debian/Ubuntu packages
|
||||||
|
|
||||||
For `buster` and `sid`, Synapse is available in the Debian repositories and
|
For `buster` and `sid`, Synapse is available in the Debian repositories and
|
||||||
|
@ -173,7 +173,7 @@ Synapse offers two database engines:
|
|||||||
* `PostgreSQL <https://www.postgresql.org>`_
|
* `PostgreSQL <https://www.postgresql.org>`_
|
||||||
|
|
||||||
By default Synapse uses SQLite in and doing so trades performance for convenience.
|
By default Synapse uses SQLite in and doing so trades performance for convenience.
|
||||||
SQLite is only recommended in Synapse for testing purposes or for servers with
|
SQLite is only recommended in Synapse for testing purposes or for servers with
|
||||||
light workloads.
|
light workloads.
|
||||||
|
|
||||||
Almost all installations should opt to use PostreSQL. Advantages include:
|
Almost all installations should opt to use PostreSQL. Advantages include:
|
||||||
@ -272,7 +272,7 @@ to install using pip and a virtualenv::
|
|||||||
|
|
||||||
virtualenv -p python3 env
|
virtualenv -p python3 env
|
||||||
source env/bin/activate
|
source env/bin/activate
|
||||||
python -m pip install -e .[all]
|
python -m pip install --no-pep-517 -e .[all]
|
||||||
|
|
||||||
This will run a process of downloading and installing all the needed
|
This will run a process of downloading and installing all the needed
|
||||||
dependencies into a virtual env.
|
dependencies into a virtual env.
|
||||||
|
1
changelog.d/4339.feature
Normal file
1
changelog.d/4339.feature
Normal file
@ -0,0 +1 @@
|
|||||||
|
Add systemd-python to the optional dependencies to enable logging to the systemd journal. Install with `pip install matrix-synapse[systemd]`.
|
1
changelog.d/4967.feature
Normal file
1
changelog.d/4967.feature
Normal file
@ -0,0 +1 @@
|
|||||||
|
Implementation of [MSC1711](https://github.com/matrix-org/matrix-doc/pull/1711) including config options for requiring valid TLS certificates for federation traffic, the ability to disable TLS validation for specific domains, and the ability to specify your own list of CA certificates.
|
1
changelog.d/4972.misc
Normal file
1
changelog.d/4972.misc
Normal file
@ -0,0 +1 @@
|
|||||||
|
Reduce log level of .well-known/matrix/client responses.
|
1
changelog.d/4992.misc
Normal file
1
changelog.d/4992.misc
Normal file
@ -0,0 +1 @@
|
|||||||
|
Remove a number of unused tables from the database schema.
|
1
changelog.d/5001.misc
Normal file
1
changelog.d/5001.misc
Normal file
@ -0,0 +1 @@
|
|||||||
|
Clean up some code in the server-key Keyring.
|
1
changelog.d/5005.misc
Normal file
1
changelog.d/5005.misc
Normal file
@ -0,0 +1 @@
|
|||||||
|
Convert SYNAPSE_NO_TLS Docker variable to boolean for user friendliness. Contributed by Gabriel Eckerson.
|
1
changelog.d/5009.bugfix
Normal file
1
changelog.d/5009.bugfix
Normal file
@ -0,0 +1 @@
|
|||||||
|
Clients timing out/disappearing while downloading from the media repository will now no longer log a spurious "Producer was not unregistered" message.
|
1
changelog.d/5010.feature
Normal file
1
changelog.d/5010.feature
Normal file
@ -0,0 +1 @@
|
|||||||
|
Add config option to block users from looking up 3PIDs.
|
1
changelog.d/5020.feature
Normal file
1
changelog.d/5020.feature
Normal file
@ -0,0 +1 @@
|
|||||||
|
Add context to phonehome stats.
|
1
changelog.d/5024.misc
Normal file
1
changelog.d/5024.misc
Normal file
@ -0,0 +1 @@
|
|||||||
|
Store the notary server name correctly in server_keys_json.
|
1
changelog.d/5027.feature
Normal file
1
changelog.d/5027.feature
Normal file
@ -0,0 +1 @@
|
|||||||
|
Add time-based account expiration.
|
1
changelog.d/5028.misc
Normal file
1
changelog.d/5028.misc
Normal file
@ -0,0 +1 @@
|
|||||||
|
Remove a number of unused tables from the database schema.
|
1
changelog.d/5030.misc
Normal file
1
changelog.d/5030.misc
Normal file
@ -0,0 +1 @@
|
|||||||
|
Rewrite Datastore.get_server_verify_keys to reduce the number of database transactions.
|
1
changelog.d/5032.bugfix
Normal file
1
changelog.d/5032.bugfix
Normal file
@ -0,0 +1 @@
|
|||||||
|
Fix "cannot import name execute_batch" error with postgres.
|
1
changelog.d/5033.misc
Normal file
1
changelog.d/5033.misc
Normal file
@ -0,0 +1 @@
|
|||||||
|
Remove a number of unused tables from the database schema.
|
1
changelog.d/5035.bugfix
Normal file
1
changelog.d/5035.bugfix
Normal file
@ -0,0 +1 @@
|
|||||||
|
Fix disappearing exceptions in manhole.
|
1
changelog.d/5046.misc
Normal file
1
changelog.d/5046.misc
Normal file
@ -0,0 +1 @@
|
|||||||
|
Remove extraneous period from copyright headers.
|
1
changelog.d/5047.feature
Normal file
1
changelog.d/5047.feature
Normal file
@ -0,0 +1 @@
|
|||||||
|
Add time-based account expiration.
|
1
changelog.d/5065.feature
Normal file
1
changelog.d/5065.feature
Normal file
@ -0,0 +1 @@
|
|||||||
|
Add support for handling /verions, /voip and /push_rules client endpoints to client_reader worker.
|
1
changelog.d/5067.misc
Normal file
1
changelog.d/5067.misc
Normal file
@ -0,0 +1 @@
|
|||||||
|
Update documentation for where to get Synapse packages.
|
1
changelog.d/5071.bugfix
Normal file
1
changelog.d/5071.bugfix
Normal file
@ -0,0 +1 @@
|
|||||||
|
Make sure we're not registering the same 3pid twice on registration.
|
1
changelog.d/5073.feature
Normal file
1
changelog.d/5073.feature
Normal file
@ -0,0 +1 @@
|
|||||||
|
Add time-based account expiration.
|
1
changelog.d/5077.bugfix
Normal file
1
changelog.d/5077.bugfix
Normal file
@ -0,0 +1 @@
|
|||||||
|
Don't crash on lack of expiry templates.
|
1
changelog.d/5098.misc
Normal file
1
changelog.d/5098.misc
Normal file
@ -0,0 +1 @@
|
|||||||
|
Add workarounds for pep-517 install errors.
|
1
changelog.d/5103.bugfix
Normal file
1
changelog.d/5103.bugfix
Normal file
@ -0,0 +1 @@
|
|||||||
|
Fix bug where presence updates were sent to all servers in a room when a new server joined, rather than to just the new server.
|
File diff suppressed because one or more lines are too long
@ -50,7 +50,9 @@ RUN apt-get update -qq -o Acquire::Languages=none \
|
|||||||
debhelper \
|
debhelper \
|
||||||
devscripts \
|
devscripts \
|
||||||
dh-systemd \
|
dh-systemd \
|
||||||
|
libsystemd-dev \
|
||||||
lsb-release \
|
lsb-release \
|
||||||
|
pkg-config \
|
||||||
python3-dev \
|
python3-dev \
|
||||||
python3-pip \
|
python3-pip \
|
||||||
python3-setuptools \
|
python3-setuptools \
|
||||||
|
@ -102,8 +102,9 @@ when ``SYNAPSE_CONFIG_PATH`` is not set.
|
|||||||
* ``SYNAPSE_SERVER_NAME`` (mandatory), the server public hostname.
|
* ``SYNAPSE_SERVER_NAME`` (mandatory), the server public hostname.
|
||||||
* ``SYNAPSE_REPORT_STATS``, (mandatory, ``yes`` or ``no``), enable anonymous
|
* ``SYNAPSE_REPORT_STATS``, (mandatory, ``yes`` or ``no``), enable anonymous
|
||||||
statistics reporting back to the Matrix project which helps us to get funding.
|
statistics reporting back to the Matrix project which helps us to get funding.
|
||||||
* ``SYNAPSE_NO_TLS``, set this variable to disable TLS in Synapse (use this if
|
* `SYNAPSE_NO_TLS`, (accepts `true`, `false`, `on`, `off`, `1`, `0`, `yes`, `no`]): disable
|
||||||
you run your own TLS-capable reverse proxy).
|
TLS in Synapse (use this if you run your own TLS-capable reverse proxy). Defaults
|
||||||
|
to `false` (ie, TLS is enabled by default).
|
||||||
* ``SYNAPSE_ENABLE_REGISTRATION``, set this variable to enable registration on
|
* ``SYNAPSE_ENABLE_REGISTRATION``, set this variable to enable registration on
|
||||||
the Synapse instance.
|
the Synapse instance.
|
||||||
* ``SYNAPSE_ALLOW_GUEST``, set this variable to allow guest joining this server.
|
* ``SYNAPSE_ALLOW_GUEST``, set this variable to allow guest joining this server.
|
||||||
|
@ -59,6 +59,18 @@ else:
|
|||||||
if not os.path.exists("/compiled"): os.mkdir("/compiled")
|
if not os.path.exists("/compiled"): os.mkdir("/compiled")
|
||||||
|
|
||||||
config_path = "/compiled/homeserver.yaml"
|
config_path = "/compiled/homeserver.yaml"
|
||||||
|
|
||||||
|
# Convert SYNAPSE_NO_TLS to boolean if exists
|
||||||
|
if "SYNAPSE_NO_TLS" in environ:
|
||||||
|
tlsanswerstring = str.lower(environ["SYNAPSE_NO_TLS"])
|
||||||
|
if tlsanswerstring in ("true", "on", "1", "yes"):
|
||||||
|
environ["SYNAPSE_NO_TLS"] = True
|
||||||
|
else:
|
||||||
|
if tlsanswerstring in ("false", "off", "0", "no"):
|
||||||
|
environ["SYNAPSE_NO_TLS"] = False
|
||||||
|
else:
|
||||||
|
print("Environment variable \"SYNAPSE_NO_TLS\" found but value \"" + tlsanswerstring + "\" unrecognized; exiting.")
|
||||||
|
sys.exit(2)
|
||||||
|
|
||||||
convert("/conf/homeserver.yaml", config_path, environ)
|
convert("/conf/homeserver.yaml", config_path, environ)
|
||||||
convert("/conf/log.config", "/compiled/log.config", environ)
|
convert("/conf/log.config", "/compiled/log.config", environ)
|
||||||
|
@ -177,7 +177,6 @@ You can do this with a `.well-known` file as follows:
|
|||||||
on `customer.example.net:8000` it correctly handles HTTP requests with
|
on `customer.example.net:8000` it correctly handles HTTP requests with
|
||||||
Host header set to `customer.example.net:8000`.
|
Host header set to `customer.example.net:8000`.
|
||||||
|
|
||||||
|
|
||||||
## FAQ
|
## FAQ
|
||||||
|
|
||||||
### Synapse 0.99.0 has just been released, what do I need to do right now?
|
### Synapse 0.99.0 has just been released, what do I need to do right now?
|
||||||
|
42
docs/admin_api/account_validity.rst
Normal file
42
docs/admin_api/account_validity.rst
Normal file
@ -0,0 +1,42 @@
|
|||||||
|
Account validity API
|
||||||
|
====================
|
||||||
|
|
||||||
|
This API allows a server administrator to manage the validity of an account. To
|
||||||
|
use it, you must enable the account validity feature (under
|
||||||
|
``account_validity``) in Synapse's configuration.
|
||||||
|
|
||||||
|
Renew account
|
||||||
|
-------------
|
||||||
|
|
||||||
|
This API extends the validity of an account by as much time as configured in the
|
||||||
|
``period`` parameter from the ``account_validity`` configuration.
|
||||||
|
|
||||||
|
The API is::
|
||||||
|
|
||||||
|
POST /_matrix/client/unstable/account_validity/send_mail
|
||||||
|
|
||||||
|
with the following body:
|
||||||
|
|
||||||
|
.. code:: json
|
||||||
|
|
||||||
|
{
|
||||||
|
"user_id": "<user ID for the account to renew>",
|
||||||
|
"expiration_ts": 0,
|
||||||
|
"enable_renewal_emails": true
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
``expiration_ts`` is an optional parameter and overrides the expiration date,
|
||||||
|
which otherwise defaults to now + validity period.
|
||||||
|
|
||||||
|
``enable_renewal_emails`` is also an optional parameter and enables/disables
|
||||||
|
sending renewal emails to the user. Defaults to true.
|
||||||
|
|
||||||
|
The API returns with the new expiration date for this account, as a timestamp in
|
||||||
|
milliseconds since epoch:
|
||||||
|
|
||||||
|
.. code:: json
|
||||||
|
|
||||||
|
{
|
||||||
|
"expiration_ts": 0
|
||||||
|
}
|
@ -236,6 +236,9 @@ listeners:
|
|||||||
# - medium: 'email'
|
# - medium: 'email'
|
||||||
# address: 'reserved_user@example.com'
|
# address: 'reserved_user@example.com'
|
||||||
|
|
||||||
|
# Used by phonehome stats to group together related servers.
|
||||||
|
#server_context: context
|
||||||
|
|
||||||
|
|
||||||
## TLS ##
|
## TLS ##
|
||||||
|
|
||||||
@ -257,6 +260,40 @@ listeners:
|
|||||||
#
|
#
|
||||||
#tls_private_key_path: "CONFDIR/SERVERNAME.tls.key"
|
#tls_private_key_path: "CONFDIR/SERVERNAME.tls.key"
|
||||||
|
|
||||||
|
# Whether to verify TLS certificates when sending federation traffic.
|
||||||
|
#
|
||||||
|
# This currently defaults to `false`, however this will change in
|
||||||
|
# Synapse 1.0 when valid federation certificates will be required.
|
||||||
|
#
|
||||||
|
#federation_verify_certificates: true
|
||||||
|
|
||||||
|
# Skip federation certificate verification on the following whitelist
|
||||||
|
# of domains.
|
||||||
|
#
|
||||||
|
# This setting should only be used in very specific cases, such as
|
||||||
|
# federation over Tor hidden services and similar. For private networks
|
||||||
|
# of homeservers, you likely want to use a private CA instead.
|
||||||
|
#
|
||||||
|
# Only effective if federation_verify_certicates is `true`.
|
||||||
|
#
|
||||||
|
#federation_certificate_verification_whitelist:
|
||||||
|
# - lon.example.com
|
||||||
|
# - *.domain.com
|
||||||
|
# - *.onion
|
||||||
|
|
||||||
|
# List of custom certificate authorities for federation traffic.
|
||||||
|
#
|
||||||
|
# This setting should only normally be used within a private network of
|
||||||
|
# homeservers.
|
||||||
|
#
|
||||||
|
# Note that this list will replace those that are provided by your
|
||||||
|
# operating environment. Certificates must be in PEM format.
|
||||||
|
#
|
||||||
|
#federation_custom_ca_list:
|
||||||
|
# - myCA1.pem
|
||||||
|
# - myCA2.pem
|
||||||
|
# - myCA3.pem
|
||||||
|
|
||||||
# ACME support: This will configure Synapse to request a valid TLS certificate
|
# ACME support: This will configure Synapse to request a valid TLS certificate
|
||||||
# for your configured `server_name` via Let's Encrypt.
|
# for your configured `server_name` via Let's Encrypt.
|
||||||
#
|
#
|
||||||
@ -643,6 +680,32 @@ uploads_path: "DATADIR/uploads"
|
|||||||
#
|
#
|
||||||
#enable_registration: false
|
#enable_registration: false
|
||||||
|
|
||||||
|
# Optional account validity configuration. This allows for accounts to be denied
|
||||||
|
# any request after a given period.
|
||||||
|
#
|
||||||
|
# ``enabled`` defines whether the account validity feature is enabled. Defaults
|
||||||
|
# to False.
|
||||||
|
#
|
||||||
|
# ``period`` allows setting the period after which an account is valid
|
||||||
|
# after its registration. When renewing the account, its validity period
|
||||||
|
# will be extended by this amount of time. This parameter is required when using
|
||||||
|
# the account validity feature.
|
||||||
|
#
|
||||||
|
# ``renew_at`` is the amount of time before an account's expiry date at which
|
||||||
|
# Synapse will send an email to the account's email address with a renewal link.
|
||||||
|
# This needs the ``email`` and ``public_baseurl`` configuration sections to be
|
||||||
|
# filled.
|
||||||
|
#
|
||||||
|
# ``renew_email_subject`` is the subject of the email sent out with the renewal
|
||||||
|
# link. ``%(app)s`` can be used as a placeholder for the ``app_name`` parameter
|
||||||
|
# from the ``email`` section.
|
||||||
|
#
|
||||||
|
#account_validity:
|
||||||
|
# enabled: True
|
||||||
|
# period: 6w
|
||||||
|
# renew_at: 1w
|
||||||
|
# renew_email_subject: "Renew your %(app)s account"
|
||||||
|
|
||||||
# The user must provide all of the below types of 3PID when registering.
|
# The user must provide all of the below types of 3PID when registering.
|
||||||
#
|
#
|
||||||
#registrations_require_3pid:
|
#registrations_require_3pid:
|
||||||
@ -665,6 +728,10 @@ uploads_path: "DATADIR/uploads"
|
|||||||
# - medium: msisdn
|
# - medium: msisdn
|
||||||
# pattern: '\+44'
|
# pattern: '\+44'
|
||||||
|
|
||||||
|
# Enable 3PIDs lookup requests to identity servers from this server.
|
||||||
|
#
|
||||||
|
#enable_3pid_lookup: true
|
||||||
|
|
||||||
# If set, allows registration of standard or admin accounts by anyone who
|
# If set, allows registration of standard or admin accounts by anyone who
|
||||||
# has the shared secret, even if registration is otherwise disabled.
|
# has the shared secret, even if registration is otherwise disabled.
|
||||||
#
|
#
|
||||||
@ -884,7 +951,7 @@ password_config:
|
|||||||
|
|
||||||
|
|
||||||
|
|
||||||
# Enable sending emails for notification events
|
# Enable sending emails for notification events or expiry notices
|
||||||
# Defining a custom URL for Riot is only needed if email notifications
|
# Defining a custom URL for Riot is only needed if email notifications
|
||||||
# should contain links to a self-hosted installation of Riot; when set
|
# should contain links to a self-hosted installation of Riot; when set
|
||||||
# the "app_name" setting is ignored.
|
# the "app_name" setting is ignored.
|
||||||
@ -906,6 +973,9 @@ password_config:
|
|||||||
# #template_dir: res/templates
|
# #template_dir: res/templates
|
||||||
# notif_template_html: notif_mail.html
|
# notif_template_html: notif_mail.html
|
||||||
# notif_template_text: notif_mail.txt
|
# notif_template_text: notif_mail.txt
|
||||||
|
# # Templates for account expiry notices.
|
||||||
|
# expiry_template_html: notice_expiry.html
|
||||||
|
# expiry_template_text: notice_expiry.txt
|
||||||
# notif_for_new_users: True
|
# notif_for_new_users: True
|
||||||
# riot_base_url: "http://localhost/riot"
|
# riot_base_url: "http://localhost/riot"
|
||||||
|
|
||||||
|
@ -58,15 +58,11 @@ BOOLEAN_COLUMNS = {
|
|||||||
|
|
||||||
|
|
||||||
APPEND_ONLY_TABLES = [
|
APPEND_ONLY_TABLES = [
|
||||||
"event_content_hashes",
|
|
||||||
"event_reference_hashes",
|
"event_reference_hashes",
|
||||||
"event_signatures",
|
|
||||||
"event_edge_hashes",
|
|
||||||
"events",
|
"events",
|
||||||
"event_json",
|
"event_json",
|
||||||
"state_events",
|
"state_events",
|
||||||
"room_memberships",
|
"room_memberships",
|
||||||
"feedback",
|
|
||||||
"topics",
|
"topics",
|
||||||
"room_names",
|
"room_names",
|
||||||
"rooms",
|
"rooms",
|
||||||
@ -88,7 +84,6 @@ APPEND_ONLY_TABLES = [
|
|||||||
"event_search",
|
"event_search",
|
||||||
"presence_stream",
|
"presence_stream",
|
||||||
"push_rules_stream",
|
"push_rules_stream",
|
||||||
"current_state_resets",
|
|
||||||
"ex_outlier_stream",
|
"ex_outlier_stream",
|
||||||
"cache_invalidation_stream",
|
"cache_invalidation_stream",
|
||||||
"public_room_list_stream",
|
"public_room_list_stream",
|
||||||
|
6
setup.py
6
setup.py
@ -86,13 +86,9 @@ long_description = read_file(("README.rst",))
|
|||||||
|
|
||||||
REQUIREMENTS = dependencies['REQUIREMENTS']
|
REQUIREMENTS = dependencies['REQUIREMENTS']
|
||||||
CONDITIONAL_REQUIREMENTS = dependencies['CONDITIONAL_REQUIREMENTS']
|
CONDITIONAL_REQUIREMENTS = dependencies['CONDITIONAL_REQUIREMENTS']
|
||||||
|
ALL_OPTIONAL_REQUIREMENTS = dependencies['ALL_OPTIONAL_REQUIREMENTS']
|
||||||
|
|
||||||
# Make `pip install matrix-synapse[all]` install all the optional dependencies.
|
# Make `pip install matrix-synapse[all]` install all the optional dependencies.
|
||||||
ALL_OPTIONAL_REQUIREMENTS = set()
|
|
||||||
|
|
||||||
for optional_deps in CONDITIONAL_REQUIREMENTS.values():
|
|
||||||
ALL_OPTIONAL_REQUIREMENTS = set(optional_deps) | ALL_OPTIONAL_REQUIREMENTS
|
|
||||||
|
|
||||||
CONDITIONAL_REQUIREMENTS["all"] = list(ALL_OPTIONAL_REQUIREMENTS)
|
CONDITIONAL_REQUIREMENTS["all"] = list(ALL_OPTIONAL_REQUIREMENTS)
|
||||||
|
|
||||||
|
|
||||||
|
@ -64,6 +64,8 @@ class Auth(object):
|
|||||||
self.token_cache = LruCache(CACHE_SIZE_FACTOR * 10000)
|
self.token_cache = LruCache(CACHE_SIZE_FACTOR * 10000)
|
||||||
register_cache("cache", "token_cache", self.token_cache)
|
register_cache("cache", "token_cache", self.token_cache)
|
||||||
|
|
||||||
|
self._account_validity = hs.config.account_validity
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def check_from_context(self, room_version, event, context, do_sig_check=True):
|
def check_from_context(self, room_version, event, context, do_sig_check=True):
|
||||||
prev_state_ids = yield context.get_prev_state_ids(self.store)
|
prev_state_ids = yield context.get_prev_state_ids(self.store)
|
||||||
@ -226,6 +228,17 @@ class Auth(object):
|
|||||||
token_id = user_info["token_id"]
|
token_id = user_info["token_id"]
|
||||||
is_guest = user_info["is_guest"]
|
is_guest = user_info["is_guest"]
|
||||||
|
|
||||||
|
# Deny the request if the user account has expired.
|
||||||
|
if self._account_validity.enabled:
|
||||||
|
user_id = user.to_string()
|
||||||
|
expiration_ts = yield self.store.get_expiration_ts_for_user(user_id)
|
||||||
|
if expiration_ts is not None and self.clock.time_msec() >= expiration_ts:
|
||||||
|
raise AuthError(
|
||||||
|
403,
|
||||||
|
"User account has expired",
|
||||||
|
errcode=Codes.EXPIRED_ACCOUNT,
|
||||||
|
)
|
||||||
|
|
||||||
# device_id may not be present if get_user_by_access_token has been
|
# device_id may not be present if get_user_by_access_token has been
|
||||||
# stubbed out.
|
# stubbed out.
|
||||||
device_id = user_info.get("device_id")
|
device_id = user_info.get("device_id")
|
||||||
|
@ -1,7 +1,7 @@
|
|||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
# Copyright 2014-2016 OpenMarket Ltd
|
# Copyright 2014-2016 OpenMarket Ltd
|
||||||
# Copyright 2017 Vector Creations Ltd
|
# Copyright 2017 Vector Creations Ltd
|
||||||
# Copyright 2018 New Vector Ltd.
|
# Copyright 2018 New Vector Ltd
|
||||||
#
|
#
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
# you may not use this file except in compliance with the License.
|
# you may not use this file except in compliance with the License.
|
||||||
|
@ -1,6 +1,6 @@
|
|||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
# Copyright 2014-2016 OpenMarket Ltd
|
# Copyright 2014-2016 OpenMarket Ltd
|
||||||
# Copyright 2018 New Vector Ltd.
|
# Copyright 2018 New Vector Ltd
|
||||||
#
|
#
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
# you may not use this file except in compliance with the License.
|
# you may not use this file except in compliance with the License.
|
||||||
@ -60,6 +60,7 @@ class Codes(object):
|
|||||||
UNSUPPORTED_ROOM_VERSION = "M_UNSUPPORTED_ROOM_VERSION"
|
UNSUPPORTED_ROOM_VERSION = "M_UNSUPPORTED_ROOM_VERSION"
|
||||||
INCOMPATIBLE_ROOM_VERSION = "M_INCOMPATIBLE_ROOM_VERSION"
|
INCOMPATIBLE_ROOM_VERSION = "M_INCOMPATIBLE_ROOM_VERSION"
|
||||||
WRONG_ROOM_KEYS_VERSION = "M_WRONG_ROOM_KEYS_VERSION"
|
WRONG_ROOM_KEYS_VERSION = "M_WRONG_ROOM_KEYS_VERSION"
|
||||||
|
EXPIRED_ACCOUNT = "ORG_MATRIX_EXPIRED_ACCOUNT"
|
||||||
|
|
||||||
|
|
||||||
class CodeMessageException(RuntimeError):
|
class CodeMessageException(RuntimeError):
|
||||||
|
@ -1,6 +1,6 @@
|
|||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
# Copyright 2014-2016 OpenMarket Ltd
|
# Copyright 2014-2016 OpenMarket Ltd
|
||||||
# Copyright 2018 New Vector Ltd.
|
# Copyright 2018 New Vector Ltd
|
||||||
#
|
#
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
# you may not use this file except in compliance with the License.
|
# you may not use this file except in compliance with the License.
|
||||||
|
@ -114,7 +114,7 @@ class ClientReaderServer(HomeServer):
|
|||||||
KeyChangesServlet(self).register(resource)
|
KeyChangesServlet(self).register(resource)
|
||||||
VoipRestServlet(self).register(resource)
|
VoipRestServlet(self).register(resource)
|
||||||
PushRuleRestServlet(self).register(resource)
|
PushRuleRestServlet(self).register(resource)
|
||||||
VersionsRestServlet(self).register(resource)
|
VersionsRestServlet().register(resource)
|
||||||
|
|
||||||
resources.update({
|
resources.update({
|
||||||
"/_matrix/client": resource,
|
"/_matrix/client": resource,
|
||||||
|
@ -518,6 +518,7 @@ def run(hs):
|
|||||||
uptime = 0
|
uptime = 0
|
||||||
|
|
||||||
stats["homeserver"] = hs.config.server_name
|
stats["homeserver"] = hs.config.server_name
|
||||||
|
stats["server_context"] = hs.config.server_context
|
||||||
stats["timestamp"] = now
|
stats["timestamp"] = now
|
||||||
stats["uptime_seconds"] = uptime
|
stats["uptime_seconds"] = uptime
|
||||||
version = sys.version_info
|
version = sys.version_info
|
||||||
@ -558,7 +559,6 @@ def run(hs):
|
|||||||
|
|
||||||
stats["database_engine"] = hs.get_datastore().database_engine_name
|
stats["database_engine"] = hs.get_datastore().database_engine_name
|
||||||
stats["database_server_version"] = hs.get_datastore().get_server_version()
|
stats["database_server_version"] = hs.get_datastore().get_server_version()
|
||||||
|
|
||||||
logger.info("Reporting stats to matrix.org: %s" % (stats,))
|
logger.info("Reporting stats to matrix.org: %s" % (stats,))
|
||||||
try:
|
try:
|
||||||
yield hs.get_simple_http_client().put_json(
|
yield hs.get_simple_http_client().put_json(
|
||||||
|
@ -71,6 +71,12 @@ class EmailConfig(Config):
|
|||||||
self.email_notif_from = email_config["notif_from"]
|
self.email_notif_from = email_config["notif_from"]
|
||||||
self.email_notif_template_html = email_config["notif_template_html"]
|
self.email_notif_template_html = email_config["notif_template_html"]
|
||||||
self.email_notif_template_text = email_config["notif_template_text"]
|
self.email_notif_template_text = email_config["notif_template_text"]
|
||||||
|
self.email_expiry_template_html = email_config.get(
|
||||||
|
"expiry_template_html", "notice_expiry.html",
|
||||||
|
)
|
||||||
|
self.email_expiry_template_text = email_config.get(
|
||||||
|
"expiry_template_text", "notice_expiry.txt",
|
||||||
|
)
|
||||||
|
|
||||||
template_dir = email_config.get("template_dir")
|
template_dir = email_config.get("template_dir")
|
||||||
# we need an absolute path, because we change directory after starting (and
|
# we need an absolute path, because we change directory after starting (and
|
||||||
@ -120,7 +126,7 @@ class EmailConfig(Config):
|
|||||||
|
|
||||||
def default_config(self, config_dir_path, server_name, **kwargs):
|
def default_config(self, config_dir_path, server_name, **kwargs):
|
||||||
return """
|
return """
|
||||||
# Enable sending emails for notification events
|
# Enable sending emails for notification events or expiry notices
|
||||||
# Defining a custom URL for Riot is only needed if email notifications
|
# Defining a custom URL for Riot is only needed if email notifications
|
||||||
# should contain links to a self-hosted installation of Riot; when set
|
# should contain links to a self-hosted installation of Riot; when set
|
||||||
# the "app_name" setting is ignored.
|
# the "app_name" setting is ignored.
|
||||||
@ -142,6 +148,9 @@ class EmailConfig(Config):
|
|||||||
# #template_dir: res/templates
|
# #template_dir: res/templates
|
||||||
# notif_template_html: notif_mail.html
|
# notif_template_html: notif_mail.html
|
||||||
# notif_template_text: notif_mail.txt
|
# notif_template_text: notif_mail.txt
|
||||||
|
# # Templates for account expiry notices.
|
||||||
|
# expiry_template_html: notice_expiry.html
|
||||||
|
# expiry_template_text: notice_expiry.txt
|
||||||
# notif_for_new_users: True
|
# notif_for_new_users: True
|
||||||
# riot_base_url: "http://localhost/riot"
|
# riot_base_url: "http://localhost/riot"
|
||||||
"""
|
"""
|
||||||
|
@ -20,6 +20,29 @@ from synapse.types import RoomAlias
|
|||||||
from synapse.util.stringutils import random_string_with_symbols
|
from synapse.util.stringutils import random_string_with_symbols
|
||||||
|
|
||||||
|
|
||||||
|
class AccountValidityConfig(Config):
|
||||||
|
def __init__(self, config, synapse_config):
|
||||||
|
self.enabled = config.get("enabled", False)
|
||||||
|
self.renew_by_email_enabled = ("renew_at" in config)
|
||||||
|
|
||||||
|
if self.enabled:
|
||||||
|
if "period" in config:
|
||||||
|
self.period = self.parse_duration(config["period"])
|
||||||
|
else:
|
||||||
|
raise ConfigError("'period' is required when using account validity")
|
||||||
|
|
||||||
|
if "renew_at" in config:
|
||||||
|
self.renew_at = self.parse_duration(config["renew_at"])
|
||||||
|
|
||||||
|
if "renew_email_subject" in config:
|
||||||
|
self.renew_email_subject = config["renew_email_subject"]
|
||||||
|
else:
|
||||||
|
self.renew_email_subject = "Renew your %(app)s account"
|
||||||
|
|
||||||
|
if self.renew_by_email_enabled and "public_baseurl" not in synapse_config:
|
||||||
|
raise ConfigError("Can't send renewal emails without 'public_baseurl'")
|
||||||
|
|
||||||
|
|
||||||
class RegistrationConfig(Config):
|
class RegistrationConfig(Config):
|
||||||
|
|
||||||
def read_config(self, config):
|
def read_config(self, config):
|
||||||
@ -31,8 +54,13 @@ class RegistrationConfig(Config):
|
|||||||
strtobool(str(config["disable_registration"]))
|
strtobool(str(config["disable_registration"]))
|
||||||
)
|
)
|
||||||
|
|
||||||
|
self.account_validity = AccountValidityConfig(
|
||||||
|
config.get("account_validity", {}), config,
|
||||||
|
)
|
||||||
|
|
||||||
self.registrations_require_3pid = config.get("registrations_require_3pid", [])
|
self.registrations_require_3pid = config.get("registrations_require_3pid", [])
|
||||||
self.allowed_local_3pids = config.get("allowed_local_3pids", [])
|
self.allowed_local_3pids = config.get("allowed_local_3pids", [])
|
||||||
|
self.enable_3pid_lookup = config.get("enable_3pid_lookup", True)
|
||||||
self.registration_shared_secret = config.get("registration_shared_secret")
|
self.registration_shared_secret = config.get("registration_shared_secret")
|
||||||
|
|
||||||
self.bcrypt_rounds = config.get("bcrypt_rounds", 12)
|
self.bcrypt_rounds = config.get("bcrypt_rounds", 12)
|
||||||
@ -75,6 +103,32 @@ class RegistrationConfig(Config):
|
|||||||
#
|
#
|
||||||
#enable_registration: false
|
#enable_registration: false
|
||||||
|
|
||||||
|
# Optional account validity configuration. This allows for accounts to be denied
|
||||||
|
# any request after a given period.
|
||||||
|
#
|
||||||
|
# ``enabled`` defines whether the account validity feature is enabled. Defaults
|
||||||
|
# to False.
|
||||||
|
#
|
||||||
|
# ``period`` allows setting the period after which an account is valid
|
||||||
|
# after its registration. When renewing the account, its validity period
|
||||||
|
# will be extended by this amount of time. This parameter is required when using
|
||||||
|
# the account validity feature.
|
||||||
|
#
|
||||||
|
# ``renew_at`` is the amount of time before an account's expiry date at which
|
||||||
|
# Synapse will send an email to the account's email address with a renewal link.
|
||||||
|
# This needs the ``email`` and ``public_baseurl`` configuration sections to be
|
||||||
|
# filled.
|
||||||
|
#
|
||||||
|
# ``renew_email_subject`` is the subject of the email sent out with the renewal
|
||||||
|
# link. ``%%(app)s`` can be used as a placeholder for the ``app_name`` parameter
|
||||||
|
# from the ``email`` section.
|
||||||
|
#
|
||||||
|
#account_validity:
|
||||||
|
# enabled: True
|
||||||
|
# period: 6w
|
||||||
|
# renew_at: 1w
|
||||||
|
# renew_email_subject: "Renew your %%(app)s account"
|
||||||
|
|
||||||
# The user must provide all of the below types of 3PID when registering.
|
# The user must provide all of the below types of 3PID when registering.
|
||||||
#
|
#
|
||||||
#registrations_require_3pid:
|
#registrations_require_3pid:
|
||||||
@ -97,6 +151,10 @@ class RegistrationConfig(Config):
|
|||||||
# - medium: msisdn
|
# - medium: msisdn
|
||||||
# pattern: '\\+44'
|
# pattern: '\\+44'
|
||||||
|
|
||||||
|
# Enable 3PIDs lookup requests to identity servers from this server.
|
||||||
|
#
|
||||||
|
#enable_3pid_lookup: true
|
||||||
|
|
||||||
# If set, allows registration of standard or admin accounts by anyone who
|
# If set, allows registration of standard or admin accounts by anyone who
|
||||||
# has the shared secret, even if registration is otherwise disabled.
|
# has the shared secret, even if registration is otherwise disabled.
|
||||||
#
|
#
|
||||||
|
@ -1,5 +1,5 @@
|
|||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
# Copyright 2018 New Vector Ltd.
|
# Copyright 2018 New Vector Ltd
|
||||||
#
|
#
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
# you may not use this file except in compliance with the License.
|
# you may not use this file except in compliance with the License.
|
||||||
|
@ -37,6 +37,7 @@ class ServerConfig(Config):
|
|||||||
|
|
||||||
def read_config(self, config):
|
def read_config(self, config):
|
||||||
self.server_name = config["server_name"]
|
self.server_name = config["server_name"]
|
||||||
|
self.server_context = config.get("server_context", None)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
parse_and_validate_server_name(self.server_name)
|
parse_and_validate_server_name(self.server_name)
|
||||||
@ -113,11 +114,13 @@ class ServerConfig(Config):
|
|||||||
# FIXME: federation_domain_whitelist needs sytests
|
# FIXME: federation_domain_whitelist needs sytests
|
||||||
self.federation_domain_whitelist = None
|
self.federation_domain_whitelist = None
|
||||||
federation_domain_whitelist = config.get(
|
federation_domain_whitelist = config.get(
|
||||||
"federation_domain_whitelist", None
|
"federation_domain_whitelist", None,
|
||||||
)
|
)
|
||||||
# turn the whitelist into a hash for speed of lookup
|
|
||||||
if federation_domain_whitelist is not None:
|
if federation_domain_whitelist is not None:
|
||||||
|
# turn the whitelist into a hash for speed of lookup
|
||||||
self.federation_domain_whitelist = {}
|
self.federation_domain_whitelist = {}
|
||||||
|
|
||||||
for domain in federation_domain_whitelist:
|
for domain in federation_domain_whitelist:
|
||||||
self.federation_domain_whitelist[domain] = True
|
self.federation_domain_whitelist[domain] = True
|
||||||
|
|
||||||
@ -484,6 +487,9 @@ class ServerConfig(Config):
|
|||||||
#mau_limit_reserved_threepids:
|
#mau_limit_reserved_threepids:
|
||||||
# - medium: 'email'
|
# - medium: 'email'
|
||||||
# address: 'reserved_user@example.com'
|
# address: 'reserved_user@example.com'
|
||||||
|
|
||||||
|
# Used by phonehome stats to group together related servers.
|
||||||
|
#server_context: context
|
||||||
""" % locals()
|
""" % locals()
|
||||||
|
|
||||||
def read_arguments(self, args):
|
def read_arguments(self, args):
|
||||||
|
@ -24,8 +24,10 @@ import six
|
|||||||
from unpaddedbase64 import encode_base64
|
from unpaddedbase64 import encode_base64
|
||||||
|
|
||||||
from OpenSSL import crypto
|
from OpenSSL import crypto
|
||||||
|
from twisted.internet._sslverify import Certificate, trustRootFromCertificates
|
||||||
|
|
||||||
from synapse.config._base import Config, ConfigError
|
from synapse.config._base import Config, ConfigError
|
||||||
|
from synapse.util import glob_to_regex
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
@ -70,6 +72,53 @@ class TlsConfig(Config):
|
|||||||
|
|
||||||
self.tls_fingerprints = list(self._original_tls_fingerprints)
|
self.tls_fingerprints = list(self._original_tls_fingerprints)
|
||||||
|
|
||||||
|
# Whether to verify certificates on outbound federation traffic
|
||||||
|
self.federation_verify_certificates = config.get(
|
||||||
|
"federation_verify_certificates", False,
|
||||||
|
)
|
||||||
|
|
||||||
|
# Whitelist of domains to not verify certificates for
|
||||||
|
fed_whitelist_entries = config.get(
|
||||||
|
"federation_certificate_verification_whitelist", [],
|
||||||
|
)
|
||||||
|
|
||||||
|
# Support globs (*) in whitelist values
|
||||||
|
self.federation_certificate_verification_whitelist = []
|
||||||
|
for entry in fed_whitelist_entries:
|
||||||
|
# Convert globs to regex
|
||||||
|
entry_regex = glob_to_regex(entry)
|
||||||
|
self.federation_certificate_verification_whitelist.append(entry_regex)
|
||||||
|
|
||||||
|
# List of custom certificate authorities for federation traffic validation
|
||||||
|
custom_ca_list = config.get(
|
||||||
|
"federation_custom_ca_list", None,
|
||||||
|
)
|
||||||
|
|
||||||
|
# Read in and parse custom CA certificates
|
||||||
|
self.federation_ca_trust_root = None
|
||||||
|
if custom_ca_list is not None:
|
||||||
|
if len(custom_ca_list) == 0:
|
||||||
|
# A trustroot cannot be generated without any CA certificates.
|
||||||
|
# Raise an error if this option has been specified without any
|
||||||
|
# corresponding certificates.
|
||||||
|
raise ConfigError("federation_custom_ca_list specified without "
|
||||||
|
"any certificate files")
|
||||||
|
|
||||||
|
certs = []
|
||||||
|
for ca_file in custom_ca_list:
|
||||||
|
logger.debug("Reading custom CA certificate file: %s", ca_file)
|
||||||
|
content = self.read_file(ca_file)
|
||||||
|
|
||||||
|
# Parse the CA certificates
|
||||||
|
try:
|
||||||
|
cert_base = Certificate.loadPEM(content)
|
||||||
|
certs.append(cert_base)
|
||||||
|
except Exception as e:
|
||||||
|
raise ConfigError("Error parsing custom CA certificate file %s: %s"
|
||||||
|
% (ca_file, e))
|
||||||
|
|
||||||
|
self.federation_ca_trust_root = trustRootFromCertificates(certs)
|
||||||
|
|
||||||
# This config option applies to non-federation HTTP clients
|
# This config option applies to non-federation HTTP clients
|
||||||
# (e.g. for talking to recaptcha, identity servers, and such)
|
# (e.g. for talking to recaptcha, identity servers, and such)
|
||||||
# It should never be used in production, and is intended for
|
# It should never be used in production, and is intended for
|
||||||
@ -99,15 +148,15 @@ class TlsConfig(Config):
|
|||||||
try:
|
try:
|
||||||
with open(self.tls_certificate_file, 'rb') as f:
|
with open(self.tls_certificate_file, 'rb') as f:
|
||||||
cert_pem = f.read()
|
cert_pem = f.read()
|
||||||
except Exception:
|
except Exception as e:
|
||||||
logger.exception("Failed to read existing certificate off disk!")
|
raise ConfigError("Failed to read existing certificate file %s: %s"
|
||||||
raise
|
% (self.tls_certificate_file, e))
|
||||||
|
|
||||||
try:
|
try:
|
||||||
tls_certificate = crypto.load_certificate(crypto.FILETYPE_PEM, cert_pem)
|
tls_certificate = crypto.load_certificate(crypto.FILETYPE_PEM, cert_pem)
|
||||||
except Exception:
|
except Exception as e:
|
||||||
logger.exception("Failed to parse existing certificate off disk!")
|
raise ConfigError("Failed to parse existing certificate file %s: %s"
|
||||||
raise
|
% (self.tls_certificate_file, e))
|
||||||
|
|
||||||
if not allow_self_signed:
|
if not allow_self_signed:
|
||||||
if tls_certificate.get_subject() == tls_certificate.get_issuer():
|
if tls_certificate.get_subject() == tls_certificate.get_issuer():
|
||||||
@ -192,6 +241,40 @@ class TlsConfig(Config):
|
|||||||
#
|
#
|
||||||
#tls_private_key_path: "%(tls_private_key_path)s"
|
#tls_private_key_path: "%(tls_private_key_path)s"
|
||||||
|
|
||||||
|
# Whether to verify TLS certificates when sending federation traffic.
|
||||||
|
#
|
||||||
|
# This currently defaults to `false`, however this will change in
|
||||||
|
# Synapse 1.0 when valid federation certificates will be required.
|
||||||
|
#
|
||||||
|
#federation_verify_certificates: true
|
||||||
|
|
||||||
|
# Skip federation certificate verification on the following whitelist
|
||||||
|
# of domains.
|
||||||
|
#
|
||||||
|
# This setting should only be used in very specific cases, such as
|
||||||
|
# federation over Tor hidden services and similar. For private networks
|
||||||
|
# of homeservers, you likely want to use a private CA instead.
|
||||||
|
#
|
||||||
|
# Only effective if federation_verify_certicates is `true`.
|
||||||
|
#
|
||||||
|
#federation_certificate_verification_whitelist:
|
||||||
|
# - lon.example.com
|
||||||
|
# - *.domain.com
|
||||||
|
# - *.onion
|
||||||
|
|
||||||
|
# List of custom certificate authorities for federation traffic.
|
||||||
|
#
|
||||||
|
# This setting should only normally be used within a private network of
|
||||||
|
# homeservers.
|
||||||
|
#
|
||||||
|
# Note that this list will replace those that are provided by your
|
||||||
|
# operating environment. Certificates must be in PEM format.
|
||||||
|
#
|
||||||
|
#federation_custom_ca_list:
|
||||||
|
# - myCA1.pem
|
||||||
|
# - myCA2.pem
|
||||||
|
# - myCA3.pem
|
||||||
|
|
||||||
# ACME support: This will configure Synapse to request a valid TLS certificate
|
# ACME support: This will configure Synapse to request a valid TLS certificate
|
||||||
# for your configured `server_name` via Let's Encrypt.
|
# for your configured `server_name` via Let's Encrypt.
|
||||||
#
|
#
|
||||||
|
@ -18,10 +18,10 @@ import logging
|
|||||||
from zope.interface import implementer
|
from zope.interface import implementer
|
||||||
|
|
||||||
from OpenSSL import SSL, crypto
|
from OpenSSL import SSL, crypto
|
||||||
from twisted.internet._sslverify import _defaultCurveName
|
from twisted.internet._sslverify import ClientTLSOptions, _defaultCurveName
|
||||||
from twisted.internet.abstract import isIPAddress, isIPv6Address
|
from twisted.internet.abstract import isIPAddress, isIPv6Address
|
||||||
from twisted.internet.interfaces import IOpenSSLClientConnectionCreator
|
from twisted.internet.interfaces import IOpenSSLClientConnectionCreator
|
||||||
from twisted.internet.ssl import CertificateOptions, ContextFactory
|
from twisted.internet.ssl import CertificateOptions, ContextFactory, platformTrust
|
||||||
from twisted.python.failure import Failure
|
from twisted.python.failure import Failure
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
@ -90,7 +90,7 @@ def _tolerateErrors(wrapped):
|
|||||||
|
|
||||||
|
|
||||||
@implementer(IOpenSSLClientConnectionCreator)
|
@implementer(IOpenSSLClientConnectionCreator)
|
||||||
class ClientTLSOptions(object):
|
class ClientTLSOptionsNoVerify(object):
|
||||||
"""
|
"""
|
||||||
Client creator for TLS without certificate identity verification. This is a
|
Client creator for TLS without certificate identity verification. This is a
|
||||||
copy of twisted.internet._sslverify.ClientTLSOptions with the identity
|
copy of twisted.internet._sslverify.ClientTLSOptions with the identity
|
||||||
@ -127,9 +127,30 @@ class ClientTLSOptionsFactory(object):
|
|||||||
to remote servers for federation."""
|
to remote servers for federation."""
|
||||||
|
|
||||||
def __init__(self, config):
|
def __init__(self, config):
|
||||||
# We don't use config options yet
|
self._config = config
|
||||||
self._options = CertificateOptions(verify=False)
|
self._options_noverify = CertificateOptions()
|
||||||
|
|
||||||
|
# Check if we're using a custom list of a CA certificates
|
||||||
|
trust_root = config.federation_ca_trust_root
|
||||||
|
if trust_root is None:
|
||||||
|
# Use CA root certs provided by OpenSSL
|
||||||
|
trust_root = platformTrust()
|
||||||
|
|
||||||
|
self._options_verify = CertificateOptions(trustRoot=trust_root)
|
||||||
|
|
||||||
def get_options(self, host):
|
def get_options(self, host):
|
||||||
# Use _makeContext so that we get a fresh OpenSSL CTX each time.
|
# Use _makeContext so that we get a fresh OpenSSL CTX each time.
|
||||||
return ClientTLSOptions(host, self._options._makeContext())
|
|
||||||
|
# Check if certificate verification has been enabled
|
||||||
|
should_verify = self._config.federation_verify_certificates
|
||||||
|
|
||||||
|
# Check if we've disabled certificate verification for this host
|
||||||
|
if should_verify:
|
||||||
|
for regex in self._config.federation_certificate_verification_whitelist:
|
||||||
|
if regex.match(host):
|
||||||
|
should_verify = False
|
||||||
|
break
|
||||||
|
|
||||||
|
if should_verify:
|
||||||
|
return ClientTLSOptions(host, self._options_verify._makeContext())
|
||||||
|
return ClientTLSOptionsNoVerify(host, self._options_noverify._makeContext())
|
||||||
|
@ -1,6 +1,6 @@
|
|||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
# Copyright 2014-2016 OpenMarket Ltd
|
# Copyright 2014-2016 OpenMarket Ltd
|
||||||
# Copyright 2017, 2018 New Vector Ltd.
|
# Copyright 2017, 2018 New Vector Ltd
|
||||||
#
|
#
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
# you may not use this file except in compliance with the License.
|
# you may not use this file except in compliance with the License.
|
||||||
@ -20,6 +20,7 @@ from collections import namedtuple
|
|||||||
from six import raise_from
|
from six import raise_from
|
||||||
from six.moves import urllib
|
from six.moves import urllib
|
||||||
|
|
||||||
|
import nacl.signing
|
||||||
from signedjson.key import (
|
from signedjson.key import (
|
||||||
decode_verify_key_bytes,
|
decode_verify_key_bytes,
|
||||||
encode_verify_key_base64,
|
encode_verify_key_base64,
|
||||||
@ -274,10 +275,6 @@ class Keyring(object):
|
|||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def do_iterations():
|
def do_iterations():
|
||||||
with Measure(self.clock, "get_server_verify_keys"):
|
with Measure(self.clock, "get_server_verify_keys"):
|
||||||
# dict[str, dict[str, VerifyKey]]: results so far.
|
|
||||||
# map server_name -> key_id -> VerifyKey
|
|
||||||
merged_results = {}
|
|
||||||
|
|
||||||
# dict[str, set(str)]: keys to fetch for each server
|
# dict[str, set(str)]: keys to fetch for each server
|
||||||
missing_keys = {}
|
missing_keys = {}
|
||||||
for verify_request in verify_requests:
|
for verify_request in verify_requests:
|
||||||
@ -287,29 +284,29 @@ class Keyring(object):
|
|||||||
|
|
||||||
for fn in key_fetch_fns:
|
for fn in key_fetch_fns:
|
||||||
results = yield fn(missing_keys.items())
|
results = yield fn(missing_keys.items())
|
||||||
merged_results.update(results)
|
|
||||||
|
|
||||||
# We now need to figure out which verify requests we have keys
|
# We now need to figure out which verify requests we have keys
|
||||||
# for and which we don't
|
# for and which we don't
|
||||||
missing_keys = {}
|
missing_keys = {}
|
||||||
requests_missing_keys = []
|
requests_missing_keys = []
|
||||||
for verify_request in verify_requests:
|
for verify_request in verify_requests:
|
||||||
server_name = verify_request.server_name
|
|
||||||
result_keys = merged_results[server_name]
|
|
||||||
|
|
||||||
if verify_request.deferred.called:
|
if verify_request.deferred.called:
|
||||||
# We've already called this deferred, which probably
|
# We've already called this deferred, which probably
|
||||||
# means that we've already found a key for it.
|
# means that we've already found a key for it.
|
||||||
continue
|
continue
|
||||||
|
|
||||||
|
server_name = verify_request.server_name
|
||||||
|
|
||||||
|
# see if any of the keys we got this time are sufficient to
|
||||||
|
# complete this VerifyKeyRequest.
|
||||||
|
result_keys = results.get(server_name, {})
|
||||||
for key_id in verify_request.key_ids:
|
for key_id in verify_request.key_ids:
|
||||||
if key_id in result_keys:
|
key = result_keys.get(key_id)
|
||||||
|
if key:
|
||||||
with PreserveLoggingContext():
|
with PreserveLoggingContext():
|
||||||
verify_request.deferred.callback((
|
verify_request.deferred.callback(
|
||||||
server_name,
|
(server_name, key_id, key)
|
||||||
key_id,
|
)
|
||||||
result_keys[key_id],
|
|
||||||
))
|
|
||||||
break
|
break
|
||||||
else:
|
else:
|
||||||
# The else block is only reached if the loop above
|
# The else block is only reached if the loop above
|
||||||
@ -343,27 +340,24 @@ class Keyring(object):
|
|||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def get_keys_from_store(self, server_name_and_key_ids):
|
def get_keys_from_store(self, server_name_and_key_ids):
|
||||||
"""
|
"""
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
server_name_and_key_ids (list[(str, iterable[str])]):
|
server_name_and_key_ids (iterable(Tuple[str, iterable[str]]):
|
||||||
list of (server_name, iterable[key_id]) tuples to fetch keys for
|
list of (server_name, iterable[key_id]) tuples to fetch keys for
|
||||||
|
|
||||||
Returns:
|
Returns:
|
||||||
Deferred: resolves to dict[str, dict[str, VerifyKey]]: map from
|
Deferred: resolves to dict[str, dict[str, VerifyKey|None]]: map from
|
||||||
server_name -> key_id -> VerifyKey
|
server_name -> key_id -> VerifyKey
|
||||||
"""
|
"""
|
||||||
res = yield logcontext.make_deferred_yieldable(defer.gatherResults(
|
keys_to_fetch = (
|
||||||
[
|
(server_name, key_id)
|
||||||
run_in_background(
|
for server_name, key_ids in server_name_and_key_ids
|
||||||
self.store.get_server_verify_keys,
|
for key_id in key_ids
|
||||||
server_name, key_ids,
|
)
|
||||||
).addCallback(lambda ks, server: (server, ks), server_name)
|
res = yield self.store.get_server_verify_keys(keys_to_fetch)
|
||||||
for server_name, key_ids in server_name_and_key_ids
|
keys = {}
|
||||||
],
|
for (server_name, key_id), key in res.items():
|
||||||
consumeErrors=True,
|
keys.setdefault(server_name, {})[key_id] = key
|
||||||
).addErrback(unwrapFirstError))
|
defer.returnValue(keys)
|
||||||
|
|
||||||
defer.returnValue(dict(res))
|
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def get_keys_from_perspectives(self, server_name_and_key_ids):
|
def get_keys_from_perspectives(self, server_name_and_key_ids):
|
||||||
@ -494,11 +488,11 @@ class Keyring(object):
|
|||||||
)
|
)
|
||||||
|
|
||||||
processed_response = yield self.process_v2_response(
|
processed_response = yield self.process_v2_response(
|
||||||
perspective_name, response, only_from_server=False
|
perspective_name, response
|
||||||
)
|
)
|
||||||
|
server_name = response["server_name"]
|
||||||
|
|
||||||
for server_name, response_keys in processed_response.items():
|
keys.setdefault(server_name, {}).update(processed_response)
|
||||||
keys.setdefault(server_name, {}).update(response_keys)
|
|
||||||
|
|
||||||
yield logcontext.make_deferred_yieldable(defer.gatherResults(
|
yield logcontext.make_deferred_yieldable(defer.gatherResults(
|
||||||
[
|
[
|
||||||
@ -517,7 +511,7 @@ class Keyring(object):
|
|||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def get_server_verify_key_v2_direct(self, server_name, key_ids):
|
def get_server_verify_key_v2_direct(self, server_name, key_ids):
|
||||||
keys = {}
|
keys = {} # type: dict[str, nacl.signing.VerifyKey]
|
||||||
|
|
||||||
for requested_key_id in key_ids:
|
for requested_key_id in key_ids:
|
||||||
if requested_key_id in keys:
|
if requested_key_id in keys:
|
||||||
@ -542,6 +536,11 @@ class Keyring(object):
|
|||||||
or server_name not in response[u"signatures"]):
|
or server_name not in response[u"signatures"]):
|
||||||
raise KeyLookupError("Key response not signed by remote server")
|
raise KeyLookupError("Key response not signed by remote server")
|
||||||
|
|
||||||
|
if response["server_name"] != server_name:
|
||||||
|
raise KeyLookupError("Expected a response for server %r not %r" % (
|
||||||
|
server_name, response["server_name"]
|
||||||
|
))
|
||||||
|
|
||||||
response_keys = yield self.process_v2_response(
|
response_keys = yield self.process_v2_response(
|
||||||
from_server=server_name,
|
from_server=server_name,
|
||||||
requested_ids=[requested_key_id],
|
requested_ids=[requested_key_id],
|
||||||
@ -550,24 +549,45 @@ class Keyring(object):
|
|||||||
|
|
||||||
keys.update(response_keys)
|
keys.update(response_keys)
|
||||||
|
|
||||||
yield logcontext.make_deferred_yieldable(defer.gatherResults(
|
yield self.store_keys(
|
||||||
[
|
server_name=server_name,
|
||||||
run_in_background(
|
from_server=server_name,
|
||||||
self.store_keys,
|
verify_keys=keys,
|
||||||
server_name=key_server_name,
|
)
|
||||||
from_server=server_name,
|
defer.returnValue({server_name: keys})
|
||||||
verify_keys=verify_keys,
|
|
||||||
)
|
|
||||||
for key_server_name, verify_keys in keys.items()
|
|
||||||
],
|
|
||||||
consumeErrors=True
|
|
||||||
).addErrback(unwrapFirstError))
|
|
||||||
|
|
||||||
defer.returnValue(keys)
|
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def process_v2_response(self, from_server, response_json,
|
def process_v2_response(
|
||||||
requested_ids=[], only_from_server=True):
|
self, from_server, response_json, requested_ids=[],
|
||||||
|
):
|
||||||
|
"""Parse a 'Server Keys' structure from the result of a /key request
|
||||||
|
|
||||||
|
This is used to parse either the entirety of the response from
|
||||||
|
GET /_matrix/key/v2/server, or a single entry from the list returned by
|
||||||
|
POST /_matrix/key/v2/query.
|
||||||
|
|
||||||
|
Checks that each signature in the response that claims to come from the origin
|
||||||
|
server is valid. (Does not check that there actually is such a signature, for
|
||||||
|
some reason.)
|
||||||
|
|
||||||
|
Stores the json in server_keys_json so that it can be used for future responses
|
||||||
|
to /_matrix/key/v2/query.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
from_server (str): the name of the server producing this result: either
|
||||||
|
the origin server for a /_matrix/key/v2/server request, or the notary
|
||||||
|
for a /_matrix/key/v2/query.
|
||||||
|
|
||||||
|
response_json (dict): the json-decoded Server Keys response object
|
||||||
|
|
||||||
|
requested_ids (iterable[str]): a list of the key IDs that were requested.
|
||||||
|
We will store the json for these key ids as well as any that are
|
||||||
|
actually in the response
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Deferred[dict[str, nacl.signing.VerifyKey]]:
|
||||||
|
map from key_id to key object
|
||||||
|
"""
|
||||||
time_now_ms = self.clock.time_msec()
|
time_now_ms = self.clock.time_msec()
|
||||||
response_keys = {}
|
response_keys = {}
|
||||||
verify_keys = {}
|
verify_keys = {}
|
||||||
@ -589,15 +609,7 @@ class Keyring(object):
|
|||||||
verify_key.time_added = time_now_ms
|
verify_key.time_added = time_now_ms
|
||||||
old_verify_keys[key_id] = verify_key
|
old_verify_keys[key_id] = verify_key
|
||||||
|
|
||||||
results = {}
|
|
||||||
server_name = response_json["server_name"]
|
server_name = response_json["server_name"]
|
||||||
if only_from_server:
|
|
||||||
if server_name != from_server:
|
|
||||||
raise KeyLookupError(
|
|
||||||
"Expected a response for server %r not %r" % (
|
|
||||||
from_server, server_name
|
|
||||||
)
|
|
||||||
)
|
|
||||||
for key_id in response_json["signatures"].get(server_name, {}):
|
for key_id in response_json["signatures"].get(server_name, {}):
|
||||||
if key_id not in response_json["verify_keys"]:
|
if key_id not in response_json["verify_keys"]:
|
||||||
raise KeyLookupError(
|
raise KeyLookupError(
|
||||||
@ -633,7 +645,7 @@ class Keyring(object):
|
|||||||
self.store.store_server_keys_json,
|
self.store.store_server_keys_json,
|
||||||
server_name=server_name,
|
server_name=server_name,
|
||||||
key_id=key_id,
|
key_id=key_id,
|
||||||
from_server=server_name,
|
from_server=from_server,
|
||||||
ts_now_ms=time_now_ms,
|
ts_now_ms=time_now_ms,
|
||||||
ts_expires_ms=ts_valid_until_ms,
|
ts_expires_ms=ts_valid_until_ms,
|
||||||
key_json_bytes=signed_key_json_bytes,
|
key_json_bytes=signed_key_json_bytes,
|
||||||
@ -643,9 +655,7 @@ class Keyring(object):
|
|||||||
consumeErrors=True,
|
consumeErrors=True,
|
||||||
).addErrback(unwrapFirstError))
|
).addErrback(unwrapFirstError))
|
||||||
|
|
||||||
results[server_name] = response_keys
|
defer.returnValue(response_keys)
|
||||||
|
|
||||||
defer.returnValue(results)
|
|
||||||
|
|
||||||
def store_keys(self, server_name, from_server, verify_keys):
|
def store_keys(self, server_name, from_server, verify_keys):
|
||||||
"""Store a collection of verify keys for a given server
|
"""Store a collection of verify keys for a given server
|
||||||
|
@ -1,5 +1,5 @@
|
|||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
# Copyright 2017 New Vector Ltd.
|
# Copyright 2017 New Vector Ltd
|
||||||
#
|
#
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
# you may not use this file except in compliance with the License.
|
# you may not use this file except in compliance with the License.
|
||||||
|
253
synapse/handlers/account_validity.py
Normal file
253
synapse/handlers/account_validity.py
Normal file
@ -0,0 +1,253 @@
|
|||||||
|
# -*- coding: utf-8 -*-
|
||||||
|
# Copyright 2019 New Vector Ltd
|
||||||
|
#
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
# you may not use this file except in compliance with the License.
|
||||||
|
# You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
# See the License for the specific language governing permissions and
|
||||||
|
# limitations under the License.
|
||||||
|
|
||||||
|
import email.mime.multipart
|
||||||
|
import email.utils
|
||||||
|
import logging
|
||||||
|
from email.mime.multipart import MIMEMultipart
|
||||||
|
from email.mime.text import MIMEText
|
||||||
|
|
||||||
|
from twisted.internet import defer
|
||||||
|
|
||||||
|
from synapse.api.errors import StoreError
|
||||||
|
from synapse.types import UserID
|
||||||
|
from synapse.util import stringutils
|
||||||
|
from synapse.util.logcontext import make_deferred_yieldable
|
||||||
|
|
||||||
|
try:
|
||||||
|
from synapse.push.mailer import load_jinja2_templates
|
||||||
|
except ImportError:
|
||||||
|
load_jinja2_templates = None
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
class AccountValidityHandler(object):
|
||||||
|
def __init__(self, hs):
|
||||||
|
self.hs = hs
|
||||||
|
self.store = self.hs.get_datastore()
|
||||||
|
self.sendmail = self.hs.get_sendmail()
|
||||||
|
self.clock = self.hs.get_clock()
|
||||||
|
|
||||||
|
self._account_validity = self.hs.config.account_validity
|
||||||
|
|
||||||
|
if self._account_validity.renew_by_email_enabled and load_jinja2_templates:
|
||||||
|
# Don't do email-specific configuration if renewal by email is disabled.
|
||||||
|
try:
|
||||||
|
app_name = self.hs.config.email_app_name
|
||||||
|
|
||||||
|
self._subject = self._account_validity.renew_email_subject % {
|
||||||
|
"app": app_name,
|
||||||
|
}
|
||||||
|
|
||||||
|
self._from_string = self.hs.config.email_notif_from % {
|
||||||
|
"app": app_name,
|
||||||
|
}
|
||||||
|
except Exception:
|
||||||
|
# If substitution failed, fall back to the bare strings.
|
||||||
|
self._subject = self._account_validity.renew_email_subject
|
||||||
|
self._from_string = self.hs.config.email_notif_from
|
||||||
|
|
||||||
|
self._raw_from = email.utils.parseaddr(self._from_string)[1]
|
||||||
|
|
||||||
|
self._template_html, self._template_text = load_jinja2_templates(
|
||||||
|
config=self.hs.config,
|
||||||
|
template_html_name=self.hs.config.email_expiry_template_html,
|
||||||
|
template_text_name=self.hs.config.email_expiry_template_text,
|
||||||
|
)
|
||||||
|
|
||||||
|
# Check the renewal emails to send and send them every 30min.
|
||||||
|
self.clock.looping_call(
|
||||||
|
self.send_renewal_emails,
|
||||||
|
30 * 60 * 1000,
|
||||||
|
)
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def send_renewal_emails(self):
|
||||||
|
"""Gets the list of users whose account is expiring in the amount of time
|
||||||
|
configured in the ``renew_at`` parameter from the ``account_validity``
|
||||||
|
configuration, and sends renewal emails to all of these users as long as they
|
||||||
|
have an email 3PID attached to their account.
|
||||||
|
"""
|
||||||
|
expiring_users = yield self.store.get_users_expiring_soon()
|
||||||
|
|
||||||
|
if expiring_users:
|
||||||
|
for user in expiring_users:
|
||||||
|
yield self._send_renewal_email(
|
||||||
|
user_id=user["user_id"],
|
||||||
|
expiration_ts=user["expiration_ts_ms"],
|
||||||
|
)
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def send_renewal_email_to_user(self, user_id):
|
||||||
|
expiration_ts = yield self.store.get_expiration_ts_for_user(user_id)
|
||||||
|
yield self._send_renewal_email(user_id, expiration_ts)
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def _send_renewal_email(self, user_id, expiration_ts):
|
||||||
|
"""Sends out a renewal email to every email address attached to the given user
|
||||||
|
with a unique link allowing them to renew their account.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
user_id (str): ID of the user to send email(s) to.
|
||||||
|
expiration_ts (int): Timestamp in milliseconds for the expiration date of
|
||||||
|
this user's account (used in the email templates).
|
||||||
|
"""
|
||||||
|
addresses = yield self._get_email_addresses_for_user(user_id)
|
||||||
|
|
||||||
|
# Stop right here if the user doesn't have at least one email address.
|
||||||
|
# In this case, they will have to ask their server admin to renew their
|
||||||
|
# account manually.
|
||||||
|
if not addresses:
|
||||||
|
return
|
||||||
|
|
||||||
|
try:
|
||||||
|
user_display_name = yield self.store.get_profile_displayname(
|
||||||
|
UserID.from_string(user_id).localpart
|
||||||
|
)
|
||||||
|
if user_display_name is None:
|
||||||
|
user_display_name = user_id
|
||||||
|
except StoreError:
|
||||||
|
user_display_name = user_id
|
||||||
|
|
||||||
|
renewal_token = yield self._get_renewal_token(user_id)
|
||||||
|
url = "%s_matrix/client/unstable/account_validity/renew?token=%s" % (
|
||||||
|
self.hs.config.public_baseurl,
|
||||||
|
renewal_token,
|
||||||
|
)
|
||||||
|
|
||||||
|
template_vars = {
|
||||||
|
"display_name": user_display_name,
|
||||||
|
"expiration_ts": expiration_ts,
|
||||||
|
"url": url,
|
||||||
|
}
|
||||||
|
|
||||||
|
html_text = self._template_html.render(**template_vars)
|
||||||
|
html_part = MIMEText(html_text, "html", "utf8")
|
||||||
|
|
||||||
|
plain_text = self._template_text.render(**template_vars)
|
||||||
|
text_part = MIMEText(plain_text, "plain", "utf8")
|
||||||
|
|
||||||
|
for address in addresses:
|
||||||
|
raw_to = email.utils.parseaddr(address)[1]
|
||||||
|
|
||||||
|
multipart_msg = MIMEMultipart('alternative')
|
||||||
|
multipart_msg['Subject'] = self._subject
|
||||||
|
multipart_msg['From'] = self._from_string
|
||||||
|
multipart_msg['To'] = address
|
||||||
|
multipart_msg['Date'] = email.utils.formatdate()
|
||||||
|
multipart_msg['Message-ID'] = email.utils.make_msgid()
|
||||||
|
multipart_msg.attach(text_part)
|
||||||
|
multipart_msg.attach(html_part)
|
||||||
|
|
||||||
|
logger.info("Sending renewal email to %s", address)
|
||||||
|
|
||||||
|
yield make_deferred_yieldable(self.sendmail(
|
||||||
|
self.hs.config.email_smtp_host,
|
||||||
|
self._raw_from, raw_to, multipart_msg.as_string().encode('utf8'),
|
||||||
|
reactor=self.hs.get_reactor(),
|
||||||
|
port=self.hs.config.email_smtp_port,
|
||||||
|
requireAuthentication=self.hs.config.email_smtp_user is not None,
|
||||||
|
username=self.hs.config.email_smtp_user,
|
||||||
|
password=self.hs.config.email_smtp_pass,
|
||||||
|
requireTransportSecurity=self.hs.config.require_transport_security
|
||||||
|
))
|
||||||
|
|
||||||
|
yield self.store.set_renewal_mail_status(
|
||||||
|
user_id=user_id,
|
||||||
|
email_sent=True,
|
||||||
|
)
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def _get_email_addresses_for_user(self, user_id):
|
||||||
|
"""Retrieve the list of email addresses attached to a user's account.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
user_id (str): ID of the user to lookup email addresses for.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
defer.Deferred[list[str]]: Email addresses for this account.
|
||||||
|
"""
|
||||||
|
threepids = yield self.store.user_get_threepids(user_id)
|
||||||
|
|
||||||
|
addresses = []
|
||||||
|
for threepid in threepids:
|
||||||
|
if threepid["medium"] == "email":
|
||||||
|
addresses.append(threepid["address"])
|
||||||
|
|
||||||
|
defer.returnValue(addresses)
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def _get_renewal_token(self, user_id):
|
||||||
|
"""Generates a 32-byte long random string that will be inserted into the
|
||||||
|
user's renewal email's unique link, then saves it into the database.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
user_id (str): ID of the user to generate a string for.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
defer.Deferred[str]: The generated string.
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
StoreError(500): Couldn't generate a unique string after 5 attempts.
|
||||||
|
"""
|
||||||
|
attempts = 0
|
||||||
|
while attempts < 5:
|
||||||
|
try:
|
||||||
|
renewal_token = stringutils.random_string(32)
|
||||||
|
yield self.store.set_renewal_token_for_user(user_id, renewal_token)
|
||||||
|
defer.returnValue(renewal_token)
|
||||||
|
except StoreError:
|
||||||
|
attempts += 1
|
||||||
|
raise StoreError(500, "Couldn't generate a unique string as refresh string.")
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def renew_account(self, renewal_token):
|
||||||
|
"""Renews the account attached to a given renewal token by pushing back the
|
||||||
|
expiration date by the current validity period in the server's configuration.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
renewal_token (str): Token sent with the renewal request.
|
||||||
|
"""
|
||||||
|
user_id = yield self.store.get_user_from_renewal_token(renewal_token)
|
||||||
|
logger.debug("Renewing an account for user %s", user_id)
|
||||||
|
yield self.renew_account_for_user(user_id)
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def renew_account_for_user(self, user_id, expiration_ts=None, email_sent=False):
|
||||||
|
"""Renews the account attached to a given user by pushing back the
|
||||||
|
expiration date by the current validity period in the server's
|
||||||
|
configuration.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
renewal_token (str): Token sent with the renewal request.
|
||||||
|
expiration_ts (int): New expiration date. Defaults to now + validity period.
|
||||||
|
email_sent (bool): Whether an email has been sent for this validity period.
|
||||||
|
Defaults to False.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
defer.Deferred[int]: New expiration date for this account, as a timestamp
|
||||||
|
in milliseconds since epoch.
|
||||||
|
"""
|
||||||
|
if expiration_ts is None:
|
||||||
|
expiration_ts = self.clock.time_msec() + self._account_validity.period
|
||||||
|
|
||||||
|
yield self.store.set_account_validity_for_user(
|
||||||
|
user_id=user_id,
|
||||||
|
expiration_ts=expiration_ts,
|
||||||
|
email_sent=email_sent,
|
||||||
|
)
|
||||||
|
|
||||||
|
defer.returnValue(expiration_ts)
|
@ -828,6 +828,11 @@ class PresenceHandler(object):
|
|||||||
if typ != EventTypes.Member:
|
if typ != EventTypes.Member:
|
||||||
continue
|
continue
|
||||||
|
|
||||||
|
if event_id is None:
|
||||||
|
# state has been deleted, so this is not a join. We only care about
|
||||||
|
# joins.
|
||||||
|
continue
|
||||||
|
|
||||||
event = yield self.store.get_event(event_id)
|
event = yield self.store.get_event(event_id)
|
||||||
if event.content.get("membership") != Membership.JOIN:
|
if event.content.get("membership") != Membership.JOIN:
|
||||||
# We only care about joins
|
# We only care about joins
|
||||||
|
@ -72,6 +72,7 @@ class RoomMemberHandler(object):
|
|||||||
self.clock = hs.get_clock()
|
self.clock = hs.get_clock()
|
||||||
self.spam_checker = hs.get_spam_checker()
|
self.spam_checker = hs.get_spam_checker()
|
||||||
self._server_notices_mxid = self.config.server_notices_mxid
|
self._server_notices_mxid = self.config.server_notices_mxid
|
||||||
|
self._enable_lookup = hs.config.enable_3pid_lookup
|
||||||
|
|
||||||
# This is only used to get at ratelimit function, and
|
# This is only used to get at ratelimit function, and
|
||||||
# maybe_kick_guest_users. It's fine there are multiple of these as
|
# maybe_kick_guest_users. It's fine there are multiple of these as
|
||||||
@ -748,6 +749,10 @@ class RoomMemberHandler(object):
|
|||||||
Returns:
|
Returns:
|
||||||
str: the matrix ID of the 3pid, or None if it is not recognized.
|
str: the matrix ID of the 3pid, or None if it is not recognized.
|
||||||
"""
|
"""
|
||||||
|
if not self._enable_lookup:
|
||||||
|
raise SynapseError(
|
||||||
|
403, "Looking up third-party identifiers is denied from this server",
|
||||||
|
)
|
||||||
try:
|
try:
|
||||||
data = yield self.simple_http_client.get_json(
|
data = yield self.simple_http_client.get_json(
|
||||||
"%s%s/_matrix/identity/api/v1/lookup" % (id_server_scheme, id_server,),
|
"%s%s/_matrix/identity/api/v1/lookup" % (id_server_scheme, id_server,),
|
||||||
|
@ -149,7 +149,7 @@ class MatrixFederationAgent(object):
|
|||||||
tls_options = None
|
tls_options = None
|
||||||
else:
|
else:
|
||||||
tls_options = self._tls_client_options_factory.get_options(
|
tls_options = self._tls_client_options_factory.get_options(
|
||||||
res.tls_server_name.decode("ascii")
|
res.tls_server_name.decode("ascii"),
|
||||||
)
|
)
|
||||||
|
|
||||||
# make sure that the Host header is set correctly
|
# make sure that the Host header is set correctly
|
||||||
|
@ -521,11 +521,11 @@ def format_ts_filter(value, format):
|
|||||||
return time.strftime(format, time.localtime(value / 1000))
|
return time.strftime(format, time.localtime(value / 1000))
|
||||||
|
|
||||||
|
|
||||||
def load_jinja2_templates(config):
|
def load_jinja2_templates(config, template_html_name, template_text_name):
|
||||||
"""Load the jinja2 email templates from disk
|
"""Load the jinja2 email templates from disk
|
||||||
|
|
||||||
Returns:
|
Returns:
|
||||||
(notif_template_html, notif_template_text)
|
(template_html, template_text)
|
||||||
"""
|
"""
|
||||||
logger.info("loading email templates from '%s'", config.email_template_dir)
|
logger.info("loading email templates from '%s'", config.email_template_dir)
|
||||||
loader = jinja2.FileSystemLoader(config.email_template_dir)
|
loader = jinja2.FileSystemLoader(config.email_template_dir)
|
||||||
@ -533,14 +533,10 @@ def load_jinja2_templates(config):
|
|||||||
env.filters["format_ts"] = format_ts_filter
|
env.filters["format_ts"] = format_ts_filter
|
||||||
env.filters["mxc_to_http"] = _create_mxc_to_http_filter(config)
|
env.filters["mxc_to_http"] = _create_mxc_to_http_filter(config)
|
||||||
|
|
||||||
notif_template_html = env.get_template(
|
template_html = env.get_template(template_html_name)
|
||||||
config.email_notif_template_html
|
template_text = env.get_template(template_text_name)
|
||||||
)
|
|
||||||
notif_template_text = env.get_template(
|
|
||||||
config.email_notif_template_text
|
|
||||||
)
|
|
||||||
|
|
||||||
return notif_template_html, notif_template_text
|
return template_html, template_text
|
||||||
|
|
||||||
|
|
||||||
def _create_mxc_to_http_filter(config):
|
def _create_mxc_to_http_filter(config):
|
||||||
|
@ -44,7 +44,11 @@ class PusherFactory(object):
|
|||||||
if hs.config.email_enable_notifs:
|
if hs.config.email_enable_notifs:
|
||||||
self.mailers = {} # app_name -> Mailer
|
self.mailers = {} # app_name -> Mailer
|
||||||
|
|
||||||
templates = load_jinja2_templates(hs.config)
|
templates = load_jinja2_templates(
|
||||||
|
config=hs.config,
|
||||||
|
template_html_name=hs.config.email_notif_template_html,
|
||||||
|
template_text_name=hs.config.email_notif_template_text,
|
||||||
|
)
|
||||||
self.notif_template_html, self.notif_template_text = templates
|
self.notif_template_html, self.notif_template_text = templates
|
||||||
|
|
||||||
self.pusher_types["email"] = self._create_email_pusher
|
self.pusher_types["email"] = self._create_email_pusher
|
||||||
|
@ -74,7 +74,9 @@ REQUIREMENTS = [
|
|||||||
CONDITIONAL_REQUIREMENTS = {
|
CONDITIONAL_REQUIREMENTS = {
|
||||||
"email.enable_notifs": ["Jinja2>=2.9", "bleach>=1.4.2"],
|
"email.enable_notifs": ["Jinja2>=2.9", "bleach>=1.4.2"],
|
||||||
"matrix-synapse-ldap3": ["matrix-synapse-ldap3>=0.1"],
|
"matrix-synapse-ldap3": ["matrix-synapse-ldap3>=0.1"],
|
||||||
"postgres": ["psycopg2>=2.6"],
|
|
||||||
|
# we use execute_batch, which arrived in psycopg 2.7.
|
||||||
|
"postgres": ["psycopg2>=2.7"],
|
||||||
|
|
||||||
# ConsentResource uses select_autoescape, which arrived in jinja 2.9
|
# ConsentResource uses select_autoescape, which arrived in jinja 2.9
|
||||||
"resources.consent": ["Jinja2>=2.9"],
|
"resources.consent": ["Jinja2>=2.9"],
|
||||||
@ -84,18 +86,22 @@ CONDITIONAL_REQUIREMENTS = {
|
|||||||
"acme": ["txacme>=0.9.2"],
|
"acme": ["txacme>=0.9.2"],
|
||||||
|
|
||||||
"saml2": ["pysaml2>=4.5.0"],
|
"saml2": ["pysaml2>=4.5.0"],
|
||||||
|
"systemd": ["systemd-python>=231"],
|
||||||
"url_preview": ["lxml>=3.5.0"],
|
"url_preview": ["lxml>=3.5.0"],
|
||||||
"test": ["mock>=2.0", "parameterized"],
|
"test": ["mock>=2.0", "parameterized"],
|
||||||
"sentry": ["sentry-sdk>=0.7.2"],
|
"sentry": ["sentry-sdk>=0.7.2"],
|
||||||
}
|
}
|
||||||
|
|
||||||
|
ALL_OPTIONAL_REQUIREMENTS = set()
|
||||||
|
|
||||||
|
for name, optional_deps in CONDITIONAL_REQUIREMENTS.items():
|
||||||
|
# Exclude systemd as it's a system-based requirement.
|
||||||
|
if name not in ["systemd"]:
|
||||||
|
ALL_OPTIONAL_REQUIREMENTS = set(optional_deps) | ALL_OPTIONAL_REQUIREMENTS
|
||||||
|
|
||||||
|
|
||||||
def list_requirements():
|
def list_requirements():
|
||||||
deps = set(REQUIREMENTS)
|
return list(set(REQUIREMENTS) | ALL_OPTIONAL_REQUIREMENTS)
|
||||||
for opt in CONDITIONAL_REQUIREMENTS.values():
|
|
||||||
deps = set(opt) | deps
|
|
||||||
|
|
||||||
return list(deps)
|
|
||||||
|
|
||||||
|
|
||||||
class DependencyException(Exception):
|
class DependencyException(Exception):
|
||||||
|
@ -13,22 +13,9 @@
|
|||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
|
||||||
from synapse.storage import DataStore
|
from synapse.storage import KeyStore
|
||||||
from synapse.storage.keys import KeyStore
|
|
||||||
|
|
||||||
from ._base import BaseSlavedStore, __func__
|
# KeyStore isn't really safe to use from a worker, but for now we do so and hope that
|
||||||
|
# the races it creates aren't too bad.
|
||||||
|
|
||||||
|
SlavedKeyStore = KeyStore
|
||||||
class SlavedKeyStore(BaseSlavedStore):
|
|
||||||
_get_server_verify_key = KeyStore.__dict__[
|
|
||||||
"_get_server_verify_key"
|
|
||||||
]
|
|
||||||
|
|
||||||
get_server_verify_keys = __func__(DataStore.get_server_verify_keys)
|
|
||||||
store_server_verify_key = __func__(DataStore.store_server_verify_key)
|
|
||||||
|
|
||||||
get_server_certificate = __func__(DataStore.get_server_certificate)
|
|
||||||
store_server_certificate = __func__(DataStore.store_server_certificate)
|
|
||||||
|
|
||||||
get_server_keys_json = __func__(DataStore.get_server_keys_json)
|
|
||||||
store_server_keys_json = __func__(DataStore.store_server_keys_json)
|
|
||||||
|
4
synapse/res/templates/mail-expiry.css
Normal file
4
synapse/res/templates/mail-expiry.css
Normal file
@ -0,0 +1,4 @@
|
|||||||
|
.noticetext {
|
||||||
|
margin-top: 10px;
|
||||||
|
margin-bottom: 10px;
|
||||||
|
}
|
43
synapse/res/templates/notice_expiry.html
Normal file
43
synapse/res/templates/notice_expiry.html
Normal file
@ -0,0 +1,43 @@
|
|||||||
|
<!doctype html>
|
||||||
|
<html lang="en">
|
||||||
|
<head>
|
||||||
|
<style type="text/css">
|
||||||
|
{% include 'mail.css' without context %}
|
||||||
|
{% include "mail-%s.css" % app_name ignore missing without context %}
|
||||||
|
{% include 'mail-expiry.css' without context %}
|
||||||
|
</style>
|
||||||
|
</head>
|
||||||
|
<body>
|
||||||
|
<table id="page">
|
||||||
|
<tr>
|
||||||
|
<td> </td>
|
||||||
|
<td id="inner">
|
||||||
|
<table class="header">
|
||||||
|
<tr>
|
||||||
|
<td>
|
||||||
|
<div class="salutation">Hi {{ display_name }},</div>
|
||||||
|
</td>
|
||||||
|
<td class="logo">
|
||||||
|
{% if app_name == "Riot" %}
|
||||||
|
<img src="http://riot.im/img/external/riot-logo-email.png" width="83" height="83" alt="[Riot]"/>
|
||||||
|
{% elif app_name == "Vector" %}
|
||||||
|
<img src="http://matrix.org/img/vector-logo-email.png" width="64" height="83" alt="[Vector]"/>
|
||||||
|
{% else %}
|
||||||
|
<img src="http://matrix.org/img/matrix-120x51.png" width="120" height="51" alt="[matrix]"/>
|
||||||
|
{% endif %}
|
||||||
|
</td>
|
||||||
|
</tr>
|
||||||
|
<tr>
|
||||||
|
<td colspan="2">
|
||||||
|
<div class="noticetext">Your account will expire on {{ expiration_ts|format_ts("%d-%m-%Y") }}. This means that you will lose access to your account after this date.</div>
|
||||||
|
<div class="noticetext">To extend the validity of your account, please click on the link bellow (or copy and paste it into a new browser tab):</div>
|
||||||
|
<div class="noticetext"><a href="{{ url }}">{{ url }}</a></div>
|
||||||
|
</td>
|
||||||
|
</tr>
|
||||||
|
</table>
|
||||||
|
</td>
|
||||||
|
<td> </td>
|
||||||
|
</tr>
|
||||||
|
</table>
|
||||||
|
</body>
|
||||||
|
</html>
|
7
synapse/res/templates/notice_expiry.txt
Normal file
7
synapse/res/templates/notice_expiry.txt
Normal file
@ -0,0 +1,7 @@
|
|||||||
|
Hi {{ display_name }},
|
||||||
|
|
||||||
|
Your account will expire on {{ expiration_ts|format_ts("%d-%m-%Y") }}. This means that you will lose access to your account after this date.
|
||||||
|
|
||||||
|
To extend the validity of your account, please click on the link bellow (or copy and paste it to a new browser tab):
|
||||||
|
|
||||||
|
{{ url }}
|
@ -33,6 +33,7 @@ from synapse.rest.client.v1 import (
|
|||||||
from synapse.rest.client.v2_alpha import (
|
from synapse.rest.client.v2_alpha import (
|
||||||
account,
|
account,
|
||||||
account_data,
|
account_data,
|
||||||
|
account_validity,
|
||||||
auth,
|
auth,
|
||||||
capabilities,
|
capabilities,
|
||||||
devices,
|
devices,
|
||||||
@ -109,3 +110,4 @@ class ClientRestResource(JsonResource):
|
|||||||
groups.register_servlets(hs, client_resource)
|
groups.register_servlets(hs, client_resource)
|
||||||
room_upgrade_rest_servlet.register_servlets(hs, client_resource)
|
room_upgrade_rest_servlet.register_servlets(hs, client_resource)
|
||||||
capabilities.register_servlets(hs, client_resource)
|
capabilities.register_servlets(hs, client_resource)
|
||||||
|
account_validity.register_servlets(hs, client_resource)
|
||||||
|
@ -809,6 +809,44 @@ class DeleteGroupAdminRestServlet(ClientV1RestServlet):
|
|||||||
defer.returnValue((200, {}))
|
defer.returnValue((200, {}))
|
||||||
|
|
||||||
|
|
||||||
|
class AccountValidityRenewServlet(ClientV1RestServlet):
|
||||||
|
PATTERNS = client_path_patterns("/admin/account_validity/validity$")
|
||||||
|
|
||||||
|
def __init__(self, hs):
|
||||||
|
"""
|
||||||
|
Args:
|
||||||
|
hs (synapse.server.HomeServer): server
|
||||||
|
"""
|
||||||
|
super(AccountValidityRenewServlet, self).__init__(hs)
|
||||||
|
|
||||||
|
self.hs = hs
|
||||||
|
self.account_activity_handler = hs.get_account_validity_handler()
|
||||||
|
self.auth = hs.get_auth()
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def on_POST(self, request):
|
||||||
|
requester = yield self.auth.get_user_by_req(request)
|
||||||
|
is_admin = yield self.auth.is_server_admin(requester.user)
|
||||||
|
|
||||||
|
if not is_admin:
|
||||||
|
raise AuthError(403, "You are not a server admin")
|
||||||
|
|
||||||
|
body = parse_json_object_from_request(request)
|
||||||
|
|
||||||
|
if "user_id" not in body:
|
||||||
|
raise SynapseError(400, "Missing property 'user_id' in the request body")
|
||||||
|
|
||||||
|
expiration_ts = yield self.account_activity_handler.renew_account_for_user(
|
||||||
|
body["user_id"], body.get("expiration_ts"),
|
||||||
|
not body.get("enable_renewal_emails", True),
|
||||||
|
)
|
||||||
|
|
||||||
|
res = {
|
||||||
|
"expiration_ts": expiration_ts,
|
||||||
|
}
|
||||||
|
defer.returnValue((200, res))
|
||||||
|
|
||||||
|
|
||||||
def register_servlets(hs, http_server):
|
def register_servlets(hs, http_server):
|
||||||
WhoisRestServlet(hs).register(http_server)
|
WhoisRestServlet(hs).register(http_server)
|
||||||
PurgeMediaCacheRestServlet(hs).register(http_server)
|
PurgeMediaCacheRestServlet(hs).register(http_server)
|
||||||
@ -825,3 +863,4 @@ def register_servlets(hs, http_server):
|
|||||||
UserRegisterServlet(hs).register(http_server)
|
UserRegisterServlet(hs).register(http_server)
|
||||||
VersionServlet(hs).register(http_server)
|
VersionServlet(hs).register(http_server)
|
||||||
DeleteGroupAdminRestServlet(hs).register(http_server)
|
DeleteGroupAdminRestServlet(hs).register(http_server)
|
||||||
|
AccountValidityRenewServlet(hs).register(http_server)
|
||||||
|
91
synapse/rest/client/v2_alpha/account_validity.py
Normal file
91
synapse/rest/client/v2_alpha/account_validity.py
Normal file
@ -0,0 +1,91 @@
|
|||||||
|
# -*- coding: utf-8 -*-
|
||||||
|
# Copyright 2019 New Vector Ltd
|
||||||
|
#
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
# you may not use this file except in compliance with the License.
|
||||||
|
# You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
# See the License for the specific language governing permissions and
|
||||||
|
# limitations under the License.
|
||||||
|
|
||||||
|
import logging
|
||||||
|
|
||||||
|
from twisted.internet import defer
|
||||||
|
|
||||||
|
from synapse.api.errors import AuthError, SynapseError
|
||||||
|
from synapse.http.server import finish_request
|
||||||
|
from synapse.http.servlet import RestServlet
|
||||||
|
|
||||||
|
from ._base import client_v2_patterns
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
class AccountValidityRenewServlet(RestServlet):
|
||||||
|
PATTERNS = client_v2_patterns("/account_validity/renew$")
|
||||||
|
SUCCESS_HTML = b"<html><body>Your account has been successfully renewed.</body><html>"
|
||||||
|
|
||||||
|
def __init__(self, hs):
|
||||||
|
"""
|
||||||
|
Args:
|
||||||
|
hs (synapse.server.HomeServer): server
|
||||||
|
"""
|
||||||
|
super(AccountValidityRenewServlet, self).__init__()
|
||||||
|
|
||||||
|
self.hs = hs
|
||||||
|
self.account_activity_handler = hs.get_account_validity_handler()
|
||||||
|
self.auth = hs.get_auth()
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def on_GET(self, request):
|
||||||
|
if b"token" not in request.args:
|
||||||
|
raise SynapseError(400, "Missing renewal token")
|
||||||
|
renewal_token = request.args[b"token"][0]
|
||||||
|
|
||||||
|
yield self.account_activity_handler.renew_account(renewal_token.decode('utf8'))
|
||||||
|
|
||||||
|
request.setResponseCode(200)
|
||||||
|
request.setHeader(b"Content-Type", b"text/html; charset=utf-8")
|
||||||
|
request.setHeader(b"Content-Length", b"%d" % (
|
||||||
|
len(AccountValidityRenewServlet.SUCCESS_HTML),
|
||||||
|
))
|
||||||
|
request.write(AccountValidityRenewServlet.SUCCESS_HTML)
|
||||||
|
finish_request(request)
|
||||||
|
defer.returnValue(None)
|
||||||
|
|
||||||
|
|
||||||
|
class AccountValiditySendMailServlet(RestServlet):
|
||||||
|
PATTERNS = client_v2_patterns("/account_validity/send_mail$")
|
||||||
|
|
||||||
|
def __init__(self, hs):
|
||||||
|
"""
|
||||||
|
Args:
|
||||||
|
hs (synapse.server.HomeServer): server
|
||||||
|
"""
|
||||||
|
super(AccountValiditySendMailServlet, self).__init__()
|
||||||
|
|
||||||
|
self.hs = hs
|
||||||
|
self.account_activity_handler = hs.get_account_validity_handler()
|
||||||
|
self.auth = hs.get_auth()
|
||||||
|
self.account_validity = self.hs.config.account_validity
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def on_POST(self, request):
|
||||||
|
if not self.account_validity.renew_by_email_enabled:
|
||||||
|
raise AuthError(403, "Account renewal via email is disabled on this server.")
|
||||||
|
|
||||||
|
requester = yield self.auth.get_user_by_req(request)
|
||||||
|
user_id = requester.user.to_string()
|
||||||
|
yield self.account_activity_handler.send_renewal_email_to_user(user_id)
|
||||||
|
|
||||||
|
defer.returnValue((200, {}))
|
||||||
|
|
||||||
|
|
||||||
|
def register_servlets(hs, http_server):
|
||||||
|
AccountValidityRenewServlet(hs).register(http_server)
|
||||||
|
AccountValiditySendMailServlet(hs).register(http_server)
|
@ -391,6 +391,13 @@ class RegisterRestServlet(RestServlet):
|
|||||||
# the user-facing checks will probably already have happened in
|
# the user-facing checks will probably already have happened in
|
||||||
# /register/email/requestToken when we requested a 3pid, but that's not
|
# /register/email/requestToken when we requested a 3pid, but that's not
|
||||||
# guaranteed.
|
# guaranteed.
|
||||||
|
#
|
||||||
|
# Also check that we're not trying to register a 3pid that's already
|
||||||
|
# been registered.
|
||||||
|
#
|
||||||
|
# This has probably happened in /register/email/requestToken as well,
|
||||||
|
# but if a user hits this endpoint twice then clicks on each link from
|
||||||
|
# the two activation emails, they would register the same 3pid twice.
|
||||||
|
|
||||||
if auth_result:
|
if auth_result:
|
||||||
for login_type in [LoginType.EMAIL_IDENTITY, LoginType.MSISDN]:
|
for login_type in [LoginType.EMAIL_IDENTITY, LoginType.MSISDN]:
|
||||||
@ -406,6 +413,17 @@ class RegisterRestServlet(RestServlet):
|
|||||||
Codes.THREEPID_DENIED,
|
Codes.THREEPID_DENIED,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
existingUid = yield self.store.get_user_id_by_threepid(
|
||||||
|
medium, address,
|
||||||
|
)
|
||||||
|
|
||||||
|
if existingUid is not None:
|
||||||
|
raise SynapseError(
|
||||||
|
400,
|
||||||
|
"%s is already in use" % medium,
|
||||||
|
Codes.THREEPID_IN_USE,
|
||||||
|
)
|
||||||
|
|
||||||
if registered_user_id is not None:
|
if registered_user_id is not None:
|
||||||
logger.info(
|
logger.info(
|
||||||
"Already registered user ID %r for this session",
|
"Already registered user ID %r for this session",
|
||||||
|
@ -1,6 +1,6 @@
|
|||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
# Copyright 2014-2016 OpenMarket Ltd
|
# Copyright 2014-2016 OpenMarket Ltd
|
||||||
# Copyright 2019 New Vector Ltd.
|
# Copyright 2019 New Vector Ltd
|
||||||
#
|
#
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
# you may not use this file except in compliance with the License.
|
# you may not use this file except in compliance with the License.
|
||||||
@ -191,6 +191,10 @@ def respond_with_responder(request, responder, media_type, file_size, upload_nam
|
|||||||
# in that case.
|
# in that case.
|
||||||
logger.warning("Failed to write to consumer: %s %s", type(e), e)
|
logger.warning("Failed to write to consumer: %s %s", type(e), e)
|
||||||
|
|
||||||
|
# Unregister the producer, if it has one, so Twisted doesn't complain
|
||||||
|
if request.producer:
|
||||||
|
request.unregisterProducer()
|
||||||
|
|
||||||
finish_request(request)
|
finish_request(request)
|
||||||
|
|
||||||
|
|
||||||
|
@ -1,5 +1,5 @@
|
|||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
# Copyright 2018 New Vector Ltd.
|
# Copyright 2018 New Vector Ltd
|
||||||
#
|
#
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
# you may not use this file except in compliance with the License.
|
# you may not use this file except in compliance with the License.
|
||||||
@ -68,6 +68,6 @@ class WellKnownResource(Resource):
|
|||||||
request.setHeader(b"Content-Type", b"text/plain")
|
request.setHeader(b"Content-Type", b"text/plain")
|
||||||
return b'.well-known not available'
|
return b'.well-known not available'
|
||||||
|
|
||||||
logger.error("returning: %s", r)
|
logger.debug("returning: %s", r)
|
||||||
request.setHeader(b"Content-Type", b"application/json")
|
request.setHeader(b"Content-Type", b"application/json")
|
||||||
return json.dumps(r).encode("utf-8")
|
return json.dumps(r).encode("utf-8")
|
||||||
|
@ -47,6 +47,7 @@ from synapse.federation.transport.client import TransportLayerClient
|
|||||||
from synapse.groups.attestations import GroupAttestationSigning, GroupAttestionRenewer
|
from synapse.groups.attestations import GroupAttestationSigning, GroupAttestionRenewer
|
||||||
from synapse.groups.groups_server import GroupsServerHandler
|
from synapse.groups.groups_server import GroupsServerHandler
|
||||||
from synapse.handlers import Handlers
|
from synapse.handlers import Handlers
|
||||||
|
from synapse.handlers.account_validity import AccountValidityHandler
|
||||||
from synapse.handlers.acme import AcmeHandler
|
from synapse.handlers.acme import AcmeHandler
|
||||||
from synapse.handlers.appservice import ApplicationServicesHandler
|
from synapse.handlers.appservice import ApplicationServicesHandler
|
||||||
from synapse.handlers.auth import AuthHandler, MacaroonGenerator
|
from synapse.handlers.auth import AuthHandler, MacaroonGenerator
|
||||||
@ -183,6 +184,7 @@ class HomeServer(object):
|
|||||||
'room_context_handler',
|
'room_context_handler',
|
||||||
'sendmail',
|
'sendmail',
|
||||||
'registration_handler',
|
'registration_handler',
|
||||||
|
'account_validity_handler',
|
||||||
]
|
]
|
||||||
|
|
||||||
REQUIRED_ON_MASTER_STARTUP = [
|
REQUIRED_ON_MASTER_STARTUP = [
|
||||||
@ -506,6 +508,9 @@ class HomeServer(object):
|
|||||||
def build_registration_handler(self):
|
def build_registration_handler(self):
|
||||||
return RegistrationHandler(self)
|
return RegistrationHandler(self)
|
||||||
|
|
||||||
|
def build_account_validity_handler(self):
|
||||||
|
return AccountValidityHandler(self)
|
||||||
|
|
||||||
def remove_pusher(self, app_id, push_key, user_id):
|
def remove_pusher(self, app_id, push_key, user_id):
|
||||||
return self.get_pusherpool().remove_pusher(app_id, push_key, user_id)
|
return self.get_pusherpool().remove_pusher(app_id, push_key, user_id)
|
||||||
|
|
||||||
|
@ -1179,14 +1179,10 @@ class EventsStore(
|
|||||||
"events",
|
"events",
|
||||||
"event_auth",
|
"event_auth",
|
||||||
"event_json",
|
"event_json",
|
||||||
"event_content_hashes",
|
|
||||||
"event_destinations",
|
|
||||||
"event_edge_hashes",
|
|
||||||
"event_edges",
|
"event_edges",
|
||||||
"event_forward_extremities",
|
"event_forward_extremities",
|
||||||
"event_reference_hashes",
|
"event_reference_hashes",
|
||||||
"event_search",
|
"event_search",
|
||||||
"event_signatures",
|
|
||||||
"event_to_state_groups",
|
"event_to_state_groups",
|
||||||
"guest_access",
|
"guest_access",
|
||||||
"history_visibility",
|
"history_visibility",
|
||||||
@ -1857,16 +1853,12 @@ class EventsStore(
|
|||||||
# Tables that should be pruned:
|
# Tables that should be pruned:
|
||||||
# event_auth
|
# event_auth
|
||||||
# event_backward_extremities
|
# event_backward_extremities
|
||||||
# event_content_hashes
|
|
||||||
# event_destinations
|
|
||||||
# event_edge_hashes
|
|
||||||
# event_edges
|
# event_edges
|
||||||
# event_forward_extremities
|
# event_forward_extremities
|
||||||
# event_json
|
# event_json
|
||||||
# event_push_actions
|
# event_push_actions
|
||||||
# event_reference_hashes
|
# event_reference_hashes
|
||||||
# event_search
|
# event_search
|
||||||
# event_signatures
|
|
||||||
# event_to_state_groups
|
# event_to_state_groups
|
||||||
# events
|
# events
|
||||||
# rejections
|
# rejections
|
||||||
@ -2065,14 +2057,10 @@ class EventsStore(
|
|||||||
"events",
|
"events",
|
||||||
"event_json",
|
"event_json",
|
||||||
"event_auth",
|
"event_auth",
|
||||||
"event_content_hashes",
|
|
||||||
"event_destinations",
|
|
||||||
"event_edge_hashes",
|
|
||||||
"event_edges",
|
"event_edges",
|
||||||
"event_forward_extremities",
|
"event_forward_extremities",
|
||||||
"event_reference_hashes",
|
"event_reference_hashes",
|
||||||
"event_search",
|
"event_search",
|
||||||
"event_signatures",
|
|
||||||
"rejections",
|
"rejections",
|
||||||
):
|
):
|
||||||
logger.info("[purge] removing events from %s", table)
|
logger.info("[purge] removing events from %s", table)
|
||||||
|
@ -1,5 +1,6 @@
|
|||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
# Copyright 2014-2016 OpenMarket Ltd
|
# Copyright 2014-2016 OpenMarket Ltd
|
||||||
|
# Copyright 2019 New Vector Ltd.
|
||||||
#
|
#
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
# you may not use this file except in compliance with the License.
|
# you may not use this file except in compliance with the License.
|
||||||
@ -13,17 +14,15 @@
|
|||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
|
||||||
import hashlib
|
import itertools
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
import six
|
import six
|
||||||
|
|
||||||
from signedjson.key import decode_verify_key_bytes
|
from signedjson.key import decode_verify_key_bytes
|
||||||
|
|
||||||
import OpenSSL
|
from synapse.util import batch_iter
|
||||||
from twisted.internet import defer
|
from synapse.util.caches.descriptors import cached, cachedList
|
||||||
|
|
||||||
from synapse.util.caches.descriptors import cachedInlineCallbacks
|
|
||||||
|
|
||||||
from ._base import SQLBaseStore
|
from ._base import SQLBaseStore
|
||||||
|
|
||||||
@ -38,83 +37,52 @@ else:
|
|||||||
|
|
||||||
|
|
||||||
class KeyStore(SQLBaseStore):
|
class KeyStore(SQLBaseStore):
|
||||||
"""Persistence for signature verification keys and tls X.509 certificates
|
"""Persistence for signature verification keys
|
||||||
"""
|
"""
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@cached()
|
||||||
def get_server_certificate(self, server_name):
|
def _get_server_verify_key(self, server_name_and_key_id):
|
||||||
"""Retrieve the TLS X.509 certificate for the given server
|
raise NotImplementedError()
|
||||||
Args:
|
|
||||||
server_name (bytes): The name of the server.
|
@cachedList(
|
||||||
Returns:
|
cached_method_name="_get_server_verify_key", list_name="server_name_and_key_ids"
|
||||||
(OpenSSL.crypto.X509): The tls certificate.
|
)
|
||||||
|
def get_server_verify_keys(self, server_name_and_key_ids):
|
||||||
"""
|
"""
|
||||||
tls_certificate_bytes, = yield self._simple_select_one(
|
|
||||||
table="server_tls_certificates",
|
|
||||||
keyvalues={"server_name": server_name},
|
|
||||||
retcols=("tls_certificate",),
|
|
||||||
desc="get_server_certificate",
|
|
||||||
)
|
|
||||||
tls_certificate = OpenSSL.crypto.load_certificate(
|
|
||||||
OpenSSL.crypto.FILETYPE_ASN1, tls_certificate_bytes
|
|
||||||
)
|
|
||||||
defer.returnValue(tls_certificate)
|
|
||||||
|
|
||||||
def store_server_certificate(
|
|
||||||
self, server_name, from_server, time_now_ms, tls_certificate
|
|
||||||
):
|
|
||||||
"""Stores the TLS X.509 certificate for the given server
|
|
||||||
Args:
|
Args:
|
||||||
server_name (str): The name of the server.
|
server_name_and_key_ids (iterable[Tuple[str, str]]):
|
||||||
from_server (str): Where the certificate was looked up
|
iterable of (server_name, key-id) tuples to fetch keys for
|
||||||
time_now_ms (int): The time now in milliseconds
|
|
||||||
tls_certificate (OpenSSL.crypto.X509): The X.509 certificate.
|
|
||||||
"""
|
|
||||||
tls_certificate_bytes = OpenSSL.crypto.dump_certificate(
|
|
||||||
OpenSSL.crypto.FILETYPE_ASN1, tls_certificate
|
|
||||||
)
|
|
||||||
fingerprint = hashlib.sha256(tls_certificate_bytes).hexdigest()
|
|
||||||
return self._simple_upsert(
|
|
||||||
table="server_tls_certificates",
|
|
||||||
keyvalues={"server_name": server_name, "fingerprint": fingerprint},
|
|
||||||
values={
|
|
||||||
"from_server": from_server,
|
|
||||||
"ts_added_ms": time_now_ms,
|
|
||||||
"tls_certificate": db_binary_type(tls_certificate_bytes),
|
|
||||||
},
|
|
||||||
desc="store_server_certificate",
|
|
||||||
)
|
|
||||||
|
|
||||||
@cachedInlineCallbacks()
|
|
||||||
def _get_server_verify_key(self, server_name, key_id):
|
|
||||||
verify_key_bytes = yield self._simple_select_one_onecol(
|
|
||||||
table="server_signature_keys",
|
|
||||||
keyvalues={"server_name": server_name, "key_id": key_id},
|
|
||||||
retcol="verify_key",
|
|
||||||
desc="_get_server_verify_key",
|
|
||||||
allow_none=True,
|
|
||||||
)
|
|
||||||
|
|
||||||
if verify_key_bytes:
|
|
||||||
defer.returnValue(decode_verify_key_bytes(key_id, bytes(verify_key_bytes)))
|
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
|
||||||
def get_server_verify_keys(self, server_name, key_ids):
|
|
||||||
"""Retrieve the NACL verification key for a given server for the given
|
|
||||||
key_ids
|
|
||||||
Args:
|
|
||||||
server_name (str): The name of the server.
|
|
||||||
key_ids (iterable[str]): key_ids to try and look up.
|
|
||||||
Returns:
|
Returns:
|
||||||
Deferred: resolves to dict[str, VerifyKey]: map from
|
Deferred: resolves to dict[Tuple[str, str], VerifyKey|None]:
|
||||||
key_id to verification key.
|
map from (server_name, key_id) -> VerifyKey, or None if the key is
|
||||||
|
unknown
|
||||||
"""
|
"""
|
||||||
keys = {}
|
keys = {}
|
||||||
for key_id in key_ids:
|
|
||||||
key = yield self._get_server_verify_key(server_name, key_id)
|
def _get_keys(txn, batch):
|
||||||
if key:
|
"""Processes a batch of keys to fetch, and adds the result to `keys`."""
|
||||||
keys[key_id] = key
|
|
||||||
defer.returnValue(keys)
|
# batch_iter always returns tuples so it's safe to do len(batch)
|
||||||
|
sql = (
|
||||||
|
"SELECT server_name, key_id, verify_key FROM server_signature_keys "
|
||||||
|
"WHERE 1=0"
|
||||||
|
) + " OR (server_name=? AND key_id=?)" * len(batch)
|
||||||
|
|
||||||
|
txn.execute(sql, tuple(itertools.chain.from_iterable(batch)))
|
||||||
|
|
||||||
|
for row in txn:
|
||||||
|
server_name, key_id, key_bytes = row
|
||||||
|
keys[(server_name, key_id)] = decode_verify_key_bytes(
|
||||||
|
key_id, bytes(key_bytes)
|
||||||
|
)
|
||||||
|
|
||||||
|
def _txn(txn):
|
||||||
|
for batch in batch_iter(server_name_and_key_ids, 50):
|
||||||
|
_get_keys(txn, batch)
|
||||||
|
return keys
|
||||||
|
|
||||||
|
return self.runInteraction("get_server_verify_keys", _txn)
|
||||||
|
|
||||||
def store_server_verify_key(
|
def store_server_verify_key(
|
||||||
self, server_name, from_server, time_now_ms, verify_key
|
self, server_name, from_server, time_now_ms, verify_key
|
||||||
@ -140,8 +108,11 @@ class KeyStore(SQLBaseStore):
|
|||||||
"verify_key": db_binary_type(verify_key.encode()),
|
"verify_key": db_binary_type(verify_key.encode()),
|
||||||
},
|
},
|
||||||
)
|
)
|
||||||
|
# invalidate takes a tuple corresponding to the params of
|
||||||
|
# _get_server_verify_key. _get_server_verify_key only takes one
|
||||||
|
# param, which is itself the 2-tuple (server_name, key_id).
|
||||||
txn.call_after(
|
txn.call_after(
|
||||||
self._get_server_verify_key.invalidate, (server_name, key_id)
|
self._get_server_verify_key.invalidate, ((server_name, key_id),)
|
||||||
)
|
)
|
||||||
|
|
||||||
return self.runInteraction("store_server_verify_key", _txn)
|
return self.runInteraction("store_server_verify_key", _txn)
|
||||||
@ -188,8 +159,8 @@ class KeyStore(SQLBaseStore):
|
|||||||
Args:
|
Args:
|
||||||
server_keys (list): List of (server_name, key_id, source) triplets.
|
server_keys (list): List of (server_name, key_id, source) triplets.
|
||||||
Returns:
|
Returns:
|
||||||
Dict mapping (server_name, key_id, source) triplets to dicts with
|
Deferred[dict[Tuple[str, str, str|None], list[dict]]]:
|
||||||
"ts_valid_until_ms" and "key_json" keys.
|
Dict mapping (server_name, key_id, source) triplets to lists of dicts
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def _get_server_keys_json_txn(txn):
|
def _get_server_keys_json_txn(txn):
|
||||||
|
@ -32,6 +32,7 @@ class RegistrationWorkerStore(SQLBaseStore):
|
|||||||
super(RegistrationWorkerStore, self).__init__(db_conn, hs)
|
super(RegistrationWorkerStore, self).__init__(db_conn, hs)
|
||||||
|
|
||||||
self.config = hs.config
|
self.config = hs.config
|
||||||
|
self.clock = hs.get_clock()
|
||||||
|
|
||||||
@cached()
|
@cached()
|
||||||
def get_user_by_id(self, user_id):
|
def get_user_by_id(self, user_id):
|
||||||
@ -86,6 +87,162 @@ class RegistrationWorkerStore(SQLBaseStore):
|
|||||||
"get_user_by_access_token", self._query_for_auth, token
|
"get_user_by_access_token", self._query_for_auth, token
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@cachedInlineCallbacks()
|
||||||
|
def get_expiration_ts_for_user(self, user_id):
|
||||||
|
"""Get the expiration timestamp for the account bearing a given user ID.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
user_id (str): The ID of the user.
|
||||||
|
Returns:
|
||||||
|
defer.Deferred: None, if the account has no expiration timestamp,
|
||||||
|
otherwise int representation of the timestamp (as a number of
|
||||||
|
milliseconds since epoch).
|
||||||
|
"""
|
||||||
|
res = yield self._simple_select_one_onecol(
|
||||||
|
table="account_validity",
|
||||||
|
keyvalues={"user_id": user_id},
|
||||||
|
retcol="expiration_ts_ms",
|
||||||
|
allow_none=True,
|
||||||
|
desc="get_expiration_ts_for_user",
|
||||||
|
)
|
||||||
|
defer.returnValue(res)
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def set_account_validity_for_user(self, user_id, expiration_ts, email_sent,
|
||||||
|
renewal_token=None):
|
||||||
|
"""Updates the account validity properties of the given account, with the
|
||||||
|
given values.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
user_id (str): ID of the account to update properties for.
|
||||||
|
expiration_ts (int): New expiration date, as a timestamp in milliseconds
|
||||||
|
since epoch.
|
||||||
|
email_sent (bool): True means a renewal email has been sent for this
|
||||||
|
account and there's no need to send another one for the current validity
|
||||||
|
period.
|
||||||
|
renewal_token (str): Renewal token the user can use to extend the validity
|
||||||
|
of their account. Defaults to no token.
|
||||||
|
"""
|
||||||
|
def set_account_validity_for_user_txn(txn):
|
||||||
|
self._simple_update_txn(
|
||||||
|
txn=txn,
|
||||||
|
table="account_validity",
|
||||||
|
keyvalues={"user_id": user_id},
|
||||||
|
updatevalues={
|
||||||
|
"expiration_ts_ms": expiration_ts,
|
||||||
|
"email_sent": email_sent,
|
||||||
|
"renewal_token": renewal_token,
|
||||||
|
},
|
||||||
|
)
|
||||||
|
self._invalidate_cache_and_stream(
|
||||||
|
txn, self.get_expiration_ts_for_user, (user_id,),
|
||||||
|
)
|
||||||
|
|
||||||
|
yield self.runInteraction(
|
||||||
|
"set_account_validity_for_user",
|
||||||
|
set_account_validity_for_user_txn,
|
||||||
|
)
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def set_renewal_token_for_user(self, user_id, renewal_token):
|
||||||
|
"""Defines a renewal token for a given user.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
user_id (str): ID of the user to set the renewal token for.
|
||||||
|
renewal_token (str): Random unique string that will be used to renew the
|
||||||
|
user's account.
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
StoreError: The provided token is already set for another user.
|
||||||
|
"""
|
||||||
|
yield self._simple_update_one(
|
||||||
|
table="account_validity",
|
||||||
|
keyvalues={"user_id": user_id},
|
||||||
|
updatevalues={"renewal_token": renewal_token},
|
||||||
|
desc="set_renewal_token_for_user",
|
||||||
|
)
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def get_user_from_renewal_token(self, renewal_token):
|
||||||
|
"""Get a user ID from a renewal token.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
renewal_token (str): The renewal token to perform the lookup with.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
defer.Deferred[str]: The ID of the user to which the token belongs.
|
||||||
|
"""
|
||||||
|
res = yield self._simple_select_one_onecol(
|
||||||
|
table="account_validity",
|
||||||
|
keyvalues={"renewal_token": renewal_token},
|
||||||
|
retcol="user_id",
|
||||||
|
desc="get_user_from_renewal_token",
|
||||||
|
)
|
||||||
|
|
||||||
|
defer.returnValue(res)
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def get_renewal_token_for_user(self, user_id):
|
||||||
|
"""Get the renewal token associated with a given user ID.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
user_id (str): The user ID to lookup a token for.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
defer.Deferred[str]: The renewal token associated with this user ID.
|
||||||
|
"""
|
||||||
|
res = yield self._simple_select_one_onecol(
|
||||||
|
table="account_validity",
|
||||||
|
keyvalues={"user_id": user_id},
|
||||||
|
retcol="renewal_token",
|
||||||
|
desc="get_renewal_token_for_user",
|
||||||
|
)
|
||||||
|
|
||||||
|
defer.returnValue(res)
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def get_users_expiring_soon(self):
|
||||||
|
"""Selects users whose account will expire in the [now, now + renew_at] time
|
||||||
|
window (see configuration for account_validity for information on what renew_at
|
||||||
|
refers to).
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Deferred: Resolves to a list[dict[user_id (str), expiration_ts_ms (int)]]
|
||||||
|
"""
|
||||||
|
def select_users_txn(txn, now_ms, renew_at):
|
||||||
|
sql = (
|
||||||
|
"SELECT user_id, expiration_ts_ms FROM account_validity"
|
||||||
|
" WHERE email_sent = ? AND (expiration_ts_ms - ?) <= ?"
|
||||||
|
)
|
||||||
|
values = [False, now_ms, renew_at]
|
||||||
|
txn.execute(sql, values)
|
||||||
|
return self.cursor_to_dict(txn)
|
||||||
|
|
||||||
|
res = yield self.runInteraction(
|
||||||
|
"get_users_expiring_soon",
|
||||||
|
select_users_txn,
|
||||||
|
self.clock.time_msec(), self.config.account_validity.renew_at,
|
||||||
|
)
|
||||||
|
|
||||||
|
defer.returnValue(res)
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def set_renewal_mail_status(self, user_id, email_sent):
|
||||||
|
"""Sets or unsets the flag that indicates whether a renewal email has been sent
|
||||||
|
to the user (and the user hasn't renewed their account yet).
|
||||||
|
|
||||||
|
Args:
|
||||||
|
user_id (str): ID of the user to set/unset the flag for.
|
||||||
|
email_sent (bool): Flag which indicates whether a renewal email has been sent
|
||||||
|
to this user.
|
||||||
|
"""
|
||||||
|
yield self._simple_update_one(
|
||||||
|
table="account_validity",
|
||||||
|
keyvalues={"user_id": user_id},
|
||||||
|
updatevalues={"email_sent": email_sent},
|
||||||
|
desc="set_renewal_mail_status",
|
||||||
|
)
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def is_server_admin(self, user):
|
def is_server_admin(self, user):
|
||||||
res = yield self._simple_select_one_onecol(
|
res = yield self._simple_select_one_onecol(
|
||||||
@ -425,6 +582,8 @@ class RegistrationStore(
|
|||||||
columns=["creation_ts"],
|
columns=["creation_ts"],
|
||||||
)
|
)
|
||||||
|
|
||||||
|
self._account_validity = hs.config.account_validity
|
||||||
|
|
||||||
# we no longer use refresh tokens, but it's possible that some people
|
# we no longer use refresh tokens, but it's possible that some people
|
||||||
# might have a background update queued to build this index. Just
|
# might have a background update queued to build this index. Just
|
||||||
# clear the background update.
|
# clear the background update.
|
||||||
@ -561,9 +720,23 @@ class RegistrationStore(
|
|||||||
"user_type": user_type,
|
"user_type": user_type,
|
||||||
},
|
},
|
||||||
)
|
)
|
||||||
|
|
||||||
except self.database_engine.module.IntegrityError:
|
except self.database_engine.module.IntegrityError:
|
||||||
raise StoreError(400, "User ID already taken.", errcode=Codes.USER_IN_USE)
|
raise StoreError(400, "User ID already taken.", errcode=Codes.USER_IN_USE)
|
||||||
|
|
||||||
|
if self._account_validity.enabled:
|
||||||
|
now_ms = self.clock.time_msec()
|
||||||
|
expiration_ts = now_ms + self._account_validity.period
|
||||||
|
self._simple_insert_txn(
|
||||||
|
txn,
|
||||||
|
"account_validity",
|
||||||
|
values={
|
||||||
|
"user_id": user_id,
|
||||||
|
"expiration_ts_ms": expiration_ts,
|
||||||
|
"email_sent": False,
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
if token:
|
if token:
|
||||||
# it's possible for this to get a conflict, but only for a single user
|
# it's possible for this to get a conflict, but only for a single user
|
||||||
# since tokens are namespaced based on their user ID
|
# since tokens are namespaced based on their user ID
|
||||||
|
@ -13,19 +13,7 @@
|
|||||||
* limitations under the License.
|
* limitations under the License.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS application_services(
|
/* We used to create a tables called application_services and
|
||||||
id INTEGER PRIMARY KEY AUTOINCREMENT,
|
* application_services_regex, but these are no longer used and are removed in
|
||||||
url TEXT,
|
* delta 54.
|
||||||
token TEXT,
|
*/
|
||||||
hs_token TEXT,
|
|
||||||
sender TEXT,
|
|
||||||
UNIQUE(token)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS application_services_regex(
|
|
||||||
id INTEGER PRIMARY KEY AUTOINCREMENT,
|
|
||||||
as_id BIGINT UNSIGNED NOT NULL,
|
|
||||||
namespace INTEGER, /* enum[room_id|room_alias|user_id] */
|
|
||||||
regex TEXT,
|
|
||||||
FOREIGN KEY(as_id) REFERENCES application_services(id)
|
|
||||||
);
|
|
||||||
|
@ -1,42 +0,0 @@
|
|||||||
# Copyright 2015, 2016 OpenMarket Ltd
|
|
||||||
#
|
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
# you may not use this file except in compliance with the License.
|
|
||||||
# You may obtain a copy of the License at
|
|
||||||
#
|
|
||||||
# http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
#
|
|
||||||
# Unless required by applicable law or agreed to in writing, software
|
|
||||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
# See the License for the specific language governing permissions and
|
|
||||||
# limitations under the License.
|
|
||||||
|
|
||||||
import logging
|
|
||||||
|
|
||||||
import simplejson as json
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur, *args, **kwargs):
|
|
||||||
cur.execute("SELECT id, regex FROM application_services_regex")
|
|
||||||
for row in cur.fetchall():
|
|
||||||
try:
|
|
||||||
logger.debug("Checking %s..." % row[0])
|
|
||||||
json.loads(row[1])
|
|
||||||
except ValueError:
|
|
||||||
# row isn't in json, make it so.
|
|
||||||
string_regex = row[1]
|
|
||||||
new_regex = json.dumps({
|
|
||||||
"regex": string_regex,
|
|
||||||
"exclusive": True
|
|
||||||
})
|
|
||||||
cur.execute(
|
|
||||||
"UPDATE application_services_regex SET regex=? WHERE id=?",
|
|
||||||
(new_regex, row[0])
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def run_upgrade(*args, **kwargs):
|
|
||||||
pass
|
|
@ -17,14 +17,6 @@ DELETE FROM room_memberships WHERE rowid not in (
|
|||||||
DROP INDEX IF EXISTS room_memberships_event_id;
|
DROP INDEX IF EXISTS room_memberships_event_id;
|
||||||
CREATE UNIQUE INDEX room_memberships_event_id ON room_memberships(event_id);
|
CREATE UNIQUE INDEX room_memberships_event_id ON room_memberships(event_id);
|
||||||
|
|
||||||
--
|
|
||||||
DELETE FROM feedback WHERE rowid not in (
|
|
||||||
SELECT MIN(rowid) FROM feedback GROUP BY event_id
|
|
||||||
);
|
|
||||||
|
|
||||||
DROP INDEX IF EXISTS feedback_event_id;
|
|
||||||
CREATE UNIQUE INDEX feedback_event_id ON feedback(event_id);
|
|
||||||
|
|
||||||
--
|
--
|
||||||
DELETE FROM topics WHERE rowid not in (
|
DELETE FROM topics WHERE rowid not in (
|
||||||
SELECT MIN(rowid) FROM topics GROUP BY event_id
|
SELECT MIN(rowid) FROM topics GROUP BY event_id
|
||||||
|
@ -1,4 +1,4 @@
|
|||||||
/* Copyright 2015, 2016 OpenMarket Ltd
|
/* Copyright 2019 New Vector Ltd
|
||||||
*
|
*
|
||||||
* Licensed under the Apache License, Version 2.0 (the "License");
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
* you may not use this file except in compliance with the License.
|
* you may not use this file except in compliance with the License.
|
||||||
@ -13,10 +13,6 @@
|
|||||||
* limitations under the License.
|
* limitations under the License.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
-- Should only ever contain one row
|
/* We used to create a table called stats_reporting, but this is no longer
|
||||||
CREATE TABLE IF NOT EXISTS stats_reporting(
|
* used and is removed in delta 54.
|
||||||
-- The stream ordering token which was most recently reported as stats
|
*/
|
||||||
reported_stream_token INTEGER,
|
|
||||||
-- The time (seconds since epoch) stats were most recently reported
|
|
||||||
reported_time BIGINT
|
|
||||||
);
|
|
@ -14,15 +14,10 @@
|
|||||||
*/
|
*/
|
||||||
|
|
||||||
|
|
||||||
/**
|
/* We used to create a table called current_state_resets, but this is no
|
||||||
* The positions in the event stream_ordering when the current_state was
|
* longer used and is removed in delta 54.
|
||||||
* replaced by the state at the event.
|
|
||||||
*/
|
*/
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS current_state_resets(
|
|
||||||
event_stream_ordering BIGINT PRIMARY KEY NOT NULL
|
|
||||||
);
|
|
||||||
|
|
||||||
/* The outlier events that have aquired a state group typically through
|
/* The outlier events that have aquired a state group typically through
|
||||||
* backfill. This is tracked separately to the events table, as assigning a
|
* backfill. This is tracked separately to the events table, as assigning a
|
||||||
* state group change the position of the existing event in the stream
|
* state group change the position of the existing event in the stream
|
||||||
|
@ -24,13 +24,9 @@ DROP INDEX IF EXISTS state_groups_id; -- Duplicate of PRIMARY KEY
|
|||||||
DROP INDEX IF EXISTS event_to_state_groups_id; -- Duplicate of PRIMARY KEY
|
DROP INDEX IF EXISTS event_to_state_groups_id; -- Duplicate of PRIMARY KEY
|
||||||
DROP INDEX IF EXISTS event_push_actions_room_id_event_id_user_id_profile_tag; -- Duplicate of UNIQUE CONSTRAINT
|
DROP INDEX IF EXISTS event_push_actions_room_id_event_id_user_id_profile_tag; -- Duplicate of UNIQUE CONSTRAINT
|
||||||
|
|
||||||
DROP INDEX IF EXISTS event_destinations_id; -- Prefix of UNIQUE CONSTRAINT
|
|
||||||
DROP INDEX IF EXISTS st_extrem_id; -- Prefix of UNIQUE CONSTRAINT
|
DROP INDEX IF EXISTS st_extrem_id; -- Prefix of UNIQUE CONSTRAINT
|
||||||
DROP INDEX IF EXISTS event_content_hashes_id; -- Prefix of UNIQUE CONSTRAINT
|
|
||||||
DROP INDEX IF EXISTS event_signatures_id; -- Prefix of UNIQUE CONSTRAINT
|
DROP INDEX IF EXISTS event_signatures_id; -- Prefix of UNIQUE CONSTRAINT
|
||||||
DROP INDEX IF EXISTS event_edge_hashes_id; -- Prefix of UNIQUE CONSTRAINT
|
|
||||||
DROP INDEX IF EXISTS redactions_event_id; -- Duplicate of UNIQUE CONSTRAINT
|
DROP INDEX IF EXISTS redactions_event_id; -- Duplicate of UNIQUE CONSTRAINT
|
||||||
DROP INDEX IF EXISTS room_hosts_room_id; -- Prefix of UNIQUE CONSTRAINT
|
|
||||||
|
|
||||||
-- The following indices were unused
|
-- The following indices were unused
|
||||||
DROP INDEX IF EXISTS remote_media_cache_thumbnails_media_id;
|
DROP INDEX IF EXISTS remote_media_cache_thumbnails_media_id;
|
||||||
|
@ -1,4 +1,4 @@
|
|||||||
/* Copyright 2014-2016 OpenMarket Ltd
|
/* Copyright 2019 New Vector Ltd
|
||||||
*
|
*
|
||||||
* Licensed under the Apache License, Version 2.0 (the "License");
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
* you may not use this file except in compliance with the License.
|
* you may not use this file except in compliance with the License.
|
||||||
@ -13,12 +13,15 @@
|
|||||||
* limitations under the License.
|
* limitations under the License.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS room_aliases(
|
DROP TABLE IF EXISTS account_validity;
|
||||||
room_alias TEXT NOT NULL,
|
|
||||||
room_id TEXT NOT NULL
|
-- Track what users are in public rooms.
|
||||||
|
CREATE TABLE IF NOT EXISTS account_validity (
|
||||||
|
user_id TEXT PRIMARY KEY,
|
||||||
|
expiration_ts_ms BIGINT NOT NULL,
|
||||||
|
email_sent BOOLEAN NOT NULL,
|
||||||
|
renewal_token TEXT
|
||||||
);
|
);
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS room_alias_servers(
|
CREATE INDEX account_validity_email_sent_idx ON account_validity(email_sent, expiration_ts_ms)
|
||||||
room_alias TEXT NOT NULL,
|
CREATE UNIQUE INDEX account_validity_renewal_string_idx ON account_validity(renewal_token)
|
||||||
server TEXT NOT NULL
|
|
||||||
);
|
|
30
synapse/storage/schema/delta/54/drop_legacy_tables.sql
Normal file
30
synapse/storage/schema/delta/54/drop_legacy_tables.sql
Normal file
@ -0,0 +1,30 @@
|
|||||||
|
/* Copyright 2019 New Vector Ltd
|
||||||
|
*
|
||||||
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
* you may not use this file except in compliance with the License.
|
||||||
|
* You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
-- we need to do this first due to foreign constraints
|
||||||
|
DROP TABLE IF EXISTS application_services_regex;
|
||||||
|
|
||||||
|
DROP TABLE IF EXISTS application_services;
|
||||||
|
DROP TABLE IF EXISTS transaction_id_to_pdu;
|
||||||
|
DROP TABLE IF EXISTS stats_reporting;
|
||||||
|
DROP TABLE IF EXISTS current_state_resets;
|
||||||
|
DROP TABLE IF EXISTS event_content_hashes;
|
||||||
|
DROP TABLE IF EXISTS event_destinations;
|
||||||
|
DROP TABLE IF EXISTS event_edge_hashes;
|
||||||
|
DROP TABLE IF EXISTS event_signatures;
|
||||||
|
DROP TABLE IF EXISTS feedback;
|
||||||
|
DROP TABLE IF EXISTS room_hosts;
|
||||||
|
DROP TABLE IF EXISTS server_tls_certificates;
|
||||||
|
DROP TABLE IF EXISTS state_forward_extremities;
|
@ -1,91 +0,0 @@
|
|||||||
/* Copyright 2014-2016 OpenMarket Ltd
|
|
||||||
*
|
|
||||||
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
* you may not use this file except in compliance with the License.
|
|
||||||
* You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing, software
|
|
||||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
* See the License for the specific language governing permissions and
|
|
||||||
* limitations under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS event_forward_extremities(
|
|
||||||
event_id TEXT NOT NULL,
|
|
||||||
room_id TEXT NOT NULL,
|
|
||||||
UNIQUE (event_id, room_id)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX ev_extrem_room ON event_forward_extremities(room_id);
|
|
||||||
CREATE INDEX ev_extrem_id ON event_forward_extremities(event_id);
|
|
||||||
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS event_backward_extremities(
|
|
||||||
event_id TEXT NOT NULL,
|
|
||||||
room_id TEXT NOT NULL,
|
|
||||||
UNIQUE (event_id, room_id)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX ev_b_extrem_room ON event_backward_extremities(room_id);
|
|
||||||
CREATE INDEX ev_b_extrem_id ON event_backward_extremities(event_id);
|
|
||||||
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS event_edges(
|
|
||||||
event_id TEXT NOT NULL,
|
|
||||||
prev_event_id TEXT NOT NULL,
|
|
||||||
room_id TEXT NOT NULL,
|
|
||||||
-- We no longer insert prev_state into this table, so all new rows will have
|
|
||||||
-- is_state as false.
|
|
||||||
is_state BOOL NOT NULL,
|
|
||||||
UNIQUE (event_id, prev_event_id, room_id, is_state)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX ev_edges_id ON event_edges(event_id);
|
|
||||||
CREATE INDEX ev_edges_prev_id ON event_edges(prev_event_id);
|
|
||||||
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS room_depth(
|
|
||||||
room_id TEXT NOT NULL,
|
|
||||||
min_depth INTEGER NOT NULL,
|
|
||||||
UNIQUE (room_id)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX room_depth_room ON room_depth(room_id);
|
|
||||||
|
|
||||||
|
|
||||||
create TABLE IF NOT EXISTS event_destinations(
|
|
||||||
event_id TEXT NOT NULL,
|
|
||||||
destination TEXT NOT NULL,
|
|
||||||
delivered_ts BIGINT DEFAULT 0, -- or 0 if not delivered
|
|
||||||
UNIQUE (event_id, destination)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX event_destinations_id ON event_destinations(event_id);
|
|
||||||
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS state_forward_extremities(
|
|
||||||
event_id TEXT NOT NULL,
|
|
||||||
room_id TEXT NOT NULL,
|
|
||||||
type TEXT NOT NULL,
|
|
||||||
state_key TEXT NOT NULL,
|
|
||||||
UNIQUE (event_id, room_id)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX st_extrem_keys ON state_forward_extremities(
|
|
||||||
room_id, type, state_key
|
|
||||||
);
|
|
||||||
CREATE INDEX st_extrem_id ON state_forward_extremities(event_id);
|
|
||||||
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS event_auth(
|
|
||||||
event_id TEXT NOT NULL,
|
|
||||||
auth_id TEXT NOT NULL,
|
|
||||||
room_id TEXT NOT NULL,
|
|
||||||
UNIQUE (event_id, auth_id, room_id)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX evauth_edges_id ON event_auth(event_id);
|
|
||||||
CREATE INDEX evauth_edges_auth_id ON event_auth(auth_id);
|
|
@ -1,55 +0,0 @@
|
|||||||
/* Copyright 2014-2016 OpenMarket Ltd
|
|
||||||
*
|
|
||||||
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
* you may not use this file except in compliance with the License.
|
|
||||||
* You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing, software
|
|
||||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
* See the License for the specific language governing permissions and
|
|
||||||
* limitations under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS event_content_hashes (
|
|
||||||
event_id TEXT,
|
|
||||||
algorithm TEXT,
|
|
||||||
hash bytea,
|
|
||||||
UNIQUE (event_id, algorithm)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX event_content_hashes_id ON event_content_hashes(event_id);
|
|
||||||
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS event_reference_hashes (
|
|
||||||
event_id TEXT,
|
|
||||||
algorithm TEXT,
|
|
||||||
hash bytea,
|
|
||||||
UNIQUE (event_id, algorithm)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX event_reference_hashes_id ON event_reference_hashes(event_id);
|
|
||||||
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS event_signatures (
|
|
||||||
event_id TEXT,
|
|
||||||
signature_name TEXT,
|
|
||||||
key_id TEXT,
|
|
||||||
signature bytea,
|
|
||||||
UNIQUE (event_id, signature_name, key_id)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX event_signatures_id ON event_signatures(event_id);
|
|
||||||
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS event_edge_hashes(
|
|
||||||
event_id TEXT,
|
|
||||||
prev_event_id TEXT,
|
|
||||||
algorithm TEXT,
|
|
||||||
hash bytea,
|
|
||||||
UNIQUE (event_id, prev_event_id, algorithm)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX event_edge_hashes_id ON event_edge_hashes(event_id);
|
|
@ -1,123 +0,0 @@
|
|||||||
/* Copyright 2014-2016 OpenMarket Ltd
|
|
||||||
*
|
|
||||||
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
* you may not use this file except in compliance with the License.
|
|
||||||
* You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing, software
|
|
||||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
* See the License for the specific language governing permissions and
|
|
||||||
* limitations under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS events(
|
|
||||||
stream_ordering INTEGER PRIMARY KEY AUTOINCREMENT,
|
|
||||||
topological_ordering BIGINT NOT NULL,
|
|
||||||
event_id TEXT NOT NULL,
|
|
||||||
type TEXT NOT NULL,
|
|
||||||
room_id TEXT NOT NULL,
|
|
||||||
content TEXT NOT NULL,
|
|
||||||
unrecognized_keys TEXT,
|
|
||||||
processed BOOL NOT NULL,
|
|
||||||
outlier BOOL NOT NULL,
|
|
||||||
depth BIGINT DEFAULT 0 NOT NULL,
|
|
||||||
UNIQUE (event_id)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX events_stream_ordering ON events (stream_ordering);
|
|
||||||
CREATE INDEX events_topological_ordering ON events (topological_ordering);
|
|
||||||
CREATE INDEX events_room_id ON events (room_id);
|
|
||||||
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS event_json(
|
|
||||||
event_id TEXT NOT NULL,
|
|
||||||
room_id TEXT NOT NULL,
|
|
||||||
internal_metadata TEXT NOT NULL,
|
|
||||||
json TEXT NOT NULL,
|
|
||||||
UNIQUE (event_id)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX event_json_room_id ON event_json(room_id);
|
|
||||||
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS state_events(
|
|
||||||
event_id TEXT NOT NULL,
|
|
||||||
room_id TEXT NOT NULL,
|
|
||||||
type TEXT NOT NULL,
|
|
||||||
state_key TEXT NOT NULL,
|
|
||||||
prev_state TEXT,
|
|
||||||
UNIQUE (event_id)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX state_events_room_id ON state_events (room_id);
|
|
||||||
CREATE INDEX state_events_type ON state_events (type);
|
|
||||||
CREATE INDEX state_events_state_key ON state_events (state_key);
|
|
||||||
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS current_state_events(
|
|
||||||
event_id TEXT NOT NULL,
|
|
||||||
room_id TEXT NOT NULL,
|
|
||||||
type TEXT NOT NULL,
|
|
||||||
state_key TEXT NOT NULL,
|
|
||||||
UNIQUE (room_id, type, state_key)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX curr_events_event_id ON current_state_events (event_id);
|
|
||||||
CREATE INDEX current_state_events_room_id ON current_state_events (room_id);
|
|
||||||
CREATE INDEX current_state_events_type ON current_state_events (type);
|
|
||||||
CREATE INDEX current_state_events_state_key ON current_state_events (state_key);
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS room_memberships(
|
|
||||||
event_id TEXT NOT NULL,
|
|
||||||
user_id TEXT NOT NULL,
|
|
||||||
sender TEXT NOT NULL,
|
|
||||||
room_id TEXT NOT NULL,
|
|
||||||
membership TEXT NOT NULL
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX room_memberships_event_id ON room_memberships (event_id);
|
|
||||||
CREATE INDEX room_memberships_room_id ON room_memberships (room_id);
|
|
||||||
CREATE INDEX room_memberships_user_id ON room_memberships (user_id);
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS feedback(
|
|
||||||
event_id TEXT NOT NULL,
|
|
||||||
feedback_type TEXT,
|
|
||||||
target_event_id TEXT,
|
|
||||||
sender TEXT,
|
|
||||||
room_id TEXT
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS topics(
|
|
||||||
event_id TEXT NOT NULL,
|
|
||||||
room_id TEXT NOT NULL,
|
|
||||||
topic TEXT NOT NULL
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX topics_event_id ON topics(event_id);
|
|
||||||
CREATE INDEX topics_room_id ON topics(room_id);
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS room_names(
|
|
||||||
event_id TEXT NOT NULL,
|
|
||||||
room_id TEXT NOT NULL,
|
|
||||||
name TEXT NOT NULL
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX room_names_event_id ON room_names(event_id);
|
|
||||||
CREATE INDEX room_names_room_id ON room_names(room_id);
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS rooms(
|
|
||||||
room_id TEXT PRIMARY KEY NOT NULL,
|
|
||||||
is_public BOOL,
|
|
||||||
creator TEXT
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS room_hosts(
|
|
||||||
room_id TEXT NOT NULL,
|
|
||||||
host TEXT NOT NULL,
|
|
||||||
UNIQUE (room_id, host)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX room_hosts_room_id ON room_hosts (room_id);
|
|
@ -1,31 +0,0 @@
|
|||||||
/* Copyright 2014-2016 OpenMarket Ltd
|
|
||||||
*
|
|
||||||
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
* you may not use this file except in compliance with the License.
|
|
||||||
* You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing, software
|
|
||||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
* See the License for the specific language governing permissions and
|
|
||||||
* limitations under the License.
|
|
||||||
*/
|
|
||||||
CREATE TABLE IF NOT EXISTS server_tls_certificates(
|
|
||||||
server_name TEXT, -- Server name.
|
|
||||||
fingerprint TEXT, -- Certificate fingerprint.
|
|
||||||
from_server TEXT, -- Which key server the certificate was fetched from.
|
|
||||||
ts_added_ms BIGINT, -- When the certifcate was added.
|
|
||||||
tls_certificate bytea, -- DER encoded x509 certificate.
|
|
||||||
UNIQUE (server_name, fingerprint)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS server_signature_keys(
|
|
||||||
server_name TEXT, -- Server name.
|
|
||||||
key_id TEXT, -- Key version.
|
|
||||||
from_server TEXT, -- Which key server the key was fetched form.
|
|
||||||
ts_added_ms BIGINT, -- When the key was added.
|
|
||||||
verify_key bytea, -- NACL verification key.
|
|
||||||
UNIQUE (server_name, key_id)
|
|
||||||
);
|
|
@ -1,65 +0,0 @@
|
|||||||
/* Copyright 2014-2016 OpenMarket Ltd
|
|
||||||
*
|
|
||||||
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
* you may not use this file except in compliance with the License.
|
|
||||||
* You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing, software
|
|
||||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
* See the License for the specific language governing permissions and
|
|
||||||
* limitations under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS local_media_repository (
|
|
||||||
media_id TEXT, -- The id used to refer to the media.
|
|
||||||
media_type TEXT, -- The MIME-type of the media.
|
|
||||||
media_length INTEGER, -- Length of the media in bytes.
|
|
||||||
created_ts BIGINT, -- When the content was uploaded in ms.
|
|
||||||
upload_name TEXT, -- The name the media was uploaded with.
|
|
||||||
user_id TEXT, -- The user who uploaded the file.
|
|
||||||
UNIQUE (media_id)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS local_media_repository_thumbnails (
|
|
||||||
media_id TEXT, -- The id used to refer to the media.
|
|
||||||
thumbnail_width INTEGER, -- The width of the thumbnail in pixels.
|
|
||||||
thumbnail_height INTEGER, -- The height of the thumbnail in pixels.
|
|
||||||
thumbnail_type TEXT, -- The MIME-type of the thumbnail.
|
|
||||||
thumbnail_method TEXT, -- The method used to make the thumbnail.
|
|
||||||
thumbnail_length INTEGER, -- The length of the thumbnail in bytes.
|
|
||||||
UNIQUE (
|
|
||||||
media_id, thumbnail_width, thumbnail_height, thumbnail_type
|
|
||||||
)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX local_media_repository_thumbnails_media_id
|
|
||||||
ON local_media_repository_thumbnails (media_id);
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS remote_media_cache (
|
|
||||||
media_origin TEXT, -- The remote HS the media came from.
|
|
||||||
media_id TEXT, -- The id used to refer to the media on that server.
|
|
||||||
media_type TEXT, -- The MIME-type of the media.
|
|
||||||
created_ts BIGINT, -- When the content was uploaded in ms.
|
|
||||||
upload_name TEXT, -- The name the media was uploaded with.
|
|
||||||
media_length INTEGER, -- Length of the media in bytes.
|
|
||||||
filesystem_id TEXT, -- The name used to store the media on disk.
|
|
||||||
UNIQUE (media_origin, media_id)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS remote_media_cache_thumbnails (
|
|
||||||
media_origin TEXT, -- The remote HS the media came from.
|
|
||||||
media_id TEXT, -- The id used to refer to the media.
|
|
||||||
thumbnail_width INTEGER, -- The width of the thumbnail in pixels.
|
|
||||||
thumbnail_height INTEGER, -- The height of the thumbnail in pixels.
|
|
||||||
thumbnail_method TEXT, -- The method used to make the thumbnail
|
|
||||||
thumbnail_type TEXT, -- The MIME-type of the thumbnail.
|
|
||||||
thumbnail_length INTEGER, -- The length of the thumbnail in bytes.
|
|
||||||
filesystem_id TEXT, -- The name used to store the media on disk.
|
|
||||||
UNIQUE (
|
|
||||||
media_origin, media_id, thumbnail_width, thumbnail_height,
|
|
||||||
thumbnail_type
|
|
||||||
)
|
|
||||||
);
|
|
@ -1,35 +0,0 @@
|
|||||||
/* Copyright 2014-2016 OpenMarket Ltd
|
|
||||||
*
|
|
||||||
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
* you may not use this file except in compliance with the License.
|
|
||||||
* You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing, software
|
|
||||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
* See the License for the specific language governing permissions and
|
|
||||||
* limitations under the License.
|
|
||||||
*/
|
|
||||||
CREATE TABLE IF NOT EXISTS presence(
|
|
||||||
user_id TEXT NOT NULL,
|
|
||||||
state VARCHAR(20),
|
|
||||||
status_msg TEXT,
|
|
||||||
mtime BIGINT -- miliseconds since last state change
|
|
||||||
);
|
|
||||||
|
|
||||||
-- For each of /my/ users which possibly-remote users are allowed to see their
|
|
||||||
-- presence state
|
|
||||||
CREATE TABLE IF NOT EXISTS presence_allow_inbound(
|
|
||||||
observed_user_id TEXT NOT NULL,
|
|
||||||
observer_user_id TEXT NOT NULL -- a UserID,
|
|
||||||
);
|
|
||||||
|
|
||||||
-- For each of /my/ users (watcher), which possibly-remote users are they
|
|
||||||
-- watching?
|
|
||||||
CREATE TABLE IF NOT EXISTS presence_list(
|
|
||||||
user_id TEXT NOT NULL,
|
|
||||||
observed_user_id TEXT NOT NULL, -- a UserID,
|
|
||||||
accepted BOOLEAN NOT NULL
|
|
||||||
);
|
|
@ -1,19 +0,0 @@
|
|||||||
/* Copyright 2014-2016 OpenMarket Ltd
|
|
||||||
*
|
|
||||||
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
* you may not use this file except in compliance with the License.
|
|
||||||
* You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing, software
|
|
||||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
* See the License for the specific language governing permissions and
|
|
||||||
* limitations under the License.
|
|
||||||
*/
|
|
||||||
CREATE TABLE IF NOT EXISTS profiles(
|
|
||||||
user_id TEXT NOT NULL,
|
|
||||||
displayname TEXT,
|
|
||||||
avatar_url TEXT
|
|
||||||
);
|
|
@ -1,22 +0,0 @@
|
|||||||
/* Copyright 2014-2016 OpenMarket Ltd
|
|
||||||
*
|
|
||||||
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
* you may not use this file except in compliance with the License.
|
|
||||||
* You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing, software
|
|
||||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
* See the License for the specific language governing permissions and
|
|
||||||
* limitations under the License.
|
|
||||||
*/
|
|
||||||
CREATE TABLE IF NOT EXISTS redactions (
|
|
||||||
event_id TEXT NOT NULL,
|
|
||||||
redacts TEXT NOT NULL,
|
|
||||||
UNIQUE (event_id)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX redactions_event_id ON redactions (event_id);
|
|
||||||
CREATE INDEX redactions_redacts ON redactions (redacts);
|
|
@ -1,40 +0,0 @@
|
|||||||
/* Copyright 2014-2016 OpenMarket Ltd
|
|
||||||
*
|
|
||||||
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
* you may not use this file except in compliance with the License.
|
|
||||||
* You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing, software
|
|
||||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
* See the License for the specific language governing permissions and
|
|
||||||
* limitations under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS state_groups(
|
|
||||||
id INTEGER PRIMARY KEY,
|
|
||||||
room_id TEXT NOT NULL,
|
|
||||||
event_id TEXT NOT NULL
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS state_groups_state(
|
|
||||||
state_group INTEGER NOT NULL,
|
|
||||||
room_id TEXT NOT NULL,
|
|
||||||
type TEXT NOT NULL,
|
|
||||||
state_key TEXT NOT NULL,
|
|
||||||
event_id TEXT NOT NULL
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS event_to_state_groups(
|
|
||||||
event_id TEXT NOT NULL,
|
|
||||||
state_group INTEGER NOT NULL,
|
|
||||||
UNIQUE (event_id)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX state_groups_id ON state_groups(id);
|
|
||||||
|
|
||||||
CREATE INDEX state_groups_state_id ON state_groups_state(state_group);
|
|
||||||
CREATE INDEX state_groups_state_tuple ON state_groups_state(room_id, type, state_key);
|
|
||||||
CREATE INDEX event_to_state_groups_id ON event_to_state_groups(event_id);
|
|
@ -1,44 +0,0 @@
|
|||||||
/* Copyright 2014-2016 OpenMarket Ltd
|
|
||||||
*
|
|
||||||
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
* you may not use this file except in compliance with the License.
|
|
||||||
* You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing, software
|
|
||||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
* See the License for the specific language governing permissions and
|
|
||||||
* limitations under the License.
|
|
||||||
*/
|
|
||||||
-- Stores what transaction ids we have received and what our response was
|
|
||||||
CREATE TABLE IF NOT EXISTS received_transactions(
|
|
||||||
transaction_id TEXT,
|
|
||||||
origin TEXT,
|
|
||||||
ts BIGINT,
|
|
||||||
response_code INTEGER,
|
|
||||||
response_json bytea,
|
|
||||||
has_been_referenced SMALLINT DEFAULT 0, -- Whether thishas been referenced by a prev_tx
|
|
||||||
UNIQUE (transaction_id, origin)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX transactions_have_ref ON received_transactions(origin, has_been_referenced);-- WHERE has_been_referenced = 0;
|
|
||||||
|
|
||||||
-- For sent transactions only.
|
|
||||||
CREATE TABLE IF NOT EXISTS transaction_id_to_pdu(
|
|
||||||
transaction_id INTEGER,
|
|
||||||
destination TEXT,
|
|
||||||
pdu_id TEXT,
|
|
||||||
pdu_origin TEXT
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX transaction_id_to_pdu_tx ON transaction_id_to_pdu(transaction_id, destination);
|
|
||||||
CREATE INDEX transaction_id_to_pdu_dest ON transaction_id_to_pdu(destination);
|
|
||||||
|
|
||||||
-- To track destination health
|
|
||||||
CREATE TABLE IF NOT EXISTS destinations(
|
|
||||||
destination TEXT PRIMARY KEY,
|
|
||||||
retry_last_ts BIGINT,
|
|
||||||
retry_interval INTEGER
|
|
||||||
);
|
|
@ -1,43 +0,0 @@
|
|||||||
/* Copyright 2014-2016 OpenMarket Ltd
|
|
||||||
*
|
|
||||||
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
* you may not use this file except in compliance with the License.
|
|
||||||
* You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing, software
|
|
||||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
* See the License for the specific language governing permissions and
|
|
||||||
* limitations under the License.
|
|
||||||
*/
|
|
||||||
CREATE TABLE IF NOT EXISTS users(
|
|
||||||
id INTEGER PRIMARY KEY AUTOINCREMENT,
|
|
||||||
name TEXT,
|
|
||||||
password_hash TEXT,
|
|
||||||
creation_ts BIGINT,
|
|
||||||
admin SMALLINT DEFAULT 0 NOT NULL,
|
|
||||||
UNIQUE(name)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS access_tokens(
|
|
||||||
id INTEGER PRIMARY KEY AUTOINCREMENT,
|
|
||||||
user_id TEXT NOT NULL,
|
|
||||||
device_id TEXT,
|
|
||||||
token TEXT NOT NULL,
|
|
||||||
last_used BIGINT,
|
|
||||||
UNIQUE(token)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS user_ips (
|
|
||||||
user TEXT NOT NULL,
|
|
||||||
access_token TEXT NOT NULL,
|
|
||||||
device_id TEXT,
|
|
||||||
ip TEXT NOT NULL,
|
|
||||||
user_agent TEXT NOT NULL,
|
|
||||||
last_seen BIGINT NOT NULL,
|
|
||||||
UNIQUE (user, access_token, ip, user_agent)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX user_ips_user ON user_ips(user);
|
|
@ -13,22 +13,11 @@
|
|||||||
* limitations under the License.
|
* limitations under the License.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS application_services(
|
/* We used to create tables called application_services and
|
||||||
id BIGINT PRIMARY KEY,
|
* application_services_regex, but these are no longer used and are removed in
|
||||||
url TEXT,
|
* delta 54.
|
||||||
token TEXT,
|
*/
|
||||||
hs_token TEXT,
|
|
||||||
sender TEXT,
|
|
||||||
UNIQUE(token)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS application_services_regex(
|
|
||||||
id BIGINT PRIMARY KEY,
|
|
||||||
as_id BIGINT NOT NULL,
|
|
||||||
namespace INTEGER, /* enum[room_id|room_alias|user_id] */
|
|
||||||
regex TEXT,
|
|
||||||
FOREIGN KEY(as_id) REFERENCES application_services(id)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS application_services_state(
|
CREATE TABLE IF NOT EXISTS application_services_state(
|
||||||
as_id TEXT PRIMARY KEY,
|
as_id TEXT PRIMARY KEY,
|
||||||
|
@ -13,6 +13,11 @@
|
|||||||
* limitations under the License.
|
* limitations under the License.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
|
/* We used to create tables called event_destinations and
|
||||||
|
* state_forward_extremities, but these are no longer used and are removed in
|
||||||
|
* delta 54.
|
||||||
|
*/
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS event_forward_extremities(
|
CREATE TABLE IF NOT EXISTS event_forward_extremities(
|
||||||
event_id TEXT NOT NULL,
|
event_id TEXT NOT NULL,
|
||||||
room_id TEXT NOT NULL,
|
room_id TEXT NOT NULL,
|
||||||
@ -54,31 +59,6 @@ CREATE TABLE IF NOT EXISTS room_depth(
|
|||||||
|
|
||||||
CREATE INDEX room_depth_room ON room_depth(room_id);
|
CREATE INDEX room_depth_room ON room_depth(room_id);
|
||||||
|
|
||||||
|
|
||||||
create TABLE IF NOT EXISTS event_destinations(
|
|
||||||
event_id TEXT NOT NULL,
|
|
||||||
destination TEXT NOT NULL,
|
|
||||||
delivered_ts BIGINT DEFAULT 0, -- or 0 if not delivered
|
|
||||||
UNIQUE (event_id, destination)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX event_destinations_id ON event_destinations(event_id);
|
|
||||||
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS state_forward_extremities(
|
|
||||||
event_id TEXT NOT NULL,
|
|
||||||
room_id TEXT NOT NULL,
|
|
||||||
type TEXT NOT NULL,
|
|
||||||
state_key TEXT NOT NULL,
|
|
||||||
UNIQUE (event_id, room_id)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX st_extrem_keys ON state_forward_extremities(
|
|
||||||
room_id, type, state_key
|
|
||||||
);
|
|
||||||
CREATE INDEX st_extrem_id ON state_forward_extremities(event_id);
|
|
||||||
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS event_auth(
|
CREATE TABLE IF NOT EXISTS event_auth(
|
||||||
event_id TEXT NOT NULL,
|
event_id TEXT NOT NULL,
|
||||||
auth_id TEXT NOT NULL,
|
auth_id TEXT NOT NULL,
|
||||||
|
@ -13,15 +13,9 @@
|
|||||||
* limitations under the License.
|
* limitations under the License.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS event_content_hashes (
|
/* We used to create tables called event_content_hashes and event_edge_hashes,
|
||||||
event_id TEXT,
|
* but these are no longer used and are removed in delta 54.
|
||||||
algorithm TEXT,
|
*/
|
||||||
hash bytea,
|
|
||||||
UNIQUE (event_id, algorithm)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX event_content_hashes_id ON event_content_hashes(event_id);
|
|
||||||
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS event_reference_hashes (
|
CREATE TABLE IF NOT EXISTS event_reference_hashes (
|
||||||
event_id TEXT,
|
event_id TEXT,
|
||||||
@ -42,14 +36,3 @@ CREATE TABLE IF NOT EXISTS event_signatures (
|
|||||||
);
|
);
|
||||||
|
|
||||||
CREATE INDEX event_signatures_id ON event_signatures(event_id);
|
CREATE INDEX event_signatures_id ON event_signatures(event_id);
|
||||||
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS event_edge_hashes(
|
|
||||||
event_id TEXT,
|
|
||||||
prev_event_id TEXT,
|
|
||||||
algorithm TEXT,
|
|
||||||
hash bytea,
|
|
||||||
UNIQUE (event_id, prev_event_id, algorithm)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX event_edge_hashes_id ON event_edge_hashes(event_id);
|
|
||||||
|
@ -13,6 +13,10 @@
|
|||||||
* limitations under the License.
|
* limitations under the License.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
|
/* We used to create tables called room_hosts and feedback,
|
||||||
|
* but these are no longer used and are removed in delta 54.
|
||||||
|
*/
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS events(
|
CREATE TABLE IF NOT EXISTS events(
|
||||||
stream_ordering INTEGER PRIMARY KEY,
|
stream_ordering INTEGER PRIMARY KEY,
|
||||||
topological_ordering BIGINT NOT NULL,
|
topological_ordering BIGINT NOT NULL,
|
||||||
@ -91,15 +95,6 @@ CREATE TABLE IF NOT EXISTS room_memberships(
|
|||||||
CREATE INDEX room_memberships_room_id ON room_memberships (room_id);
|
CREATE INDEX room_memberships_room_id ON room_memberships (room_id);
|
||||||
CREATE INDEX room_memberships_user_id ON room_memberships (user_id);
|
CREATE INDEX room_memberships_user_id ON room_memberships (user_id);
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS feedback(
|
|
||||||
event_id TEXT NOT NULL,
|
|
||||||
feedback_type TEXT,
|
|
||||||
target_event_id TEXT,
|
|
||||||
sender TEXT,
|
|
||||||
room_id TEXT,
|
|
||||||
UNIQUE (event_id)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS topics(
|
CREATE TABLE IF NOT EXISTS topics(
|
||||||
event_id TEXT NOT NULL,
|
event_id TEXT NOT NULL,
|
||||||
room_id TEXT NOT NULL,
|
room_id TEXT NOT NULL,
|
||||||
@ -123,11 +118,3 @@ CREATE TABLE IF NOT EXISTS rooms(
|
|||||||
is_public BOOL,
|
is_public BOOL,
|
||||||
creator TEXT
|
creator TEXT
|
||||||
);
|
);
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS room_hosts(
|
|
||||||
room_id TEXT NOT NULL,
|
|
||||||
host TEXT NOT NULL,
|
|
||||||
UNIQUE (room_id, host)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE INDEX room_hosts_room_id ON room_hosts (room_id);
|
|
||||||
|
@ -12,14 +12,9 @@
|
|||||||
* See the License for the specific language governing permissions and
|
* See the License for the specific language governing permissions and
|
||||||
* limitations under the License.
|
* limitations under the License.
|
||||||
*/
|
*/
|
||||||
CREATE TABLE IF NOT EXISTS server_tls_certificates(
|
|
||||||
server_name TEXT, -- Server name.
|
-- we used to create a table called server_tls_certificates, but this is no
|
||||||
fingerprint TEXT, -- Certificate fingerprint.
|
-- longer used, and is removed in delta 54.
|
||||||
from_server TEXT, -- Which key server the certificate was fetched from.
|
|
||||||
ts_added_ms BIGINT, -- When the certifcate was added.
|
|
||||||
tls_certificate bytea, -- DER encoded x509 certificate.
|
|
||||||
UNIQUE (server_name, fingerprint)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS server_signature_keys(
|
CREATE TABLE IF NOT EXISTS server_signature_keys(
|
||||||
server_name TEXT, -- Server name.
|
server_name TEXT, -- Server name.
|
||||||
|
@ -28,5 +28,5 @@ CREATE TABLE IF NOT EXISTS presence_allow_inbound(
|
|||||||
UNIQUE (observed_user_id, observer_user_id)
|
UNIQUE (observed_user_id, observer_user_id)
|
||||||
);
|
);
|
||||||
|
|
||||||
-- We used to create a table called presence_list, but this is no longer used
|
-- We used to create a table called presence_list, but this is no longer used
|
||||||
-- and is removed in delta 54.
|
-- and is removed in delta 54.
|
@ -22,6 +22,24 @@ logger = logging.getLogger(__name__)
|
|||||||
|
|
||||||
class StateDeltasStore(SQLBaseStore):
|
class StateDeltasStore(SQLBaseStore):
|
||||||
def get_current_state_deltas(self, prev_stream_id):
|
def get_current_state_deltas(self, prev_stream_id):
|
||||||
|
"""Fetch a list of room state changes since the given stream id
|
||||||
|
|
||||||
|
Each entry in the result contains the following fields:
|
||||||
|
- stream_id (int)
|
||||||
|
- room_id (str)
|
||||||
|
- type (str): event type
|
||||||
|
- state_key (str):
|
||||||
|
- event_id (str|None): new event_id for this state key. None if the
|
||||||
|
state has been deleted.
|
||||||
|
- prev_event_id (str|None): previous event_id for this state key. None
|
||||||
|
if it's new state.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
prev_stream_id (int): point to get changes since (exclusive)
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Deferred[list[dict]]: results
|
||||||
|
"""
|
||||||
prev_stream_id = int(prev_stream_id)
|
prev_stream_id = int(prev_stream_id)
|
||||||
if not self._curr_state_delta_stream_cache.has_any_entity_changed(
|
if not self._curr_state_delta_stream_cache.has_any_entity_changed(
|
||||||
prev_stream_id
|
prev_stream_id
|
||||||
|
@ -1,6 +1,6 @@
|
|||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
# Copyright 2014-2016 OpenMarket Ltd
|
# Copyright 2014-2016 OpenMarket Ltd
|
||||||
# Copyright 2018 New Vector Ltd.
|
# Copyright 2018 New Vector Ltd
|
||||||
#
|
#
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
# you may not use this file except in compliance with the License.
|
# you may not use this file except in compliance with the License.
|
||||||
|
@ -1,4 +1,5 @@
|
|||||||
# Copyright 2016 OpenMarket Ltd
|
# Copyright 2016 OpenMarket Ltd
|
||||||
|
# Copyright 2019 New Vector Ltd.
|
||||||
#
|
#
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
# you may not use this file except in compliance with the License.
|
# you may not use this file except in compliance with the License.
|
||||||
@ -11,10 +12,12 @@
|
|||||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
import sys
|
||||||
|
import traceback
|
||||||
|
|
||||||
from twisted.conch import manhole_ssh
|
from twisted.conch import manhole_ssh
|
||||||
from twisted.conch.insults import insults
|
from twisted.conch.insults import insults
|
||||||
from twisted.conch.manhole import ColoredManhole
|
from twisted.conch.manhole import ColoredManhole, ManholeInterpreter
|
||||||
from twisted.conch.ssh.keys import Key
|
from twisted.conch.ssh.keys import Key
|
||||||
from twisted.cred import checkers, portal
|
from twisted.cred import checkers, portal
|
||||||
|
|
||||||
@ -79,7 +82,7 @@ def manhole(username, password, globals):
|
|||||||
|
|
||||||
rlm = manhole_ssh.TerminalRealm()
|
rlm = manhole_ssh.TerminalRealm()
|
||||||
rlm.chainedProtocolFactory = lambda: insults.ServerProtocol(
|
rlm.chainedProtocolFactory = lambda: insults.ServerProtocol(
|
||||||
ColoredManhole,
|
SynapseManhole,
|
||||||
dict(globals, __name__="__console__")
|
dict(globals, __name__="__console__")
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -88,3 +91,55 @@ def manhole(username, password, globals):
|
|||||||
factory.privateKeys[b'ssh-rsa'] = Key.fromString(PRIVATE_KEY)
|
factory.privateKeys[b'ssh-rsa'] = Key.fromString(PRIVATE_KEY)
|
||||||
|
|
||||||
return factory
|
return factory
|
||||||
|
|
||||||
|
|
||||||
|
class SynapseManhole(ColoredManhole):
|
||||||
|
"""Overrides connectionMade to create our own ManholeInterpreter"""
|
||||||
|
def connectionMade(self):
|
||||||
|
super(SynapseManhole, self).connectionMade()
|
||||||
|
|
||||||
|
# replace the manhole interpreter with our own impl
|
||||||
|
self.interpreter = SynapseManholeInterpreter(self, self.namespace)
|
||||||
|
|
||||||
|
# this would also be a good place to add more keyHandlers.
|
||||||
|
|
||||||
|
|
||||||
|
class SynapseManholeInterpreter(ManholeInterpreter):
|
||||||
|
def showsyntaxerror(self, filename=None):
|
||||||
|
"""Display the syntax error that just occurred.
|
||||||
|
|
||||||
|
Overrides the base implementation, ignoring sys.excepthook. We always want
|
||||||
|
any syntax errors to be sent to the terminal, rather than sentry.
|
||||||
|
"""
|
||||||
|
type, value, tb = sys.exc_info()
|
||||||
|
sys.last_type = type
|
||||||
|
sys.last_value = value
|
||||||
|
sys.last_traceback = tb
|
||||||
|
if filename and type is SyntaxError:
|
||||||
|
# Work hard to stuff the correct filename in the exception
|
||||||
|
try:
|
||||||
|
msg, (dummy_filename, lineno, offset, line) = value.args
|
||||||
|
except ValueError:
|
||||||
|
# Not the format we expect; leave it alone
|
||||||
|
pass
|
||||||
|
else:
|
||||||
|
# Stuff in the right filename
|
||||||
|
value = SyntaxError(msg, (filename, lineno, offset, line))
|
||||||
|
sys.last_value = value
|
||||||
|
lines = traceback.format_exception_only(type, value)
|
||||||
|
self.write(''.join(lines))
|
||||||
|
|
||||||
|
def showtraceback(self):
|
||||||
|
"""Display the exception that just occurred.
|
||||||
|
|
||||||
|
Overrides the base implementation, ignoring sys.excepthook. We always want
|
||||||
|
any syntax errors to be sent to the terminal, rather than sentry.
|
||||||
|
"""
|
||||||
|
sys.last_type, sys.last_value, last_tb = ei = sys.exc_info()
|
||||||
|
sys.last_traceback = last_tb
|
||||||
|
try:
|
||||||
|
# We remove the first stack item because it is our own code.
|
||||||
|
lines = traceback.format_exception(ei[0], ei[1], last_tb.tb_next)
|
||||||
|
self.write(''.join(lines))
|
||||||
|
finally:
|
||||||
|
last_tb = ei = None
|
||||||
|
@ -1,5 +1,5 @@
|
|||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
# Copyright 2017 New Vector Ltd.
|
# Copyright 2017 New Vector Ltd
|
||||||
#
|
#
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
# you may not use this file except in compliance with the License.
|
# you may not use this file except in compliance with the License.
|
||||||
@ -16,6 +16,7 @@ import time
|
|||||||
|
|
||||||
from mock import Mock
|
from mock import Mock
|
||||||
|
|
||||||
|
import canonicaljson
|
||||||
import signedjson.key
|
import signedjson.key
|
||||||
import signedjson.sign
|
import signedjson.sign
|
||||||
|
|
||||||
@ -23,6 +24,7 @@ from twisted.internet import defer
|
|||||||
|
|
||||||
from synapse.api.errors import SynapseError
|
from synapse.api.errors import SynapseError
|
||||||
from synapse.crypto import keyring
|
from synapse.crypto import keyring
|
||||||
|
from synapse.crypto.keyring import KeyLookupError
|
||||||
from synapse.util import logcontext
|
from synapse.util import logcontext
|
||||||
from synapse.util.logcontext import LoggingContext
|
from synapse.util.logcontext import LoggingContext
|
||||||
|
|
||||||
@ -48,6 +50,9 @@ class MockPerspectiveServer(object):
|
|||||||
key_id: {"key": signedjson.key.encode_verify_key_base64(verify_key)}
|
key_id: {"key": signedjson.key.encode_verify_key_base64(verify_key)}
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
return self.get_signed_response(res)
|
||||||
|
|
||||||
|
def get_signed_response(self, res):
|
||||||
signedjson.sign.sign_json(res, self.server_name, self.key)
|
signedjson.sign.sign_json(res, self.server_name, self.key)
|
||||||
return res
|
return res
|
||||||
|
|
||||||
@ -202,6 +207,132 @@ class KeyringTestCase(unittest.HomeserverTestCase):
|
|||||||
self.assertFalse(d.called)
|
self.assertFalse(d.called)
|
||||||
self.get_success(d)
|
self.get_success(d)
|
||||||
|
|
||||||
|
def test_get_keys_from_server(self):
|
||||||
|
# arbitrarily advance the clock a bit
|
||||||
|
self.reactor.advance(100)
|
||||||
|
|
||||||
|
SERVER_NAME = "server2"
|
||||||
|
kr = keyring.Keyring(self.hs)
|
||||||
|
testkey = signedjson.key.generate_signing_key("ver1")
|
||||||
|
testverifykey = signedjson.key.get_verify_key(testkey)
|
||||||
|
testverifykey_id = "ed25519:ver1"
|
||||||
|
VALID_UNTIL_TS = 1000
|
||||||
|
|
||||||
|
# valid response
|
||||||
|
response = {
|
||||||
|
"server_name": SERVER_NAME,
|
||||||
|
"old_verify_keys": {},
|
||||||
|
"valid_until_ts": VALID_UNTIL_TS,
|
||||||
|
"verify_keys": {
|
||||||
|
testverifykey_id: {
|
||||||
|
"key": signedjson.key.encode_verify_key_base64(testverifykey)
|
||||||
|
}
|
||||||
|
},
|
||||||
|
}
|
||||||
|
signedjson.sign.sign_json(response, SERVER_NAME, testkey)
|
||||||
|
|
||||||
|
def get_json(destination, path, **kwargs):
|
||||||
|
self.assertEqual(destination, SERVER_NAME)
|
||||||
|
self.assertEqual(path, "/_matrix/key/v2/server/key1")
|
||||||
|
return response
|
||||||
|
|
||||||
|
self.http_client.get_json.side_effect = get_json
|
||||||
|
|
||||||
|
server_name_and_key_ids = [(SERVER_NAME, ("key1",))]
|
||||||
|
keys = self.get_success(kr.get_keys_from_server(server_name_and_key_ids))
|
||||||
|
k = keys[SERVER_NAME][testverifykey_id]
|
||||||
|
self.assertEqual(k, testverifykey)
|
||||||
|
self.assertEqual(k.alg, "ed25519")
|
||||||
|
self.assertEqual(k.version, "ver1")
|
||||||
|
|
||||||
|
# check that the perspectives store is correctly updated
|
||||||
|
lookup_triplet = (SERVER_NAME, testverifykey_id, None)
|
||||||
|
key_json = self.get_success(
|
||||||
|
self.hs.get_datastore().get_server_keys_json([lookup_triplet])
|
||||||
|
)
|
||||||
|
res = key_json[lookup_triplet]
|
||||||
|
self.assertEqual(len(res), 1)
|
||||||
|
res = res[0]
|
||||||
|
self.assertEqual(res["key_id"], testverifykey_id)
|
||||||
|
self.assertEqual(res["from_server"], SERVER_NAME)
|
||||||
|
self.assertEqual(res["ts_added_ms"], self.reactor.seconds() * 1000)
|
||||||
|
self.assertEqual(res["ts_valid_until_ms"], VALID_UNTIL_TS)
|
||||||
|
|
||||||
|
# we expect it to be encoded as canonical json *before* it hits the db
|
||||||
|
self.assertEqual(
|
||||||
|
bytes(res["key_json"]), canonicaljson.encode_canonical_json(response)
|
||||||
|
)
|
||||||
|
|
||||||
|
# change the server name: it should cause a rejection
|
||||||
|
response["server_name"] = "OTHER_SERVER"
|
||||||
|
self.get_failure(
|
||||||
|
kr.get_keys_from_server(server_name_and_key_ids), KeyLookupError
|
||||||
|
)
|
||||||
|
|
||||||
|
def test_get_keys_from_perspectives(self):
|
||||||
|
# arbitrarily advance the clock a bit
|
||||||
|
self.reactor.advance(100)
|
||||||
|
|
||||||
|
SERVER_NAME = "server2"
|
||||||
|
kr = keyring.Keyring(self.hs)
|
||||||
|
testkey = signedjson.key.generate_signing_key("ver1")
|
||||||
|
testverifykey = signedjson.key.get_verify_key(testkey)
|
||||||
|
testverifykey_id = "ed25519:ver1"
|
||||||
|
VALID_UNTIL_TS = 200 * 1000
|
||||||
|
|
||||||
|
# valid response
|
||||||
|
response = {
|
||||||
|
"server_name": SERVER_NAME,
|
||||||
|
"old_verify_keys": {},
|
||||||
|
"valid_until_ts": VALID_UNTIL_TS,
|
||||||
|
"verify_keys": {
|
||||||
|
testverifykey_id: {
|
||||||
|
"key": signedjson.key.encode_verify_key_base64(testverifykey)
|
||||||
|
}
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
persp_resp = {
|
||||||
|
"server_keys": [self.mock_perspective_server.get_signed_response(response)]
|
||||||
|
}
|
||||||
|
|
||||||
|
def post_json(destination, path, data, **kwargs):
|
||||||
|
self.assertEqual(destination, self.mock_perspective_server.server_name)
|
||||||
|
self.assertEqual(path, "/_matrix/key/v2/query")
|
||||||
|
|
||||||
|
# check that the request is for the expected key
|
||||||
|
q = data["server_keys"]
|
||||||
|
self.assertEqual(list(q[SERVER_NAME].keys()), ["key1"])
|
||||||
|
return persp_resp
|
||||||
|
|
||||||
|
self.http_client.post_json.side_effect = post_json
|
||||||
|
|
||||||
|
server_name_and_key_ids = [(SERVER_NAME, ("key1",))]
|
||||||
|
keys = self.get_success(kr.get_keys_from_perspectives(server_name_and_key_ids))
|
||||||
|
self.assertIn(SERVER_NAME, keys)
|
||||||
|
k = keys[SERVER_NAME][testverifykey_id]
|
||||||
|
self.assertEqual(k, testverifykey)
|
||||||
|
self.assertEqual(k.alg, "ed25519")
|
||||||
|
self.assertEqual(k.version, "ver1")
|
||||||
|
|
||||||
|
# check that the perspectives store is correctly updated
|
||||||
|
lookup_triplet = (SERVER_NAME, testverifykey_id, None)
|
||||||
|
key_json = self.get_success(
|
||||||
|
self.hs.get_datastore().get_server_keys_json([lookup_triplet])
|
||||||
|
)
|
||||||
|
res = key_json[lookup_triplet]
|
||||||
|
self.assertEqual(len(res), 1)
|
||||||
|
res = res[0]
|
||||||
|
self.assertEqual(res["key_id"], testverifykey_id)
|
||||||
|
self.assertEqual(res["from_server"], self.mock_perspective_server.server_name)
|
||||||
|
self.assertEqual(res["ts_added_ms"], self.reactor.seconds() * 1000)
|
||||||
|
self.assertEqual(res["ts_valid_until_ms"], VALID_UNTIL_TS)
|
||||||
|
|
||||||
|
self.assertEqual(
|
||||||
|
bytes(res["key_json"]),
|
||||||
|
canonicaljson.encode_canonical_json(persp_resp["server_keys"][0]),
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def run_in_context(f, *args, **kwargs):
|
def run_in_context(f, *args, **kwargs):
|
||||||
|
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in New Issue
Block a user