2015-04-14 09:50:29 -04:00
|
|
|
#
|
2023-11-21 15:29:58 -05:00
|
|
|
# This file is licensed under the Affero General Public License (AGPL) version 3.
|
|
|
|
#
|
2024-01-23 06:26:48 -05:00
|
|
|
# Copyright 2015, 2016 OpenMarket Ltd
|
2023-11-21 15:29:58 -05:00
|
|
|
# Copyright (C) 2023 New Vector, Ltd
|
|
|
|
#
|
|
|
|
# This program is free software: you can redistribute it and/or modify
|
|
|
|
# it under the terms of the GNU Affero General Public License as
|
|
|
|
# published by the Free Software Foundation, either version 3 of the
|
|
|
|
# License, or (at your option) any later version.
|
|
|
|
#
|
|
|
|
# See the GNU Affero General Public License for more details:
|
|
|
|
# <https://www.gnu.org/licenses/agpl-3.0.html>.
|
|
|
|
#
|
|
|
|
# Originally licensed under the Apache License, Version 2.0:
|
|
|
|
# <http://www.apache.org/licenses/LICENSE-2.0>.
|
|
|
|
#
|
|
|
|
# [This file includes modifications made by New Vector Limited]
|
2015-04-14 09:50:29 -04:00
|
|
|
#
|
|
|
|
#
|
|
|
|
|
2020-01-28 08:44:21 -05:00
|
|
|
import logging
|
2022-05-16 19:34:38 -04:00
|
|
|
from typing import TYPE_CHECKING, Any, Mapping, NoReturn, Optional, Tuple, cast
|
2020-01-28 08:44:21 -05:00
|
|
|
|
2022-06-07 12:33:55 -04:00
|
|
|
import psycopg2.extensions
|
|
|
|
|
2022-01-25 09:14:46 -05:00
|
|
|
from synapse.storage.engines._base import (
|
|
|
|
BaseDatabaseEngine,
|
|
|
|
IncorrectDatabaseSetup,
|
|
|
|
IsolationLevel,
|
|
|
|
)
|
2022-05-16 19:34:38 -04:00
|
|
|
from synapse.storage.types import Cursor
|
|
|
|
|
|
|
|
if TYPE_CHECKING:
|
|
|
|
from synapse.storage.database import LoggingDatabaseConnection
|
|
|
|
|
2015-04-29 06:42:28 -04:00
|
|
|
|
2020-01-28 08:44:21 -05:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
2015-04-14 09:50:29 -04:00
|
|
|
|
2022-09-26 13:28:32 -04:00
|
|
|
class PostgresEngine(
|
|
|
|
BaseDatabaseEngine[psycopg2.extensions.connection, psycopg2.extensions.cursor]
|
|
|
|
):
|
2022-05-16 19:34:38 -04:00
|
|
|
def __init__(self, database_config: Mapping[str, Any]):
|
|
|
|
super().__init__(psycopg2, database_config)
|
|
|
|
psycopg2.extensions.register_type(psycopg2.extensions.UNICODE)
|
2019-10-08 11:21:17 -04:00
|
|
|
|
2023-11-15 08:02:11 -05:00
|
|
|
# Disables passing `bytes` to txn.execute, c.f.
|
|
|
|
# https://github.com/matrix-org/synapse/issues/6186. If you do
|
2019-10-08 11:21:17 -04:00
|
|
|
# actually want to use bytes than wrap it in `bytearray`.
|
2022-05-16 19:34:38 -04:00
|
|
|
def _disable_bytes_adapter(_: bytes) -> NoReturn:
|
2019-10-08 11:21:17 -04:00
|
|
|
raise Exception("Passing bytes to DB is disabled.")
|
|
|
|
|
2022-05-16 19:34:38 -04:00
|
|
|
psycopg2.extensions.register_adapter(bytes, _disable_bytes_adapter)
|
|
|
|
self.synchronous_commit: bool = database_config.get("synchronous_commit", True)
|
2023-07-03 06:38:57 -04:00
|
|
|
# Set the statement timeout to 1 hour by default.
|
|
|
|
# Any query taking more than 1 hour should probably be considered a bug;
|
|
|
|
# most of the time this is a sign that work needs to be split up or that
|
|
|
|
# some degenerate query plan has been created and the client has probably
|
|
|
|
# timed out/walked off anyway.
|
|
|
|
# This is in milliseconds.
|
|
|
|
self.statement_timeout: Optional[int] = database_config.get(
|
|
|
|
"statement_timeout", 60 * 60 * 1000
|
|
|
|
)
|
2022-05-16 19:34:38 -04:00
|
|
|
self._version: Optional[int] = None # unknown as yet
|
2015-04-14 09:50:29 -04:00
|
|
|
|
2022-01-25 09:14:46 -05:00
|
|
|
self.isolation_level_map: Mapping[int, int] = {
|
2022-05-16 19:34:38 -04:00
|
|
|
IsolationLevel.READ_COMMITTED: psycopg2.extensions.ISOLATION_LEVEL_READ_COMMITTED,
|
|
|
|
IsolationLevel.REPEATABLE_READ: psycopg2.extensions.ISOLATION_LEVEL_REPEATABLE_READ,
|
|
|
|
IsolationLevel.SERIALIZABLE: psycopg2.extensions.ISOLATION_LEVEL_SERIALIZABLE,
|
2022-01-25 09:14:46 -05:00
|
|
|
}
|
|
|
|
self.default_isolation_level = (
|
2022-05-16 19:34:38 -04:00
|
|
|
psycopg2.extensions.ISOLATION_LEVEL_REPEATABLE_READ
|
2022-01-25 09:14:46 -05:00
|
|
|
)
|
2022-03-23 13:23:05 -04:00
|
|
|
self.config = database_config
|
2022-01-25 09:14:46 -05:00
|
|
|
|
2020-02-27 06:53:40 -05:00
|
|
|
@property
|
|
|
|
def single_threaded(self) -> bool:
|
|
|
|
return False
|
|
|
|
|
2022-05-16 19:34:38 -04:00
|
|
|
def get_db_locale(self, txn: Cursor) -> Tuple[str, str]:
|
2022-03-23 13:23:05 -04:00
|
|
|
txn.execute(
|
|
|
|
"SELECT datcollate, datctype FROM pg_database WHERE datname = current_database()"
|
|
|
|
)
|
2022-05-16 19:34:38 -04:00
|
|
|
collation, ctype = cast(Tuple[str, str], txn.fetchone())
|
2022-03-23 13:23:05 -04:00
|
|
|
return collation, ctype
|
|
|
|
|
2022-05-16 19:34:38 -04:00
|
|
|
def check_database(
|
2022-06-07 12:33:55 -04:00
|
|
|
self,
|
|
|
|
db_conn: psycopg2.extensions.connection,
|
|
|
|
allow_outdated_version: bool = False,
|
2022-05-16 19:34:38 -04:00
|
|
|
) -> None:
|
2020-01-09 12:33:41 -05:00
|
|
|
# Get the version of PostgreSQL that we're using. As per the psycopg2
|
|
|
|
# docs: The number is formed by converting the major, minor, and
|
|
|
|
# revision numbers into two-decimal-digit numbers and appending them
|
|
|
|
# together. For example, version 8.1.5 will be returned as 80105
|
2022-12-12 16:25:07 -05:00
|
|
|
self._version = db_conn.server_version
|
2022-03-23 13:23:05 -04:00
|
|
|
allow_unsafe_locale = self.config.get("allow_unsafe_locale", False)
|
2020-01-09 12:33:41 -05:00
|
|
|
|
|
|
|
# Are we on a supported PostgreSQL version?
|
2022-11-08 12:17:13 -05:00
|
|
|
if not allow_outdated_version and self._version < 110000:
|
|
|
|
raise RuntimeError("Synapse requires PostgreSQL 11 or above.")
|
2020-01-09 12:33:41 -05:00
|
|
|
|
2020-01-09 12:21:30 -05:00
|
|
|
with db_conn.cursor() as txn:
|
|
|
|
txn.execute("SHOW SERVER_ENCODING")
|
|
|
|
rows = txn.fetchall()
|
|
|
|
if rows and rows[0][0] != "UTF8":
|
|
|
|
raise IncorrectDatabaseSetup(
|
|
|
|
"Database has incorrect encoding: '%s' instead of 'UTF8'\n"
|
2020-02-26 10:17:03 -05:00
|
|
|
"See docs/postgres.md for more information." % (rows[0][0],)
|
2020-01-09 12:21:30 -05:00
|
|
|
)
|
2015-04-29 06:42:28 -04:00
|
|
|
|
2022-03-23 13:23:05 -04:00
|
|
|
collation, ctype = self.get_db_locale(txn)
|
2020-01-28 08:44:21 -05:00
|
|
|
if collation != "C":
|
|
|
|
logger.warning(
|
2022-03-23 13:23:05 -04:00
|
|
|
"Database has incorrect collation of %r. Should be 'C'",
|
2020-02-26 10:17:03 -05:00
|
|
|
collation,
|
2020-01-28 08:44:21 -05:00
|
|
|
)
|
2022-03-23 13:23:05 -04:00
|
|
|
if not allow_unsafe_locale:
|
|
|
|
raise IncorrectDatabaseSetup(
|
|
|
|
"Database has incorrect collation of %r. Should be 'C'\n"
|
|
|
|
"See docs/postgres.md for more information. You can override this check by"
|
|
|
|
"setting 'allow_unsafe_locale' to true in the database config.",
|
|
|
|
collation,
|
|
|
|
)
|
2020-01-28 08:44:21 -05:00
|
|
|
|
|
|
|
if ctype != "C":
|
2022-03-23 13:23:05 -04:00
|
|
|
if not allow_unsafe_locale:
|
|
|
|
logger.warning(
|
|
|
|
"Database has incorrect ctype of %r. Should be 'C'",
|
|
|
|
ctype,
|
|
|
|
)
|
|
|
|
raise IncorrectDatabaseSetup(
|
|
|
|
"Database has incorrect ctype of %r. Should be 'C'\n"
|
|
|
|
"See docs/postgres.md for more information. You can override this check by"
|
|
|
|
"setting 'allow_unsafe_locale' to true in the database config.",
|
|
|
|
ctype,
|
|
|
|
)
|
2020-01-28 08:44:21 -05:00
|
|
|
|
2022-05-16 19:34:38 -04:00
|
|
|
def check_new_database(self, txn: Cursor) -> None:
|
2020-01-28 08:44:21 -05:00
|
|
|
"""Gets called when setting up a brand new database. This allows us to
|
|
|
|
apply stricter checks on new databases versus existing database.
|
|
|
|
"""
|
|
|
|
|
2022-03-23 13:23:05 -04:00
|
|
|
collation, ctype = self.get_db_locale(txn)
|
2020-01-28 08:44:21 -05:00
|
|
|
|
|
|
|
errors = []
|
|
|
|
|
|
|
|
if collation != "C":
|
|
|
|
errors.append(" - 'COLLATE' is set to %r. Should be 'C'" % (collation,))
|
|
|
|
|
|
|
|
if ctype != "C":
|
2020-07-01 08:56:16 -04:00
|
|
|
errors.append(" - 'CTYPE' is set to %r. Should be 'C'" % (ctype,))
|
2020-01-28 08:44:21 -05:00
|
|
|
|
|
|
|
if errors:
|
|
|
|
raise IncorrectDatabaseSetup(
|
|
|
|
"Database is incorrectly configured:\n\n%s\n\n"
|
|
|
|
"See docs/postgres.md for more information." % ("\n".join(errors))
|
|
|
|
)
|
|
|
|
|
2022-05-16 19:34:38 -04:00
|
|
|
def convert_param_style(self, sql: str) -> str:
|
2015-04-14 09:50:29 -04:00
|
|
|
return sql.replace("?", "%s")
|
|
|
|
|
2022-05-16 19:34:38 -04:00
|
|
|
def on_new_connection(self, db_conn: "LoggingDatabaseConnection") -> None:
|
2022-01-25 09:14:46 -05:00
|
|
|
db_conn.set_isolation_level(self.default_isolation_level)
|
2018-08-30 10:19:58 -04:00
|
|
|
|
|
|
|
# Set the bytea output to escape, vs the default of hex
|
|
|
|
cursor = db_conn.cursor()
|
|
|
|
cursor.execute("SET bytea_output TO escape")
|
|
|
|
|
2016-06-20 12:53:38 -04:00
|
|
|
# Asynchronous commit, don't wait for the server to call fsync before
|
|
|
|
# ending the transaction.
|
|
|
|
# https://www.postgresql.org/docs/current/static/wal-async-commit.html
|
|
|
|
if not self.synchronous_commit:
|
|
|
|
cursor.execute("SET synchronous_commit TO OFF")
|
2018-08-30 10:19:58 -04:00
|
|
|
|
2023-07-03 06:38:57 -04:00
|
|
|
# Abort really long-running statements and turn them into errors.
|
|
|
|
if self.statement_timeout is not None:
|
|
|
|
cursor.execute("SET statement_timeout TO ?", (self.statement_timeout,))
|
|
|
|
|
2018-08-30 10:19:58 -04:00
|
|
|
cursor.close()
|
2020-10-07 10:15:57 -04:00
|
|
|
db_conn.commit()
|
2015-04-14 09:50:29 -04:00
|
|
|
|
2019-10-10 10:35:46 -04:00
|
|
|
@property
|
2022-05-16 19:34:38 -04:00
|
|
|
def supports_using_any_list(self) -> bool:
|
2019-10-10 10:35:46 -04:00
|
|
|
"""Do we support using `a = ANY(?)` and passing a list"""
|
|
|
|
return True
|
|
|
|
|
2021-06-30 07:07:16 -04:00
|
|
|
@property
|
|
|
|
def supports_returning(self) -> bool:
|
|
|
|
"""Do we support the `RETURNING` clause in insert/update/delete?"""
|
|
|
|
return True
|
|
|
|
|
2022-05-16 19:34:38 -04:00
|
|
|
def is_deadlock(self, error: Exception) -> bool:
|
|
|
|
if isinstance(error, psycopg2.DatabaseError):
|
2016-06-20 12:53:38 -04:00
|
|
|
# https://www.postgresql.org/docs/current/static/errcodes-appendix.html
|
|
|
|
# "40001" serialization_failure
|
|
|
|
# "40P01" deadlock_detected
|
2015-04-15 05:23:42 -04:00
|
|
|
return error.pgcode in ["40001", "40P01"]
|
2015-04-14 09:50:29 -04:00
|
|
|
return False
|
2015-04-27 07:40:49 -04:00
|
|
|
|
2022-06-07 12:33:55 -04:00
|
|
|
def is_connection_closed(self, conn: psycopg2.extensions.connection) -> bool:
|
2015-05-01 05:24:24 -04:00
|
|
|
return bool(conn.closed)
|
2015-04-27 08:22:30 -04:00
|
|
|
|
2022-05-16 19:34:38 -04:00
|
|
|
def lock_table(self, txn: Cursor, table: str) -> None:
|
2015-04-27 08:22:30 -04:00
|
|
|
txn.execute("LOCK TABLE %s in EXCLUSIVE MODE" % (table,))
|
2018-02-06 09:31:24 -05:00
|
|
|
|
2019-02-27 05:21:49 -05:00
|
|
|
@property
|
2022-05-16 19:34:38 -04:00
|
|
|
def server_version(self) -> str:
|
|
|
|
"""Returns a string giving the server version. For example: '8.1.5'."""
|
2020-01-09 12:33:41 -05:00
|
|
|
# note that this is a bit of a hack because it relies on check_database
|
|
|
|
# having been called. Still, that should be a safe bet here.
|
2019-02-27 05:21:49 -05:00
|
|
|
numver = self._version
|
|
|
|
assert numver is not None
|
|
|
|
|
|
|
|
# https://www.postgresql.org/docs/current/libpq-status.html#LIBPQ-PQSERVERVERSION
|
|
|
|
if numver >= 100000:
|
2019-04-03 05:07:29 -04:00
|
|
|
return "%i.%i" % (numver / 10000, numver % 10000)
|
2019-02-27 05:21:49 -05:00
|
|
|
else:
|
2019-04-03 05:07:29 -04:00
|
|
|
return "%i.%i.%i" % (numver / 10000, (numver % 10000) / 100, numver % 100)
|
2020-10-07 10:15:57 -04:00
|
|
|
|
2023-09-06 03:30:53 -04:00
|
|
|
@property
|
|
|
|
def row_id_name(self) -> str:
|
|
|
|
return "ctid"
|
|
|
|
|
2022-06-07 12:33:55 -04:00
|
|
|
def in_transaction(self, conn: psycopg2.extensions.connection) -> bool:
|
2022-05-16 19:34:38 -04:00
|
|
|
return conn.status != psycopg2.extensions.STATUS_READY
|
2020-10-07 10:15:57 -04:00
|
|
|
|
2022-05-16 19:34:38 -04:00
|
|
|
def attempt_to_set_autocommit(
|
2022-06-07 12:33:55 -04:00
|
|
|
self, conn: psycopg2.extensions.connection, autocommit: bool
|
2022-05-16 19:34:38 -04:00
|
|
|
) -> None:
|
|
|
|
return conn.set_session(autocommit=autocommit)
|
2022-01-25 09:14:46 -05:00
|
|
|
|
|
|
|
def attempt_to_set_isolation_level(
|
2022-06-07 12:33:55 -04:00
|
|
|
self, conn: psycopg2.extensions.connection, isolation_level: Optional[int]
|
2022-05-16 19:34:38 -04:00
|
|
|
) -> None:
|
2022-01-25 09:14:46 -05:00
|
|
|
if isolation_level is None:
|
|
|
|
isolation_level = self.default_isolation_level
|
|
|
|
else:
|
|
|
|
isolation_level = self.isolation_level_map[isolation_level]
|
2022-05-16 19:34:38 -04:00
|
|
|
return conn.set_isolation_level(isolation_level)
|
2022-09-26 13:28:32 -04:00
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def executescript(cursor: psycopg2.extensions.cursor, script: str) -> None:
|
|
|
|
"""Execute a chunk of SQL containing multiple semicolon-delimited statements.
|
|
|
|
|
|
|
|
Psycopg2 seems happy to do this in DBAPI2's `execute()` function.
|
2023-01-31 06:03:55 -05:00
|
|
|
|
|
|
|
For consistency with SQLite, any ongoing transaction is committed before
|
|
|
|
executing the script in its own transaction. The script transaction is
|
|
|
|
left open and it is the responsibility of the caller to commit it.
|
2022-09-26 13:28:32 -04:00
|
|
|
"""
|
2023-01-31 06:03:55 -05:00
|
|
|
cursor.execute(f"COMMIT; BEGIN TRANSACTION; {script}")
|