2019-05-21 12:36:50 -04:00
|
|
|
|
# Copyright 2018, 2019 New Vector Ltd
|
2019-09-04 08:04:27 -04:00
|
|
|
|
# Copyright 2019 The Matrix.org Foundation C.I.C.
|
2019-05-21 12:36:50 -04:00
|
|
|
|
#
|
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
|
#
|
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
|
#
|
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
|
# limitations under the License.
|
|
|
|
|
|
|
|
|
|
import logging
|
2020-11-05 13:59:12 -05:00
|
|
|
|
from enum import Enum
|
2019-09-04 08:04:27 -04:00
|
|
|
|
from itertools import chain
|
2022-06-29 13:12:45 -04:00
|
|
|
|
from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union, cast
|
2019-05-21 12:36:50 -04:00
|
|
|
|
|
2021-01-26 10:50:21 -05:00
|
|
|
|
from typing_extensions import Counter
|
|
|
|
|
|
2019-09-04 08:04:27 -04:00
|
|
|
|
from twisted.internet.defer import DeferredLock
|
2019-05-21 12:36:50 -04:00
|
|
|
|
|
2021-09-08 10:00:43 -04:00
|
|
|
|
from synapse.api.constants import EventContentFields, EventTypes, Membership
|
2020-11-05 13:59:12 -05:00
|
|
|
|
from synapse.api.errors import StoreError
|
2021-12-29 08:01:13 -05:00
|
|
|
|
from synapse.storage.database import (
|
|
|
|
|
DatabasePool,
|
|
|
|
|
LoggingDatabaseConnection,
|
|
|
|
|
LoggingTransaction,
|
|
|
|
|
)
|
2023-01-23 06:36:20 -05:00
|
|
|
|
from synapse.storage.databases.main.events_worker import InvalidEventError
|
2020-08-05 16:38:57 -04:00
|
|
|
|
from synapse.storage.databases.main.state_deltas import StateDeltasStore
|
2020-11-05 13:59:12 -05:00
|
|
|
|
from synapse.types import JsonDict
|
2019-05-21 12:36:50 -04:00
|
|
|
|
from synapse.util.caches.descriptors import cached
|
|
|
|
|
|
2021-10-22 13:15:41 -04:00
|
|
|
|
if TYPE_CHECKING:
|
|
|
|
|
from synapse.server import HomeServer
|
|
|
|
|
|
2019-05-21 12:36:50 -04:00
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
|
# these fields track absolutes (e.g. total number of rooms on the server)
|
2019-09-04 08:04:27 -04:00
|
|
|
|
# You can think of these as Prometheus Gauges.
|
|
|
|
|
# You can draw these stats on a line graph.
|
|
|
|
|
# Example: number of users in a room
|
2019-05-21 12:36:50 -04:00
|
|
|
|
ABSOLUTE_STATS_FIELDS = {
|
|
|
|
|
"room": (
|
|
|
|
|
"current_state_events",
|
|
|
|
|
"joined_members",
|
|
|
|
|
"invited_members",
|
2021-06-09 14:39:51 -04:00
|
|
|
|
"knocked_members",
|
2019-05-21 12:36:50 -04:00
|
|
|
|
"left_members",
|
|
|
|
|
"banned_members",
|
2019-09-04 08:04:27 -04:00
|
|
|
|
"local_users_in_room",
|
2019-05-21 12:36:50 -04:00
|
|
|
|
),
|
2019-09-04 08:04:27 -04:00
|
|
|
|
"user": ("joined_rooms",),
|
2019-05-21 12:36:50 -04:00
|
|
|
|
}
|
|
|
|
|
|
2019-09-04 08:04:27 -04:00
|
|
|
|
TYPE_TO_TABLE = {"room": ("room_stats", "room_id"), "user": ("user_stats", "user_id")}
|
2019-05-21 12:36:50 -04:00
|
|
|
|
|
2019-09-04 08:04:27 -04:00
|
|
|
|
# these are the tables (& ID columns) which contain our actual subjects
|
|
|
|
|
TYPE_TO_ORIGIN_TABLE = {"room": ("rooms", "room_id"), "user": ("users", "name")}
|
2019-05-21 12:36:50 -04:00
|
|
|
|
|
|
|
|
|
|
2020-11-05 13:59:12 -05:00
|
|
|
|
class UserSortOrder(Enum):
|
|
|
|
|
"""
|
|
|
|
|
Enum to define the sorting method used when returning users
|
2021-04-01 06:28:53 -04:00
|
|
|
|
with get_users_paginate in __init__.py
|
|
|
|
|
and get_users_media_usage_paginate in stats.py
|
2020-11-05 13:59:12 -05:00
|
|
|
|
|
2021-04-01 06:28:53 -04:00
|
|
|
|
When moves this to __init__.py gets `builtins.ImportError` with
|
|
|
|
|
`most likely due to a circular import`
|
|
|
|
|
|
|
|
|
|
MEDIA_LENGTH = ordered by size of uploaded media.
|
|
|
|
|
MEDIA_COUNT = ordered by number of uploaded media.
|
2020-11-05 13:59:12 -05:00
|
|
|
|
USER_ID = ordered alphabetically by `user_id`.
|
2021-04-01 06:28:53 -04:00
|
|
|
|
NAME = ordered alphabetically by `user_id`. This is for compatibility reasons,
|
|
|
|
|
as the user_id is returned in the name field in the response in list users admin API.
|
2020-11-05 13:59:12 -05:00
|
|
|
|
DISPLAYNAME = ordered alphabetically by `displayname`
|
2021-04-01 06:28:53 -04:00
|
|
|
|
GUEST = ordered by `is_guest`
|
|
|
|
|
ADMIN = ordered by `admin`
|
|
|
|
|
DEACTIVATED = ordered by `deactivated`
|
|
|
|
|
USER_TYPE = ordered alphabetically by `user_type`
|
|
|
|
|
AVATAR_URL = ordered alphabetically by `avatar_url`
|
|
|
|
|
SHADOW_BANNED = ordered by `shadow_banned`
|
2021-07-22 10:05:16 -04:00
|
|
|
|
CREATION_TS = ordered by `creation_ts`
|
2020-11-05 13:59:12 -05:00
|
|
|
|
"""
|
|
|
|
|
|
|
|
|
|
MEDIA_LENGTH = "media_length"
|
|
|
|
|
MEDIA_COUNT = "media_count"
|
|
|
|
|
USER_ID = "user_id"
|
2021-04-01 06:28:53 -04:00
|
|
|
|
NAME = "name"
|
2020-11-05 13:59:12 -05:00
|
|
|
|
DISPLAYNAME = "displayname"
|
2021-04-01 06:28:53 -04:00
|
|
|
|
GUEST = "is_guest"
|
|
|
|
|
ADMIN = "admin"
|
|
|
|
|
DEACTIVATED = "deactivated"
|
|
|
|
|
USER_TYPE = "user_type"
|
|
|
|
|
AVATAR_URL = "avatar_url"
|
|
|
|
|
SHADOW_BANNED = "shadow_banned"
|
2021-07-22 10:05:16 -04:00
|
|
|
|
CREATION_TS = "creation_ts"
|
2020-11-05 13:59:12 -05:00
|
|
|
|
|
|
|
|
|
|
2019-05-21 12:36:50 -04:00
|
|
|
|
class StatsStore(StateDeltasStore):
|
2021-12-13 12:05:00 -05:00
|
|
|
|
def __init__(
|
|
|
|
|
self,
|
|
|
|
|
database: DatabasePool,
|
|
|
|
|
db_conn: LoggingDatabaseConnection,
|
|
|
|
|
hs: "HomeServer",
|
|
|
|
|
):
|
2020-09-18 09:56:44 -04:00
|
|
|
|
super().__init__(database, db_conn, hs)
|
2019-05-21 12:36:50 -04:00
|
|
|
|
|
2022-03-28 14:11:14 -04:00
|
|
|
|
self.server_name: str = hs.hostname
|
2019-05-21 12:36:50 -04:00
|
|
|
|
self.clock = self.hs.get_clock()
|
2021-09-24 07:25:21 -04:00
|
|
|
|
self.stats_enabled = hs.config.stats.stats_enabled
|
2019-05-21 12:36:50 -04:00
|
|
|
|
|
2019-09-04 08:04:27 -04:00
|
|
|
|
self.stats_delta_processing_lock = DeferredLock()
|
|
|
|
|
|
2020-08-05 16:38:57 -04:00
|
|
|
|
self.db_pool.updates.register_background_update_handler(
|
2019-05-21 12:36:50 -04:00
|
|
|
|
"populate_stats_process_rooms", self._populate_stats_process_rooms
|
|
|
|
|
)
|
2020-08-05 16:38:57 -04:00
|
|
|
|
self.db_pool.updates.register_background_update_handler(
|
2019-09-04 08:04:27 -04:00
|
|
|
|
"populate_stats_process_users", self._populate_stats_process_users
|
2019-05-21 12:36:50 -04:00
|
|
|
|
)
|
|
|
|
|
|
2021-12-29 08:01:13 -05:00
|
|
|
|
async def _populate_stats_process_users(
|
|
|
|
|
self, progress: JsonDict, batch_size: int
|
|
|
|
|
) -> int:
|
2019-05-21 12:36:50 -04:00
|
|
|
|
"""
|
2019-09-04 08:04:27 -04:00
|
|
|
|
This is a background update which regenerates statistics for users.
|
2019-05-21 12:36:50 -04:00
|
|
|
|
"""
|
|
|
|
|
if not self.stats_enabled:
|
2020-08-05 16:38:57 -04:00
|
|
|
|
await self.db_pool.updates._end_background_update(
|
|
|
|
|
"populate_stats_process_users"
|
|
|
|
|
)
|
2019-07-23 09:00:55 -04:00
|
|
|
|
return 1
|
2019-05-21 12:36:50 -04:00
|
|
|
|
|
2019-09-04 08:04:27 -04:00
|
|
|
|
last_user_id = progress.get("last_user_id", "")
|
|
|
|
|
|
2021-12-29 08:01:13 -05:00
|
|
|
|
def _get_next_batch(txn: LoggingTransaction) -> List[str]:
|
2019-09-04 08:04:27 -04:00
|
|
|
|
sql = """
|
|
|
|
|
SELECT DISTINCT name FROM users
|
|
|
|
|
WHERE name > ?
|
|
|
|
|
ORDER BY name ASC
|
|
|
|
|
LIMIT ?
|
|
|
|
|
"""
|
|
|
|
|
txn.execute(sql, (last_user_id, batch_size))
|
|
|
|
|
return [r for r, in txn]
|
|
|
|
|
|
2020-08-05 16:38:57 -04:00
|
|
|
|
users_to_work_on = await self.db_pool.runInteraction(
|
2019-09-04 08:04:27 -04:00
|
|
|
|
"_populate_stats_process_users", _get_next_batch
|
2019-05-21 12:36:50 -04:00
|
|
|
|
)
|
|
|
|
|
|
2019-09-04 08:04:27 -04:00
|
|
|
|
# No more rooms -- complete the transaction.
|
|
|
|
|
if not users_to_work_on:
|
2020-08-05 16:38:57 -04:00
|
|
|
|
await self.db_pool.updates._end_background_update(
|
|
|
|
|
"populate_stats_process_users"
|
|
|
|
|
)
|
2019-09-04 08:04:27 -04:00
|
|
|
|
return 1
|
2019-05-21 12:36:50 -04:00
|
|
|
|
|
2019-09-04 08:04:27 -04:00
|
|
|
|
for user_id in users_to_work_on:
|
2020-07-30 07:20:41 -04:00
|
|
|
|
await self._calculate_and_set_initial_state_for_user(user_id)
|
2019-09-04 08:04:27 -04:00
|
|
|
|
progress["last_user_id"] = user_id
|
2019-05-21 12:36:50 -04:00
|
|
|
|
|
2020-08-05 16:38:57 -04:00
|
|
|
|
await self.db_pool.runInteraction(
|
2019-09-04 08:04:27 -04:00
|
|
|
|
"populate_stats_process_users",
|
2020-08-05 16:38:57 -04:00
|
|
|
|
self.db_pool.updates._background_update_progress_txn,
|
2019-09-04 08:04:27 -04:00
|
|
|
|
"populate_stats_process_users",
|
|
|
|
|
progress,
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
return len(users_to_work_on)
|
2019-05-21 12:36:50 -04:00
|
|
|
|
|
2021-12-29 08:01:13 -05:00
|
|
|
|
async def _populate_stats_process_rooms(
|
|
|
|
|
self, progress: JsonDict, batch_size: int
|
|
|
|
|
) -> int:
|
2020-09-08 06:05:59 -04:00
|
|
|
|
"""This is a background update which regenerates statistics for rooms."""
|
2019-05-21 12:36:50 -04:00
|
|
|
|
if not self.stats_enabled:
|
2020-08-05 16:38:57 -04:00
|
|
|
|
await self.db_pool.updates._end_background_update(
|
2020-09-08 06:05:59 -04:00
|
|
|
|
"populate_stats_process_rooms"
|
2020-08-03 16:54:24 -04:00
|
|
|
|
)
|
2019-07-23 09:00:55 -04:00
|
|
|
|
return 1
|
2019-05-21 12:36:50 -04:00
|
|
|
|
|
2019-09-04 08:04:27 -04:00
|
|
|
|
last_room_id = progress.get("last_room_id", "")
|
2019-05-21 12:36:50 -04:00
|
|
|
|
|
2021-12-29 08:01:13 -05:00
|
|
|
|
def _get_next_batch(txn: LoggingTransaction) -> List[str]:
|
2019-05-21 12:36:50 -04:00
|
|
|
|
sql = """
|
2019-09-04 08:04:27 -04:00
|
|
|
|
SELECT DISTINCT room_id FROM current_state_events
|
|
|
|
|
WHERE room_id > ?
|
|
|
|
|
ORDER BY room_id ASC
|
|
|
|
|
LIMIT ?
|
|
|
|
|
"""
|
|
|
|
|
txn.execute(sql, (last_room_id, batch_size))
|
|
|
|
|
return [r for r, in txn]
|
2019-05-21 12:36:50 -04:00
|
|
|
|
|
2020-08-05 16:38:57 -04:00
|
|
|
|
rooms_to_work_on = await self.db_pool.runInteraction(
|
2020-09-08 06:05:59 -04:00
|
|
|
|
"populate_stats_rooms_get_batch", _get_next_batch
|
2019-05-21 12:36:50 -04:00
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
# No more rooms -- complete the transaction.
|
|
|
|
|
if not rooms_to_work_on:
|
2020-08-05 16:38:57 -04:00
|
|
|
|
await self.db_pool.updates._end_background_update(
|
2020-09-08 06:05:59 -04:00
|
|
|
|
"populate_stats_process_rooms"
|
2020-08-03 16:54:24 -04:00
|
|
|
|
)
|
2019-07-23 09:00:55 -04:00
|
|
|
|
return 1
|
2019-05-21 12:36:50 -04:00
|
|
|
|
|
2019-09-04 08:04:27 -04:00
|
|
|
|
for room_id in rooms_to_work_on:
|
2020-07-30 07:20:41 -04:00
|
|
|
|
await self._calculate_and_set_initial_state_for_room(room_id)
|
2019-09-04 08:04:27 -04:00
|
|
|
|
progress["last_room_id"] = room_id
|
2019-05-21 12:36:50 -04:00
|
|
|
|
|
2020-08-05 16:38:57 -04:00
|
|
|
|
await self.db_pool.runInteraction(
|
2020-09-08 06:05:59 -04:00
|
|
|
|
"_populate_stats_process_rooms",
|
2020-08-05 16:38:57 -04:00
|
|
|
|
self.db_pool.updates._background_update_progress_txn,
|
2020-09-08 06:05:59 -04:00
|
|
|
|
"populate_stats_process_rooms",
|
2019-09-04 08:04:27 -04:00
|
|
|
|
progress,
|
|
|
|
|
)
|
2019-05-21 12:36:50 -04:00
|
|
|
|
|
2019-09-04 08:04:27 -04:00
|
|
|
|
return len(rooms_to_work_on)
|
2019-05-21 12:36:50 -04:00
|
|
|
|
|
2020-08-26 07:19:32 -04:00
|
|
|
|
async def get_stats_positions(self) -> int:
|
2019-05-21 12:36:50 -04:00
|
|
|
|
"""
|
2019-09-04 08:04:27 -04:00
|
|
|
|
Returns the stats processor positions.
|
2019-05-21 12:36:50 -04:00
|
|
|
|
"""
|
2020-08-26 07:19:32 -04:00
|
|
|
|
return await self.db_pool.simple_select_one_onecol(
|
2019-09-04 08:04:27 -04:00
|
|
|
|
table="stats_incremental_position",
|
2019-05-21 12:36:50 -04:00
|
|
|
|
keyvalues={},
|
|
|
|
|
retcol="stream_id",
|
2019-09-04 08:04:27 -04:00
|
|
|
|
desc="stats_incremental_position",
|
2019-05-21 12:36:50 -04:00
|
|
|
|
)
|
|
|
|
|
|
2020-08-27 13:38:41 -04:00
|
|
|
|
async def update_room_state(self, room_id: str, fields: Dict[str, Any]) -> None:
|
2020-09-02 07:59:39 -04:00
|
|
|
|
"""Update the state of a room.
|
|
|
|
|
|
|
|
|
|
fields can contain the following keys with string values:
|
|
|
|
|
* join_rules
|
|
|
|
|
* history_visibility
|
|
|
|
|
* encryption
|
|
|
|
|
* name
|
|
|
|
|
* topic
|
|
|
|
|
* avatar
|
|
|
|
|
* canonical_alias
|
2020-09-22 14:39:29 -04:00
|
|
|
|
* guest_access
|
2022-06-29 13:12:45 -04:00
|
|
|
|
* room_type
|
2020-09-02 07:59:39 -04:00
|
|
|
|
|
|
|
|
|
A is_federatable key can also be included with a boolean value.
|
|
|
|
|
|
2019-05-21 12:36:50 -04:00
|
|
|
|
Args:
|
2020-09-02 07:59:39 -04:00
|
|
|
|
room_id: The room ID to update the state of.
|
|
|
|
|
fields: The fields to update. This can include a partial list of the
|
|
|
|
|
above fields to only update some room information.
|
2019-05-21 12:36:50 -04:00
|
|
|
|
"""
|
2020-09-02 07:59:39 -04:00
|
|
|
|
# Ensure that the values to update are valid, they should be strings and
|
|
|
|
|
# not contain any null bytes.
|
|
|
|
|
#
|
|
|
|
|
# Invalid data gets overwritten with null.
|
|
|
|
|
#
|
|
|
|
|
# Note that a missing value should not be overwritten (it keeps the
|
|
|
|
|
# previous value).
|
|
|
|
|
sentinel = object()
|
2019-06-03 12:06:54 -04:00
|
|
|
|
for col in (
|
|
|
|
|
"join_rules",
|
|
|
|
|
"history_visibility",
|
|
|
|
|
"encryption",
|
|
|
|
|
"name",
|
|
|
|
|
"topic",
|
|
|
|
|
"avatar",
|
2019-06-20 05:32:02 -04:00
|
|
|
|
"canonical_alias",
|
2020-09-22 14:39:29 -04:00
|
|
|
|
"guest_access",
|
2022-06-29 13:12:45 -04:00
|
|
|
|
"room_type",
|
2019-06-03 12:06:54 -04:00
|
|
|
|
):
|
2020-09-02 07:59:39 -04:00
|
|
|
|
field = fields.get(col, sentinel)
|
|
|
|
|
if field is not sentinel and (not isinstance(field, str) or "\0" in field):
|
2019-06-03 12:06:54 -04:00
|
|
|
|
fields[col] = None
|
|
|
|
|
|
2020-08-27 13:38:41 -04:00
|
|
|
|
await self.db_pool.simple_upsert(
|
2019-09-04 08:04:27 -04:00
|
|
|
|
table="room_stats_state",
|
2019-05-21 12:36:50 -04:00
|
|
|
|
keyvalues={"room_id": room_id},
|
|
|
|
|
values=fields,
|
|
|
|
|
desc="update_room_state",
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
@cached()
|
2021-01-26 10:50:21 -05:00
|
|
|
|
async def get_earliest_token_for_stats(
|
|
|
|
|
self, stats_type: str, id: str
|
|
|
|
|
) -> Optional[int]:
|
2019-05-21 12:36:50 -04:00
|
|
|
|
"""
|
|
|
|
|
Fetch the "earliest token". This is used by the room stats delta
|
|
|
|
|
processor to ignore deltas that have been processed between the
|
|
|
|
|
start of the background task and any particular room's stats
|
|
|
|
|
being calculated.
|
|
|
|
|
|
|
|
|
|
Returns:
|
2020-08-26 07:19:32 -04:00
|
|
|
|
The earliest token.
|
2019-05-21 12:36:50 -04:00
|
|
|
|
"""
|
2019-09-04 08:04:27 -04:00
|
|
|
|
table, id_col = TYPE_TO_TABLE[stats_type]
|
|
|
|
|
|
2020-08-26 07:19:32 -04:00
|
|
|
|
return await self.db_pool.simple_select_one_onecol(
|
2019-09-04 08:04:27 -04:00
|
|
|
|
"%s_current" % (table,),
|
|
|
|
|
keyvalues={id_col: id},
|
|
|
|
|
retcol="completed_delta_stream_id",
|
2019-05-21 12:36:50 -04:00
|
|
|
|
allow_none=True,
|
2022-06-16 13:50:46 -04:00
|
|
|
|
desc="get_earliest_token_for_stats",
|
2019-05-21 12:36:50 -04:00
|
|
|
|
)
|
|
|
|
|
|
2020-08-27 17:24:37 -04:00
|
|
|
|
async def bulk_update_stats_delta(
|
2021-01-26 10:50:21 -05:00
|
|
|
|
self, ts: int, updates: Dict[str, Dict[str, Counter[str]]], stream_id: int
|
2020-08-27 17:24:37 -04:00
|
|
|
|
) -> None:
|
2019-09-04 08:04:27 -04:00
|
|
|
|
"""Bulk update stats tables for a given stream_id and updates the stats
|
|
|
|
|
incremental position.
|
|
|
|
|
|
|
|
|
|
Args:
|
2020-08-27 17:24:37 -04:00
|
|
|
|
ts: Current timestamp in ms
|
|
|
|
|
updates: The updates to commit as a mapping of
|
|
|
|
|
stats_type -> stats_id -> field -> delta.
|
|
|
|
|
stream_id: Current position.
|
2019-09-04 08:04:27 -04:00
|
|
|
|
"""
|
|
|
|
|
|
2021-12-29 08:01:13 -05:00
|
|
|
|
def _bulk_update_stats_delta_txn(txn: LoggingTransaction) -> None:
|
2019-09-04 08:04:27 -04:00
|
|
|
|
for stats_type, stats_updates in updates.items():
|
|
|
|
|
for stats_id, fields in stats_updates.items():
|
2019-10-24 12:55:53 -04:00
|
|
|
|
logger.debug(
|
2019-10-07 09:41:39 -04:00
|
|
|
|
"Updating %s stats for %s: %s", stats_type, stats_id, fields
|
|
|
|
|
)
|
2019-09-04 08:04:27 -04:00
|
|
|
|
self._update_stats_delta_txn(
|
|
|
|
|
txn,
|
|
|
|
|
ts=ts,
|
|
|
|
|
stats_type=stats_type,
|
|
|
|
|
stats_id=stats_id,
|
|
|
|
|
fields=fields,
|
|
|
|
|
complete_with_stream_id=stream_id,
|
|
|
|
|
)
|
|
|
|
|
|
2020-08-05 16:38:57 -04:00
|
|
|
|
self.db_pool.simple_update_one_txn(
|
2019-09-04 08:04:27 -04:00
|
|
|
|
txn,
|
|
|
|
|
table="stats_incremental_position",
|
|
|
|
|
keyvalues={},
|
|
|
|
|
updatevalues={"stream_id": stream_id},
|
|
|
|
|
)
|
|
|
|
|
|
2020-08-27 17:24:37 -04:00
|
|
|
|
await self.db_pool.runInteraction(
|
2019-09-04 08:04:27 -04:00
|
|
|
|
"bulk_update_stats_delta", _bulk_update_stats_delta_txn
|
2019-05-21 12:36:50 -04:00
|
|
|
|
)
|
|
|
|
|
|
2020-08-27 17:24:37 -04:00
|
|
|
|
async def update_stats_delta(
|
2019-09-04 08:04:27 -04:00
|
|
|
|
self,
|
2020-08-27 17:24:37 -04:00
|
|
|
|
ts: int,
|
|
|
|
|
stats_type: str,
|
|
|
|
|
stats_id: str,
|
|
|
|
|
fields: Dict[str, int],
|
2021-12-29 08:01:13 -05:00
|
|
|
|
complete_with_stream_id: int,
|
2020-08-27 17:24:37 -04:00
|
|
|
|
absolute_field_overrides: Optional[Dict[str, int]] = None,
|
|
|
|
|
) -> None:
|
2019-09-04 08:04:27 -04:00
|
|
|
|
"""
|
|
|
|
|
Updates the statistics for a subject, with a delta (difference/relative
|
|
|
|
|
change).
|
|
|
|
|
|
|
|
|
|
Args:
|
2020-08-27 17:24:37 -04:00
|
|
|
|
ts: timestamp of the change
|
|
|
|
|
stats_type: "room" or "user" – the kind of subject
|
|
|
|
|
stats_id: the subject's ID (room ID or user ID)
|
|
|
|
|
fields: Deltas of stats values.
|
|
|
|
|
complete_with_stream_id:
|
2019-09-04 08:04:27 -04:00
|
|
|
|
If supplied, converts an incomplete row into a complete row,
|
|
|
|
|
with the supplied stream_id marked as the stream_id where the
|
|
|
|
|
row was completed.
|
2020-08-27 17:24:37 -04:00
|
|
|
|
absolute_field_overrides: Current stats values (i.e. not deltas) of
|
|
|
|
|
absolute fields. Does not work with per-slice fields.
|
2019-09-04 08:04:27 -04:00
|
|
|
|
"""
|
|
|
|
|
|
2020-08-27 17:24:37 -04:00
|
|
|
|
await self.db_pool.runInteraction(
|
2019-09-04 08:04:27 -04:00
|
|
|
|
"update_stats_delta",
|
|
|
|
|
self._update_stats_delta_txn,
|
|
|
|
|
ts,
|
|
|
|
|
stats_type,
|
|
|
|
|
stats_id,
|
|
|
|
|
fields,
|
|
|
|
|
complete_with_stream_id=complete_with_stream_id,
|
|
|
|
|
absolute_field_overrides=absolute_field_overrides,
|
2019-05-21 12:36:50 -04:00
|
|
|
|
)
|
|
|
|
|
|
2019-09-04 08:04:27 -04:00
|
|
|
|
def _update_stats_delta_txn(
|
|
|
|
|
self,
|
2021-12-29 08:01:13 -05:00
|
|
|
|
txn: LoggingTransaction,
|
|
|
|
|
ts: int,
|
|
|
|
|
stats_type: str,
|
|
|
|
|
stats_id: str,
|
|
|
|
|
fields: Dict[str, int],
|
|
|
|
|
complete_with_stream_id: int,
|
|
|
|
|
absolute_field_overrides: Optional[Dict[str, int]] = None,
|
|
|
|
|
) -> None:
|
2019-09-04 08:04:27 -04:00
|
|
|
|
if absolute_field_overrides is None:
|
|
|
|
|
absolute_field_overrides = {}
|
|
|
|
|
|
|
|
|
|
table, id_col = TYPE_TO_TABLE[stats_type]
|
|
|
|
|
|
|
|
|
|
# Lets be paranoid and check that all the given field names are known
|
|
|
|
|
abs_field_names = ABSOLUTE_STATS_FIELDS[stats_type]
|
|
|
|
|
for field in chain(fields.keys(), absolute_field_overrides.keys()):
|
2021-07-08 11:57:13 -04:00
|
|
|
|
if field not in abs_field_names:
|
2019-09-04 08:04:27 -04:00
|
|
|
|
# guard against potential SQL injection dodginess
|
|
|
|
|
raise ValueError(
|
|
|
|
|
"%s is not a recognised field"
|
|
|
|
|
" for stats type %s" % (field, stats_type)
|
2019-05-21 12:36:50 -04:00
|
|
|
|
)
|
2019-09-04 08:04:27 -04:00
|
|
|
|
|
|
|
|
|
# Per slice fields do not get added to the _current table
|
|
|
|
|
|
|
|
|
|
# This calculates the deltas (`field = field + ?` values)
|
|
|
|
|
# for absolute fields,
|
|
|
|
|
# * defaulting to 0 if not specified
|
|
|
|
|
# (required for the INSERT part of upserting to work)
|
|
|
|
|
# * omitting overrides specified in `absolute_field_overrides`
|
|
|
|
|
deltas_of_absolute_fields = {
|
|
|
|
|
key: fields.get(key, 0)
|
|
|
|
|
for key in abs_field_names
|
|
|
|
|
if key not in absolute_field_overrides
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
# Keep the delta stream ID field up to date
|
|
|
|
|
absolute_field_overrides = absolute_field_overrides.copy()
|
|
|
|
|
absolute_field_overrides["completed_delta_stream_id"] = complete_with_stream_id
|
|
|
|
|
|
|
|
|
|
# first upsert the `_current` table
|
|
|
|
|
self._upsert_with_additive_relatives_txn(
|
|
|
|
|
txn=txn,
|
|
|
|
|
table=table + "_current",
|
|
|
|
|
keyvalues={id_col: stats_id},
|
|
|
|
|
absolutes=absolute_field_overrides,
|
|
|
|
|
additive_relatives=deltas_of_absolute_fields,
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
def _upsert_with_additive_relatives_txn(
|
2021-12-29 08:01:13 -05:00
|
|
|
|
self,
|
|
|
|
|
txn: LoggingTransaction,
|
|
|
|
|
table: str,
|
|
|
|
|
keyvalues: Dict[str, Any],
|
|
|
|
|
absolutes: Dict[str, Any],
|
|
|
|
|
additive_relatives: Dict[str, int],
|
|
|
|
|
) -> None:
|
2019-09-04 08:04:27 -04:00
|
|
|
|
"""Used to update values in the stats tables.
|
|
|
|
|
|
|
|
|
|
This is basically a slightly convoluted upsert that *adds* to any
|
|
|
|
|
existing rows.
|
|
|
|
|
|
|
|
|
|
Args:
|
2021-12-29 08:01:13 -05:00
|
|
|
|
table: Table name
|
|
|
|
|
keyvalues: Row-identifying key values
|
|
|
|
|
absolutes: Absolute (set) fields
|
|
|
|
|
additive_relatives: Fields that will be added onto if existing row present.
|
2019-09-04 08:04:27 -04:00
|
|
|
|
"""
|
2022-09-09 06:14:10 -04:00
|
|
|
|
absolute_updates = [
|
|
|
|
|
"%(field)s = EXCLUDED.%(field)s" % {"field": field}
|
|
|
|
|
for field in absolutes.keys()
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
relative_updates = [
|
|
|
|
|
"%(field)s = EXCLUDED.%(field)s + COALESCE(%(table)s.%(field)s, 0)"
|
|
|
|
|
% {"table": table, "field": field}
|
|
|
|
|
for field in additive_relatives.keys()
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
insert_cols = []
|
|
|
|
|
qargs = []
|
|
|
|
|
|
|
|
|
|
for (key, val) in chain(
|
|
|
|
|
keyvalues.items(), absolutes.items(), additive_relatives.items()
|
|
|
|
|
):
|
|
|
|
|
insert_cols.append(key)
|
|
|
|
|
qargs.append(val)
|
|
|
|
|
|
|
|
|
|
sql = """
|
|
|
|
|
INSERT INTO %(table)s (%(insert_cols_cs)s)
|
|
|
|
|
VALUES (%(insert_vals_qs)s)
|
|
|
|
|
ON CONFLICT (%(key_columns)s) DO UPDATE SET %(updates)s
|
|
|
|
|
""" % {
|
|
|
|
|
"table": table,
|
|
|
|
|
"insert_cols_cs": ", ".join(insert_cols),
|
|
|
|
|
"insert_vals_qs": ", ".join(
|
|
|
|
|
["?"] * (len(keyvalues) + len(absolutes) + len(additive_relatives))
|
|
|
|
|
),
|
|
|
|
|
"key_columns": ", ".join(keyvalues),
|
|
|
|
|
"updates": ", ".join(chain(absolute_updates, relative_updates)),
|
|
|
|
|
}
|
2019-09-04 08:04:27 -04:00
|
|
|
|
|
2022-09-09 06:14:10 -04:00
|
|
|
|
txn.execute(sql, qargs)
|
2019-09-04 08:04:27 -04:00
|
|
|
|
|
2021-12-29 08:01:13 -05:00
|
|
|
|
async def _calculate_and_set_initial_state_for_room(self, room_id: str) -> None:
|
2019-09-04 08:04:27 -04:00
|
|
|
|
"""Calculate and insert an entry into room_stats_current.
|
|
|
|
|
|
|
|
|
|
Args:
|
2020-07-30 07:20:41 -04:00
|
|
|
|
room_id: The room ID under calculation.
|
2019-09-04 08:04:27 -04:00
|
|
|
|
"""
|
|
|
|
|
|
2021-12-29 08:01:13 -05:00
|
|
|
|
def _fetch_current_state_stats(
|
|
|
|
|
txn: LoggingTransaction,
|
|
|
|
|
) -> Tuple[List[str], Dict[str, int], int, List[str], int]:
|
|
|
|
|
pos = self.get_room_max_stream_ordering() # type: ignore[attr-defined]
|
2019-09-04 08:04:27 -04:00
|
|
|
|
|
2020-08-05 16:38:57 -04:00
|
|
|
|
rows = self.db_pool.simple_select_many_txn(
|
2019-09-04 08:04:27 -04:00
|
|
|
|
txn,
|
|
|
|
|
table="current_state_events",
|
|
|
|
|
column="type",
|
|
|
|
|
iterable=[
|
|
|
|
|
EventTypes.Create,
|
|
|
|
|
EventTypes.JoinRules,
|
|
|
|
|
EventTypes.RoomHistoryVisibility,
|
2020-02-04 12:25:54 -05:00
|
|
|
|
EventTypes.RoomEncryption,
|
2019-09-04 08:04:27 -04:00
|
|
|
|
EventTypes.Name,
|
|
|
|
|
EventTypes.Topic,
|
|
|
|
|
EventTypes.RoomAvatar,
|
|
|
|
|
EventTypes.CanonicalAlias,
|
|
|
|
|
],
|
|
|
|
|
keyvalues={"room_id": room_id, "state_key": ""},
|
|
|
|
|
retcols=["event_id"],
|
|
|
|
|
)
|
|
|
|
|
|
2021-12-29 08:01:13 -05:00
|
|
|
|
event_ids = cast(List[str], [row["event_id"] for row in rows])
|
2019-09-04 08:04:27 -04:00
|
|
|
|
|
|
|
|
|
txn.execute(
|
|
|
|
|
"""
|
|
|
|
|
SELECT membership, count(*) FROM current_state_events
|
|
|
|
|
WHERE room_id = ? AND type = 'm.room.member'
|
|
|
|
|
GROUP BY membership
|
|
|
|
|
""",
|
|
|
|
|
(room_id,),
|
|
|
|
|
)
|
|
|
|
|
membership_counts = {membership: cnt for membership, cnt in txn}
|
|
|
|
|
|
|
|
|
|
txn.execute(
|
|
|
|
|
"""
|
2021-12-14 07:34:30 -05:00
|
|
|
|
SELECT COUNT(*) FROM current_state_events
|
2019-09-04 08:04:27 -04:00
|
|
|
|
WHERE room_id = ?
|
|
|
|
|
""",
|
|
|
|
|
(room_id,),
|
|
|
|
|
)
|
|
|
|
|
|
2021-12-29 08:01:13 -05:00
|
|
|
|
current_state_events_count = cast(Tuple[int], txn.fetchone())[0]
|
2019-09-04 08:04:27 -04:00
|
|
|
|
|
2021-12-29 08:01:13 -05:00
|
|
|
|
users_in_room = self.get_users_in_room_txn(txn, room_id) # type: ignore[attr-defined]
|
2019-09-04 08:04:27 -04:00
|
|
|
|
|
|
|
|
|
return (
|
|
|
|
|
event_ids,
|
|
|
|
|
membership_counts,
|
|
|
|
|
current_state_events_count,
|
|
|
|
|
users_in_room,
|
|
|
|
|
pos,
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
(
|
|
|
|
|
event_ids,
|
|
|
|
|
membership_counts,
|
|
|
|
|
current_state_events_count,
|
|
|
|
|
users_in_room,
|
|
|
|
|
pos,
|
2020-08-05 16:38:57 -04:00
|
|
|
|
) = await self.db_pool.runInteraction(
|
2019-09-04 08:04:27 -04:00
|
|
|
|
"get_initial_state_for_room", _fetch_current_state_stats
|
|
|
|
|
)
|
|
|
|
|
|
2023-01-23 06:36:20 -05:00
|
|
|
|
try:
|
|
|
|
|
state_event_map = await self.get_events(event_ids, get_prev_content=False) # type: ignore[attr-defined]
|
|
|
|
|
except InvalidEventError as e:
|
|
|
|
|
# If an exception occurs fetching events then the room is broken;
|
|
|
|
|
# skip process it to avoid being stuck on a room.
|
|
|
|
|
logger.warning(
|
|
|
|
|
"Failed to fetch events for room %s, skipping stats calculation: %r.",
|
|
|
|
|
room_id,
|
|
|
|
|
e,
|
|
|
|
|
)
|
|
|
|
|
return
|
2019-09-04 08:04:27 -04:00
|
|
|
|
|
2022-06-29 13:12:45 -04:00
|
|
|
|
room_state: Dict[str, Union[None, bool, str]] = {
|
2019-09-04 08:04:27 -04:00
|
|
|
|
"join_rules": None,
|
|
|
|
|
"history_visibility": None,
|
|
|
|
|
"encryption": None,
|
|
|
|
|
"name": None,
|
|
|
|
|
"topic": None,
|
|
|
|
|
"avatar": None,
|
|
|
|
|
"canonical_alias": None,
|
|
|
|
|
"is_federatable": True,
|
2022-06-29 13:12:45 -04:00
|
|
|
|
"room_type": None,
|
2019-09-04 08:04:27 -04:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
for event in state_event_map.values():
|
|
|
|
|
if event.type == EventTypes.JoinRules:
|
|
|
|
|
room_state["join_rules"] = event.content.get("join_rule")
|
|
|
|
|
elif event.type == EventTypes.RoomHistoryVisibility:
|
|
|
|
|
room_state["history_visibility"] = event.content.get(
|
|
|
|
|
"history_visibility"
|
2019-05-21 12:36:50 -04:00
|
|
|
|
)
|
2020-02-04 12:25:54 -05:00
|
|
|
|
elif event.type == EventTypes.RoomEncryption:
|
2019-09-04 08:04:27 -04:00
|
|
|
|
room_state["encryption"] = event.content.get("algorithm")
|
|
|
|
|
elif event.type == EventTypes.Name:
|
|
|
|
|
room_state["name"] = event.content.get("name")
|
|
|
|
|
elif event.type == EventTypes.Topic:
|
|
|
|
|
room_state["topic"] = event.content.get("topic")
|
|
|
|
|
elif event.type == EventTypes.RoomAvatar:
|
|
|
|
|
room_state["avatar"] = event.content.get("url")
|
|
|
|
|
elif event.type == EventTypes.CanonicalAlias:
|
|
|
|
|
room_state["canonical_alias"] = event.content.get("alias")
|
|
|
|
|
elif event.type == EventTypes.Create:
|
2019-09-06 09:09:16 -04:00
|
|
|
|
room_state["is_federatable"] = (
|
2021-09-08 10:00:43 -04:00
|
|
|
|
event.content.get(EventContentFields.FEDERATE, True) is True
|
2019-09-06 09:09:16 -04:00
|
|
|
|
)
|
2022-06-29 13:12:45 -04:00
|
|
|
|
room_type = event.content.get(EventContentFields.ROOM_TYPE)
|
|
|
|
|
if isinstance(room_type, str):
|
|
|
|
|
room_state["room_type"] = room_type
|
2019-09-04 08:04:27 -04:00
|
|
|
|
|
2020-07-30 07:20:41 -04:00
|
|
|
|
await self.update_room_state(room_id, room_state)
|
2019-09-04 08:04:27 -04:00
|
|
|
|
|
|
|
|
|
local_users_in_room = [u for u in users_in_room if self.hs.is_mine_id(u)]
|
|
|
|
|
|
2020-07-30 07:20:41 -04:00
|
|
|
|
await self.update_stats_delta(
|
2019-09-04 08:04:27 -04:00
|
|
|
|
ts=self.clock.time_msec(),
|
|
|
|
|
stats_type="room",
|
|
|
|
|
stats_id=room_id,
|
|
|
|
|
fields={},
|
|
|
|
|
complete_with_stream_id=pos,
|
|
|
|
|
absolute_field_overrides={
|
|
|
|
|
"current_state_events": current_state_events_count,
|
|
|
|
|
"joined_members": membership_counts.get(Membership.JOIN, 0),
|
|
|
|
|
"invited_members": membership_counts.get(Membership.INVITE, 0),
|
|
|
|
|
"left_members": membership_counts.get(Membership.LEAVE, 0),
|
|
|
|
|
"banned_members": membership_counts.get(Membership.BAN, 0),
|
2021-07-09 09:12:47 -04:00
|
|
|
|
"knocked_members": membership_counts.get(Membership.KNOCK, 0),
|
2019-09-04 08:04:27 -04:00
|
|
|
|
"local_users_in_room": len(local_users_in_room),
|
|
|
|
|
},
|
|
|
|
|
)
|
|
|
|
|
|
2021-12-29 08:01:13 -05:00
|
|
|
|
async def _calculate_and_set_initial_state_for_user(self, user_id: str) -> None:
|
|
|
|
|
def _calculate_and_set_initial_state_for_user_txn(
|
|
|
|
|
txn: LoggingTransaction,
|
|
|
|
|
) -> Tuple[int, int]:
|
2019-09-04 08:04:27 -04:00
|
|
|
|
pos = self._get_max_stream_id_in_current_state_deltas_txn(txn)
|
2019-05-21 12:36:50 -04:00
|
|
|
|
|
2019-09-04 08:04:27 -04:00
|
|
|
|
txn.execute(
|
|
|
|
|
"""
|
|
|
|
|
SELECT COUNT(distinct room_id) FROM current_state_events
|
|
|
|
|
WHERE type = 'm.room.member' AND state_key = ?
|
|
|
|
|
AND membership = 'join'
|
|
|
|
|
""",
|
|
|
|
|
(user_id,),
|
|
|
|
|
)
|
2021-12-29 08:01:13 -05:00
|
|
|
|
count = cast(Tuple[int], txn.fetchone())[0]
|
2019-09-04 08:04:27 -04:00
|
|
|
|
return count, pos
|
|
|
|
|
|
2020-08-05 16:38:57 -04:00
|
|
|
|
joined_rooms, pos = await self.db_pool.runInteraction(
|
2019-09-04 08:04:27 -04:00
|
|
|
|
"calculate_and_set_initial_state_for_user",
|
|
|
|
|
_calculate_and_set_initial_state_for_user_txn,
|
|
|
|
|
)
|
|
|
|
|
|
2020-07-30 07:20:41 -04:00
|
|
|
|
await self.update_stats_delta(
|
2019-09-04 08:04:27 -04:00
|
|
|
|
ts=self.clock.time_msec(),
|
|
|
|
|
stats_type="user",
|
|
|
|
|
stats_id=user_id,
|
|
|
|
|
fields={},
|
|
|
|
|
complete_with_stream_id=pos,
|
|
|
|
|
absolute_field_overrides={"joined_rooms": joined_rooms},
|
|
|
|
|
)
|
2020-11-05 13:59:12 -05:00
|
|
|
|
|
|
|
|
|
async def get_users_media_usage_paginate(
|
|
|
|
|
self,
|
|
|
|
|
start: int,
|
|
|
|
|
limit: int,
|
|
|
|
|
from_ts: Optional[int] = None,
|
|
|
|
|
until_ts: Optional[int] = None,
|
2021-07-21 09:47:56 -04:00
|
|
|
|
order_by: Optional[str] = UserSortOrder.USER_ID.value,
|
2020-11-05 13:59:12 -05:00
|
|
|
|
direction: Optional[str] = "f",
|
|
|
|
|
search_term: Optional[str] = None,
|
2021-07-21 14:12:22 -04:00
|
|
|
|
) -> Tuple[List[JsonDict], int]:
|
2020-11-05 13:59:12 -05:00
|
|
|
|
"""Function to retrieve a paginated list of users and their uploaded local media
|
|
|
|
|
(size and number). This will return a json list of users and the
|
|
|
|
|
total number of users matching the filter criteria.
|
|
|
|
|
|
|
|
|
|
Args:
|
|
|
|
|
start: offset to begin the query from
|
|
|
|
|
limit: number of rows to retrieve
|
|
|
|
|
from_ts: request only media that are created later than this timestamp (ms)
|
|
|
|
|
until_ts: request only media that are created earlier than this timestamp (ms)
|
|
|
|
|
order_by: the sort order of the returned list
|
|
|
|
|
direction: sort ascending or descending
|
|
|
|
|
search_term: a string to filter user names by
|
|
|
|
|
Returns:
|
|
|
|
|
A list of user dicts and an integer representing the total number of
|
|
|
|
|
users that exist given this query
|
|
|
|
|
"""
|
|
|
|
|
|
2021-12-29 08:01:13 -05:00
|
|
|
|
def get_users_media_usage_paginate_txn(
|
|
|
|
|
txn: LoggingTransaction,
|
|
|
|
|
) -> Tuple[List[JsonDict], int]:
|
2020-11-05 13:59:12 -05:00
|
|
|
|
filters = []
|
2021-09-13 13:07:12 -04:00
|
|
|
|
args = [self.hs.config.server.server_name]
|
2020-11-05 13:59:12 -05:00
|
|
|
|
|
|
|
|
|
if search_term:
|
|
|
|
|
filters.append("(lmr.user_id LIKE ? OR displayname LIKE ?)")
|
|
|
|
|
args.extend(["@%" + search_term + "%:%", "%" + search_term + "%"])
|
|
|
|
|
|
|
|
|
|
if from_ts:
|
|
|
|
|
filters.append("created_ts >= ?")
|
|
|
|
|
args.extend([from_ts])
|
|
|
|
|
if until_ts:
|
|
|
|
|
filters.append("created_ts <= ?")
|
|
|
|
|
args.extend([until_ts])
|
|
|
|
|
|
|
|
|
|
# Set ordering
|
|
|
|
|
if UserSortOrder(order_by) == UserSortOrder.MEDIA_LENGTH:
|
|
|
|
|
order_by_column = "media_length"
|
|
|
|
|
elif UserSortOrder(order_by) == UserSortOrder.MEDIA_COUNT:
|
|
|
|
|
order_by_column = "media_count"
|
|
|
|
|
elif UserSortOrder(order_by) == UserSortOrder.USER_ID:
|
|
|
|
|
order_by_column = "lmr.user_id"
|
|
|
|
|
elif UserSortOrder(order_by) == UserSortOrder.DISPLAYNAME:
|
|
|
|
|
order_by_column = "displayname"
|
|
|
|
|
else:
|
|
|
|
|
raise StoreError(
|
|
|
|
|
500, "Incorrect value for order_by provided: %s" % order_by
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
if direction == "b":
|
|
|
|
|
order = "DESC"
|
|
|
|
|
else:
|
|
|
|
|
order = "ASC"
|
|
|
|
|
|
|
|
|
|
where_clause = "WHERE " + " AND ".join(filters) if len(filters) > 0 else ""
|
|
|
|
|
|
|
|
|
|
sql_base = """
|
|
|
|
|
FROM local_media_repository as lmr
|
|
|
|
|
LEFT JOIN profiles AS p ON lmr.user_id = '@' || p.user_id || ':' || ?
|
|
|
|
|
{}
|
|
|
|
|
GROUP BY lmr.user_id, displayname
|
|
|
|
|
""".format(
|
|
|
|
|
where_clause
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
# SQLite does not support SELECT COUNT(*) OVER()
|
|
|
|
|
sql = """
|
|
|
|
|
SELECT COUNT(*) FROM (
|
|
|
|
|
SELECT lmr.user_id
|
|
|
|
|
{sql_base}
|
|
|
|
|
) AS count_user_ids
|
|
|
|
|
""".format(
|
|
|
|
|
sql_base=sql_base,
|
|
|
|
|
)
|
|
|
|
|
txn.execute(sql, args)
|
2021-12-29 08:01:13 -05:00
|
|
|
|
count = cast(Tuple[int], txn.fetchone())[0]
|
2020-11-05 13:59:12 -05:00
|
|
|
|
|
|
|
|
|
sql = """
|
|
|
|
|
SELECT
|
|
|
|
|
lmr.user_id,
|
|
|
|
|
displayname,
|
|
|
|
|
COUNT(lmr.user_id) as media_count,
|
|
|
|
|
SUM(media_length) as media_length
|
|
|
|
|
{sql_base}
|
|
|
|
|
ORDER BY {order_by_column} {order}
|
|
|
|
|
LIMIT ? OFFSET ?
|
|
|
|
|
""".format(
|
2021-02-16 17:32:34 -05:00
|
|
|
|
sql_base=sql_base,
|
|
|
|
|
order_by_column=order_by_column,
|
|
|
|
|
order=order,
|
2020-11-05 13:59:12 -05:00
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
args += [limit, start]
|
|
|
|
|
txn.execute(sql, args)
|
|
|
|
|
users = self.db_pool.cursor_to_dict(txn)
|
|
|
|
|
|
|
|
|
|
return users, count
|
|
|
|
|
|
|
|
|
|
return await self.db_pool.runInteraction(
|
|
|
|
|
"get_users_media_usage_paginate_txn", get_users_media_usage_paginate_txn
|
|
|
|
|
)
|