mirror of
https://git.anonymousland.org/anonymousland/synapse.git
synced 2024-12-29 13:26:12 -05:00
748d8fdc7b
Some streams will occaisonally advance their positions without actually having any new rows to send over federation. Currently this means that the token will not advance on the workers, leading to them repeatedly sending a slightly out of date token. This in turns requires the master to hit the DB to check if there are any new rows, rather than hitting the no op logic where we check if the given token matches the current token. This commit changes the API to always return an entry if the position for a stream has changed, allowing workers to advance their tokens correctly.
300 lines
10 KiB
Python
300 lines
10 KiB
Python
#!/usr/bin/env python
|
|
# -*- coding: utf-8 -*-
|
|
# Copyright 2016 OpenMarket Ltd
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
|
|
import synapse
|
|
|
|
from synapse.server import HomeServer
|
|
from synapse.config._base import ConfigError
|
|
from synapse.config.logger import setup_logging
|
|
from synapse.config.homeserver import HomeServerConfig
|
|
from synapse.http.site import SynapseSite
|
|
from synapse.metrics.resource import MetricsResource, METRICS_PREFIX
|
|
from synapse.storage.roommember import RoomMemberStore
|
|
from synapse.replication.slave.storage.events import SlavedEventStore
|
|
from synapse.replication.slave.storage.pushers import SlavedPusherStore
|
|
from synapse.replication.slave.storage.receipts import SlavedReceiptsStore
|
|
from synapse.replication.slave.storage.account_data import SlavedAccountDataStore
|
|
from synapse.storage.engines import create_engine
|
|
from synapse.storage import DataStore
|
|
from synapse.util.async import sleep
|
|
from synapse.util.httpresourcetree import create_resource_tree
|
|
from synapse.util.logcontext import LoggingContext, preserve_fn
|
|
from synapse.util.manhole import manhole
|
|
from synapse.util.rlimit import change_resource_limit
|
|
from synapse.util.versionstring import get_version_string
|
|
|
|
from twisted.internet import reactor, defer
|
|
from twisted.web.resource import Resource
|
|
|
|
from daemonize import Daemonize
|
|
|
|
import sys
|
|
import logging
|
|
import gc
|
|
|
|
logger = logging.getLogger("synapse.app.pusher")
|
|
|
|
|
|
class PusherSlaveStore(
|
|
SlavedEventStore, SlavedPusherStore, SlavedReceiptsStore,
|
|
SlavedAccountDataStore
|
|
):
|
|
update_pusher_last_stream_ordering_and_success = (
|
|
DataStore.update_pusher_last_stream_ordering_and_success.__func__
|
|
)
|
|
|
|
update_pusher_failing_since = (
|
|
DataStore.update_pusher_failing_since.__func__
|
|
)
|
|
|
|
update_pusher_last_stream_ordering = (
|
|
DataStore.update_pusher_last_stream_ordering.__func__
|
|
)
|
|
|
|
get_throttle_params_by_room = (
|
|
DataStore.get_throttle_params_by_room.__func__
|
|
)
|
|
|
|
set_throttle_params = (
|
|
DataStore.set_throttle_params.__func__
|
|
)
|
|
|
|
get_time_of_last_push_action_before = (
|
|
DataStore.get_time_of_last_push_action_before.__func__
|
|
)
|
|
|
|
get_profile_displayname = (
|
|
DataStore.get_profile_displayname.__func__
|
|
)
|
|
|
|
who_forgot_in_room = (
|
|
RoomMemberStore.__dict__["who_forgot_in_room"]
|
|
)
|
|
|
|
|
|
class PusherServer(HomeServer):
|
|
|
|
def get_db_conn(self, run_new_connection=True):
|
|
# Any param beginning with cp_ is a parameter for adbapi, and should
|
|
# not be passed to the database engine.
|
|
db_params = {
|
|
k: v for k, v in self.db_config.get("args", {}).items()
|
|
if not k.startswith("cp_")
|
|
}
|
|
db_conn = self.database_engine.module.connect(**db_params)
|
|
|
|
if run_new_connection:
|
|
self.database_engine.on_new_connection(db_conn)
|
|
return db_conn
|
|
|
|
def setup(self):
|
|
logger.info("Setting up.")
|
|
self.datastore = PusherSlaveStore(self.get_db_conn(), self)
|
|
logger.info("Finished setting up.")
|
|
|
|
def remove_pusher(self, app_id, push_key, user_id):
|
|
http_client = self.get_simple_http_client()
|
|
replication_url = self.config.worker_replication_url
|
|
url = replication_url + "/remove_pushers"
|
|
return http_client.post_json_get_json(url, {
|
|
"remove": [{
|
|
"app_id": app_id,
|
|
"push_key": push_key,
|
|
"user_id": user_id,
|
|
}]
|
|
})
|
|
|
|
def _listen_http(self, listener_config):
|
|
port = listener_config["port"]
|
|
bind_address = listener_config.get("bind_address", "")
|
|
site_tag = listener_config.get("tag", port)
|
|
resources = {}
|
|
for res in listener_config["resources"]:
|
|
for name in res["names"]:
|
|
if name == "metrics":
|
|
resources[METRICS_PREFIX] = MetricsResource(self)
|
|
|
|
root_resource = create_resource_tree(resources, Resource())
|
|
reactor.listenTCP(
|
|
port,
|
|
SynapseSite(
|
|
"synapse.access.http.%s" % (site_tag,),
|
|
site_tag,
|
|
listener_config,
|
|
root_resource,
|
|
),
|
|
interface=bind_address
|
|
)
|
|
logger.info("Synapse pusher now listening on port %d", port)
|
|
|
|
def start_listening(self, listeners):
|
|
for listener in listeners:
|
|
if listener["type"] == "http":
|
|
self._listen_http(listener)
|
|
elif listener["type"] == "manhole":
|
|
reactor.listenTCP(
|
|
listener["port"],
|
|
manhole(
|
|
username="matrix",
|
|
password="rabbithole",
|
|
globals={"hs": self},
|
|
),
|
|
interface=listener.get("bind_address", '127.0.0.1')
|
|
)
|
|
else:
|
|
logger.warn("Unrecognized listener type: %s", listener["type"])
|
|
|
|
@defer.inlineCallbacks
|
|
def replicate(self):
|
|
http_client = self.get_simple_http_client()
|
|
store = self.get_datastore()
|
|
replication_url = self.config.worker_replication_url
|
|
pusher_pool = self.get_pusherpool()
|
|
|
|
def stop_pusher(user_id, app_id, pushkey):
|
|
key = "%s:%s" % (app_id, pushkey)
|
|
pushers_for_user = pusher_pool.pushers.get(user_id, {})
|
|
pusher = pushers_for_user.pop(key, None)
|
|
if pusher is None:
|
|
return
|
|
logger.info("Stopping pusher %r / %r", user_id, key)
|
|
pusher.on_stop()
|
|
|
|
def start_pusher(user_id, app_id, pushkey):
|
|
key = "%s:%s" % (app_id, pushkey)
|
|
logger.info("Starting pusher %r / %r", user_id, key)
|
|
return pusher_pool._refresh_pusher(app_id, pushkey, user_id)
|
|
|
|
@defer.inlineCallbacks
|
|
def poke_pushers(results):
|
|
pushers_rows = set(
|
|
map(tuple, results.get("pushers", {}).get("rows", []))
|
|
)
|
|
deleted_pushers_rows = set(
|
|
map(tuple, results.get("deleted_pushers", {}).get("rows", []))
|
|
)
|
|
for row in sorted(pushers_rows | deleted_pushers_rows):
|
|
if row in deleted_pushers_rows:
|
|
user_id, app_id, pushkey = row[1:4]
|
|
stop_pusher(user_id, app_id, pushkey)
|
|
elif row in pushers_rows:
|
|
user_id = row[1]
|
|
app_id = row[5]
|
|
pushkey = row[8]
|
|
yield start_pusher(user_id, app_id, pushkey)
|
|
|
|
stream = results.get("events")
|
|
if stream and stream["rows"]:
|
|
min_stream_id = stream["rows"][0][0]
|
|
max_stream_id = stream["position"]
|
|
preserve_fn(pusher_pool.on_new_notifications)(
|
|
min_stream_id, max_stream_id
|
|
)
|
|
|
|
stream = results.get("receipts")
|
|
if stream and stream["rows"]:
|
|
rows = stream["rows"]
|
|
affected_room_ids = set(row[1] for row in rows)
|
|
min_stream_id = rows[0][0]
|
|
max_stream_id = stream["position"]
|
|
preserve_fn(pusher_pool.on_new_receipts)(
|
|
min_stream_id, max_stream_id, affected_room_ids
|
|
)
|
|
|
|
while True:
|
|
try:
|
|
args = store.stream_positions()
|
|
args["timeout"] = 30000
|
|
result = yield http_client.get_json(replication_url, args=args)
|
|
yield store.process_replication(result)
|
|
poke_pushers(result)
|
|
except:
|
|
logger.exception("Error replicating from %r", replication_url)
|
|
yield sleep(30)
|
|
|
|
|
|
def start(config_options):
|
|
try:
|
|
config = HomeServerConfig.load_config(
|
|
"Synapse pusher", config_options
|
|
)
|
|
except ConfigError as e:
|
|
sys.stderr.write("\n" + e.message + "\n")
|
|
sys.exit(1)
|
|
|
|
assert config.worker_app == "synapse.app.pusher"
|
|
|
|
setup_logging(config.worker_log_config, config.worker_log_file)
|
|
|
|
if config.start_pushers:
|
|
sys.stderr.write(
|
|
"\nThe pushers must be disabled in the main synapse process"
|
|
"\nbefore they can be run in a separate worker."
|
|
"\nPlease add ``start_pushers: false`` to the main config"
|
|
"\n"
|
|
)
|
|
sys.exit(1)
|
|
|
|
# Force the pushers to start since they will be disabled in the main config
|
|
config.start_pushers = True
|
|
|
|
database_engine = create_engine(config.database_config)
|
|
|
|
ps = PusherServer(
|
|
config.server_name,
|
|
db_config=config.database_config,
|
|
config=config,
|
|
version_string="Synapse/" + get_version_string(synapse),
|
|
database_engine=database_engine,
|
|
)
|
|
|
|
ps.setup()
|
|
ps.start_listening(config.worker_listeners)
|
|
|
|
def run():
|
|
with LoggingContext("run"):
|
|
logger.info("Running")
|
|
change_resource_limit(config.soft_file_limit)
|
|
if config.gc_thresholds:
|
|
gc.set_threshold(*config.gc_thresholds)
|
|
reactor.run()
|
|
|
|
def start():
|
|
ps.replicate()
|
|
ps.get_pusherpool().start()
|
|
ps.get_datastore().start_profiling()
|
|
ps.get_state_handler().start_caching()
|
|
|
|
reactor.callWhenRunning(start)
|
|
|
|
if config.worker_daemonize:
|
|
daemon = Daemonize(
|
|
app="synapse-pusher",
|
|
pid=config.worker_pid_file,
|
|
action=run,
|
|
auto_close_fds=False,
|
|
verbose=True,
|
|
logger=logger,
|
|
)
|
|
daemon.start()
|
|
else:
|
|
run()
|
|
|
|
|
|
if __name__ == '__main__':
|
|
with LoggingContext("main"):
|
|
ps = start(sys.argv[1:])
|