2016-01-06 23:26:29 -05:00
|
|
|
# Copyright 2015, 2016 OpenMarket Ltd
|
2015-02-03 09:44:16 -05:00
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
2018-07-09 02:09:20 -04:00
|
|
|
import logging
|
2021-12-14 12:02:46 -05:00
|
|
|
import urllib.parse
|
|
|
|
from typing import TYPE_CHECKING, Dict, Iterable, List, Optional, Tuple
|
2018-07-09 02:09:20 -04:00
|
|
|
|
|
|
|
from prometheus_client import Counter
|
|
|
|
|
2021-06-09 14:39:51 -04:00
|
|
|
from synapse.api.constants import EventTypes, Membership, ThirdPartyEntityKind
|
2015-02-04 12:32:44 -05:00
|
|
|
from synapse.api.errors import CodeMessageException
|
2020-10-15 12:33:28 -04:00
|
|
|
from synapse.events import EventBase
|
2015-02-05 08:42:35 -05:00
|
|
|
from synapse.events.utils import serialize_event
|
2018-07-09 02:09:20 -04:00
|
|
|
from synapse.http.client import SimpleHttpClient
|
2020-09-01 11:03:49 -04:00
|
|
|
from synapse.types import JsonDict, ThirdPartyInstanceID
|
2018-07-09 02:09:20 -04:00
|
|
|
from synapse.util.caches.response_cache import ResponseCache
|
2018-06-05 08:17:55 -04:00
|
|
|
|
2020-09-01 11:03:49 -04:00
|
|
|
if TYPE_CHECKING:
|
|
|
|
from synapse.appservice import ApplicationService
|
2021-10-22 13:15:41 -04:00
|
|
|
from synapse.server import HomeServer
|
2020-09-01 11:03:49 -04:00
|
|
|
|
2015-02-04 11:44:53 -05:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
2018-06-05 08:17:55 -04:00
|
|
|
sent_transactions_counter = Counter(
|
|
|
|
"synapse_appservice_api_sent_transactions",
|
|
|
|
"Number of /transactions/ requests sent",
|
|
|
|
["service"],
|
|
|
|
)
|
|
|
|
|
|
|
|
failed_transactions_counter = Counter(
|
|
|
|
"synapse_appservice_api_failed_transactions",
|
|
|
|
"Number of /transactions/ requests that failed to send",
|
|
|
|
["service"],
|
|
|
|
)
|
|
|
|
|
|
|
|
sent_events_counter = Counter(
|
|
|
|
"synapse_appservice_api_sent_events", "Number of events sent to the AS", ["service"]
|
|
|
|
)
|
2015-02-04 11:44:53 -05:00
|
|
|
|
2016-08-25 10:56:27 -04:00
|
|
|
HOUR_IN_MS = 60 * 60 * 1000
|
|
|
|
|
|
|
|
|
2016-08-25 13:06:29 -04:00
|
|
|
APP_SERVICE_PREFIX = "/_matrix/app/unstable"
|
|
|
|
|
|
|
|
|
2021-12-14 12:02:46 -05:00
|
|
|
def _is_valid_3pe_metadata(info: JsonDict) -> bool:
|
2016-09-09 10:09:46 -04:00
|
|
|
if "instances" not in info:
|
|
|
|
return False
|
|
|
|
if not isinstance(info["instances"], list):
|
|
|
|
return False
|
|
|
|
return True
|
|
|
|
|
|
|
|
|
2021-12-14 12:02:46 -05:00
|
|
|
def _is_valid_3pe_result(r: JsonDict, field: str) -> bool:
|
2016-08-18 12:33:56 -04:00
|
|
|
if not isinstance(r, dict):
|
|
|
|
return False
|
|
|
|
|
|
|
|
for k in (field, "protocol"):
|
|
|
|
if k not in r:
|
|
|
|
return False
|
|
|
|
if not isinstance(r[k], str):
|
|
|
|
return False
|
|
|
|
|
|
|
|
if "fields" not in r:
|
|
|
|
return False
|
|
|
|
fields = r["fields"]
|
|
|
|
if not isinstance(fields, dict):
|
|
|
|
return False
|
|
|
|
|
|
|
|
return True
|
|
|
|
|
|
|
|
|
2015-02-04 11:44:53 -05:00
|
|
|
class ApplicationServiceApi(SimpleHttpClient):
|
2015-02-04 06:19:18 -05:00
|
|
|
"""This class manages HS -> AS communications, including querying and
|
|
|
|
pushing.
|
|
|
|
"""
|
|
|
|
|
2021-10-22 13:15:41 -04:00
|
|
|
def __init__(self, hs: "HomeServer"):
|
2020-09-18 09:56:44 -04:00
|
|
|
super().__init__(hs)
|
2015-02-05 08:42:35 -05:00
|
|
|
self.clock = hs.get_clock()
|
2015-02-04 06:19:18 -05:00
|
|
|
|
2021-07-15 06:02:43 -04:00
|
|
|
self.protocol_meta_cache: ResponseCache[Tuple[str, str]] = ResponseCache(
|
2021-03-08 14:00:07 -05:00
|
|
|
hs.get_clock(), "as_protocol_meta", timeout_ms=HOUR_IN_MS
|
2021-07-15 06:02:43 -04:00
|
|
|
)
|
2016-08-25 10:56:27 -04:00
|
|
|
|
2021-12-14 12:02:46 -05:00
|
|
|
async def query_user(self, service: "ApplicationService", user_id: str) -> bool:
|
2016-08-30 12:16:00 -04:00
|
|
|
if service.url is None:
|
2019-07-23 09:00:55 -04:00
|
|
|
return False
|
2021-12-14 12:02:46 -05:00
|
|
|
|
|
|
|
# This is required by the configuration.
|
|
|
|
assert service.hs_token is not None
|
|
|
|
|
2018-09-05 10:10:47 -04:00
|
|
|
uri = service.url + ("/users/%s" % urllib.parse.quote(user_id))
|
2015-02-04 11:44:53 -05:00
|
|
|
try:
|
2020-07-30 07:27:39 -04:00
|
|
|
response = await self.get_json(uri, {"access_token": service.hs_token})
|
2015-02-09 10:01:28 -05:00
|
|
|
if response is not None: # just an empty json object
|
2019-07-23 09:00:55 -04:00
|
|
|
return True
|
2015-02-04 12:32:44 -05:00
|
|
|
except CodeMessageException as e:
|
|
|
|
if e.code == 404:
|
2019-07-23 09:00:55 -04:00
|
|
|
return False
|
2015-02-04 12:32:44 -05:00
|
|
|
logger.warning("query_user to %s received %s", uri, e.code)
|
2015-02-05 08:19:46 -05:00
|
|
|
except Exception as ex:
|
|
|
|
logger.warning("query_user to %s threw exception %s", uri, ex)
|
2019-07-23 09:00:55 -04:00
|
|
|
return False
|
2015-02-04 06:19:18 -05:00
|
|
|
|
2021-12-14 12:02:46 -05:00
|
|
|
async def query_alias(self, service: "ApplicationService", alias: str) -> bool:
|
2016-08-30 12:16:00 -04:00
|
|
|
if service.url is None:
|
2019-07-23 09:00:55 -04:00
|
|
|
return False
|
2021-12-14 12:02:46 -05:00
|
|
|
|
|
|
|
# This is required by the configuration.
|
|
|
|
assert service.hs_token is not None
|
|
|
|
|
2018-09-05 10:10:47 -04:00
|
|
|
uri = service.url + ("/rooms/%s" % urllib.parse.quote(alias))
|
2015-02-04 11:44:53 -05:00
|
|
|
try:
|
2020-07-30 07:27:39 -04:00
|
|
|
response = await self.get_json(uri, {"access_token": service.hs_token})
|
2015-02-09 10:01:28 -05:00
|
|
|
if response is not None: # just an empty json object
|
2019-07-23 09:00:55 -04:00
|
|
|
return True
|
2015-02-04 12:32:44 -05:00
|
|
|
except CodeMessageException as e:
|
2015-02-09 10:01:28 -05:00
|
|
|
logger.warning("query_alias to %s received %s", uri, e.code)
|
2015-02-04 12:32:44 -05:00
|
|
|
if e.code == 404:
|
2019-07-23 09:00:55 -04:00
|
|
|
return False
|
2015-02-05 08:19:46 -05:00
|
|
|
except Exception as ex:
|
|
|
|
logger.warning("query_alias to %s threw exception %s", uri, ex)
|
2019-07-23 09:00:55 -04:00
|
|
|
return False
|
2015-02-05 08:19:46 -05:00
|
|
|
|
2021-12-14 12:02:46 -05:00
|
|
|
async def query_3pe(
|
|
|
|
self,
|
|
|
|
service: "ApplicationService",
|
|
|
|
kind: str,
|
|
|
|
protocol: str,
|
|
|
|
fields: Dict[bytes, List[bytes]],
|
|
|
|
) -> List[JsonDict]:
|
2016-08-18 12:19:55 -04:00
|
|
|
if kind == ThirdPartyEntityKind.USER:
|
2016-08-18 12:33:56 -04:00
|
|
|
required_field = "userid"
|
2016-08-18 12:19:55 -04:00
|
|
|
elif kind == ThirdPartyEntityKind.LOCATION:
|
2016-08-18 12:33:56 -04:00
|
|
|
required_field = "alias"
|
2016-08-18 12:19:55 -04:00
|
|
|
else:
|
|
|
|
raise ValueError("Unrecognised 'kind' argument %r to query_3pe()", kind)
|
2016-08-30 12:16:00 -04:00
|
|
|
if service.url is None:
|
2019-07-23 09:00:55 -04:00
|
|
|
return []
|
2016-08-17 08:15:06 -04:00
|
|
|
|
2016-08-25 13:06:29 -04:00
|
|
|
uri = "%s%s/thirdparty/%s/%s" % (
|
|
|
|
service.url,
|
|
|
|
APP_SERVICE_PREFIX,
|
2016-08-25 13:35:38 -04:00
|
|
|
kind,
|
2018-09-05 10:10:47 -04:00
|
|
|
urllib.parse.quote(protocol),
|
2016-08-25 13:06:29 -04:00
|
|
|
)
|
2016-08-18 11:09:50 -04:00
|
|
|
try:
|
2020-07-30 07:27:39 -04:00
|
|
|
response = await self.get_json(uri, fields)
|
2016-08-18 12:33:56 -04:00
|
|
|
if not isinstance(response, list):
|
|
|
|
logger.warning(
|
|
|
|
"query_3pe to %s returned an invalid response %r", uri, response
|
|
|
|
)
|
2019-07-23 09:00:55 -04:00
|
|
|
return []
|
2016-08-18 12:33:56 -04:00
|
|
|
|
|
|
|
ret = []
|
|
|
|
for r in response:
|
|
|
|
if _is_valid_3pe_result(r, field=required_field):
|
|
|
|
ret.append(r)
|
|
|
|
else:
|
|
|
|
logger.warning(
|
|
|
|
"query_3pe to %s returned an invalid result %r", uri, r
|
|
|
|
)
|
|
|
|
|
2019-07-23 09:00:55 -04:00
|
|
|
return ret
|
2016-08-18 11:09:50 -04:00
|
|
|
except Exception as ex:
|
2016-08-18 12:19:55 -04:00
|
|
|
logger.warning("query_3pe to %s threw exception %s", uri, ex)
|
2019-07-23 09:00:55 -04:00
|
|
|
return []
|
2016-08-18 11:09:50 -04:00
|
|
|
|
2020-09-01 11:03:49 -04:00
|
|
|
async def get_3pe_protocol(
|
|
|
|
self, service: "ApplicationService", protocol: str
|
|
|
|
) -> Optional[JsonDict]:
|
2016-08-30 12:16:00 -04:00
|
|
|
if service.url is None:
|
2019-07-23 09:00:55 -04:00
|
|
|
return {}
|
2016-08-30 11:30:12 -04:00
|
|
|
|
2020-09-01 11:03:49 -04:00
|
|
|
async def _get() -> Optional[JsonDict]:
|
2016-08-25 13:06:29 -04:00
|
|
|
uri = "%s%s/thirdparty/protocol/%s" % (
|
|
|
|
service.url,
|
|
|
|
APP_SERVICE_PREFIX,
|
2018-09-05 10:10:47 -04:00
|
|
|
urllib.parse.quote(protocol),
|
2016-08-25 13:06:29 -04:00
|
|
|
)
|
2016-08-25 10:56:27 -04:00
|
|
|
try:
|
2020-09-24 10:47:20 -04:00
|
|
|
info = await self.get_json(uri)
|
2016-09-09 10:07:04 -04:00
|
|
|
|
2016-09-09 10:09:46 -04:00
|
|
|
if not _is_valid_3pe_metadata(info):
|
|
|
|
logger.warning(
|
2019-11-21 07:00:14 -05:00
|
|
|
"query_3pe_protocol to %s did not return a valid result", uri
|
2016-09-09 10:09:46 -04:00
|
|
|
)
|
2019-07-23 09:00:55 -04:00
|
|
|
return None
|
2016-09-09 10:07:04 -04:00
|
|
|
|
2016-12-06 05:43:48 -05:00
|
|
|
for instance in info.get("instances", []):
|
|
|
|
network_id = instance.get("network_id", None)
|
|
|
|
if network_id is not None:
|
2016-12-12 09:46:13 -05:00
|
|
|
instance["instance_id"] = ThirdPartyInstanceID(
|
2016-12-06 05:43:48 -05:00
|
|
|
service.id, network_id
|
|
|
|
).to_string()
|
|
|
|
|
2019-07-23 09:00:55 -04:00
|
|
|
return info
|
2016-08-25 10:56:27 -04:00
|
|
|
except Exception as ex:
|
|
|
|
logger.warning("query_3pe_protocol to %s threw exception %s", uri, ex)
|
2019-07-23 09:00:55 -04:00
|
|
|
return None
|
2016-08-25 10:56:27 -04:00
|
|
|
|
|
|
|
key = (service.id, protocol)
|
2020-09-01 11:03:49 -04:00
|
|
|
return await self.protocol_meta_cache.wrap(key, _get)
|
2016-08-25 10:10:06 -04:00
|
|
|
|
2020-10-15 12:33:28 -04:00
|
|
|
async def push_bulk(
|
|
|
|
self,
|
|
|
|
service: "ApplicationService",
|
|
|
|
events: List[EventBase],
|
|
|
|
ephemeral: List[JsonDict],
|
2022-02-01 09:13:38 -05:00
|
|
|
to_device_messages: List[JsonDict],
|
2020-10-15 12:33:28 -04:00
|
|
|
txn_id: Optional[int] = None,
|
2021-12-14 12:02:46 -05:00
|
|
|
) -> bool:
|
2022-02-01 09:13:38 -05:00
|
|
|
"""
|
|
|
|
Push data to an application service.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
service: The application service to send to.
|
|
|
|
events: The persistent events to send.
|
|
|
|
ephemeral: The ephemeral events to send.
|
|
|
|
to_device_messages: The to-device messages to send.
|
|
|
|
txn_id: An unique ID to assign to this transaction. Application services should
|
|
|
|
deduplicate transactions received with identitical IDs.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
True if the task succeeded, False if it failed.
|
|
|
|
"""
|
2016-08-30 12:16:00 -04:00
|
|
|
if service.url is None:
|
2019-07-23 09:00:55 -04:00
|
|
|
return True
|
2016-08-30 11:21:16 -04:00
|
|
|
|
2021-12-14 12:02:46 -05:00
|
|
|
# This is required by the configuration.
|
|
|
|
assert service.hs_token is not None
|
|
|
|
|
|
|
|
serialized_events = self._serialize(service, events)
|
2015-02-05 08:42:35 -05:00
|
|
|
|
2015-03-05 10:40:07 -05:00
|
|
|
if txn_id is None:
|
|
|
|
logger.warning(
|
|
|
|
"push_bulk: Missing txn ID sending events to %s", service.url
|
|
|
|
)
|
2020-10-15 12:33:28 -04:00
|
|
|
txn_id = 0
|
|
|
|
|
|
|
|
uri = service.url + ("/transactions/%s" % urllib.parse.quote(str(txn_id)))
|
|
|
|
|
|
|
|
# Never send ephemeral events to appservices that do not support it
|
2022-02-01 09:13:38 -05:00
|
|
|
body: Dict[str, List[JsonDict]] = {"events": serialized_events}
|
2020-10-15 12:33:28 -04:00
|
|
|
if service.supports_ephemeral:
|
2022-02-01 09:13:38 -05:00
|
|
|
body.update(
|
|
|
|
{
|
|
|
|
# TODO: Update to stable prefixes once MSC2409 completes FCP merge.
|
|
|
|
"de.sorunome.msc2409.ephemeral": ephemeral,
|
|
|
|
"de.sorunome.msc2409.to_device": to_device_messages,
|
|
|
|
}
|
|
|
|
)
|
2015-03-05 10:40:07 -05:00
|
|
|
|
2015-02-05 04:43:22 -05:00
|
|
|
try:
|
2020-07-30 07:27:39 -04:00
|
|
|
await self.put_json(
|
2020-10-15 12:33:28 -04:00
|
|
|
uri=uri,
|
|
|
|
json_body=body,
|
|
|
|
args={"access_token": service.hs_token},
|
2015-02-05 04:43:22 -05:00
|
|
|
)
|
2021-11-18 15:16:08 -05:00
|
|
|
if logger.isEnabledFor(logging.DEBUG):
|
|
|
|
logger.debug(
|
|
|
|
"push_bulk to %s succeeded! events=%s",
|
|
|
|
uri,
|
|
|
|
[event.get("event_id") for event in events],
|
|
|
|
)
|
2018-06-05 12:30:45 -04:00
|
|
|
sent_transactions_counter.labels(service.id).inc()
|
2021-12-14 12:02:46 -05:00
|
|
|
sent_events_counter.labels(service.id).inc(len(serialized_events))
|
2019-07-23 09:00:55 -04:00
|
|
|
return True
|
2015-02-05 04:43:22 -05:00
|
|
|
except CodeMessageException as e:
|
2021-11-18 15:16:08 -05:00
|
|
|
logger.warning(
|
|
|
|
"push_bulk to %s received code=%s msg=%s",
|
|
|
|
uri,
|
|
|
|
e.code,
|
|
|
|
e.msg,
|
|
|
|
exc_info=logger.isEnabledFor(logging.DEBUG),
|
|
|
|
)
|
2015-02-05 08:19:46 -05:00
|
|
|
except Exception as ex:
|
2021-11-18 15:16:08 -05:00
|
|
|
logger.warning(
|
|
|
|
"push_bulk to %s threw exception(%s) %s args=%s",
|
|
|
|
uri,
|
|
|
|
type(ex).__name__,
|
|
|
|
ex,
|
|
|
|
ex.args,
|
|
|
|
exc_info=logger.isEnabledFor(logging.DEBUG),
|
|
|
|
)
|
2018-06-05 12:30:45 -04:00
|
|
|
failed_transactions_counter.labels(service.id).inc()
|
2019-07-23 09:00:55 -04:00
|
|
|
return False
|
2015-02-04 06:19:18 -05:00
|
|
|
|
2021-12-14 12:02:46 -05:00
|
|
|
def _serialize(
|
|
|
|
self, service: "ApplicationService", events: Iterable[EventBase]
|
|
|
|
) -> List[JsonDict]:
|
2015-02-05 08:42:35 -05:00
|
|
|
time_now = self.clock.time_msec()
|
2020-07-10 13:44:56 -04:00
|
|
|
return [
|
|
|
|
serialize_event(
|
|
|
|
e,
|
|
|
|
time_now,
|
|
|
|
as_client_event=True,
|
2021-06-09 14:39:51 -04:00
|
|
|
# If this is an invite or a knock membership event, and we're interested
|
|
|
|
# in this user, then include any stripped state alongside the event.
|
|
|
|
include_stripped_room_state=(
|
2020-07-10 13:44:56 -04:00
|
|
|
e.type == EventTypes.Member
|
2021-06-09 14:39:51 -04:00
|
|
|
and (
|
|
|
|
e.membership == Membership.INVITE
|
|
|
|
or e.membership == Membership.KNOCK
|
|
|
|
)
|
2020-07-10 13:44:56 -04:00
|
|
|
and service.is_interested_in_user(e.state_key)
|
|
|
|
),
|
|
|
|
)
|
|
|
|
for e in events
|
|
|
|
]
|