2014-09-30 10:15:10 -04:00
|
|
|
# -*- coding: utf-8 -*-
|
2016-01-06 23:26:29 -05:00
|
|
|
# Copyright 2014-2016 OpenMarket Ltd
|
2014-09-30 10:15:10 -04:00
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
|
|
|
from synapse.crypto.keyclient import fetch_server_key
|
2015-08-24 11:17:38 -04:00
|
|
|
from synapse.api.errors import SynapseError, Codes
|
|
|
|
from synapse.util.retryutils import get_retry_limiter
|
|
|
|
from synapse.util import unwrapFirstError
|
|
|
|
from synapse.util.async import ObservableDeferred
|
2016-02-04 05:22:44 -05:00
|
|
|
from synapse.util.logcontext import (
|
|
|
|
preserve_context_over_deferred, preserve_context_over_fn, PreserveLoggingContext,
|
|
|
|
preserve_fn
|
|
|
|
)
|
2015-08-24 11:17:38 -04:00
|
|
|
|
2014-09-30 10:15:10 -04:00
|
|
|
from twisted.internet import defer
|
2015-08-24 11:17:38 -04:00
|
|
|
|
|
|
|
from signedjson.sign import (
|
2015-04-22 09:21:08 -04:00
|
|
|
verify_signed_json, signature_ids, sign_json, encode_canonical_json
|
|
|
|
)
|
2015-08-24 11:17:38 -04:00
|
|
|
from signedjson.key import (
|
2014-09-30 10:15:10 -04:00
|
|
|
is_signing_algorithm_supported, decode_verify_key_bytes
|
|
|
|
)
|
2015-08-24 11:17:38 -04:00
|
|
|
from unpaddedbase64 import decode_base64, encode_base64
|
2015-06-26 06:25:00 -04:00
|
|
|
|
2014-09-30 10:15:10 -04:00
|
|
|
from OpenSSL import crypto
|
|
|
|
|
2015-06-26 04:52:24 -04:00
|
|
|
from collections import namedtuple
|
2015-04-22 09:21:08 -04:00
|
|
|
import urllib
|
|
|
|
import hashlib
|
2014-09-30 10:15:10 -04:00
|
|
|
import logging
|
|
|
|
|
|
|
|
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
2015-06-26 04:52:24 -04:00
|
|
|
KeyGroup = namedtuple("KeyGroup", ("server_name", "group_id", "key_ids"))
|
|
|
|
|
|
|
|
|
2014-09-30 10:15:10 -04:00
|
|
|
class Keyring(object):
|
|
|
|
def __init__(self, hs):
|
|
|
|
self.store = hs.get_datastore()
|
|
|
|
self.clock = hs.get_clock()
|
2015-04-20 11:23:47 -04:00
|
|
|
self.client = hs.get_http_client()
|
2015-04-22 09:21:08 -04:00
|
|
|
self.config = hs.get_config()
|
2015-04-24 12:01:34 -04:00
|
|
|
self.perspective_servers = self.config.perspectives
|
2014-09-30 10:15:10 -04:00
|
|
|
self.hs = hs
|
|
|
|
|
2015-04-27 09:20:26 -04:00
|
|
|
self.key_downloads = {}
|
|
|
|
|
2014-09-30 10:15:10 -04:00
|
|
|
def verify_json_for_server(self, server_name, json_object):
|
2015-06-26 04:52:24 -04:00
|
|
|
return self.verify_json_objects_for_server(
|
|
|
|
[(server_name, json_object)]
|
|
|
|
)[0]
|
2015-03-05 12:09:13 -05:00
|
|
|
|
2015-06-26 04:52:24 -04:00
|
|
|
def verify_json_objects_for_server(self, server_and_json):
|
|
|
|
"""Bulk verfies signatures of json objects, bulk fetching keys as
|
|
|
|
necessary.
|
2014-09-30 10:15:10 -04:00
|
|
|
|
|
|
|
Args:
|
2015-06-26 04:52:24 -04:00
|
|
|
server_and_json (list): List of pairs of (server_name, json_object)
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
list of deferreds indicating success or failure to verify each
|
|
|
|
json object's signature for the given server_name.
|
2014-09-30 10:15:10 -04:00
|
|
|
"""
|
2015-06-26 04:52:24 -04:00
|
|
|
group_id_to_json = {}
|
|
|
|
group_id_to_group = {}
|
|
|
|
group_ids = []
|
|
|
|
|
|
|
|
next_group_id = 0
|
|
|
|
deferreds = {}
|
|
|
|
|
|
|
|
for server_name, json_object in server_and_json:
|
|
|
|
logger.debug("Verifying for %s", server_name)
|
|
|
|
group_id = next_group_id
|
|
|
|
next_group_id += 1
|
|
|
|
group_ids.append(group_id)
|
|
|
|
|
|
|
|
key_ids = signature_ids(json_object, server_name)
|
|
|
|
if not key_ids:
|
|
|
|
deferreds[group_id] = defer.fail(SynapseError(
|
|
|
|
400,
|
|
|
|
"Not signed with a supported algorithm",
|
|
|
|
Codes.UNAUTHORIZED,
|
|
|
|
))
|
2015-06-26 06:25:00 -04:00
|
|
|
else:
|
|
|
|
deferreds[group_id] = defer.Deferred()
|
2015-06-26 04:52:24 -04:00
|
|
|
|
|
|
|
group = KeyGroup(server_name, group_id, key_ids)
|
2014-09-30 10:15:10 -04:00
|
|
|
|
2015-06-26 04:52:24 -04:00
|
|
|
group_id_to_group[group_id] = group
|
|
|
|
group_id_to_json[group_id] = json_object
|
2014-09-30 10:15:10 -04:00
|
|
|
|
2015-06-26 04:52:24 -04:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def handle_key_deferred(group, deferred):
|
|
|
|
server_name = group.server_name
|
|
|
|
try:
|
|
|
|
_, _, key_id, verify_key = yield deferred
|
|
|
|
except IOError as e:
|
|
|
|
logger.warn(
|
|
|
|
"Got IOError when downloading keys for %s: %s %s",
|
|
|
|
server_name, type(e).__name__, str(e.message),
|
|
|
|
)
|
|
|
|
raise SynapseError(
|
|
|
|
502,
|
|
|
|
"Error downloading keys for %s" % (server_name,),
|
|
|
|
Codes.UNAUTHORIZED,
|
|
|
|
)
|
|
|
|
except Exception as e:
|
|
|
|
logger.exception(
|
|
|
|
"Got Exception when downloading keys for %s: %s %s",
|
|
|
|
server_name, type(e).__name__, str(e.message),
|
|
|
|
)
|
|
|
|
raise SynapseError(
|
|
|
|
401,
|
|
|
|
"No key for %s with id %s" % (server_name, key_ids),
|
|
|
|
Codes.UNAUTHORIZED,
|
|
|
|
)
|
|
|
|
|
|
|
|
json_object = group_id_to_json[group.group_id]
|
2014-09-30 10:15:10 -04:00
|
|
|
|
2015-06-26 04:52:24 -04:00
|
|
|
try:
|
|
|
|
verify_signed_json(json_object, server_name, verify_key)
|
|
|
|
except:
|
|
|
|
raise SynapseError(
|
|
|
|
401,
|
|
|
|
"Invalid signature for server %s with key %s:%s" % (
|
|
|
|
server_name, verify_key.alg, verify_key.version
|
|
|
|
),
|
|
|
|
Codes.UNAUTHORIZED,
|
|
|
|
)
|
|
|
|
|
2015-06-26 06:25:00 -04:00
|
|
|
server_to_deferred = {
|
|
|
|
server_name: defer.Deferred()
|
|
|
|
for server_name, _ in server_and_json
|
|
|
|
}
|
|
|
|
|
2016-02-04 05:22:44 -05:00
|
|
|
with PreserveLoggingContext():
|
2015-06-26 06:25:00 -04:00
|
|
|
|
2016-02-04 05:22:44 -05:00
|
|
|
# We want to wait for any previous lookups to complete before
|
|
|
|
# proceeding.
|
|
|
|
wait_on_deferred = self.wait_for_previous_lookups(
|
|
|
|
[server_name for server_name, _ in server_and_json],
|
|
|
|
server_to_deferred,
|
|
|
|
)
|
2015-06-26 06:25:00 -04:00
|
|
|
|
2016-02-04 05:22:44 -05:00
|
|
|
# Actually start fetching keys.
|
|
|
|
wait_on_deferred.addBoth(
|
|
|
|
lambda _: self.get_server_verify_keys(group_id_to_group, deferreds)
|
|
|
|
)
|
|
|
|
|
|
|
|
# When we've finished fetching all the keys for a given server_name,
|
|
|
|
# resolve the deferred passed to `wait_for_previous_lookups` so that
|
|
|
|
# any lookups waiting will proceed.
|
|
|
|
server_to_gids = {}
|
2015-06-26 06:25:00 -04:00
|
|
|
|
2016-02-04 05:22:44 -05:00
|
|
|
def remove_deferreds(res, server_name, group_id):
|
|
|
|
server_to_gids[server_name].discard(group_id)
|
|
|
|
if not server_to_gids[server_name]:
|
|
|
|
d = server_to_deferred.pop(server_name, None)
|
|
|
|
if d:
|
|
|
|
d.callback(None)
|
|
|
|
return res
|
2015-06-26 04:52:24 -04:00
|
|
|
|
2016-02-04 05:22:44 -05:00
|
|
|
for g_id, deferred in deferreds.items():
|
|
|
|
server_name = group_id_to_group[g_id].server_name
|
|
|
|
server_to_gids.setdefault(server_name, set()).add(g_id)
|
|
|
|
deferred.addBoth(remove_deferreds, server_name, g_id)
|
2015-06-26 06:25:00 -04:00
|
|
|
|
|
|
|
# Pass those keys to handle_key_deferred so that the json object
|
|
|
|
# signatures can be verified
|
2015-06-26 04:52:24 -04:00
|
|
|
return [
|
2016-02-04 05:22:44 -05:00
|
|
|
preserve_context_over_fn(
|
|
|
|
handle_key_deferred,
|
2015-06-26 04:52:24 -04:00
|
|
|
group_id_to_group[g_id],
|
|
|
|
deferreds[g_id],
|
2015-05-08 11:27:36 -04:00
|
|
|
)
|
2015-06-26 04:52:24 -04:00
|
|
|
for g_id in group_ids
|
|
|
|
]
|
2015-04-27 09:37:24 -04:00
|
|
|
|
2015-06-26 06:25:00 -04:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def wait_for_previous_lookups(self, server_names, server_to_deferred):
|
|
|
|
"""Waits for any previous key lookups for the given servers to finish.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
server_names (list): list of server_names we want to lookup
|
|
|
|
server_to_deferred (dict): server_name to deferred which gets
|
|
|
|
resolved once we've finished looking up keys for that server
|
|
|
|
"""
|
|
|
|
while True:
|
|
|
|
wait_on = [
|
|
|
|
self.key_downloads[server_name]
|
|
|
|
for server_name in server_names
|
|
|
|
if server_name in self.key_downloads
|
|
|
|
]
|
|
|
|
if wait_on:
|
2016-02-04 05:22:44 -05:00
|
|
|
with PreserveLoggingContext():
|
|
|
|
yield defer.DeferredList(wait_on)
|
2015-06-26 06:25:00 -04:00
|
|
|
else:
|
|
|
|
break
|
|
|
|
|
2015-09-09 12:02:39 -04:00
|
|
|
for server_name, deferred in server_to_deferred.items():
|
2016-02-04 05:22:44 -05:00
|
|
|
d = ObservableDeferred(preserve_context_over_deferred(deferred))
|
2015-09-09 12:02:39 -04:00
|
|
|
self.key_downloads[server_name] = d
|
|
|
|
|
|
|
|
def rm(r, server_name):
|
|
|
|
self.key_downloads.pop(server_name, None)
|
|
|
|
return r
|
|
|
|
|
|
|
|
d.addBoth(rm, server_name)
|
2015-06-26 06:25:00 -04:00
|
|
|
|
|
|
|
def get_server_verify_keys(self, group_id_to_group, group_id_to_deferred):
|
2015-06-26 04:52:24 -04:00
|
|
|
"""Takes a dict of KeyGroups and tries to find at least one key for
|
|
|
|
each group.
|
|
|
|
"""
|
|
|
|
|
|
|
|
# These are functions that produce keys given a list of key ids
|
|
|
|
key_fetch_fns = (
|
|
|
|
self.get_keys_from_store, # First try the local store
|
|
|
|
self.get_keys_from_perspectives, # Then try via perspectives
|
|
|
|
self.get_keys_from_server, # Then try directly
|
|
|
|
)
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def do_iterations():
|
|
|
|
merged_results = {}
|
|
|
|
|
2015-09-11 06:51:25 -04:00
|
|
|
missing_keys = {}
|
|
|
|
for group in group_id_to_group.values():
|
2015-12-18 16:36:42 -05:00
|
|
|
missing_keys.setdefault(group.server_name, set()).update(
|
|
|
|
group.key_ids
|
|
|
|
)
|
2015-06-26 04:52:24 -04:00
|
|
|
|
|
|
|
for fn in key_fetch_fns:
|
|
|
|
results = yield fn(missing_keys.items())
|
|
|
|
merged_results.update(results)
|
|
|
|
|
|
|
|
# We now need to figure out which groups we have keys for
|
|
|
|
# and which we don't
|
|
|
|
missing_groups = {}
|
|
|
|
for group in group_id_to_group.values():
|
|
|
|
for key_id in group.key_ids:
|
|
|
|
if key_id in merged_results[group.server_name]:
|
2016-02-04 05:22:44 -05:00
|
|
|
with PreserveLoggingContext():
|
|
|
|
group_id_to_deferred[group.group_id].callback((
|
|
|
|
group.group_id,
|
|
|
|
group.server_name,
|
|
|
|
key_id,
|
|
|
|
merged_results[group.server_name][key_id],
|
|
|
|
))
|
2015-06-26 04:52:24 -04:00
|
|
|
break
|
|
|
|
else:
|
|
|
|
missing_groups.setdefault(
|
|
|
|
group.server_name, []
|
|
|
|
).append(group)
|
|
|
|
|
|
|
|
if not missing_groups:
|
|
|
|
break
|
|
|
|
|
|
|
|
missing_keys = {
|
|
|
|
server_name: set(
|
|
|
|
key_id for group in groups for key_id in group.key_ids
|
|
|
|
)
|
|
|
|
for server_name, groups in missing_groups.items()
|
|
|
|
}
|
2015-04-27 09:37:24 -04:00
|
|
|
|
2015-06-26 04:52:24 -04:00
|
|
|
for group in missing_groups.values():
|
2015-06-26 06:25:00 -04:00
|
|
|
group_id_to_deferred[group.group_id].errback(SynapseError(
|
2015-06-26 04:52:24 -04:00
|
|
|
401,
|
|
|
|
"No key for %s with id %s" % (
|
|
|
|
group.server_name, group.key_ids,
|
|
|
|
),
|
|
|
|
Codes.UNAUTHORIZED,
|
|
|
|
))
|
|
|
|
|
|
|
|
def on_err(err):
|
2015-06-26 06:25:00 -04:00
|
|
|
for deferred in group_id_to_deferred.values():
|
|
|
|
if not deferred.called:
|
|
|
|
deferred.errback(err)
|
2015-06-26 04:52:24 -04:00
|
|
|
|
|
|
|
do_iterations().addErrback(on_err)
|
|
|
|
|
2015-06-26 06:25:00 -04:00
|
|
|
return group_id_to_deferred
|
2014-09-30 10:15:10 -04:00
|
|
|
|
2015-04-27 09:37:24 -04:00
|
|
|
@defer.inlineCallbacks
|
2015-06-26 04:52:24 -04:00
|
|
|
def get_keys_from_store(self, server_name_and_key_ids):
|
|
|
|
res = yield defer.gatherResults(
|
|
|
|
[
|
2015-09-09 12:02:39 -04:00
|
|
|
self.store.get_server_verify_keys(
|
|
|
|
server_name, key_ids
|
|
|
|
).addCallback(lambda ks, server: (server, ks), server_name)
|
2015-06-26 04:52:24 -04:00
|
|
|
for server_name, key_ids in server_name_and_key_ids
|
|
|
|
],
|
|
|
|
consumeErrors=True,
|
|
|
|
).addErrback(unwrapFirstError)
|
|
|
|
|
2015-09-09 12:02:39 -04:00
|
|
|
defer.returnValue(dict(res))
|
2015-04-29 08:31:14 -04:00
|
|
|
|
2015-06-26 04:52:24 -04:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def get_keys_from_perspectives(self, server_name_and_key_ids):
|
2015-05-19 06:56:18 -04:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def get_key(perspective_name, perspective_keys):
|
|
|
|
try:
|
|
|
|
result = yield self.get_server_verify_key_v2_indirect(
|
2015-06-26 04:52:24 -04:00
|
|
|
server_name_and_key_ids, perspective_name, perspective_keys
|
2015-05-19 06:56:18 -04:00
|
|
|
)
|
|
|
|
defer.returnValue(result)
|
|
|
|
except Exception as e:
|
2015-06-26 04:52:24 -04:00
|
|
|
logger.exception(
|
|
|
|
"Unable to get key from %r: %s %s",
|
|
|
|
perspective_name,
|
2015-05-19 06:56:18 -04:00
|
|
|
type(e).__name__, str(e.message),
|
|
|
|
)
|
2015-06-26 04:52:24 -04:00
|
|
|
defer.returnValue({})
|
2015-04-29 08:31:14 -04:00
|
|
|
|
2015-06-26 04:52:24 -04:00
|
|
|
results = yield defer.gatherResults(
|
|
|
|
[
|
|
|
|
get_key(p_name, p_keys)
|
|
|
|
for p_name, p_keys in self.perspective_servers.items()
|
|
|
|
],
|
|
|
|
consumeErrors=True,
|
|
|
|
).addErrback(unwrapFirstError)
|
2015-04-29 08:31:14 -04:00
|
|
|
|
2015-06-26 04:52:24 -04:00
|
|
|
union_of_keys = {}
|
|
|
|
for result in results:
|
|
|
|
for server_name, keys in result.items():
|
|
|
|
union_of_keys.setdefault(server_name, {}).update(keys)
|
2014-09-30 10:15:10 -04:00
|
|
|
|
2015-06-26 04:52:24 -04:00
|
|
|
defer.returnValue(union_of_keys)
|
2014-09-30 10:15:10 -04:00
|
|
|
|
2015-06-26 04:52:24 -04:00
|
|
|
@defer.inlineCallbacks
|
|
|
|
def get_keys_from_server(self, server_name_and_key_ids):
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def get_key(server_name, key_ids):
|
|
|
|
limiter = yield get_retry_limiter(
|
|
|
|
server_name,
|
|
|
|
self.clock,
|
|
|
|
self.store,
|
|
|
|
)
|
|
|
|
with limiter:
|
|
|
|
keys = None
|
2015-04-20 11:23:47 -04:00
|
|
|
try:
|
|
|
|
keys = yield self.get_server_verify_key_v2_direct(
|
|
|
|
server_name, key_ids
|
|
|
|
)
|
2015-05-19 08:43:34 -04:00
|
|
|
except Exception as e:
|
2015-06-26 04:52:24 -04:00
|
|
|
logger.info(
|
2015-05-19 08:43:34 -04:00
|
|
|
"Unable to getting key %r for %r directly: %s %s",
|
|
|
|
key_ids, server_name,
|
|
|
|
type(e).__name__, str(e.message),
|
|
|
|
)
|
2015-04-20 11:23:47 -04:00
|
|
|
|
2015-06-26 04:52:24 -04:00
|
|
|
if not keys:
|
|
|
|
keys = yield self.get_server_verify_key_v1_direct(
|
|
|
|
server_name, key_ids
|
|
|
|
)
|
|
|
|
|
|
|
|
keys = {server_name: keys}
|
|
|
|
|
|
|
|
defer.returnValue(keys)
|
|
|
|
|
|
|
|
results = yield defer.gatherResults(
|
|
|
|
[
|
|
|
|
get_key(server_name, key_ids)
|
|
|
|
for server_name, key_ids in server_name_and_key_ids
|
|
|
|
],
|
|
|
|
consumeErrors=True,
|
|
|
|
).addErrback(unwrapFirstError)
|
2015-04-20 11:23:47 -04:00
|
|
|
|
2015-06-26 04:52:24 -04:00
|
|
|
merged = {}
|
|
|
|
for result in results:
|
|
|
|
merged.update(result)
|
|
|
|
|
|
|
|
defer.returnValue({
|
|
|
|
server_name: keys
|
|
|
|
for server_name, keys in merged.items()
|
|
|
|
if keys
|
|
|
|
})
|
2015-04-20 11:23:47 -04:00
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
2015-06-26 04:52:24 -04:00
|
|
|
def get_server_verify_key_v2_indirect(self, server_names_and_key_ids,
|
2015-04-20 11:23:47 -04:00
|
|
|
perspective_name,
|
|
|
|
perspective_keys):
|
2015-11-20 12:15:44 -05:00
|
|
|
# TODO(mark): Set the minimum_valid_until_ts to that needed by
|
|
|
|
# the events being validated or the current time if validating
|
|
|
|
# an incoming request.
|
|
|
|
query_response = yield self.client.post_json(
|
|
|
|
destination=perspective_name,
|
|
|
|
path=b"/_matrix/key/v2/query",
|
|
|
|
data={
|
|
|
|
u"server_keys": {
|
|
|
|
server_name: {
|
|
|
|
key_id: {
|
|
|
|
u"minimum_valid_until_ts": 0
|
|
|
|
} for key_id in key_ids
|
2015-04-29 07:14:08 -04:00
|
|
|
}
|
2015-11-20 12:15:44 -05:00
|
|
|
for server_name, key_ids in server_names_and_key_ids
|
|
|
|
}
|
|
|
|
},
|
2015-11-20 12:17:58 -05:00
|
|
|
long_retries=True,
|
2015-11-20 12:15:44 -05:00
|
|
|
)
|
2015-04-20 11:23:47 -04:00
|
|
|
|
2015-04-22 09:21:08 -04:00
|
|
|
keys = {}
|
2015-04-20 11:23:47 -04:00
|
|
|
|
2015-05-19 05:23:02 -04:00
|
|
|
responses = query_response["server_keys"]
|
|
|
|
|
2015-04-20 11:23:47 -04:00
|
|
|
for response in responses:
|
|
|
|
if (u"signatures" not in response
|
2015-04-22 09:21:08 -04:00
|
|
|
or perspective_name not in response[u"signatures"]):
|
2015-04-20 11:23:47 -04:00
|
|
|
raise ValueError(
|
|
|
|
"Key response not signed by perspective server"
|
|
|
|
" %r" % (perspective_name,)
|
|
|
|
)
|
|
|
|
|
|
|
|
verified = False
|
|
|
|
for key_id in response[u"signatures"][perspective_name]:
|
|
|
|
if key_id in perspective_keys:
|
|
|
|
verify_signed_json(
|
|
|
|
response,
|
|
|
|
perspective_name,
|
|
|
|
perspective_keys[key_id]
|
|
|
|
)
|
|
|
|
verified = True
|
|
|
|
|
|
|
|
if not verified:
|
|
|
|
logging.info(
|
|
|
|
"Response from perspective server %r not signed with a"
|
|
|
|
" known key, signed with: %r, known keys: %r",
|
|
|
|
perspective_name,
|
|
|
|
list(response[u"signatures"][perspective_name]),
|
|
|
|
list(perspective_keys)
|
|
|
|
)
|
|
|
|
raise ValueError(
|
|
|
|
"Response not signed with a known key for perspective"
|
|
|
|
" server %r" % (perspective_name,)
|
|
|
|
)
|
|
|
|
|
2015-06-26 04:52:24 -04:00
|
|
|
processed_response = yield self.process_v2_response(
|
2016-07-26 14:50:11 -04:00
|
|
|
perspective_name, response, only_from_server=False
|
2015-04-22 09:21:08 -04:00
|
|
|
)
|
2015-04-20 11:23:47 -04:00
|
|
|
|
2015-06-26 04:52:24 -04:00
|
|
|
for server_name, response_keys in processed_response.items():
|
|
|
|
keys.setdefault(server_name, {}).update(response_keys)
|
2015-04-20 11:23:47 -04:00
|
|
|
|
2015-06-26 04:52:24 -04:00
|
|
|
yield defer.gatherResults(
|
|
|
|
[
|
|
|
|
self.store_keys(
|
|
|
|
server_name=server_name,
|
|
|
|
from_server=perspective_name,
|
|
|
|
verify_keys=response_keys,
|
|
|
|
)
|
|
|
|
for server_name, response_keys in keys.items()
|
|
|
|
],
|
|
|
|
consumeErrors=True
|
|
|
|
).addErrback(unwrapFirstError)
|
2015-04-20 11:23:47 -04:00
|
|
|
|
|
|
|
defer.returnValue(keys)
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def get_server_verify_key_v2_direct(self, server_name, key_ids):
|
|
|
|
keys = {}
|
|
|
|
|
|
|
|
for requested_key_id in key_ids:
|
|
|
|
if requested_key_id in keys:
|
|
|
|
continue
|
|
|
|
|
2015-04-22 09:21:08 -04:00
|
|
|
(response, tls_certificate) = yield fetch_server_key(
|
2015-09-09 07:02:07 -04:00
|
|
|
server_name, self.hs.tls_server_context_factory,
|
2015-04-22 09:21:08 -04:00
|
|
|
path=(b"/_matrix/key/v2/server/%s" % (
|
2015-04-20 11:23:47 -04:00
|
|
|
urllib.quote(requested_key_id),
|
2015-04-22 09:21:08 -04:00
|
|
|
)).encode("ascii"),
|
2015-04-20 11:23:47 -04:00
|
|
|
)
|
|
|
|
|
|
|
|
if (u"signatures" not in response
|
2015-04-22 09:21:08 -04:00
|
|
|
or server_name not in response[u"signatures"]):
|
2015-04-20 11:23:47 -04:00
|
|
|
raise ValueError("Key response not signed by remote server")
|
|
|
|
|
|
|
|
if "tls_fingerprints" not in response:
|
|
|
|
raise ValueError("Key response missing TLS fingerprints")
|
|
|
|
|
|
|
|
certificate_bytes = crypto.dump_certificate(
|
|
|
|
crypto.FILETYPE_ASN1, tls_certificate
|
|
|
|
)
|
|
|
|
sha256_fingerprint = hashlib.sha256(certificate_bytes).digest()
|
|
|
|
sha256_fingerprint_b64 = encode_base64(sha256_fingerprint)
|
|
|
|
|
|
|
|
response_sha256_fingerprints = set()
|
2015-04-22 09:21:08 -04:00
|
|
|
for fingerprint in response[u"tls_fingerprints"]:
|
2015-04-20 11:23:47 -04:00
|
|
|
if u"sha256" in fingerprint:
|
|
|
|
response_sha256_fingerprints.add(fingerprint[u"sha256"])
|
|
|
|
|
2015-04-22 09:21:08 -04:00
|
|
|
if sha256_fingerprint_b64 not in response_sha256_fingerprints:
|
2015-04-20 11:23:47 -04:00
|
|
|
raise ValueError("TLS certificate not allowed by fingerprints")
|
|
|
|
|
|
|
|
response_keys = yield self.process_v2_response(
|
|
|
|
from_server=server_name,
|
2015-06-26 04:52:24 -04:00
|
|
|
requested_ids=[requested_key_id],
|
2015-04-22 09:21:08 -04:00
|
|
|
response_json=response,
|
2015-04-20 11:23:47 -04:00
|
|
|
)
|
|
|
|
|
|
|
|
keys.update(response_keys)
|
|
|
|
|
2015-06-26 04:52:24 -04:00
|
|
|
yield defer.gatherResults(
|
|
|
|
[
|
2016-02-04 05:22:44 -05:00
|
|
|
preserve_fn(self.store_keys)(
|
2015-06-26 04:52:24 -04:00
|
|
|
server_name=key_server_name,
|
|
|
|
from_server=server_name,
|
|
|
|
verify_keys=verify_keys,
|
|
|
|
)
|
|
|
|
for key_server_name, verify_keys in keys.items()
|
|
|
|
],
|
|
|
|
consumeErrors=True
|
|
|
|
).addErrback(unwrapFirstError)
|
2015-04-20 11:23:47 -04:00
|
|
|
|
2015-04-22 09:21:08 -04:00
|
|
|
defer.returnValue(keys)
|
2015-04-20 11:23:47 -04:00
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
2015-06-26 04:52:24 -04:00
|
|
|
def process_v2_response(self, from_server, response_json,
|
2016-07-26 14:50:11 -04:00
|
|
|
requested_ids=[], only_from_server=True):
|
2015-04-22 09:21:08 -04:00
|
|
|
time_now_ms = self.clock.time_msec()
|
2015-04-20 11:23:47 -04:00
|
|
|
response_keys = {}
|
|
|
|
verify_keys = {}
|
2015-05-19 08:12:41 -04:00
|
|
|
for key_id, key_data in response_json["verify_keys"].items():
|
2015-04-20 11:23:47 -04:00
|
|
|
if is_signing_algorithm_supported(key_id):
|
2015-05-19 08:12:41 -04:00
|
|
|
key_base64 = key_data["key"]
|
2015-04-20 11:23:47 -04:00
|
|
|
key_bytes = decode_base64(key_base64)
|
|
|
|
verify_key = decode_verify_key_bytes(key_id, key_bytes)
|
2015-04-29 06:57:26 -04:00
|
|
|
verify_key.time_added = time_now_ms
|
2015-04-20 11:23:47 -04:00
|
|
|
verify_keys[key_id] = verify_key
|
|
|
|
|
|
|
|
old_verify_keys = {}
|
2015-04-22 09:21:08 -04:00
|
|
|
for key_id, key_data in response_json["old_verify_keys"].items():
|
2015-04-20 11:23:47 -04:00
|
|
|
if is_signing_algorithm_supported(key_id):
|
|
|
|
key_base64 = key_data["key"]
|
|
|
|
key_bytes = decode_base64(key_base64)
|
|
|
|
verify_key = decode_verify_key_bytes(key_id, key_bytes)
|
2015-04-23 11:39:13 -04:00
|
|
|
verify_key.expired = key_data["expired_ts"]
|
2015-04-20 11:23:47 -04:00
|
|
|
verify_key.time_added = time_now_ms
|
|
|
|
old_verify_keys[key_id] = verify_key
|
|
|
|
|
2015-06-26 04:52:24 -04:00
|
|
|
results = {}
|
|
|
|
server_name = response_json["server_name"]
|
2016-07-26 14:50:11 -04:00
|
|
|
if only_from_server:
|
|
|
|
if server_name != from_server:
|
|
|
|
raise ValueError(
|
|
|
|
"Expected a response for server %r not %r" % (
|
|
|
|
from_server, server_name
|
|
|
|
)
|
|
|
|
)
|
2015-05-19 07:49:38 -04:00
|
|
|
for key_id in response_json["signatures"].get(server_name, {}):
|
2015-04-22 09:21:08 -04:00
|
|
|
if key_id not in response_json["verify_keys"]:
|
2015-04-20 11:23:47 -04:00
|
|
|
raise ValueError(
|
|
|
|
"Key response must include verification keys for all"
|
|
|
|
" signatures"
|
|
|
|
)
|
|
|
|
if key_id in verify_keys:
|
|
|
|
verify_signed_json(
|
2015-04-22 09:21:08 -04:00
|
|
|
response_json,
|
2015-04-20 11:23:47 -04:00
|
|
|
server_name,
|
|
|
|
verify_keys[key_id]
|
|
|
|
)
|
|
|
|
|
|
|
|
signed_key_json = sign_json(
|
2015-04-22 09:21:08 -04:00
|
|
|
response_json,
|
2015-04-20 11:23:47 -04:00
|
|
|
self.config.server_name,
|
|
|
|
self.config.signing_key[0],
|
|
|
|
)
|
|
|
|
|
|
|
|
signed_key_json_bytes = encode_canonical_json(signed_key_json)
|
2015-04-23 11:39:13 -04:00
|
|
|
ts_valid_until_ms = signed_key_json[u"valid_until_ts"]
|
2015-04-20 11:23:47 -04:00
|
|
|
|
2015-06-26 04:52:24 -04:00
|
|
|
updated_key_ids = set(requested_ids)
|
2015-04-20 11:23:47 -04:00
|
|
|
updated_key_ids.update(verify_keys)
|
|
|
|
updated_key_ids.update(old_verify_keys)
|
|
|
|
|
|
|
|
response_keys.update(verify_keys)
|
|
|
|
response_keys.update(old_verify_keys)
|
|
|
|
|
2015-06-26 04:52:24 -04:00
|
|
|
yield defer.gatherResults(
|
|
|
|
[
|
2016-02-04 05:22:44 -05:00
|
|
|
preserve_fn(self.store.store_server_keys_json)(
|
2015-06-26 04:52:24 -04:00
|
|
|
server_name=server_name,
|
|
|
|
key_id=key_id,
|
|
|
|
from_server=server_name,
|
|
|
|
ts_now_ms=time_now_ms,
|
|
|
|
ts_expires_ms=ts_valid_until_ms,
|
|
|
|
key_json_bytes=signed_key_json_bytes,
|
|
|
|
)
|
|
|
|
for key_id in updated_key_ids
|
|
|
|
],
|
|
|
|
consumeErrors=True,
|
|
|
|
).addErrback(unwrapFirstError)
|
2014-09-30 10:15:10 -04:00
|
|
|
|
2015-06-26 04:52:24 -04:00
|
|
|
results[server_name] = response_keys
|
2015-04-20 11:23:47 -04:00
|
|
|
|
2015-06-26 04:52:24 -04:00
|
|
|
defer.returnValue(results)
|
2015-04-20 11:23:47 -04:00
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def get_server_verify_key_v1_direct(self, server_name, key_ids):
|
|
|
|
"""Finds a verification key for the server with one of the key ids.
|
|
|
|
Args:
|
|
|
|
server_name (str): The name of the server to fetch a key for.
|
|
|
|
keys_ids (list of str): The key_ids to check for.
|
|
|
|
"""
|
|
|
|
|
|
|
|
# Try to fetch the key from the remote server.
|
|
|
|
|
|
|
|
(response, tls_certificate) = yield fetch_server_key(
|
2015-09-09 07:02:07 -04:00
|
|
|
server_name, self.hs.tls_server_context_factory
|
2015-04-20 11:23:47 -04:00
|
|
|
)
|
|
|
|
|
2015-02-17 13:13:34 -05:00
|
|
|
# Check the response.
|
2014-09-30 10:15:10 -04:00
|
|
|
|
2015-02-17 13:13:34 -05:00
|
|
|
x509_certificate_bytes = crypto.dump_certificate(
|
|
|
|
crypto.FILETYPE_ASN1, tls_certificate
|
|
|
|
)
|
2014-09-30 10:15:10 -04:00
|
|
|
|
2015-02-17 13:13:34 -05:00
|
|
|
if ("signatures" not in response
|
|
|
|
or server_name not in response["signatures"]):
|
|
|
|
raise ValueError("Key response not signed by remote server")
|
|
|
|
|
|
|
|
if "tls_certificate" not in response:
|
|
|
|
raise ValueError("Key response missing TLS certificate")
|
2014-09-30 10:15:10 -04:00
|
|
|
|
2015-02-17 13:13:34 -05:00
|
|
|
tls_certificate_b64 = response["tls_certificate"]
|
|
|
|
|
|
|
|
if encode_base64(x509_certificate_bytes) != tls_certificate_b64:
|
|
|
|
raise ValueError("TLS certificate doesn't match")
|
|
|
|
|
2015-04-20 11:23:47 -04:00
|
|
|
# Cache the result in the datastore.
|
|
|
|
|
|
|
|
time_now_ms = self.clock.time_msec()
|
|
|
|
|
2015-02-17 13:13:34 -05:00
|
|
|
verify_keys = {}
|
|
|
|
for key_id, key_base64 in response["verify_keys"].items():
|
|
|
|
if is_signing_algorithm_supported(key_id):
|
|
|
|
key_bytes = decode_base64(key_base64)
|
|
|
|
verify_key = decode_verify_key_bytes(key_id, key_bytes)
|
2015-04-20 11:23:47 -04:00
|
|
|
verify_key.time_added = time_now_ms
|
2015-02-17 13:13:34 -05:00
|
|
|
verify_keys[key_id] = verify_key
|
|
|
|
|
|
|
|
for key_id in response["signatures"][server_name]:
|
|
|
|
if key_id not in response["verify_keys"]:
|
|
|
|
raise ValueError(
|
|
|
|
"Key response must include verification keys for all"
|
|
|
|
" signatures"
|
|
|
|
)
|
|
|
|
if key_id in verify_keys:
|
|
|
|
verify_signed_json(
|
|
|
|
response,
|
|
|
|
server_name,
|
|
|
|
verify_keys[key_id]
|
2014-09-30 10:15:10 -04:00
|
|
|
)
|
|
|
|
|
2015-02-17 13:13:34 -05:00
|
|
|
yield self.store.store_server_certificate(
|
|
|
|
server_name,
|
|
|
|
server_name,
|
|
|
|
time_now_ms,
|
|
|
|
tls_certificate,
|
|
|
|
)
|
|
|
|
|
2015-04-20 11:23:47 -04:00
|
|
|
yield self.store_keys(
|
|
|
|
server_name=server_name,
|
|
|
|
from_server=server_name,
|
|
|
|
verify_keys=verify_keys,
|
|
|
|
)
|
|
|
|
|
|
|
|
defer.returnValue(verify_keys)
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def store_keys(self, server_name, from_server, verify_keys):
|
|
|
|
"""Store a collection of verify keys for a given server
|
|
|
|
Args:
|
|
|
|
server_name(str): The name of the server the keys are for.
|
|
|
|
from_server(str): The server the keys were downloaded from.
|
|
|
|
verify_keys(dict): A mapping of key_id to VerifyKey.
|
|
|
|
Returns:
|
|
|
|
A deferred that completes when the keys are stored.
|
|
|
|
"""
|
2015-06-26 04:52:24 -04:00
|
|
|
# TODO(markjh): Store whether the keys have expired.
|
|
|
|
yield defer.gatherResults(
|
|
|
|
[
|
2016-02-04 05:22:44 -05:00
|
|
|
preserve_fn(self.store.store_server_verify_key)(
|
2015-06-26 04:52:24 -04:00
|
|
|
server_name, server_name, key.time_added, key
|
|
|
|
)
|
|
|
|
for key_id, key in verify_keys.items()
|
|
|
|
],
|
|
|
|
consumeErrors=True,
|
|
|
|
).addErrback(unwrapFirstError)
|