2014-08-12 10:10:52 -04:00
|
|
|
# -*- coding: utf-8 -*-
|
2016-01-06 23:26:29 -05:00
|
|
|
# Copyright 2014-2016 OpenMarket Ltd
|
2014-08-12 10:10:52 -04:00
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
2017-09-22 15:26:47 -04:00
|
|
|
import socket
|
2014-08-12 22:14:34 -04:00
|
|
|
|
2016-12-11 04:46:43 -05:00
|
|
|
from twisted.internet.endpoints import HostnameEndpoint, wrapClientTLS
|
2016-12-28 19:09:33 -05:00
|
|
|
from twisted.internet import defer, reactor
|
2014-08-12 10:10:52 -04:00
|
|
|
from twisted.internet.error import ConnectError
|
|
|
|
from twisted.names import client, dns
|
2016-01-20 06:34:09 -05:00
|
|
|
from twisted.names.error import DNSNameError, DomainError
|
2014-08-12 10:10:52 -04:00
|
|
|
|
|
|
|
import collections
|
|
|
|
import logging
|
|
|
|
import random
|
2016-03-31 05:04:28 -04:00
|
|
|
import time
|
2014-08-12 10:10:52 -04:00
|
|
|
|
|
|
|
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
2016-01-20 06:34:09 -05:00
|
|
|
SERVER_CACHE = {}
|
|
|
|
|
2017-09-22 15:26:47 -04:00
|
|
|
# our record of an individual server which can be tried to reach a destination.
|
|
|
|
#
|
|
|
|
# "host" is actually a dotted-quad or ipv6 address string. Except when there's
|
|
|
|
# no SRV record, in which case it is the original hostname.
|
2016-01-20 06:34:09 -05:00
|
|
|
_Server = collections.namedtuple(
|
2016-03-31 05:04:28 -04:00
|
|
|
"_Server", "priority weight host port expires"
|
2016-01-20 06:34:09 -05:00
|
|
|
)
|
|
|
|
|
|
|
|
|
2014-11-20 12:41:56 -05:00
|
|
|
def matrix_federation_endpoint(reactor, destination, ssl_context_factory=None,
|
2014-11-20 13:00:10 -05:00
|
|
|
timeout=None):
|
2014-08-12 10:10:52 -04:00
|
|
|
"""Construct an endpoint for the given matrix destination.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
reactor: Twisted reactor.
|
|
|
|
destination (bytes): The name of the server to connect to.
|
|
|
|
ssl_context_factory (twisted.internet.ssl.ContextFactory): Factory
|
|
|
|
which generates SSL contexts to use for TLS.
|
|
|
|
timeout (int): connection timeout in seconds
|
|
|
|
"""
|
|
|
|
|
|
|
|
domain_port = destination.split(":")
|
|
|
|
domain = domain_port[0]
|
|
|
|
port = int(domain_port[1]) if domain_port[1:] else None
|
|
|
|
|
|
|
|
endpoint_kw_args = {}
|
|
|
|
|
|
|
|
if timeout is not None:
|
|
|
|
endpoint_kw_args.update(timeout=timeout)
|
|
|
|
|
|
|
|
if ssl_context_factory is None:
|
2016-12-11 04:46:43 -05:00
|
|
|
transport_endpoint = HostnameEndpoint
|
2014-09-02 11:56:57 -04:00
|
|
|
default_port = 8008
|
2014-08-12 10:10:52 -04:00
|
|
|
else:
|
2016-12-12 10:19:54 -05:00
|
|
|
def transport_endpoint(reactor, host, port, timeout):
|
|
|
|
return wrapClientTLS(
|
|
|
|
ssl_context_factory,
|
|
|
|
HostnameEndpoint(reactor, host, port, timeout=timeout))
|
2014-09-02 11:56:57 -04:00
|
|
|
default_port = 8448
|
2014-08-12 10:10:52 -04:00
|
|
|
|
|
|
|
if port is None:
|
2016-12-28 19:10:49 -05:00
|
|
|
return _WrappingEndpointFac(SRVClientEndpoint(
|
2014-08-12 10:10:52 -04:00
|
|
|
reactor, "matrix", domain, protocol="tcp",
|
|
|
|
default_port=default_port, endpoint=transport_endpoint,
|
|
|
|
endpoint_kw_args=endpoint_kw_args
|
2016-12-28 17:49:31 -05:00
|
|
|
))
|
2014-08-12 10:10:52 -04:00
|
|
|
else:
|
2016-12-28 19:10:49 -05:00
|
|
|
return _WrappingEndpointFac(transport_endpoint(
|
2016-12-28 19:09:33 -05:00
|
|
|
reactor, domain, port, **endpoint_kw_args
|
|
|
|
))
|
2016-12-28 17:49:31 -05:00
|
|
|
|
|
|
|
|
2016-12-28 19:10:49 -05:00
|
|
|
class _WrappingEndpointFac(object):
|
2016-12-28 17:49:31 -05:00
|
|
|
def __init__(self, endpoint_fac):
|
|
|
|
self.endpoint_fac = endpoint_fac
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def connect(self, protocolFactory):
|
|
|
|
conn = yield self.endpoint_fac.connect(protocolFactory)
|
|
|
|
conn = _WrappedConnection(conn)
|
|
|
|
defer.returnValue(conn)
|
|
|
|
|
|
|
|
|
|
|
|
class _WrappedConnection(object):
|
2016-12-29 10:51:04 -05:00
|
|
|
"""Wraps a connection and calls abort on it if it hasn't seen any action
|
|
|
|
for 2.5-3 minutes.
|
2016-12-28 17:49:31 -05:00
|
|
|
"""
|
|
|
|
__slots__ = ["conn", "last_request"]
|
|
|
|
|
|
|
|
def __init__(self, conn):
|
|
|
|
object.__setattr__(self, "conn", conn)
|
|
|
|
object.__setattr__(self, "last_request", time.time())
|
|
|
|
|
|
|
|
def __getattr__(self, name):
|
|
|
|
return getattr(self.conn, name)
|
|
|
|
|
|
|
|
def __setattr__(self, name, value):
|
|
|
|
setattr(self.conn, name, value)
|
|
|
|
|
|
|
|
def _time_things_out_maybe(self):
|
2016-12-28 19:10:49 -05:00
|
|
|
# We use a slightly shorter timeout here just in case the callLater is
|
|
|
|
# triggered early. Paranoia ftw.
|
2016-12-29 10:51:04 -05:00
|
|
|
# TODO: Cancel the previous callLater rather than comparing time.time()?
|
2016-12-28 19:10:49 -05:00
|
|
|
if time.time() - self.last_request >= 2.5 * 60:
|
2016-12-28 17:49:31 -05:00
|
|
|
self.abort()
|
2016-12-29 10:51:04 -05:00
|
|
|
# Abort the underlying TLS connection. The abort() method calls
|
|
|
|
# loseConnection() on the underlying TLS connection which tries to
|
|
|
|
# shutdown the connection cleanly. We call abortConnection()
|
|
|
|
# since that will promptly close the underlying TCP connection.
|
|
|
|
self.transport.abortConnection()
|
2016-12-28 17:49:31 -05:00
|
|
|
|
|
|
|
def request(self, request):
|
|
|
|
self.last_request = time.time()
|
|
|
|
|
|
|
|
# Time this connection out if we haven't send a request in the last
|
|
|
|
# N minutes
|
2016-12-29 10:51:04 -05:00
|
|
|
# TODO: Cancel the previous callLater?
|
2016-12-28 17:49:31 -05:00
|
|
|
reactor.callLater(3 * 60, self._time_things_out_maybe)
|
|
|
|
|
|
|
|
d = self.conn.request(request)
|
|
|
|
|
|
|
|
def update_request_time(res):
|
|
|
|
self.last_request = time.time()
|
2016-12-29 10:51:04 -05:00
|
|
|
# TODO: Cancel the previous callLater?
|
2016-12-28 17:49:31 -05:00
|
|
|
reactor.callLater(3 * 60, self._time_things_out_maybe)
|
|
|
|
return res
|
|
|
|
|
|
|
|
d.addCallback(update_request_time)
|
|
|
|
|
|
|
|
return d
|
2014-08-12 10:10:52 -04:00
|
|
|
|
|
|
|
|
2016-04-08 13:37:15 -04:00
|
|
|
class SpiderEndpoint(object):
|
|
|
|
"""An endpoint which refuses to connect to blacklisted IP addresses
|
|
|
|
Implements twisted.internet.interfaces.IStreamClientEndpoint.
|
|
|
|
"""
|
2016-05-01 07:44:24 -04:00
|
|
|
def __init__(self, reactor, host, port, blacklist, whitelist,
|
2016-12-12 10:19:54 -05:00
|
|
|
endpoint=HostnameEndpoint, endpoint_kw_args={}):
|
2016-04-08 13:37:15 -04:00
|
|
|
self.reactor = reactor
|
|
|
|
self.host = host
|
|
|
|
self.port = port
|
|
|
|
self.blacklist = blacklist
|
2016-05-01 07:44:24 -04:00
|
|
|
self.whitelist = whitelist
|
2016-04-08 13:37:15 -04:00
|
|
|
self.endpoint = endpoint
|
|
|
|
self.endpoint_kw_args = endpoint_kw_args
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def connect(self, protocolFactory):
|
|
|
|
address = yield self.reactor.resolve(self.host)
|
|
|
|
|
|
|
|
from netaddr import IPAddress
|
2016-05-01 07:44:24 -04:00
|
|
|
ip_address = IPAddress(address)
|
|
|
|
|
|
|
|
if ip_address in self.blacklist:
|
|
|
|
if self.whitelist is None or ip_address not in self.whitelist:
|
|
|
|
raise ConnectError(
|
|
|
|
"Refusing to spider blacklisted IP address %s" % address
|
|
|
|
)
|
2016-04-08 13:37:15 -04:00
|
|
|
|
|
|
|
logger.info("Connecting to %s:%s", address, self.port)
|
|
|
|
endpoint = self.endpoint(
|
|
|
|
self.reactor, address, self.port, **self.endpoint_kw_args
|
|
|
|
)
|
|
|
|
connection = yield endpoint.connect(protocolFactory)
|
|
|
|
defer.returnValue(connection)
|
|
|
|
|
|
|
|
|
2014-08-12 10:10:52 -04:00
|
|
|
class SRVClientEndpoint(object):
|
|
|
|
"""An endpoint which looks up SRV records for a service.
|
|
|
|
Cycles through the list of servers starting with each call to connect
|
|
|
|
picking the next server.
|
|
|
|
Implements twisted.internet.interfaces.IStreamClientEndpoint.
|
|
|
|
"""
|
|
|
|
|
|
|
|
def __init__(self, reactor, service, domain, protocol="tcp",
|
2016-12-12 10:19:54 -05:00
|
|
|
default_port=None, endpoint=HostnameEndpoint,
|
2014-08-12 10:10:52 -04:00
|
|
|
endpoint_kw_args={}):
|
|
|
|
self.reactor = reactor
|
|
|
|
self.service_name = "_%s._%s.%s" % (service, protocol, domain)
|
|
|
|
|
|
|
|
if default_port is not None:
|
2016-01-21 09:02:14 -05:00
|
|
|
self.default_server = _Server(
|
2014-08-12 10:10:52 -04:00
|
|
|
host=domain,
|
|
|
|
port=default_port,
|
|
|
|
priority=0,
|
2016-03-31 05:04:28 -04:00
|
|
|
weight=0,
|
|
|
|
expires=0,
|
2014-08-12 10:10:52 -04:00
|
|
|
)
|
|
|
|
else:
|
|
|
|
self.default_server = None
|
|
|
|
|
|
|
|
self.endpoint = endpoint
|
|
|
|
self.endpoint_kw_args = endpoint_kw_args
|
|
|
|
|
|
|
|
self.servers = None
|
|
|
|
self.used_servers = None
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def fetch_servers(self):
|
|
|
|
self.used_servers = []
|
2016-01-20 06:34:09 -05:00
|
|
|
self.servers = yield resolve_service(self.service_name)
|
2014-08-12 10:10:52 -04:00
|
|
|
|
|
|
|
def pick_server(self):
|
|
|
|
if not self.servers:
|
|
|
|
if self.used_servers:
|
|
|
|
self.servers = self.used_servers
|
|
|
|
self.used_servers = []
|
|
|
|
self.servers.sort()
|
|
|
|
elif self.default_server:
|
|
|
|
return self.default_server
|
|
|
|
else:
|
|
|
|
raise ConnectError(
|
2017-09-22 15:26:47 -04:00
|
|
|
"No server available for %s" % self.service_name
|
2014-08-12 10:10:52 -04:00
|
|
|
)
|
|
|
|
|
2017-09-22 15:26:47 -04:00
|
|
|
# look for all servers with the same priority
|
2014-08-12 10:10:52 -04:00
|
|
|
min_priority = self.servers[0].priority
|
|
|
|
weight_indexes = list(
|
|
|
|
(index, server.weight + 1)
|
|
|
|
for index, server in enumerate(self.servers)
|
|
|
|
if server.priority == min_priority
|
|
|
|
)
|
|
|
|
|
|
|
|
total_weight = sum(weight for index, weight in weight_indexes)
|
|
|
|
target_weight = random.randint(0, total_weight)
|
|
|
|
for index, weight in weight_indexes:
|
|
|
|
target_weight -= weight
|
|
|
|
if target_weight <= 0:
|
|
|
|
server = self.servers[index]
|
2017-09-22 15:26:47 -04:00
|
|
|
# XXX: this looks totally dubious:
|
|
|
|
#
|
|
|
|
# (a) we never reuse a server until we have been through
|
|
|
|
# all of the servers at the same priority, so if the
|
|
|
|
# weights are A: 100, B:1, we always do ABABAB instead of
|
|
|
|
# AAAA...AAAB (approximately).
|
|
|
|
#
|
|
|
|
# (b) After using all the servers at the lowest priority,
|
|
|
|
# we move onto the next priority. We should only use the
|
|
|
|
# second priority if servers at the top priority are
|
|
|
|
# unreachable.
|
|
|
|
#
|
2014-08-12 10:10:52 -04:00
|
|
|
del self.servers[index]
|
|
|
|
self.used_servers.append(server)
|
|
|
|
return server
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def connect(self, protocolFactory):
|
|
|
|
if self.servers is None:
|
|
|
|
yield self.fetch_servers()
|
|
|
|
server = self.pick_server()
|
|
|
|
logger.info("Connecting to %s:%s", server.host, server.port)
|
|
|
|
endpoint = self.endpoint(
|
|
|
|
self.reactor, server.host, server.port, **self.endpoint_kw_args
|
|
|
|
)
|
|
|
|
connection = yield endpoint.connect(protocolFactory)
|
|
|
|
defer.returnValue(connection)
|
2016-01-20 06:34:09 -05:00
|
|
|
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
2016-03-31 05:23:48 -04:00
|
|
|
def resolve_service(service_name, dns_client=client, cache=SERVER_CACHE, clock=time):
|
2016-03-31 05:04:28 -04:00
|
|
|
cache_entry = cache.get(service_name, None)
|
|
|
|
if cache_entry:
|
2016-03-31 05:23:48 -04:00
|
|
|
if all(s.expires > int(clock.time()) for s in cache_entry):
|
2016-03-31 05:04:28 -04:00
|
|
|
servers = list(cache_entry)
|
|
|
|
defer.returnValue(servers)
|
|
|
|
|
2016-01-20 06:34:09 -05:00
|
|
|
servers = []
|
|
|
|
|
|
|
|
try:
|
|
|
|
try:
|
|
|
|
answers, _, _ = yield dns_client.lookupService(service_name)
|
|
|
|
except DNSNameError:
|
|
|
|
defer.returnValue([])
|
|
|
|
|
|
|
|
if (len(answers) == 1
|
|
|
|
and answers[0].type == dns.SRV
|
|
|
|
and answers[0].payload
|
|
|
|
and answers[0].payload.target == dns.Name('.')):
|
2016-04-08 13:37:15 -04:00
|
|
|
raise ConnectError("Service %s unavailable" % service_name)
|
2016-01-20 06:34:09 -05:00
|
|
|
|
|
|
|
for answer in answers:
|
|
|
|
if answer.type != dns.SRV or not answer.payload:
|
|
|
|
continue
|
|
|
|
|
|
|
|
payload = answer.payload
|
|
|
|
|
2017-09-22 15:26:47 -04:00
|
|
|
hosts = yield _get_hosts_for_srv_record(
|
|
|
|
dns_client, str(payload.target)
|
|
|
|
)
|
2016-01-20 06:34:09 -05:00
|
|
|
|
2017-09-22 15:26:47 -04:00
|
|
|
for (ip, ttl) in hosts:
|
|
|
|
host_ttl = min(answer.ttl, ttl)
|
2016-03-31 05:04:28 -04:00
|
|
|
|
2017-09-22 15:26:47 -04:00
|
|
|
servers.append(_Server(
|
|
|
|
host=ip,
|
|
|
|
port=int(payload.port),
|
|
|
|
priority=int(payload.priority),
|
|
|
|
weight=int(payload.weight),
|
|
|
|
expires=int(clock.time()) + host_ttl,
|
|
|
|
))
|
2016-01-20 06:34:09 -05:00
|
|
|
|
|
|
|
servers.sort()
|
|
|
|
cache[service_name] = list(servers)
|
|
|
|
except DomainError as e:
|
|
|
|
# We failed to resolve the name (other than a NameError)
|
|
|
|
# Try something in the cache, else rereaise
|
|
|
|
cache_entry = cache.get(service_name, None)
|
|
|
|
if cache_entry:
|
|
|
|
logger.warn(
|
|
|
|
"Failed to resolve %r, falling back to cache. %r",
|
|
|
|
service_name, e
|
|
|
|
)
|
|
|
|
servers = list(cache_entry)
|
|
|
|
else:
|
|
|
|
raise e
|
|
|
|
|
|
|
|
defer.returnValue(servers)
|
2017-09-22 15:26:47 -04:00
|
|
|
|
|
|
|
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def _get_hosts_for_srv_record(dns_client, host):
|
|
|
|
"""Look up each of the hosts in a SRV record
|
|
|
|
|
|
|
|
Args:
|
|
|
|
dns_client (twisted.names.dns.IResolver):
|
|
|
|
host (basestring): host to look up
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
Deferred[list[(str, int)]]: a list of (host, ttl) pairs
|
|
|
|
|
|
|
|
"""
|
|
|
|
ip4_servers = []
|
|
|
|
ip6_servers = []
|
|
|
|
|
|
|
|
def cb(res):
|
|
|
|
# lookupAddress and lookupIP6Address return a three-tuple
|
|
|
|
# giving the answer, authority, and additional sections of the
|
|
|
|
# response.
|
|
|
|
#
|
|
|
|
# we only care about the answers.
|
|
|
|
|
|
|
|
return res[0]
|
|
|
|
|
2017-09-28 10:24:00 -04:00
|
|
|
def eb(res, record_type):
|
|
|
|
if res.check(DNSNameError):
|
|
|
|
return []
|
|
|
|
logger.warn("Error looking up %s for %s: %s",
|
|
|
|
record_type, host, res, res.value)
|
|
|
|
return res
|
2017-09-22 15:26:47 -04:00
|
|
|
|
|
|
|
# no logcontexts here, so we can safely fire these off and gatherResults
|
2017-11-24 11:44:56 -05:00
|
|
|
d1 = dns_client.lookupAddress(host).addCallbacks(
|
|
|
|
cb, eb, errbackArgs=("A", ))
|
|
|
|
d2 = dns_client.lookupIPV6Address(host).addCallbacks(
|
|
|
|
cb, eb, errbackArgs=("AAAA", ))
|
2017-09-28 10:24:00 -04:00
|
|
|
results = yield defer.DeferredList(
|
|
|
|
[d1, d2], consumeErrors=True)
|
|
|
|
|
|
|
|
# if all of the lookups failed, raise an exception rather than blowing out
|
|
|
|
# the cache with an empty result.
|
|
|
|
if results and all(s == defer.FAILURE for (s, _) in results):
|
|
|
|
defer.returnValue(results[0][1])
|
|
|
|
|
|
|
|
for (success, result) in results:
|
|
|
|
if success == defer.FAILURE:
|
|
|
|
continue
|
2017-09-22 15:26:47 -04:00
|
|
|
|
|
|
|
for answer in result:
|
|
|
|
if not answer.payload:
|
|
|
|
continue
|
|
|
|
|
|
|
|
try:
|
|
|
|
if answer.type == dns.A:
|
|
|
|
ip = answer.payload.dottedQuad()
|
|
|
|
ip4_servers.append((ip, answer.ttl))
|
|
|
|
elif answer.type == dns.AAAA:
|
|
|
|
ip = socket.inet_ntop(
|
|
|
|
socket.AF_INET6, answer.payload.address,
|
|
|
|
)
|
|
|
|
ip6_servers.append((ip, answer.ttl))
|
|
|
|
else:
|
|
|
|
# the most likely candidate here is a CNAME record.
|
|
|
|
# rfc2782 says srvs may not point to aliases.
|
|
|
|
logger.warn(
|
|
|
|
"Ignoring unexpected DNS record type %s for %s",
|
|
|
|
answer.type, host,
|
|
|
|
)
|
|
|
|
continue
|
|
|
|
except Exception as e:
|
|
|
|
logger.warn("Ignoring invalid DNS response for %s: %s",
|
|
|
|
host, e)
|
|
|
|
continue
|
|
|
|
|
|
|
|
# keep the ipv4 results before the ipv6 results, mostly to match historical
|
|
|
|
# behaviour.
|
|
|
|
defer.returnValue(ip4_servers + ip6_servers)
|