2020-05-08 08:30:40 -04:00
|
|
|
#
|
2023-11-21 15:29:58 -05:00
|
|
|
# This file is licensed under the Affero General Public License (AGPL) version 3.
|
|
|
|
#
|
|
|
|
# Copyright (C) 2023 New Vector, Ltd
|
|
|
|
#
|
|
|
|
# This program is free software: you can redistribute it and/or modify
|
|
|
|
# it under the terms of the GNU Affero General Public License as
|
|
|
|
# published by the Free Software Foundation, either version 3 of the
|
|
|
|
# License, or (at your option) any later version.
|
|
|
|
#
|
|
|
|
# See the GNU Affero General Public License for more details:
|
|
|
|
# <https://www.gnu.org/licenses/agpl-3.0.html>.
|
|
|
|
#
|
|
|
|
# Originally licensed under the Apache License, Version 2.0:
|
|
|
|
# <http://www.apache.org/licenses/LICENSE-2.0>.
|
|
|
|
#
|
|
|
|
# [This file includes modifications made by New Vector Limited]
|
2020-05-08 08:30:40 -04:00
|
|
|
#
|
|
|
|
#
|
2022-10-31 13:07:30 -04:00
|
|
|
import binascii
|
2020-11-25 10:04:22 -05:00
|
|
|
import inspect
|
2022-10-31 13:07:30 -04:00
|
|
|
import json
|
2020-05-08 08:30:40 -04:00
|
|
|
import logging
|
2022-10-31 13:07:30 -04:00
|
|
|
from typing import (
|
|
|
|
TYPE_CHECKING,
|
|
|
|
Any,
|
|
|
|
Dict,
|
|
|
|
Generic,
|
|
|
|
List,
|
|
|
|
Optional,
|
|
|
|
Type,
|
|
|
|
TypeVar,
|
|
|
|
Union,
|
|
|
|
)
|
2021-04-23 13:22:47 -04:00
|
|
|
from urllib.parse import urlencode, urlparse
|
2020-05-08 08:30:40 -04:00
|
|
|
|
|
|
|
import attr
|
2022-10-31 13:07:30 -04:00
|
|
|
import unpaddedbase64
|
2020-05-08 08:30:40 -04:00
|
|
|
from authlib.common.security import generate_token
|
2022-10-31 13:07:30 -04:00
|
|
|
from authlib.jose import JsonWebToken, JWTClaims
|
|
|
|
from authlib.jose.errors import InvalidClaimError, JoseError, MissingClaimError
|
2020-05-08 08:30:40 -04:00
|
|
|
from authlib.oauth2.auth import ClientAuth
|
|
|
|
from authlib.oauth2.rfc6749.parameters import prepare_grant_uri
|
2023-01-04 14:58:08 -05:00
|
|
|
from authlib.oauth2.rfc7636.challenge import create_s256_code_challenge
|
2021-12-06 12:43:06 -05:00
|
|
|
from authlib.oidc.core import CodeIDToken, UserInfo
|
2020-05-08 08:30:40 -04:00
|
|
|
from authlib.oidc.discovery import OpenIDProviderMetadata, get_well_known_url
|
|
|
|
from jinja2 import Environment, Template
|
|
|
|
from pymacaroons.exceptions import (
|
|
|
|
MacaroonDeserializationException,
|
2021-03-16 07:29:35 -04:00
|
|
|
MacaroonInitException,
|
2020-05-08 08:30:40 -04:00
|
|
|
MacaroonInvalidSignatureException,
|
|
|
|
)
|
|
|
|
from typing_extensions import TypedDict
|
|
|
|
|
|
|
|
from twisted.web.client import readBody
|
2021-03-12 11:37:57 -05:00
|
|
|
from twisted.web.http_headers import Headers
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2022-10-31 13:07:30 -04:00
|
|
|
from synapse.api.errors import SynapseError
|
2020-05-08 08:30:40 -04:00
|
|
|
from synapse.config import ConfigError
|
2021-04-20 14:55:20 -04:00
|
|
|
from synapse.config.oidc import OidcProviderClientSecretJwtKey, OidcProviderConfig
|
2020-11-25 10:04:22 -05:00
|
|
|
from synapse.handlers.sso import MappingException, UserAttributes
|
2022-10-31 13:07:30 -04:00
|
|
|
from synapse.http.server import finish_request
|
|
|
|
from synapse.http.servlet import parse_string
|
2020-05-08 08:30:40 -04:00
|
|
|
from synapse.http.site import SynapseRequest
|
2020-06-03 16:13:17 -04:00
|
|
|
from synapse.logging.context import make_deferred_yieldable
|
2020-12-02 07:45:42 -05:00
|
|
|
from synapse.types import JsonDict, UserID, map_username_to_mxid_localpart
|
2021-03-09 10:03:37 -05:00
|
|
|
from synapse.util import Clock, json_decoder
|
2021-02-16 11:27:38 -05:00
|
|
|
from synapse.util.caches.cached_call import RetryOnExceptionCachedCall
|
2022-06-14 09:12:08 -04:00
|
|
|
from synapse.util.macaroons import MacaroonGenerator, OidcSessionData
|
2022-03-11 10:15:11 -05:00
|
|
|
from synapse.util.templates import _localpart_from_email_filter
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2020-08-11 13:00:17 -04:00
|
|
|
if TYPE_CHECKING:
|
|
|
|
from synapse.server import HomeServer
|
|
|
|
|
2020-05-08 08:30:40 -04:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
2021-02-17 05:15:14 -05:00
|
|
|
# we want the cookie to be returned to us even when the request is the POSTed
|
|
|
|
# result of a form on another domain, as is used with `response_mode=form_post`.
|
|
|
|
#
|
|
|
|
# Modern browsers will not do so unless we set SameSite=None; however *older*
|
|
|
|
# browsers (including all versions of Safari on iOS 12?) don't support
|
|
|
|
# SameSite=None, and interpret it as SameSite=Strict:
|
|
|
|
# https://bugs.webkit.org/show_bug.cgi?id=198181
|
|
|
|
#
|
|
|
|
# As a rather painful workaround, we set *two* cookies, one with SameSite=None
|
|
|
|
# and one with no SameSite, in the hope that at least one of them will get
|
|
|
|
# back to us.
|
|
|
|
#
|
|
|
|
# Secure is necessary for SameSite=None (and, empirically, also breaks things
|
|
|
|
# on iOS 12.)
|
|
|
|
#
|
|
|
|
# Here we have the names of the cookies, and the options we use to set them.
|
|
|
|
_SESSION_COOKIES = [
|
2021-04-23 13:22:47 -04:00
|
|
|
(b"oidc_session", b"HttpOnly; Secure; SameSite=None"),
|
|
|
|
(b"oidc_session_no_samesite", b"HttpOnly"),
|
2021-02-17 05:15:14 -05:00
|
|
|
]
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2021-07-19 10:28:05 -04:00
|
|
|
|
2020-05-08 08:30:40 -04:00
|
|
|
#: A token exchanged from the token endpoint, as per RFC6749 sec 5.1. and
|
|
|
|
#: OpenID.Core sec 3.1.3.3.
|
2021-07-19 10:28:05 -04:00
|
|
|
class Token(TypedDict):
|
|
|
|
access_token: str
|
|
|
|
token_type: str
|
|
|
|
id_token: Optional[str]
|
|
|
|
refresh_token: Optional[str]
|
|
|
|
expires_in: int
|
|
|
|
scope: Optional[str]
|
|
|
|
|
2020-05-08 08:30:40 -04:00
|
|
|
|
|
|
|
#: A JWK, as per RFC7517 sec 4. The type could be more precise than that, but
|
|
|
|
#: there is no real point of doing this in our case.
|
|
|
|
JWK = Dict[str, str]
|
|
|
|
|
2022-10-31 13:07:30 -04:00
|
|
|
C = TypeVar("C")
|
|
|
|
|
2021-07-19 10:28:05 -04:00
|
|
|
|
2020-05-08 08:30:40 -04:00
|
|
|
#: A JWK Set, as per RFC7517 sec 5.
|
2021-07-19 10:28:05 -04:00
|
|
|
class JWKS(TypedDict):
|
|
|
|
keys: List[JWK]
|
2020-05-08 08:30:40 -04:00
|
|
|
|
|
|
|
|
2021-01-14 08:29:17 -05:00
|
|
|
class OidcHandler:
|
|
|
|
"""Handles requests related to the OpenID Connect login flow."""
|
|
|
|
|
|
|
|
def __init__(self, hs: "HomeServer"):
|
|
|
|
self._sso_handler = hs.get_sso_handler()
|
|
|
|
|
2021-01-15 11:55:29 -05:00
|
|
|
provider_confs = hs.config.oidc.oidc_providers
|
2021-01-14 08:29:17 -05:00
|
|
|
# we should not have been instantiated if there is no configured provider.
|
2021-01-15 11:55:29 -05:00
|
|
|
assert provider_confs
|
2021-01-14 08:29:17 -05:00
|
|
|
|
2022-06-14 09:12:08 -04:00
|
|
|
self._macaroon_generator = hs.get_macaroon_generator()
|
2021-07-16 13:22:36 -04:00
|
|
|
self._providers: Dict[str, "OidcProvider"] = {
|
2022-06-14 09:12:08 -04:00
|
|
|
p.idp_id: OidcProvider(hs, self._macaroon_generator, p)
|
|
|
|
for p in provider_confs
|
2021-07-16 13:22:36 -04:00
|
|
|
}
|
2021-01-14 08:29:17 -05:00
|
|
|
|
|
|
|
async def load_metadata(self) -> None:
|
|
|
|
"""Validate the config and load the metadata from the remote endpoint.
|
|
|
|
|
|
|
|
Called at startup to ensure we have everything we need.
|
|
|
|
"""
|
2021-01-15 11:55:29 -05:00
|
|
|
for idp_id, p in self._providers.items():
|
|
|
|
try:
|
|
|
|
await p.load_metadata()
|
2021-12-06 12:43:06 -05:00
|
|
|
if not p._uses_userinfo:
|
|
|
|
await p.load_jwks()
|
2021-01-15 11:55:29 -05:00
|
|
|
except Exception as e:
|
|
|
|
raise Exception(
|
|
|
|
"Error while initialising OIDC provider %r" % (idp_id,)
|
|
|
|
) from e
|
2021-01-14 08:29:17 -05:00
|
|
|
|
|
|
|
async def handle_oidc_callback(self, request: SynapseRequest) -> None:
|
2021-02-01 17:56:01 -05:00
|
|
|
"""Handle an incoming request to /_synapse/client/oidc/callback
|
2021-01-14 08:29:17 -05:00
|
|
|
|
|
|
|
Since we might want to display OIDC-related errors in a user-friendly
|
|
|
|
way, we don't raise SynapseError from here. Instead, we call
|
|
|
|
``self._sso_handler.render_error`` which displays an HTML page for the error.
|
|
|
|
|
|
|
|
Most of the OpenID Connect logic happens here:
|
|
|
|
|
|
|
|
- first, we check if there was any error returned by the provider and
|
|
|
|
display it
|
|
|
|
- then we fetch the session cookie, decode and verify it
|
|
|
|
- the ``state`` query parameter should match with the one stored in the
|
|
|
|
session cookie
|
|
|
|
|
|
|
|
Once we know the session is legit, we then delegate to the OIDC Provider
|
|
|
|
implementation, which will exchange the code with the provider and complete the
|
|
|
|
login/authentication.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
request: the incoming request from the browser.
|
|
|
|
"""
|
2021-03-26 12:49:46 -04:00
|
|
|
# This will always be set by the time Twisted calls us.
|
|
|
|
assert request.args is not None
|
|
|
|
|
2021-01-14 08:29:17 -05:00
|
|
|
# The provider might redirect with an error.
|
|
|
|
# In that case, just display it as-is.
|
|
|
|
if b"error" in request.args:
|
|
|
|
# error response from the auth server. see:
|
|
|
|
# https://tools.ietf.org/html/rfc6749#section-4.1.2.1
|
|
|
|
# https://openid.net/specs/openid-connect-core-1_0.html#AuthError
|
|
|
|
error = request.args[b"error"][0].decode()
|
|
|
|
description = request.args.get(b"error_description", [b""])[0].decode()
|
|
|
|
|
|
|
|
# Most of the errors returned by the provider could be due by
|
|
|
|
# either the provider misbehaving or Synapse being misconfigured.
|
|
|
|
# The only exception of that is "access_denied", where the user
|
|
|
|
# probably cancelled the login flow. In other cases, log those errors.
|
2021-02-03 14:45:34 -05:00
|
|
|
logger.log(
|
|
|
|
logging.INFO if error == "access_denied" else logging.ERROR,
|
|
|
|
"Received OIDC callback with error: %s %s",
|
|
|
|
error,
|
|
|
|
description,
|
|
|
|
)
|
2021-01-14 08:29:17 -05:00
|
|
|
|
|
|
|
self._sso_handler.render_error(request, error, description)
|
|
|
|
return
|
|
|
|
|
|
|
|
# otherwise, it is presumably a successful response. see:
|
|
|
|
# https://tools.ietf.org/html/rfc6749#section-4.1.2
|
|
|
|
|
2021-02-17 05:15:14 -05:00
|
|
|
# Fetch the session cookie. See the comments on SESSION_COOKIES for why there
|
|
|
|
# are two.
|
|
|
|
|
|
|
|
for cookie_name, _ in _SESSION_COOKIES:
|
2021-07-16 13:22:36 -04:00
|
|
|
session: Optional[bytes] = request.getCookie(cookie_name)
|
2021-02-17 05:15:14 -05:00
|
|
|
if session is not None:
|
|
|
|
break
|
|
|
|
else:
|
2021-02-03 14:45:34 -05:00
|
|
|
logger.info("Received OIDC callback, with no session cookie")
|
2021-01-14 08:29:17 -05:00
|
|
|
self._sso_handler.render_error(
|
|
|
|
request, "missing_session", "No session cookie found"
|
|
|
|
)
|
|
|
|
return
|
|
|
|
|
2021-02-17 05:15:14 -05:00
|
|
|
# Remove the cookies. There is a good chance that if the callback failed
|
2021-01-14 08:29:17 -05:00
|
|
|
# once, it will fail next time and the code will already be exchanged.
|
2021-02-17 05:15:14 -05:00
|
|
|
# Removing the cookies early avoids spamming the provider with token requests.
|
|
|
|
#
|
|
|
|
# we have to build the header by hand rather than calling request.addCookie
|
|
|
|
# because the latter does not support SameSite=None
|
|
|
|
# (https://twistedmatrix.com/trac/ticket/10088)
|
|
|
|
|
|
|
|
for cookie_name, options in _SESSION_COOKIES:
|
|
|
|
request.cookies.append(
|
|
|
|
b"%s=; Expires=Thu, Jan 01 1970 00:00:00 UTC; %s"
|
|
|
|
% (cookie_name, options)
|
|
|
|
)
|
2021-01-14 08:29:17 -05:00
|
|
|
|
|
|
|
# Check for the state query parameter
|
|
|
|
if b"state" not in request.args:
|
2021-02-03 14:45:34 -05:00
|
|
|
logger.info("Received OIDC callback, with no state parameter")
|
2021-01-14 08:29:17 -05:00
|
|
|
self._sso_handler.render_error(
|
|
|
|
request, "invalid_request", "State parameter is missing"
|
|
|
|
)
|
|
|
|
return
|
|
|
|
|
|
|
|
state = request.args[b"state"][0].decode()
|
|
|
|
|
|
|
|
# Deserialize the session token and verify it.
|
|
|
|
try:
|
2022-06-14 09:12:08 -04:00
|
|
|
session_data = self._macaroon_generator.verify_oidc_session_token(
|
2021-01-14 08:29:17 -05:00
|
|
|
session, state
|
|
|
|
)
|
2021-03-16 07:29:35 -04:00
|
|
|
except (MacaroonInitException, MacaroonDeserializationException, KeyError) as e:
|
2021-02-03 14:45:34 -05:00
|
|
|
logger.exception("Invalid session for OIDC callback")
|
2021-01-14 08:29:17 -05:00
|
|
|
self._sso_handler.render_error(request, "invalid_session", str(e))
|
|
|
|
return
|
|
|
|
except MacaroonInvalidSignatureException as e:
|
2022-05-18 15:10:21 -04:00
|
|
|
logger.warning("Could not verify session for OIDC callback: %s", e)
|
2021-01-14 08:29:17 -05:00
|
|
|
self._sso_handler.render_error(request, "mismatching_session", str(e))
|
|
|
|
return
|
|
|
|
|
2021-02-03 14:45:34 -05:00
|
|
|
logger.info("Received OIDC callback for IdP %s", session_data.idp_id)
|
|
|
|
|
2021-01-15 11:55:29 -05:00
|
|
|
oidc_provider = self._providers.get(session_data.idp_id)
|
|
|
|
if not oidc_provider:
|
|
|
|
logger.error("OIDC session uses unknown IdP %r", oidc_provider)
|
|
|
|
self._sso_handler.render_error(request, "unknown_idp", "Unknown IdP")
|
|
|
|
return
|
|
|
|
|
2021-01-14 08:29:17 -05:00
|
|
|
if b"code" not in request.args:
|
|
|
|
logger.info("Code parameter is missing")
|
|
|
|
self._sso_handler.render_error(
|
|
|
|
request, "invalid_request", "Code parameter is missing"
|
|
|
|
)
|
|
|
|
return
|
|
|
|
|
|
|
|
code = request.args[b"code"][0].decode()
|
|
|
|
|
2021-01-15 11:55:29 -05:00
|
|
|
await oidc_provider.handle_oidc_callback(request, session_data, code)
|
2021-01-14 08:29:17 -05:00
|
|
|
|
2022-10-31 13:07:30 -04:00
|
|
|
async def handle_backchannel_logout(self, request: SynapseRequest) -> None:
|
|
|
|
"""Handle an incoming request to /_synapse/client/oidc/backchannel_logout
|
|
|
|
|
|
|
|
This extracts the logout_token from the request and tries to figure out
|
|
|
|
which OpenID Provider it is comming from. This works by matching the iss claim
|
|
|
|
with the issuer and the aud claim with the client_id.
|
|
|
|
|
|
|
|
Since at this point we don't know who signed the JWT, we can't just
|
|
|
|
decode it using authlib since it will always verifies the signature. We
|
|
|
|
have to decode it manually without validating the signature. The actual JWT
|
|
|
|
verification is done in the `OidcProvider.handler_backchannel_logout` method,
|
|
|
|
once we figured out which provider sent the request.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
request: the incoming request from the browser.
|
|
|
|
"""
|
|
|
|
logout_token = parse_string(request, "logout_token")
|
|
|
|
if logout_token is None:
|
|
|
|
raise SynapseError(400, "Missing logout_token in request")
|
|
|
|
|
|
|
|
# A JWT looks like this:
|
|
|
|
# header.payload.signature
|
|
|
|
# where all parts are encoded with urlsafe base64.
|
|
|
|
# The aud and iss claims we care about are in the payload part, which
|
|
|
|
# is a JSON object.
|
|
|
|
try:
|
|
|
|
# By destructuring the list after splitting, we ensure that we have
|
|
|
|
# exactly 3 segments
|
|
|
|
_, payload, _ = logout_token.split(".")
|
|
|
|
except ValueError:
|
|
|
|
raise SynapseError(400, "Invalid logout_token in request")
|
|
|
|
|
|
|
|
try:
|
|
|
|
payload_bytes = unpaddedbase64.decode_base64(payload)
|
|
|
|
claims = json_decoder.decode(payload_bytes.decode("utf-8"))
|
|
|
|
except (json.JSONDecodeError, binascii.Error, UnicodeError):
|
|
|
|
raise SynapseError(400, "Invalid logout_token payload in request")
|
|
|
|
|
|
|
|
try:
|
|
|
|
# Let's extract the iss and aud claims
|
|
|
|
iss = claims["iss"]
|
|
|
|
aud = claims["aud"]
|
|
|
|
# The aud claim can be either a string or a list of string. Here we
|
|
|
|
# normalize it as a list of strings.
|
|
|
|
if isinstance(aud, str):
|
|
|
|
aud = [aud]
|
|
|
|
|
|
|
|
# Check that we have the right types for the aud and the iss claims
|
|
|
|
if not isinstance(iss, str) or not isinstance(aud, list):
|
|
|
|
raise TypeError()
|
|
|
|
for a in aud:
|
|
|
|
if not isinstance(a, str):
|
|
|
|
raise TypeError()
|
|
|
|
|
|
|
|
# At this point we properly checked both claims types
|
|
|
|
issuer: str = iss
|
|
|
|
audience: List[str] = aud
|
|
|
|
except (TypeError, KeyError):
|
|
|
|
raise SynapseError(400, "Invalid issuer/audience in logout_token")
|
|
|
|
|
|
|
|
# Now that we know the audience and the issuer, we can figure out from
|
|
|
|
# what provider it is coming from
|
|
|
|
oidc_provider: Optional[OidcProvider] = None
|
|
|
|
for provider in self._providers.values():
|
|
|
|
if provider.issuer == issuer and provider.client_id in audience:
|
|
|
|
oidc_provider = provider
|
|
|
|
break
|
|
|
|
|
|
|
|
if oidc_provider is None:
|
|
|
|
raise SynapseError(400, "Could not find the OP that issued this event")
|
|
|
|
|
|
|
|
# Ask the provider to handle the logout request.
|
|
|
|
await oidc_provider.handle_backchannel_logout(request, logout_token)
|
|
|
|
|
2021-01-14 08:29:17 -05:00
|
|
|
|
2020-05-08 08:30:40 -04:00
|
|
|
class OidcError(Exception):
|
|
|
|
"""Used to catch errors when calling the token_endpoint"""
|
|
|
|
|
2021-09-20 08:56:23 -04:00
|
|
|
def __init__(self, error: str, error_description: Optional[str] = None):
|
2020-05-08 08:30:40 -04:00
|
|
|
self.error = error
|
|
|
|
self.error_description = error_description
|
|
|
|
|
2021-09-20 08:56:23 -04:00
|
|
|
def __str__(self) -> str:
|
2020-05-08 08:30:40 -04:00
|
|
|
if self.error_description:
|
2021-07-19 10:28:05 -04:00
|
|
|
return f"{self.error}: {self.error_description}"
|
2020-05-08 08:30:40 -04:00
|
|
|
return self.error
|
|
|
|
|
|
|
|
|
2021-01-14 08:29:17 -05:00
|
|
|
class OidcProvider:
|
|
|
|
"""Wraps the config for a single OIDC IdentityProvider
|
|
|
|
|
|
|
|
Provides methods for handling redirect requests and callbacks via that particular
|
|
|
|
IdP.
|
2020-05-08 08:30:40 -04:00
|
|
|
"""
|
|
|
|
|
2021-01-14 08:29:17 -05:00
|
|
|
def __init__(
|
|
|
|
self,
|
|
|
|
hs: "HomeServer",
|
2022-06-14 09:12:08 -04:00
|
|
|
macaroon_generator: MacaroonGenerator,
|
2021-01-14 08:29:17 -05:00
|
|
|
provider: OidcProviderConfig,
|
|
|
|
):
|
2022-02-23 06:04:02 -05:00
|
|
|
self._store = hs.get_datastores().main
|
2022-10-25 10:25:02 -04:00
|
|
|
self._clock = hs.get_clock()
|
2021-01-13 05:26:12 -05:00
|
|
|
|
2022-06-14 09:12:08 -04:00
|
|
|
self._macaroon_generaton = macaroon_generator
|
2021-01-13 05:26:12 -05:00
|
|
|
|
2021-02-16 11:27:38 -05:00
|
|
|
self._config = provider
|
2021-09-23 12:03:01 -04:00
|
|
|
self._callback_url: str = hs.config.oidc.oidc_callback_url
|
2020-12-18 07:13:03 -05:00
|
|
|
|
2021-04-23 13:22:47 -04:00
|
|
|
# Calculate the prefix for OIDC callback paths based on the public_baseurl.
|
|
|
|
# We'll insert this into the Path= parameter of any session cookies we set.
|
|
|
|
public_baseurl_path = urlparse(hs.config.server.public_baseurl).path
|
|
|
|
self._callback_path_prefix = (
|
|
|
|
public_baseurl_path.encode("utf-8") + b"_synapse/client/oidc"
|
|
|
|
)
|
|
|
|
|
2021-03-16 11:46:07 -04:00
|
|
|
self._oidc_attribute_requirements = provider.attribute_requirements
|
2020-12-18 07:13:03 -05:00
|
|
|
self._scopes = provider.scopes
|
|
|
|
self._user_profile_method = provider.user_profile_method
|
2021-03-09 10:03:37 -05:00
|
|
|
|
2021-07-16 13:22:36 -04:00
|
|
|
client_secret: Optional[Union[str, JwtClientSecret]] = None
|
2021-03-09 10:03:37 -05:00
|
|
|
if provider.client_secret:
|
|
|
|
client_secret = provider.client_secret
|
|
|
|
elif provider.client_secret_jwt_key:
|
|
|
|
client_secret = JwtClientSecret(
|
|
|
|
provider.client_secret_jwt_key,
|
|
|
|
provider.client_id,
|
|
|
|
provider.issuer,
|
|
|
|
hs.get_clock(),
|
|
|
|
)
|
|
|
|
|
2020-05-08 08:30:40 -04:00
|
|
|
self._client_auth = ClientAuth(
|
2020-12-18 07:13:03 -05:00
|
|
|
provider.client_id,
|
2021-03-09 10:03:37 -05:00
|
|
|
client_secret,
|
2020-12-18 07:13:03 -05:00
|
|
|
provider.client_auth_method,
|
2021-07-16 13:22:36 -04:00
|
|
|
)
|
2020-12-18 07:13:03 -05:00
|
|
|
self._client_auth_method = provider.client_auth_method
|
2021-02-16 11:27:38 -05:00
|
|
|
|
|
|
|
# cache of metadata for the identity provider (endpoint uris, mostly). This is
|
|
|
|
# loaded on-demand from the discovery endpoint (if discovery is enabled), with
|
|
|
|
# possible overrides from the config. Access via `load_metadata`.
|
|
|
|
self._provider_metadata = RetryOnExceptionCachedCall(self._load_metadata)
|
|
|
|
|
|
|
|
# cache of JWKs used by the identity provider to sign tokens. Loaded on demand
|
|
|
|
# from the IdP's jwks_uri, if required.
|
|
|
|
self._jwks = RetryOnExceptionCachedCall(self._load_jwks)
|
|
|
|
|
2020-12-18 07:13:03 -05:00
|
|
|
self._user_mapping_provider = provider.user_mapping_provider_class(
|
|
|
|
provider.user_mapping_provider_config
|
|
|
|
)
|
|
|
|
self._skip_verification = provider.skip_verification
|
|
|
|
self._allow_existing_users = provider.allow_existing_users
|
2020-05-08 08:30:40 -04:00
|
|
|
|
|
|
|
self._http_client = hs.get_proxied_http_client()
|
2021-09-13 13:07:12 -04:00
|
|
|
self._server_name: str = hs.config.server.server_name
|
2020-05-08 08:30:40 -04:00
|
|
|
|
|
|
|
# identifier for the external_ids table
|
2021-01-15 08:22:12 -05:00
|
|
|
self.idp_id = provider.idp_id
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2021-01-05 06:25:28 -05:00
|
|
|
# user-facing name of this auth provider
|
2021-01-15 08:22:12 -05:00
|
|
|
self.idp_name = provider.idp_name
|
2021-01-05 06:25:28 -05:00
|
|
|
|
2021-01-20 08:15:14 -05:00
|
|
|
# MXC URI for icon for this auth provider
|
|
|
|
self.idp_icon = provider.idp_icon
|
|
|
|
|
2021-01-27 16:31:45 -05:00
|
|
|
# optional brand identifier for this auth provider
|
|
|
|
self.idp_brand = provider.idp_brand
|
|
|
|
|
2020-11-17 09:46:23 -05:00
|
|
|
self._sso_handler = hs.get_sso_handler()
|
2022-10-31 13:07:30 -04:00
|
|
|
self._device_handler = hs.get_device_handler()
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2021-01-04 13:13:49 -05:00
|
|
|
self._sso_handler.register_identity_provider(self)
|
|
|
|
|
2021-02-16 11:27:38 -05:00
|
|
|
def _validate_metadata(self, m: OpenIDProviderMetadata) -> None:
|
2020-05-08 08:30:40 -04:00
|
|
|
"""Verifies the provider metadata.
|
|
|
|
|
|
|
|
This checks the validity of the currently loaded provider. Not
|
|
|
|
everything is checked, only:
|
|
|
|
|
|
|
|
- ``issuer``
|
|
|
|
- ``authorization_endpoint``
|
|
|
|
- ``token_endpoint``
|
|
|
|
- ``response_types_supported`` (checks if "code" is in it)
|
|
|
|
- ``jwks_uri``
|
|
|
|
|
|
|
|
Raises:
|
|
|
|
ValueError: if something in the provider is not valid
|
|
|
|
"""
|
|
|
|
# Skip verification to allow non-compliant providers (e.g. issuers not running on a secure origin)
|
|
|
|
if self._skip_verification is True:
|
|
|
|
return
|
|
|
|
|
|
|
|
m.validate_issuer()
|
|
|
|
m.validate_authorization_endpoint()
|
|
|
|
m.validate_token_endpoint()
|
|
|
|
|
|
|
|
if m.get("token_endpoint_auth_methods_supported") is not None:
|
|
|
|
m.validate_token_endpoint_auth_methods_supported()
|
|
|
|
if (
|
|
|
|
self._client_auth_method
|
|
|
|
not in m["token_endpoint_auth_methods_supported"]
|
|
|
|
):
|
|
|
|
raise ValueError(
|
|
|
|
'"{auth_method}" not in "token_endpoint_auth_methods_supported" ({supported!r})'.format(
|
|
|
|
auth_method=self._client_auth_method,
|
|
|
|
supported=m["token_endpoint_auth_methods_supported"],
|
|
|
|
)
|
|
|
|
)
|
|
|
|
|
2023-01-04 14:58:08 -05:00
|
|
|
# If PKCE support is advertised ensure the wanted method is available.
|
|
|
|
if m.get("code_challenge_methods_supported") is not None:
|
|
|
|
m.validate_code_challenge_methods_supported()
|
|
|
|
if "S256" not in m["code_challenge_methods_supported"]:
|
|
|
|
raise ValueError(
|
|
|
|
'"S256" not in "code_challenge_methods_supported" ({supported!r})'.format(
|
|
|
|
supported=m["code_challenge_methods_supported"],
|
|
|
|
)
|
|
|
|
)
|
|
|
|
|
2020-05-08 08:30:40 -04:00
|
|
|
if m.get("response_types_supported") is not None:
|
|
|
|
m.validate_response_types_supported()
|
|
|
|
|
|
|
|
if "code" not in m["response_types_supported"]:
|
|
|
|
raise ValueError(
|
|
|
|
'"code" not in "response_types_supported" (%r)'
|
|
|
|
% (m["response_types_supported"],)
|
|
|
|
)
|
|
|
|
|
2020-10-01 13:54:35 -04:00
|
|
|
# Ensure there's a userinfo endpoint to fetch from if it is required.
|
2020-05-08 08:30:40 -04:00
|
|
|
if self._uses_userinfo:
|
|
|
|
if m.get("userinfo_endpoint") is None:
|
|
|
|
raise ValueError(
|
2020-10-01 13:54:35 -04:00
|
|
|
'provider has no "userinfo_endpoint", even though it is required'
|
2020-05-08 08:30:40 -04:00
|
|
|
)
|
|
|
|
else:
|
|
|
|
# If we're not using userinfo, we need a valid jwks to validate the ID token
|
2021-02-16 11:27:38 -05:00
|
|
|
m.validate_jwks_uri()
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2022-10-31 13:07:30 -04:00
|
|
|
if self._config.backchannel_logout_enabled:
|
|
|
|
if not m.get("backchannel_logout_supported", False):
|
|
|
|
logger.warning(
|
|
|
|
"OIDC Back-Channel Logout is enabled for issuer %r"
|
|
|
|
"but it does not advertise support for it",
|
|
|
|
self.issuer,
|
|
|
|
)
|
|
|
|
|
|
|
|
elif not m.get("backchannel_logout_session_supported", False):
|
|
|
|
logger.warning(
|
|
|
|
"OIDC Back-Channel Logout is enabled and supported "
|
|
|
|
"by issuer %r but it might not send a session ID with "
|
|
|
|
"logout tokens, which is required for the logouts to work",
|
|
|
|
self.issuer,
|
|
|
|
)
|
|
|
|
|
|
|
|
if not self._config.backchannel_logout_ignore_sub:
|
|
|
|
# If OIDC backchannel logouts are enabled, the provider mapping provider
|
|
|
|
# should use the `sub` claim. We verify that by mapping a dumb user and
|
|
|
|
# see if we get back the sub claim
|
|
|
|
user = UserInfo({"sub": "thisisasubject"})
|
|
|
|
try:
|
|
|
|
subject = self._user_mapping_provider.get_remote_user_id(user)
|
|
|
|
if subject != user["sub"]:
|
|
|
|
raise ValueError("Unexpected subject")
|
|
|
|
except Exception:
|
|
|
|
logger.warning(
|
|
|
|
f"OIDC Back-Channel Logout is enabled for issuer {self.issuer!r} "
|
|
|
|
"but it looks like the configured `user_mapping_provider` "
|
|
|
|
"does not use the `sub` claim as subject. If it is the case, "
|
|
|
|
"and you want Synapse to ignore the `sub` claim in OIDC "
|
|
|
|
"Back-Channel Logouts, set `backchannel_logout_ignore_sub` "
|
|
|
|
"to `true` in the issuer config."
|
|
|
|
)
|
|
|
|
|
2020-05-08 08:30:40 -04:00
|
|
|
@property
|
|
|
|
def _uses_userinfo(self) -> bool:
|
|
|
|
"""Returns True if the ``userinfo_endpoint`` should be used.
|
|
|
|
|
|
|
|
This is based on the requested scopes: if the scopes include
|
|
|
|
``openid``, the provider should give use an ID token containing the
|
2020-10-23 12:38:40 -04:00
|
|
|
user information. If not, we should fetch them using the
|
2020-05-08 08:30:40 -04:00
|
|
|
``access_token`` with the ``userinfo_endpoint``.
|
|
|
|
"""
|
|
|
|
|
2020-10-01 13:54:35 -04:00
|
|
|
return (
|
|
|
|
"openid" not in self._scopes
|
|
|
|
or self._user_profile_method == "userinfo_endpoint"
|
|
|
|
)
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2022-10-31 13:07:30 -04:00
|
|
|
@property
|
|
|
|
def issuer(self) -> str:
|
|
|
|
"""The issuer identifying this provider."""
|
|
|
|
return self._config.issuer
|
|
|
|
|
|
|
|
@property
|
|
|
|
def client_id(self) -> str:
|
|
|
|
"""The client_id used when interacting with this provider."""
|
|
|
|
return self._config.client_id
|
|
|
|
|
2021-02-16 11:27:38 -05:00
|
|
|
async def load_metadata(self, force: bool = False) -> OpenIDProviderMetadata:
|
|
|
|
"""Return the provider metadata.
|
|
|
|
|
|
|
|
If this is the first call, the metadata is built from the config and from the
|
|
|
|
metadata discovery endpoint (if enabled), and then validated. If the metadata
|
|
|
|
is successfully validated, it is then cached for future use.
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2021-02-16 11:27:38 -05:00
|
|
|
Args:
|
|
|
|
force: If true, any cached metadata is discarded to force a reload.
|
2020-05-08 08:30:40 -04:00
|
|
|
|
|
|
|
Raises:
|
|
|
|
ValueError: if something in the provider is not valid
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
The provider's metadata.
|
|
|
|
"""
|
2021-02-16 11:27:38 -05:00
|
|
|
if force:
|
|
|
|
# reset the cached call to ensure we get a new result
|
|
|
|
self._provider_metadata = RetryOnExceptionCachedCall(self._load_metadata)
|
|
|
|
|
|
|
|
return await self._provider_metadata.get()
|
|
|
|
|
|
|
|
async def _load_metadata(self) -> OpenIDProviderMetadata:
|
2021-02-16 17:33:09 -05:00
|
|
|
# start out with just the issuer (unlike the other settings, discovered issuer
|
|
|
|
# takes precedence over configured issuer, because configured issuer is
|
|
|
|
# required for discovery to take place.)
|
|
|
|
#
|
|
|
|
metadata = OpenIDProviderMetadata(issuer=self._config.issuer)
|
2021-02-16 11:27:38 -05:00
|
|
|
|
|
|
|
# load any data from the discovery endpoint, if enabled
|
|
|
|
if self._config.discover:
|
|
|
|
url = get_well_known_url(self._config.issuer, external=True)
|
2020-05-08 08:30:40 -04:00
|
|
|
metadata_response = await self._http_client.get_json(url)
|
2021-02-16 11:27:38 -05:00
|
|
|
metadata.update(metadata_response)
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2021-02-16 17:33:09 -05:00
|
|
|
# override any discovered data with any settings in our config
|
|
|
|
if self._config.authorization_endpoint:
|
|
|
|
metadata["authorization_endpoint"] = self._config.authorization_endpoint
|
|
|
|
|
|
|
|
if self._config.token_endpoint:
|
|
|
|
metadata["token_endpoint"] = self._config.token_endpoint
|
|
|
|
|
|
|
|
if self._config.userinfo_endpoint:
|
|
|
|
metadata["userinfo_endpoint"] = self._config.userinfo_endpoint
|
|
|
|
|
|
|
|
if self._config.jwks_uri:
|
|
|
|
metadata["jwks_uri"] = self._config.jwks_uri
|
|
|
|
|
2023-01-04 14:58:08 -05:00
|
|
|
if self._config.pkce_method == "always":
|
|
|
|
metadata["code_challenge_methods_supported"] = ["S256"]
|
|
|
|
elif self._config.pkce_method == "never":
|
|
|
|
metadata.pop("code_challenge_methods_supported", None)
|
|
|
|
|
2021-02-16 11:27:38 -05:00
|
|
|
self._validate_metadata(metadata)
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2021-02-16 11:27:38 -05:00
|
|
|
return metadata
|
2020-05-08 08:30:40 -04:00
|
|
|
|
|
|
|
async def load_jwks(self, force: bool = False) -> JWKS:
|
|
|
|
"""Load the JSON Web Key Set used to sign ID tokens.
|
|
|
|
|
|
|
|
If we're not using the ``userinfo_endpoint``, user infos are extracted
|
|
|
|
from the ID token, which is a JWT signed by keys given by the provider.
|
|
|
|
The keys are then cached.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
force: Force reloading the keys.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
The key set
|
|
|
|
|
|
|
|
Looks like this::
|
|
|
|
|
|
|
|
{
|
|
|
|
'keys': [
|
|
|
|
{
|
|
|
|
'kid': 'abcdef',
|
|
|
|
'kty': 'RSA',
|
|
|
|
'alg': 'RS256',
|
|
|
|
'use': 'sig',
|
|
|
|
'e': 'XXXX',
|
|
|
|
'n': 'XXXX',
|
|
|
|
}
|
|
|
|
]
|
|
|
|
}
|
|
|
|
"""
|
2021-02-16 11:27:38 -05:00
|
|
|
if force:
|
|
|
|
# reset the cached call to ensure we get a new result
|
|
|
|
self._jwks = RetryOnExceptionCachedCall(self._load_jwks)
|
|
|
|
return await self._jwks.get()
|
|
|
|
|
|
|
|
async def _load_jwks(self) -> JWKS:
|
2020-05-08 08:30:40 -04:00
|
|
|
metadata = await self.load_metadata()
|
|
|
|
|
2021-02-16 11:27:38 -05:00
|
|
|
# Load the JWKS using the `jwks_uri` metadata.
|
2020-05-08 08:30:40 -04:00
|
|
|
uri = metadata.get("jwks_uri")
|
|
|
|
if not uri:
|
2021-02-16 11:27:38 -05:00
|
|
|
# this should be unreachable: load_metadata validates that
|
|
|
|
# there is a jwks_uri in the metadata if _uses_userinfo is unset
|
2020-05-08 08:30:40 -04:00
|
|
|
raise RuntimeError('Missing "jwks_uri" in metadata')
|
|
|
|
|
|
|
|
jwk_set = await self._http_client.get_json(uri)
|
|
|
|
|
|
|
|
return jwk_set
|
|
|
|
|
2023-01-04 14:58:08 -05:00
|
|
|
async def _exchange_code(self, code: str, code_verifier: str) -> Token:
|
2020-05-08 08:30:40 -04:00
|
|
|
"""Exchange an authorization code for a token.
|
|
|
|
|
|
|
|
This calls the ``token_endpoint`` with the authorization code we
|
|
|
|
received in the callback to exchange it for a token. The call uses the
|
|
|
|
``ClientAuth`` to authenticate with the client with its ID and secret.
|
|
|
|
|
2020-06-03 16:13:17 -04:00
|
|
|
See:
|
|
|
|
https://tools.ietf.org/html/rfc6749#section-3.2
|
|
|
|
https://openid.net/specs/openid-connect-core-1_0.html#TokenEndpoint
|
|
|
|
|
2020-05-08 08:30:40 -04:00
|
|
|
Args:
|
2020-05-15 12:26:02 -04:00
|
|
|
code: The authorization code we got from the callback.
|
2023-01-04 14:58:08 -05:00
|
|
|
code_verifier: The PKCE code verifier to send, blank if unused.
|
2020-05-08 08:30:40 -04:00
|
|
|
|
|
|
|
Returns:
|
|
|
|
A dict containing various tokens.
|
|
|
|
|
|
|
|
May look like this::
|
|
|
|
|
|
|
|
{
|
|
|
|
'token_type': 'bearer',
|
|
|
|
'access_token': 'abcdef',
|
|
|
|
'expires_in': 3599,
|
|
|
|
'id_token': 'ghijkl',
|
|
|
|
'refresh_token': 'mnopqr',
|
|
|
|
}
|
|
|
|
|
|
|
|
Raises:
|
|
|
|
OidcError: when the ``token_endpoint`` returned an error.
|
|
|
|
"""
|
|
|
|
metadata = await self.load_metadata()
|
|
|
|
token_endpoint = metadata.get("token_endpoint")
|
2022-02-11 07:20:16 -05:00
|
|
|
raw_headers: Dict[str, str] = {
|
2020-05-08 08:30:40 -04:00
|
|
|
"Content-Type": "application/x-www-form-urlencoded",
|
2022-02-11 07:20:16 -05:00
|
|
|
"User-Agent": self._http_client.user_agent.decode("ascii"),
|
2020-05-08 08:30:40 -04:00
|
|
|
"Accept": "application/json",
|
|
|
|
}
|
|
|
|
|
|
|
|
args = {
|
|
|
|
"grant_type": "authorization_code",
|
|
|
|
"code": code,
|
|
|
|
"redirect_uri": self._callback_url,
|
|
|
|
}
|
2023-01-04 14:58:08 -05:00
|
|
|
if code_verifier:
|
|
|
|
args["code_verifier"] = code_verifier
|
2020-05-08 08:30:40 -04:00
|
|
|
body = urlencode(args, True)
|
|
|
|
|
|
|
|
# Fill the body/headers with credentials
|
2021-03-12 11:37:57 -05:00
|
|
|
uri, raw_headers, body = self._client_auth.prepare(
|
|
|
|
method="POST", uri=token_endpoint, headers=raw_headers, body=body
|
2020-05-08 08:30:40 -04:00
|
|
|
)
|
2021-03-12 11:37:57 -05:00
|
|
|
headers = Headers({k: [v] for (k, v) in raw_headers.items()})
|
2020-05-08 08:30:40 -04:00
|
|
|
|
|
|
|
# Do the actual request
|
|
|
|
# We're not using the SimpleHttpClient util methods as we don't want to
|
|
|
|
# check the HTTP status code and we do the body encoding ourself.
|
|
|
|
response = await self._http_client.request(
|
|
|
|
method="POST",
|
|
|
|
uri=uri,
|
|
|
|
data=body.encode("utf-8"),
|
|
|
|
headers=headers,
|
|
|
|
)
|
|
|
|
|
|
|
|
# This is used in multiple error messages below
|
|
|
|
status = "{code} {phrase}".format(
|
|
|
|
code=response.code, phrase=response.phrase.decode("utf-8")
|
|
|
|
)
|
|
|
|
|
2020-06-03 16:13:17 -04:00
|
|
|
resp_body = await make_deferred_yieldable(readBody(response))
|
2020-05-08 08:30:40 -04:00
|
|
|
|
|
|
|
if response.code >= 500:
|
|
|
|
# In case of a server error, we should first try to decode the body
|
|
|
|
# and check for an error field. If not, we respond with a generic
|
|
|
|
# error message.
|
|
|
|
try:
|
2020-08-19 07:26:03 -04:00
|
|
|
resp = json_decoder.decode(resp_body.decode("utf-8"))
|
2020-05-08 08:30:40 -04:00
|
|
|
error = resp["error"]
|
|
|
|
description = resp.get("error_description", error)
|
|
|
|
except (ValueError, KeyError):
|
|
|
|
# Catch ValueError for the JSON decoding and KeyError for the "error" field
|
|
|
|
error = "server_error"
|
|
|
|
description = (
|
|
|
|
(
|
|
|
|
'Authorization server responded with a "{status}" error '
|
|
|
|
"while exchanging the authorization code."
|
|
|
|
).format(status=status),
|
|
|
|
)
|
|
|
|
|
|
|
|
raise OidcError(error, description)
|
|
|
|
|
|
|
|
# Since it is a not a 5xx code, body should be a valid JSON. It will
|
|
|
|
# raise if not.
|
2020-08-19 07:26:03 -04:00
|
|
|
resp = json_decoder.decode(resp_body.decode("utf-8"))
|
2020-05-08 08:30:40 -04:00
|
|
|
|
|
|
|
if "error" in resp:
|
|
|
|
error = resp["error"]
|
|
|
|
# In case the authorization server responded with an error field,
|
|
|
|
# it should be a 4xx code. If not, warn about it but don't do
|
|
|
|
# anything special and report the original error message.
|
|
|
|
if response.code < 400:
|
|
|
|
logger.debug(
|
|
|
|
"Invalid response from the authorization server: "
|
|
|
|
'responded with a "{status}" '
|
|
|
|
"but body has an error field: {error!r}".format(
|
|
|
|
status=status, error=resp["error"]
|
|
|
|
)
|
|
|
|
)
|
|
|
|
|
|
|
|
description = resp.get("error_description", error)
|
|
|
|
raise OidcError(error, description)
|
|
|
|
|
|
|
|
# Now, this should not be an error. According to RFC6749 sec 5.1, it
|
|
|
|
# should be a 200 code. We're a bit more flexible than that, and will
|
|
|
|
# only throw on a 4xx code.
|
|
|
|
if response.code >= 400:
|
|
|
|
description = (
|
|
|
|
'Authorization server responded with a "{status}" error '
|
|
|
|
'but did not include an "error" field in its response.'.format(
|
|
|
|
status=status
|
|
|
|
)
|
|
|
|
)
|
|
|
|
logger.warning(description)
|
|
|
|
# Body was still valid JSON. Might be useful to log it for debugging.
|
2021-07-19 10:28:05 -04:00
|
|
|
logger.warning("Code exchange response: %r", resp)
|
2020-05-08 08:30:40 -04:00
|
|
|
raise OidcError("server_error", description)
|
|
|
|
|
|
|
|
return resp
|
|
|
|
|
|
|
|
async def _fetch_userinfo(self, token: Token) -> UserInfo:
|
2020-10-23 12:38:40 -04:00
|
|
|
"""Fetch user information from the ``userinfo_endpoint``.
|
2020-05-08 08:30:40 -04:00
|
|
|
|
|
|
|
Args:
|
|
|
|
token: the token given by the ``token_endpoint``.
|
|
|
|
Must include an ``access_token`` field.
|
|
|
|
|
|
|
|
Returns:
|
2022-11-16 10:25:24 -05:00
|
|
|
an object representing the user.
|
2020-05-08 08:30:40 -04:00
|
|
|
"""
|
2021-02-03 14:45:34 -05:00
|
|
|
logger.debug("Using the OAuth2 access_token to request userinfo")
|
2020-05-08 08:30:40 -04:00
|
|
|
metadata = await self.load_metadata()
|
|
|
|
|
|
|
|
resp = await self._http_client.get_json(
|
|
|
|
metadata["userinfo_endpoint"],
|
|
|
|
headers={"Authorization": ["Bearer {}".format(token["access_token"])]},
|
|
|
|
)
|
|
|
|
|
2021-02-03 14:45:34 -05:00
|
|
|
logger.debug("Retrieved user info from userinfo endpoint: %r", resp)
|
|
|
|
|
2020-05-08 08:30:40 -04:00
|
|
|
return UserInfo(resp)
|
|
|
|
|
2022-10-31 13:07:30 -04:00
|
|
|
async def _verify_jwt(
|
|
|
|
self,
|
|
|
|
alg_values: List[str],
|
|
|
|
token: str,
|
|
|
|
claims_cls: Type[C],
|
|
|
|
claims_options: Optional[dict] = None,
|
|
|
|
claims_params: Optional[dict] = None,
|
|
|
|
) -> C:
|
|
|
|
"""Decode and validate a JWT, re-fetching the JWKS as needed.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
alg_values: list of `alg` values allowed when verifying the JWT.
|
|
|
|
token: the JWT.
|
|
|
|
claims_cls: the JWTClaims class to use to validate the claims.
|
|
|
|
claims_options: dict of options passed to the `claims_cls` constructor.
|
|
|
|
claims_params: dict of params passed to the `claims_cls` constructor.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
The decoded claims in the JWT.
|
|
|
|
"""
|
|
|
|
jwt = JsonWebToken(alg_values)
|
|
|
|
|
|
|
|
logger.debug("Attempting to decode JWT (%s) %r", claims_cls.__name__, token)
|
|
|
|
|
|
|
|
# Try to decode the keys in cache first, then retry by forcing the keys
|
|
|
|
# to be reloaded
|
|
|
|
jwk_set = await self.load_jwks()
|
|
|
|
try:
|
|
|
|
claims = jwt.decode(
|
|
|
|
token,
|
|
|
|
key=jwk_set,
|
|
|
|
claims_cls=claims_cls,
|
|
|
|
claims_options=claims_options,
|
|
|
|
claims_params=claims_params,
|
|
|
|
)
|
|
|
|
except ValueError:
|
|
|
|
logger.info("Reloading JWKS after decode error")
|
|
|
|
jwk_set = await self.load_jwks(force=True) # try reloading the jwks
|
|
|
|
claims = jwt.decode(
|
|
|
|
token,
|
|
|
|
key=jwk_set,
|
|
|
|
claims_cls=claims_cls,
|
|
|
|
claims_options=claims_options,
|
|
|
|
claims_params=claims_params,
|
|
|
|
)
|
|
|
|
|
|
|
|
logger.debug("Decoded JWT (%s) %r; validating", claims_cls.__name__, claims)
|
|
|
|
|
|
|
|
claims.validate(
|
|
|
|
now=self._clock.time(), leeway=120
|
|
|
|
) # allows 2 min of clock skew
|
|
|
|
return claims
|
|
|
|
|
2021-12-06 12:43:06 -05:00
|
|
|
async def _parse_id_token(self, token: Token, nonce: str) -> CodeIDToken:
|
2020-05-08 08:30:40 -04:00
|
|
|
"""Return an instance of UserInfo from token's ``id_token``.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
token: the token given by the ``token_endpoint``.
|
|
|
|
Must include an ``id_token`` field.
|
|
|
|
nonce: the nonce value originally sent in the initial authorization
|
|
|
|
request. This value should match the one inside the token.
|
|
|
|
|
|
|
|
Returns:
|
2021-12-06 12:43:06 -05:00
|
|
|
The decoded claims in the ID token.
|
2020-05-08 08:30:40 -04:00
|
|
|
"""
|
2022-10-25 10:25:02 -04:00
|
|
|
id_token = token.get("id_token")
|
|
|
|
|
|
|
|
# That has been theoritically been checked by the caller, so even though
|
|
|
|
# assertion are not enabled in production, it is mainly here to appease mypy
|
|
|
|
assert id_token is not None
|
|
|
|
|
2020-05-08 08:30:40 -04:00
|
|
|
metadata = await self.load_metadata()
|
2022-10-31 13:07:30 -04:00
|
|
|
|
2020-05-08 08:30:40 -04:00
|
|
|
claims_params = {
|
|
|
|
"nonce": nonce,
|
|
|
|
"client_id": self._client_auth.client_id,
|
|
|
|
}
|
|
|
|
if "access_token" in token:
|
|
|
|
# If we got an `access_token`, there should be an `at_hash` claim
|
|
|
|
# in the `id_token` that we can check against.
|
|
|
|
claims_params["access_token"] = token["access_token"]
|
|
|
|
|
2022-10-31 13:07:30 -04:00
|
|
|
claims_options = {"iss": {"values": [metadata["issuer"]]}}
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2022-10-31 13:07:30 -04:00
|
|
|
alg_values = metadata.get("id_token_signing_alg_values_supported", ["RS256"])
|
2021-02-03 14:45:34 -05:00
|
|
|
|
2022-10-31 13:07:30 -04:00
|
|
|
claims = await self._verify_jwt(
|
|
|
|
alg_values=alg_values,
|
|
|
|
token=id_token,
|
|
|
|
claims_cls=CodeIDToken,
|
|
|
|
claims_options=claims_options,
|
|
|
|
claims_params=claims_params,
|
|
|
|
)
|
2021-12-06 12:43:06 -05:00
|
|
|
|
|
|
|
return claims
|
2020-05-08 08:30:40 -04:00
|
|
|
|
|
|
|
async def handle_redirect_request(
|
2020-05-15 12:26:02 -04:00
|
|
|
self,
|
|
|
|
request: SynapseRequest,
|
2021-01-04 13:13:49 -05:00
|
|
|
client_redirect_url: Optional[bytes],
|
2020-05-15 12:26:02 -04:00
|
|
|
ui_auth_session_id: Optional[str] = None,
|
|
|
|
) -> str:
|
2020-05-08 08:30:40 -04:00
|
|
|
"""Handle an incoming request to /login/sso/redirect
|
|
|
|
|
2020-05-15 12:26:02 -04:00
|
|
|
It returns a redirect to the authorization endpoint with a few
|
2020-05-08 08:30:40 -04:00
|
|
|
parameters:
|
|
|
|
|
|
|
|
- ``client_id``: the client ID set in ``oidc_config.client_id``
|
|
|
|
- ``response_type``: ``code``
|
2021-02-01 17:56:01 -05:00
|
|
|
- ``redirect_uri``: the callback URL ; ``{base url}/_synapse/client/oidc/callback``
|
2020-05-08 08:30:40 -04:00
|
|
|
- ``scope``: the list of scopes set in ``oidc_config.scopes``
|
|
|
|
- ``state``: a random string
|
|
|
|
- ``nonce``: a random string
|
2023-01-04 14:58:08 -05:00
|
|
|
- ``code_challenge``: a RFC7636 code challenge (if PKCE is supported)
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2023-01-04 14:58:08 -05:00
|
|
|
In addition to generating a redirect URL, we are setting a cookie with
|
|
|
|
a signed macaroon token containing the state, the nonce, the
|
|
|
|
client_redirect_url, and (optionally) the code_verifier params. The state,
|
|
|
|
nonce, and client_redirect_url are then checked when the client comes back
|
|
|
|
from the provider. The code_verifier is passed back to the server during
|
|
|
|
the token exchange and compared to the code_challenge sent in this request.
|
2020-05-08 08:30:40 -04:00
|
|
|
|
|
|
|
Args:
|
|
|
|
request: the incoming request from the browser.
|
|
|
|
We'll respond to it with a redirect and a cookie.
|
|
|
|
client_redirect_url: the URL that we should redirect the client to
|
2021-01-04 13:13:49 -05:00
|
|
|
when everything is done (or None for UI Auth)
|
2020-05-15 12:26:02 -04:00
|
|
|
ui_auth_session_id: The session ID of the ongoing UI Auth (or
|
|
|
|
None if this is a login).
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
The redirect URL to the authorization endpoint.
|
|
|
|
|
2020-05-08 08:30:40 -04:00
|
|
|
"""
|
|
|
|
|
|
|
|
state = generate_token()
|
|
|
|
nonce = generate_token()
|
2023-01-04 14:58:08 -05:00
|
|
|
code_verifier = ""
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2021-01-04 13:13:49 -05:00
|
|
|
if not client_redirect_url:
|
|
|
|
client_redirect_url = b""
|
|
|
|
|
2023-01-04 14:58:08 -05:00
|
|
|
metadata = await self.load_metadata()
|
|
|
|
|
|
|
|
# Automatically enable PKCE if it is supported.
|
|
|
|
extra_grant_values = {}
|
|
|
|
if metadata.get("code_challenge_methods_supported"):
|
|
|
|
code_verifier = generate_token(48)
|
|
|
|
|
|
|
|
# Note that we verified the server supports S256 earlier (in
|
|
|
|
# OidcProvider._validate_metadata).
|
|
|
|
extra_grant_values = {
|
|
|
|
"code_challenge_method": "S256",
|
|
|
|
"code_challenge": create_s256_code_challenge(code_verifier),
|
|
|
|
}
|
|
|
|
|
2022-06-14 09:12:08 -04:00
|
|
|
cookie = self._macaroon_generaton.generate_oidc_session_token(
|
2020-05-15 12:26:02 -04:00
|
|
|
state=state,
|
2021-01-13 05:26:12 -05:00
|
|
|
session_data=OidcSessionData(
|
2021-01-15 08:22:12 -05:00
|
|
|
idp_id=self.idp_id,
|
2021-01-13 05:26:12 -05:00
|
|
|
nonce=nonce,
|
|
|
|
client_redirect_url=client_redirect_url.decode(),
|
2021-03-04 09:44:22 -05:00
|
|
|
ui_auth_session_id=ui_auth_session_id or "",
|
2023-01-04 14:58:08 -05:00
|
|
|
code_verifier=code_verifier,
|
2021-01-13 05:26:12 -05:00
|
|
|
),
|
2020-05-08 08:30:40 -04:00
|
|
|
)
|
2021-02-17 05:15:14 -05:00
|
|
|
|
|
|
|
# Set the cookies. See the comments on _SESSION_COOKIES for why there are two.
|
|
|
|
#
|
|
|
|
# we have to build the header by hand rather than calling request.addCookie
|
|
|
|
# because the latter does not support SameSite=None
|
|
|
|
# (https://twistedmatrix.com/trac/ticket/10088)
|
|
|
|
|
|
|
|
for cookie_name, options in _SESSION_COOKIES:
|
|
|
|
request.cookies.append(
|
2021-04-23 13:22:47 -04:00
|
|
|
b"%s=%s; Max-Age=3600; Path=%s; %s"
|
|
|
|
% (
|
|
|
|
cookie_name,
|
|
|
|
cookie.encode("utf-8"),
|
|
|
|
self._callback_path_prefix,
|
|
|
|
options,
|
|
|
|
)
|
2021-02-17 05:15:14 -05:00
|
|
|
)
|
2020-05-08 08:30:40 -04:00
|
|
|
|
|
|
|
authorization_endpoint = metadata.get("authorization_endpoint")
|
2020-05-15 12:26:02 -04:00
|
|
|
return prepare_grant_uri(
|
2020-05-08 08:30:40 -04:00
|
|
|
authorization_endpoint,
|
|
|
|
client_id=self._client_auth.client_id,
|
|
|
|
response_type="code",
|
|
|
|
redirect_uri=self._callback_url,
|
|
|
|
scope=self._scopes,
|
|
|
|
state=state,
|
|
|
|
nonce=nonce,
|
2023-01-04 14:58:08 -05:00
|
|
|
**extra_grant_values,
|
2020-05-08 08:30:40 -04:00
|
|
|
)
|
|
|
|
|
2021-01-14 08:29:17 -05:00
|
|
|
async def handle_oidc_callback(
|
|
|
|
self, request: SynapseRequest, session_data: "OidcSessionData", code: str
|
|
|
|
) -> None:
|
2021-02-01 17:56:01 -05:00
|
|
|
"""Handle an incoming request to /_synapse/client/oidc/callback
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2021-01-14 08:29:17 -05:00
|
|
|
By this time we have already validated the session on the synapse side, and
|
|
|
|
now need to do the provider-specific operations. This includes:
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2021-01-14 08:29:17 -05:00
|
|
|
- exchange the code with the provider using the ``token_endpoint`` (see
|
|
|
|
``_exchange_code``)
|
2020-05-08 08:30:40 -04:00
|
|
|
- once we have the token, use it to either extract the UserInfo from
|
|
|
|
the ``id_token`` (``_parse_id_token``), or use the ``access_token``
|
|
|
|
to fetch UserInfo from the ``userinfo_endpoint``
|
|
|
|
(``_fetch_userinfo``)
|
|
|
|
- map those UserInfo to a Matrix user (``_map_userinfo_to_user``) and
|
|
|
|
finish the login
|
|
|
|
|
|
|
|
Args:
|
|
|
|
request: the incoming request from the browser.
|
2021-01-14 08:29:17 -05:00
|
|
|
session_data: the session data, extracted from our cookie
|
|
|
|
code: The authorization code we got from the callback.
|
2020-05-08 08:30:40 -04:00
|
|
|
"""
|
|
|
|
# Exchange the code with the provider
|
|
|
|
try:
|
2021-02-03 14:45:34 -05:00
|
|
|
logger.debug("Exchanging OAuth2 code for a token")
|
2023-01-04 14:58:08 -05:00
|
|
|
token = await self._exchange_code(
|
|
|
|
code, code_verifier=session_data.code_verifier
|
|
|
|
)
|
2020-05-08 08:30:40 -04:00
|
|
|
except OidcError as e:
|
2022-05-18 15:10:21 -04:00
|
|
|
logger.warning("Could not exchange OAuth2 code: %s", e)
|
2020-11-17 09:46:23 -05:00
|
|
|
self._sso_handler.render_error(request, e.error, e.error_description)
|
2020-05-08 08:30:40 -04:00
|
|
|
return
|
|
|
|
|
2021-02-03 14:45:34 -05:00
|
|
|
logger.debug("Successfully obtained OAuth2 token data: %r", token)
|
2020-06-03 16:13:17 -04:00
|
|
|
|
2021-12-06 12:43:06 -05:00
|
|
|
# If there is an id_token, it should be validated, regardless of the
|
|
|
|
# userinfo endpoint is used or not.
|
|
|
|
if token.get("id_token") is not None:
|
|
|
|
try:
|
|
|
|
id_token = await self._parse_id_token(token, nonce=session_data.nonce)
|
|
|
|
sid = id_token.get("sid")
|
|
|
|
except Exception as e:
|
|
|
|
logger.exception("Invalid id_token")
|
|
|
|
self._sso_handler.render_error(request, "invalid_token", str(e))
|
|
|
|
return
|
|
|
|
else:
|
|
|
|
id_token = None
|
|
|
|
sid = None
|
|
|
|
|
|
|
|
# Now that we have a token, get the userinfo either from the `id_token`
|
|
|
|
# claims or by fetching the `userinfo_endpoint`.
|
2020-05-08 08:30:40 -04:00
|
|
|
if self._uses_userinfo:
|
|
|
|
try:
|
|
|
|
userinfo = await self._fetch_userinfo(token)
|
|
|
|
except Exception as e:
|
|
|
|
logger.exception("Could not fetch userinfo")
|
2020-11-17 09:46:23 -05:00
|
|
|
self._sso_handler.render_error(request, "fetch_error", str(e))
|
2020-05-08 08:30:40 -04:00
|
|
|
return
|
2021-12-06 12:43:06 -05:00
|
|
|
elif id_token is not None:
|
|
|
|
userinfo = UserInfo(id_token)
|
2020-05-08 08:30:40 -04:00
|
|
|
else:
|
2021-12-06 12:43:06 -05:00
|
|
|
logger.error("Missing id_token in token response")
|
|
|
|
self._sso_handler.render_error(
|
|
|
|
request, "invalid_token", "Missing id_token in token response"
|
|
|
|
)
|
|
|
|
return
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2020-12-08 09:03:38 -05:00
|
|
|
# first check if we're doing a UIA
|
2021-01-13 05:26:12 -05:00
|
|
|
if session_data.ui_auth_session_id:
|
2020-12-08 09:03:38 -05:00
|
|
|
try:
|
|
|
|
remote_user_id = self._remote_id_from_userinfo(userinfo)
|
|
|
|
except Exception as e:
|
|
|
|
logger.exception("Could not extract remote user id")
|
|
|
|
self._sso_handler.render_error(request, "mapping_error", str(e))
|
|
|
|
return
|
|
|
|
|
|
|
|
return await self._sso_handler.complete_sso_ui_auth_request(
|
2021-01-13 05:26:12 -05:00
|
|
|
self.idp_id, remote_user_id, session_data.ui_auth_session_id, request
|
2020-12-08 09:03:38 -05:00
|
|
|
)
|
|
|
|
|
|
|
|
# otherwise, it's a login
|
2021-03-16 11:46:07 -04:00
|
|
|
logger.debug("Userinfo for OIDC login: %s", userinfo)
|
|
|
|
|
|
|
|
# Ensure that the attributes of the logged in user meet the required
|
|
|
|
# attributes by checking the userinfo against attribute_requirements
|
|
|
|
# In order to deal with the fact that OIDC userinfo can contain many
|
|
|
|
# types of data, we wrap non-list values in lists.
|
|
|
|
if not self._sso_handler.check_required_attributes(
|
|
|
|
request,
|
|
|
|
{k: v if isinstance(v, list) else [v] for k, v in userinfo.items()},
|
|
|
|
self._oidc_attribute_requirements,
|
|
|
|
):
|
|
|
|
return
|
2020-12-08 09:03:38 -05:00
|
|
|
|
2020-05-08 08:30:40 -04:00
|
|
|
# Call the mapper to register/login the user
|
|
|
|
try:
|
2020-12-16 15:01:53 -05:00
|
|
|
await self._complete_oidc_login(
|
2021-12-06 12:43:06 -05:00
|
|
|
userinfo, token, request, session_data.client_redirect_url, sid
|
2020-08-20 15:42:58 -04:00
|
|
|
)
|
2020-05-08 08:30:40 -04:00
|
|
|
except MappingException as e:
|
|
|
|
logger.exception("Could not map user")
|
2020-11-17 09:46:23 -05:00
|
|
|
self._sso_handler.render_error(request, "mapping_error", str(e))
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2020-12-16 15:01:53 -05:00
|
|
|
async def _complete_oidc_login(
|
|
|
|
self,
|
|
|
|
userinfo: UserInfo,
|
|
|
|
token: Token,
|
|
|
|
request: SynapseRequest,
|
|
|
|
client_redirect_url: str,
|
2021-12-06 12:43:06 -05:00
|
|
|
sid: Optional[str],
|
2020-12-16 15:01:53 -05:00
|
|
|
) -> None:
|
|
|
|
"""Given a UserInfo response, complete the login flow
|
2020-05-08 08:30:40 -04:00
|
|
|
|
|
|
|
UserInfo should have a claim that uniquely identifies users. This claim
|
|
|
|
is usually `sub`, but can be configured with `oidc_config.subject_claim`.
|
|
|
|
It is then used as an `external_id`.
|
|
|
|
|
|
|
|
If we don't find the user that way, we should register the user,
|
|
|
|
mapping the localpart and the display name from the UserInfo.
|
|
|
|
|
2020-09-25 07:01:45 -04:00
|
|
|
If a user already exists with the mxid we've mapped and allow_existing_users
|
|
|
|
is disabled, raise an exception.
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2020-12-16 15:01:53 -05:00
|
|
|
Otherwise, render a redirect back to the client_redirect_url with a loginToken.
|
|
|
|
|
2020-05-08 08:30:40 -04:00
|
|
|
Args:
|
|
|
|
userinfo: an object representing the user
|
|
|
|
token: a dict with the tokens obtained from the provider
|
2020-12-16 15:01:53 -05:00
|
|
|
request: The request to respond to
|
|
|
|
client_redirect_url: The redirect URL passed in by the client.
|
2020-05-08 08:30:40 -04:00
|
|
|
|
|
|
|
Raises:
|
|
|
|
MappingException: if there was an error while mapping some properties
|
|
|
|
"""
|
|
|
|
try:
|
2020-12-08 09:03:38 -05:00
|
|
|
remote_user_id = self._remote_id_from_userinfo(userinfo)
|
2020-05-08 08:30:40 -04:00
|
|
|
except Exception as e:
|
|
|
|
raise MappingException(
|
|
|
|
"Failed to extract subject from OIDC response: %s" % (e,)
|
|
|
|
)
|
|
|
|
|
2020-11-25 10:04:22 -05:00
|
|
|
# Older mapping providers don't accept the `failures` argument, so we
|
|
|
|
# try and detect support.
|
|
|
|
mapper_signature = inspect.signature(
|
|
|
|
self._user_mapping_provider.map_user_attributes
|
2020-05-08 08:30:40 -04:00
|
|
|
)
|
2020-11-25 10:04:22 -05:00
|
|
|
supports_failures = "failures" in mapper_signature.parameters
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2020-11-25 10:04:22 -05:00
|
|
|
async def oidc_response_to_user_attributes(failures: int) -> UserAttributes:
|
|
|
|
"""
|
|
|
|
Call the mapping provider to map the OIDC userinfo and token to user attributes.
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2020-11-25 10:04:22 -05:00
|
|
|
This is backwards compatibility for abstraction for the SSO handler.
|
|
|
|
"""
|
|
|
|
if supports_failures:
|
|
|
|
attributes = await self._user_mapping_provider.map_user_attributes(
|
|
|
|
userinfo, token, failures
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
# If the mapping provider does not support processing failures,
|
|
|
|
# do not continually generate the same Matrix ID since it will
|
|
|
|
# continue to already be in use. Note that the error raised is
|
|
|
|
# arbitrary and will get turned into a MappingException.
|
|
|
|
if failures:
|
2020-12-04 08:25:15 -05:00
|
|
|
raise MappingException(
|
2020-11-25 10:04:22 -05:00
|
|
|
"Mapping provider does not support de-duplicating Matrix IDs"
|
|
|
|
)
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2022-04-27 09:03:44 -04:00
|
|
|
attributes = await self._user_mapping_provider.map_user_attributes(
|
2020-11-25 10:04:22 -05:00
|
|
|
userinfo, token
|
|
|
|
)
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2020-11-25 10:04:22 -05:00
|
|
|
return UserAttributes(**attributes)
|
2020-11-17 09:46:23 -05:00
|
|
|
|
2020-12-02 07:45:42 -05:00
|
|
|
async def grandfather_existing_users() -> Optional[str]:
|
|
|
|
if self._allow_existing_users:
|
|
|
|
# If allowing existing users we want to generate a single localpart
|
|
|
|
# and attempt to match it.
|
|
|
|
attributes = await oidc_response_to_user_attributes(failures=0)
|
|
|
|
|
2021-04-23 10:46:29 -04:00
|
|
|
if attributes.localpart is None:
|
|
|
|
# If no localpart is returned then we will generate one, so
|
|
|
|
# there is no need to search for existing users.
|
|
|
|
return None
|
|
|
|
|
2021-01-13 05:26:12 -05:00
|
|
|
user_id = UserID(attributes.localpart, self._server_name).to_string()
|
|
|
|
users = await self._store.get_users_by_id_case_insensitive(user_id)
|
2020-12-02 07:45:42 -05:00
|
|
|
if users:
|
|
|
|
# If an existing matrix ID is returned, then use it.
|
|
|
|
if len(users) == 1:
|
|
|
|
previously_registered_user_id = next(iter(users))
|
|
|
|
elif user_id in users:
|
|
|
|
previously_registered_user_id = user_id
|
|
|
|
else:
|
|
|
|
# Do not attempt to continue generating Matrix IDs.
|
|
|
|
raise MappingException(
|
|
|
|
"Attempted to login as '{}' but it matches more than one user inexactly: {}".format(
|
|
|
|
user_id, users
|
|
|
|
)
|
|
|
|
)
|
|
|
|
|
|
|
|
return previously_registered_user_id
|
|
|
|
|
|
|
|
return None
|
|
|
|
|
2020-12-16 15:01:53 -05:00
|
|
|
# Mapping providers might not have get_extra_attributes: only call this
|
|
|
|
# method if it exists.
|
|
|
|
extra_attributes = None
|
|
|
|
get_extra_attributes = getattr(
|
|
|
|
self._user_mapping_provider, "get_extra_attributes", None
|
|
|
|
)
|
|
|
|
if get_extra_attributes:
|
|
|
|
extra_attributes = await get_extra_attributes(userinfo, token)
|
|
|
|
|
|
|
|
await self._sso_handler.complete_sso_login_request(
|
2021-01-04 13:13:49 -05:00
|
|
|
self.idp_id,
|
2020-11-25 10:04:22 -05:00
|
|
|
remote_user_id,
|
2020-12-16 15:01:53 -05:00
|
|
|
request,
|
|
|
|
client_redirect_url,
|
2020-11-25 10:04:22 -05:00
|
|
|
oidc_response_to_user_attributes,
|
2020-12-02 07:45:42 -05:00
|
|
|
grandfather_existing_users,
|
2020-12-16 15:01:53 -05:00
|
|
|
extra_attributes,
|
2021-12-06 12:43:06 -05:00
|
|
|
auth_provider_session_id=sid,
|
2023-03-30 07:09:41 -04:00
|
|
|
registration_enabled=self._config.enable_registration,
|
2020-05-08 08:30:40 -04:00
|
|
|
)
|
|
|
|
|
2020-12-08 09:03:38 -05:00
|
|
|
def _remote_id_from_userinfo(self, userinfo: UserInfo) -> str:
|
|
|
|
"""Extract the unique remote id from an OIDC UserInfo block
|
|
|
|
|
|
|
|
Args:
|
|
|
|
userinfo: An object representing the user given by the OIDC provider
|
|
|
|
Returns:
|
|
|
|
remote user id
|
|
|
|
"""
|
|
|
|
remote_user_id = self._user_mapping_provider.get_remote_user_id(userinfo)
|
|
|
|
# Some OIDC providers use integer IDs, but Synapse expects external IDs
|
|
|
|
# to be strings.
|
|
|
|
return str(remote_user_id)
|
|
|
|
|
2022-10-31 13:07:30 -04:00
|
|
|
async def handle_backchannel_logout(
|
|
|
|
self, request: SynapseRequest, logout_token: str
|
|
|
|
) -> None:
|
|
|
|
"""Handle an incoming request to /_synapse/client/oidc/backchannel_logout
|
|
|
|
|
|
|
|
The OIDC Provider posts a logout token to this endpoint when a user
|
|
|
|
session ends. That token is a JWT signed with the same keys as
|
|
|
|
ID tokens. The OpenID Connect Back-Channel Logout draft explains how to
|
|
|
|
validate the JWT and figure out what session to end.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
request: The request to respond to
|
|
|
|
logout_token: The logout token (a JWT) extracted from the request body
|
|
|
|
"""
|
|
|
|
# Back-Channel Logout can be disabled in the config, hence this check.
|
|
|
|
# This is not that important for now since Synapse is registered
|
|
|
|
# manually to the OP, so not specifying the backchannel-logout URI is
|
|
|
|
# as effective than disabling it here. It might make more sense if we
|
|
|
|
# support dynamic registration in Synapse at some point.
|
|
|
|
if not self._config.backchannel_logout_enabled:
|
|
|
|
logger.warning(
|
|
|
|
f"Received an OIDC Back-Channel Logout request from issuer {self.issuer!r} but it is disabled in config"
|
|
|
|
)
|
|
|
|
|
|
|
|
# TODO: this responds with a 400 status code, which is what the OIDC
|
|
|
|
# Back-Channel Logout spec expects, but spec also suggests answering with
|
|
|
|
# a JSON object, with the `error` and `error_description` fields set, which
|
|
|
|
# we are not doing here.
|
|
|
|
# See https://openid.net/specs/openid-connect-backchannel-1_0.html#BCResponse
|
|
|
|
raise SynapseError(
|
|
|
|
400, "OpenID Connect Back-Channel Logout is disabled for this provider"
|
|
|
|
)
|
|
|
|
|
|
|
|
metadata = await self.load_metadata()
|
|
|
|
|
|
|
|
# As per OIDC Back-Channel Logout 1.0 sec. 2.4:
|
|
|
|
# A Logout Token MUST be signed and MAY also be encrypted. The same
|
|
|
|
# keys are used to sign and encrypt Logout Tokens as are used for ID
|
|
|
|
# Tokens. If the Logout Token is encrypted, it SHOULD replicate the
|
|
|
|
# iss (issuer) claim in the JWT Header Parameters, as specified in
|
|
|
|
# Section 5.3 of [JWT].
|
|
|
|
alg_values = metadata.get("id_token_signing_alg_values_supported", ["RS256"])
|
|
|
|
|
|
|
|
# As per sec. 2.6:
|
|
|
|
# 3. Validate the iss, aud, and iat Claims in the same way they are
|
|
|
|
# validated in ID Tokens.
|
|
|
|
# Which means the audience should contain Synapse's client_id and the
|
|
|
|
# issuer should be the IdP issuer
|
|
|
|
claims_options = {
|
|
|
|
"iss": {"values": [metadata["issuer"]]},
|
|
|
|
"aud": {"values": [self.client_id]},
|
|
|
|
}
|
|
|
|
|
|
|
|
try:
|
|
|
|
claims = await self._verify_jwt(
|
|
|
|
alg_values=alg_values,
|
|
|
|
token=logout_token,
|
|
|
|
claims_cls=LogoutToken,
|
|
|
|
claims_options=claims_options,
|
|
|
|
)
|
|
|
|
except JoseError:
|
|
|
|
logger.exception("Invalid logout_token")
|
|
|
|
raise SynapseError(400, "Invalid logout_token")
|
|
|
|
|
|
|
|
# As per sec. 2.6:
|
|
|
|
# 4. Verify that the Logout Token contains a sub Claim, a sid Claim,
|
|
|
|
# or both.
|
|
|
|
# 5. Verify that the Logout Token contains an events Claim whose
|
|
|
|
# value is JSON object containing the member name
|
|
|
|
# http://schemas.openid.net/event/backchannel-logout.
|
|
|
|
# 6. Verify that the Logout Token does not contain a nonce Claim.
|
|
|
|
# This is all verified by the LogoutToken claims class, so at this
|
|
|
|
# point the `sid` claim exists and is a string.
|
|
|
|
sid: str = claims.get("sid")
|
|
|
|
|
|
|
|
# If the `sub` claim was included in the logout token, we check that it matches
|
|
|
|
# that it matches the right user. We can have cases where the `sub` claim is not
|
|
|
|
# the ID saved in database, so we let admins disable this check in config.
|
|
|
|
sub: Optional[str] = claims.get("sub")
|
|
|
|
expected_user_id: Optional[str] = None
|
|
|
|
if sub is not None and not self._config.backchannel_logout_ignore_sub:
|
|
|
|
expected_user_id = await self._store.get_user_by_external_id(
|
|
|
|
self.idp_id, sub
|
|
|
|
)
|
|
|
|
|
|
|
|
# Invalidate any running user-mapping sessions, in-flight login tokens and
|
|
|
|
# active devices
|
|
|
|
await self._sso_handler.revoke_sessions_for_provider_session_id(
|
|
|
|
auth_provider_id=self.idp_id,
|
|
|
|
auth_provider_session_id=sid,
|
|
|
|
expected_user_id=expected_user_id,
|
|
|
|
)
|
|
|
|
|
|
|
|
request.setResponseCode(200)
|
|
|
|
request.setHeader(b"Cache-Control", b"no-cache, no-store")
|
|
|
|
request.setHeader(b"Pragma", b"no-cache")
|
|
|
|
finish_request(request)
|
|
|
|
|
|
|
|
|
2023-05-31 07:18:29 -04:00
|
|
|
class LogoutToken(JWTClaims): # type: ignore[misc]
|
2022-10-31 13:07:30 -04:00
|
|
|
"""
|
|
|
|
Holds and verify claims of a logout token, as per
|
|
|
|
https://openid.net/specs/openid-connect-backchannel-1_0.html#LogoutToken
|
|
|
|
"""
|
|
|
|
|
|
|
|
REGISTERED_CLAIMS = ["iss", "sub", "aud", "iat", "jti", "events", "sid"]
|
|
|
|
|
|
|
|
def validate(self, now: Optional[int] = None, leeway: int = 0) -> None:
|
|
|
|
"""Validate everything in claims payload."""
|
|
|
|
super().validate(now, leeway)
|
|
|
|
self.validate_sid()
|
|
|
|
self.validate_events()
|
|
|
|
self.validate_nonce()
|
|
|
|
|
|
|
|
def validate_sid(self) -> None:
|
|
|
|
"""Ensure the sid claim is present"""
|
|
|
|
sid = self.get("sid")
|
|
|
|
if not sid:
|
|
|
|
raise MissingClaimError("sid")
|
|
|
|
|
|
|
|
if not isinstance(sid, str):
|
|
|
|
raise InvalidClaimError("sid")
|
|
|
|
|
|
|
|
def validate_nonce(self) -> None:
|
|
|
|
"""Ensure the nonce claim is absent"""
|
|
|
|
if "nonce" in self:
|
|
|
|
raise InvalidClaimError("nonce")
|
|
|
|
|
|
|
|
def validate_events(self) -> None:
|
|
|
|
"""Ensure the events claim is present and with the right value"""
|
|
|
|
events = self.get("events")
|
|
|
|
if not events:
|
|
|
|
raise MissingClaimError("events")
|
|
|
|
|
|
|
|
if not isinstance(events, dict):
|
|
|
|
raise InvalidClaimError("events")
|
|
|
|
|
|
|
|
if "http://schemas.openid.net/event/backchannel-logout" not in events:
|
|
|
|
raise InvalidClaimError("events")
|
|
|
|
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2021-03-09 10:03:37 -05:00
|
|
|
# number of seconds a newly-generated client secret should be valid for
|
|
|
|
CLIENT_SECRET_VALIDITY_SECONDS = 3600
|
|
|
|
|
|
|
|
# minimum remaining validity on a client secret before we should generate a new one
|
|
|
|
CLIENT_SECRET_MIN_VALIDITY_SECONDS = 600
|
|
|
|
|
|
|
|
|
|
|
|
class JwtClientSecret:
|
|
|
|
"""A class which generates a new client secret on demand, based on a JWK
|
|
|
|
|
|
|
|
This implementation is designed to comply with the requirements for Apple Sign in:
|
|
|
|
https://developer.apple.com/documentation/sign_in_with_apple/generate_and_validate_tokens#3262048
|
|
|
|
|
|
|
|
It looks like those requirements are based on https://tools.ietf.org/html/rfc7523,
|
|
|
|
but it's worth noting that we still put the generated secret in the "client_secret"
|
|
|
|
field (or rather, whereever client_auth_method puts it) rather than in a
|
|
|
|
client_assertion field in the body as that RFC seems to require.
|
|
|
|
"""
|
|
|
|
|
|
|
|
def __init__(
|
|
|
|
self,
|
|
|
|
key: OidcProviderClientSecretJwtKey,
|
|
|
|
oauth_client_id: str,
|
|
|
|
oauth_issuer: str,
|
|
|
|
clock: Clock,
|
|
|
|
):
|
|
|
|
self._key = key
|
|
|
|
self._oauth_client_id = oauth_client_id
|
|
|
|
self._oauth_issuer = oauth_issuer
|
|
|
|
self._clock = clock
|
|
|
|
self._cached_secret = b""
|
|
|
|
self._cached_secret_replacement_time = 0
|
|
|
|
|
2021-09-20 08:56:23 -04:00
|
|
|
def __str__(self) -> str:
|
2021-03-09 10:03:37 -05:00
|
|
|
# if client_auth_method is client_secret_basic, then ClientAuth.prepare calls
|
|
|
|
# encode_client_secret_basic, which calls "{}".format(secret), which ends up
|
|
|
|
# here.
|
|
|
|
return self._get_secret().decode("ascii")
|
|
|
|
|
2021-09-20 08:56:23 -04:00
|
|
|
def __bytes__(self) -> bytes:
|
2021-03-09 10:03:37 -05:00
|
|
|
# if client_auth_method is client_secret_post, then ClientAuth.prepare calls
|
|
|
|
# encode_client_secret_post, which ends up here.
|
|
|
|
return self._get_secret()
|
|
|
|
|
|
|
|
def _get_secret(self) -> bytes:
|
|
|
|
now = self._clock.time()
|
|
|
|
|
|
|
|
# if we have enough validity on our existing secret, use it
|
|
|
|
if now < self._cached_secret_replacement_time:
|
|
|
|
return self._cached_secret
|
|
|
|
|
|
|
|
issued_at = int(now)
|
|
|
|
expires_at = issued_at + CLIENT_SECRET_VALIDITY_SECONDS
|
|
|
|
|
|
|
|
# we copy the configured header because jwt.encode modifies it.
|
|
|
|
header = dict(self._key.jwt_header)
|
|
|
|
|
|
|
|
# see https://tools.ietf.org/html/rfc7523#section-3
|
|
|
|
payload = {
|
|
|
|
"sub": self._oauth_client_id,
|
|
|
|
"aud": self._oauth_issuer,
|
|
|
|
"iat": issued_at,
|
|
|
|
"exp": expires_at,
|
|
|
|
**self._key.jwt_payload,
|
|
|
|
}
|
|
|
|
logger.info(
|
|
|
|
"Generating new JWT for %s: %s %s", self._oauth_issuer, header, payload
|
|
|
|
)
|
2022-10-31 13:07:30 -04:00
|
|
|
jwt = JsonWebToken(header["alg"])
|
2021-03-09 10:03:37 -05:00
|
|
|
self._cached_secret = jwt.encode(header, payload, self._key.key)
|
|
|
|
self._cached_secret_replacement_time = (
|
|
|
|
expires_at - CLIENT_SECRET_MIN_VALIDITY_SECONDS
|
|
|
|
)
|
|
|
|
return self._cached_secret
|
|
|
|
|
|
|
|
|
2021-07-19 10:28:05 -04:00
|
|
|
class UserAttributeDict(TypedDict):
|
|
|
|
localpart: Optional[str]
|
2022-03-11 08:20:00 -05:00
|
|
|
confirm_localpart: bool
|
2021-07-19 10:28:05 -04:00
|
|
|
display_name: Optional[str]
|
2022-11-25 10:16:50 -05:00
|
|
|
picture: Optional[str] # may be omitted by older `OidcMappingProviders`
|
2021-07-19 10:28:05 -04:00
|
|
|
emails: List[str]
|
|
|
|
|
|
|
|
|
2020-05-08 08:30:40 -04:00
|
|
|
class OidcMappingProvider(Generic[C]):
|
|
|
|
"""A mapping provider maps a UserInfo object to user attributes.
|
|
|
|
|
|
|
|
It should provide the API described by this class.
|
|
|
|
"""
|
|
|
|
|
|
|
|
def __init__(self, config: C):
|
|
|
|
"""
|
|
|
|
Args:
|
|
|
|
config: A custom config object from this module, parsed by ``parse_config()``
|
|
|
|
"""
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def parse_config(config: dict) -> C:
|
|
|
|
"""Parse the dict provided by the homeserver's config
|
|
|
|
|
|
|
|
Args:
|
|
|
|
config: A dictionary containing configuration options for this provider
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
A custom config object for this module
|
|
|
|
"""
|
|
|
|
raise NotImplementedError()
|
|
|
|
|
|
|
|
def get_remote_user_id(self, userinfo: UserInfo) -> str:
|
|
|
|
"""Get a unique user ID for this user.
|
|
|
|
|
|
|
|
Usually, in an OIDC-compliant scenario, it should be the ``sub`` claim from the UserInfo object.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
userinfo: An object representing the user given by the OIDC provider
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
A unique user ID
|
|
|
|
"""
|
|
|
|
raise NotImplementedError()
|
|
|
|
|
|
|
|
async def map_user_attributes(
|
2020-11-25 10:04:22 -05:00
|
|
|
self, userinfo: UserInfo, token: Token, failures: int
|
|
|
|
) -> UserAttributeDict:
|
2020-09-30 13:02:43 -04:00
|
|
|
"""Map a `UserInfo` object into user attributes.
|
2020-05-08 08:30:40 -04:00
|
|
|
|
|
|
|
Args:
|
|
|
|
userinfo: An object representing the user given by the OIDC provider
|
|
|
|
token: A dict with the tokens returned by the provider
|
2020-11-25 10:04:22 -05:00
|
|
|
failures: How many times a call to this function with this
|
|
|
|
UserInfo has resulted in a failure.
|
2020-05-08 08:30:40 -04:00
|
|
|
|
|
|
|
Returns:
|
|
|
|
A dict containing the ``localpart`` and (optionally) the ``display_name``
|
|
|
|
"""
|
|
|
|
raise NotImplementedError()
|
|
|
|
|
2020-09-30 13:02:43 -04:00
|
|
|
async def get_extra_attributes(self, userinfo: UserInfo, token: Token) -> JsonDict:
|
|
|
|
"""Map a `UserInfo` object into additional attributes passed to the client during login.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
userinfo: An object representing the user given by the OIDC provider
|
|
|
|
token: A dict with the tokens returned by the provider
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
A dict containing additional attributes. Must be JSON serializable.
|
|
|
|
"""
|
|
|
|
return {}
|
|
|
|
|
2020-05-08 08:30:40 -04:00
|
|
|
|
|
|
|
# Used to clear out "None" values in templates
|
2021-09-20 08:56:23 -04:00
|
|
|
def jinja_finalize(thing: Any) -> Any:
|
2020-05-08 08:30:40 -04:00
|
|
|
return thing if thing is not None else ""
|
|
|
|
|
|
|
|
|
|
|
|
env = Environment(finalize=jinja_finalize)
|
2022-03-11 10:15:11 -05:00
|
|
|
env.filters.update(
|
|
|
|
{
|
|
|
|
"localpart_from_email": _localpart_from_email_filter,
|
|
|
|
}
|
|
|
|
)
|
2020-05-08 08:30:40 -04:00
|
|
|
|
|
|
|
|
2021-09-20 08:56:23 -04:00
|
|
|
@attr.s(slots=True, frozen=True, auto_attribs=True)
|
2020-05-08 08:30:40 -04:00
|
|
|
class JinjaOidcMappingConfig:
|
2023-01-04 08:26:10 -05:00
|
|
|
subject_template: Template
|
|
|
|
picture_template: Template
|
2021-09-20 08:56:23 -04:00
|
|
|
localpart_template: Optional[Template]
|
|
|
|
display_name_template: Optional[Template]
|
|
|
|
email_template: Optional[Template]
|
|
|
|
extra_attributes: Dict[str, Template]
|
2022-03-11 08:20:00 -05:00
|
|
|
confirm_localpart: bool = False
|
2020-05-08 08:30:40 -04:00
|
|
|
|
|
|
|
|
|
|
|
class JinjaOidcMappingProvider(OidcMappingProvider[JinjaOidcMappingConfig]):
|
|
|
|
"""An implementation of a mapping provider based on Jinja templates.
|
|
|
|
|
|
|
|
This is the default mapping provider.
|
|
|
|
"""
|
|
|
|
|
|
|
|
def __init__(self, config: JinjaOidcMappingConfig):
|
|
|
|
self._config = config
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def parse_config(config: dict) -> JinjaOidcMappingConfig:
|
2023-01-04 08:26:10 -05:00
|
|
|
def parse_template_config_with_claim(
|
|
|
|
option_name: str, default_claim: str
|
|
|
|
) -> Template:
|
|
|
|
template_name = f"{option_name}_template"
|
|
|
|
template = config.get(template_name)
|
|
|
|
if not template:
|
|
|
|
# Convert the legacy subject_claim into a template.
|
|
|
|
claim = config.get(f"{option_name}_claim", default_claim)
|
|
|
|
template = "{{ user.%s }}" % (claim,)
|
|
|
|
|
|
|
|
try:
|
|
|
|
return env.from_string(template)
|
|
|
|
except Exception as e:
|
|
|
|
raise ConfigError("invalid jinja template", path=[template_name]) from e
|
|
|
|
|
|
|
|
subject_template = parse_template_config_with_claim("subject", "sub")
|
|
|
|
picture_template = parse_template_config_with_claim("picture", "picture")
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2021-01-27 16:28:59 -05:00
|
|
|
def parse_template_config(option_name: str) -> Optional[Template]:
|
|
|
|
if option_name not in config:
|
|
|
|
return None
|
2020-12-18 09:19:46 -05:00
|
|
|
try:
|
2021-01-27 16:28:59 -05:00
|
|
|
return env.from_string(config[option_name])
|
2020-12-18 09:19:46 -05:00
|
|
|
except Exception as e:
|
2021-01-27 16:28:59 -05:00
|
|
|
raise ConfigError("invalid jinja template", path=[option_name]) from e
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2021-01-27 16:28:59 -05:00
|
|
|
localpart_template = parse_template_config("localpart_template")
|
|
|
|
display_name_template = parse_template_config("display_name_template")
|
|
|
|
email_template = parse_template_config("email_template")
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2020-09-30 13:02:43 -04:00
|
|
|
extra_attributes = {} # type Dict[str, Template]
|
|
|
|
if "extra_attributes" in config:
|
|
|
|
extra_attributes_config = config.get("extra_attributes") or {}
|
|
|
|
if not isinstance(extra_attributes_config, dict):
|
2020-12-18 09:19:46 -05:00
|
|
|
raise ConfigError("must be a dict", path=["extra_attributes"])
|
2020-09-30 13:02:43 -04:00
|
|
|
|
|
|
|
for key, value in extra_attributes_config.items():
|
|
|
|
try:
|
|
|
|
extra_attributes[key] = env.from_string(value)
|
|
|
|
except Exception as e:
|
|
|
|
raise ConfigError(
|
2020-12-18 09:19:46 -05:00
|
|
|
"invalid jinja template", path=["extra_attributes", key]
|
|
|
|
) from e
|
2020-09-30 13:02:43 -04:00
|
|
|
|
2022-03-11 08:20:00 -05:00
|
|
|
confirm_localpart = config.get("confirm_localpart") or False
|
|
|
|
if not isinstance(confirm_localpart, bool):
|
|
|
|
raise ConfigError("must be a bool", path=["confirm_localpart"])
|
|
|
|
|
2020-05-08 08:30:40 -04:00
|
|
|
return JinjaOidcMappingConfig(
|
2023-01-04 08:26:10 -05:00
|
|
|
subject_template=subject_template,
|
|
|
|
picture_template=picture_template,
|
2020-05-08 08:30:40 -04:00
|
|
|
localpart_template=localpart_template,
|
|
|
|
display_name_template=display_name_template,
|
2021-01-27 16:28:59 -05:00
|
|
|
email_template=email_template,
|
2020-09-30 13:02:43 -04:00
|
|
|
extra_attributes=extra_attributes,
|
2022-03-11 08:20:00 -05:00
|
|
|
confirm_localpart=confirm_localpart,
|
2020-05-08 08:30:40 -04:00
|
|
|
)
|
|
|
|
|
|
|
|
def get_remote_user_id(self, userinfo: UserInfo) -> str:
|
2023-01-04 08:26:10 -05:00
|
|
|
return self._config.subject_template.render(user=userinfo).strip()
|
2020-05-08 08:30:40 -04:00
|
|
|
|
|
|
|
async def map_user_attributes(
|
2020-11-25 10:04:22 -05:00
|
|
|
self, userinfo: UserInfo, token: Token, failures: int
|
|
|
|
) -> UserAttributeDict:
|
2020-12-18 09:19:46 -05:00
|
|
|
localpart = None
|
|
|
|
|
|
|
|
if self._config.localpart_template:
|
|
|
|
localpart = self._config.localpart_template.render(user=userinfo).strip()
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2020-12-18 09:19:46 -05:00
|
|
|
# Ensure only valid characters are included in the MXID.
|
|
|
|
localpart = map_username_to_mxid_localpart(localpart)
|
2020-11-19 14:25:17 -05:00
|
|
|
|
2020-12-18 09:19:46 -05:00
|
|
|
# Append suffix integer if last call to this function failed to produce
|
|
|
|
# a usable mxid.
|
|
|
|
localpart += str(failures) if failures else ""
|
2020-11-25 10:04:22 -05:00
|
|
|
|
2021-01-27 16:28:59 -05:00
|
|
|
def render_template_field(template: Optional[Template]) -> Optional[str]:
|
|
|
|
if template is None:
|
|
|
|
return None
|
|
|
|
return template.render(user=userinfo).strip()
|
|
|
|
|
|
|
|
display_name = render_template_field(self._config.display_name_template)
|
|
|
|
if display_name == "":
|
|
|
|
display_name = None
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2021-07-16 13:22:36 -04:00
|
|
|
emails: List[str] = []
|
2021-01-27 16:28:59 -05:00
|
|
|
email = render_template_field(self._config.email_template)
|
|
|
|
if email:
|
|
|
|
emails.append(email)
|
2020-05-08 08:30:40 -04:00
|
|
|
|
2023-01-04 08:26:10 -05:00
|
|
|
picture = self._config.picture_template.render(user=userinfo).strip()
|
2022-11-25 10:16:50 -05:00
|
|
|
|
2021-01-27 16:28:59 -05:00
|
|
|
return UserAttributeDict(
|
2022-03-11 08:20:00 -05:00
|
|
|
localpart=localpart,
|
|
|
|
display_name=display_name,
|
|
|
|
emails=emails,
|
2022-11-25 10:16:50 -05:00
|
|
|
picture=picture,
|
2022-03-11 08:20:00 -05:00
|
|
|
confirm_localpart=self._config.confirm_localpart,
|
2021-01-27 16:28:59 -05:00
|
|
|
)
|
2020-09-30 13:02:43 -04:00
|
|
|
|
|
|
|
async def get_extra_attributes(self, userinfo: UserInfo, token: Token) -> JsonDict:
|
2021-07-16 13:22:36 -04:00
|
|
|
extras: Dict[str, str] = {}
|
2020-09-30 13:02:43 -04:00
|
|
|
for key, template in self._config.extra_attributes.items():
|
|
|
|
try:
|
|
|
|
extras[key] = template.render(user=userinfo).strip()
|
|
|
|
except Exception as e:
|
|
|
|
# Log an error and skip this value (don't break login for this).
|
|
|
|
logger.error("Failed to render OIDC extra attribute %s: %s" % (key, e))
|
|
|
|
return extras
|