mirror of
https://mau.dev/maunium/synapse.git
synced 2024-12-13 14:43:30 +01:00
Revert "Federation outbound proxy" (#15910)
Revert "Federation outbound proxy (#15773)"
This reverts commit b07b14b494
.
This commit is contained in:
parent
a704a35dd7
commit
c9bf644fa0
29 changed files with 90 additions and 890 deletions
|
@ -1 +0,0 @@
|
||||||
Allow configuring the set of workers to proxy outbound federation traffic through via `outbound_federation_restricted_to`.
|
|
|
@ -3930,14 +3930,13 @@ federation_sender_instances:
|
||||||
---
|
---
|
||||||
### `instance_map`
|
### `instance_map`
|
||||||
|
|
||||||
When using workers this should be a map from [`worker_name`](#worker_name) to the HTTP
|
When using workers this should be a map from [`worker_name`](#worker_name) to the
|
||||||
replication listener of the worker, if configured, and to the main process. Each worker
|
HTTP replication listener of the worker, if configured, and to the main process.
|
||||||
declared under [`stream_writers`](../../workers.md#stream-writers) and
|
Each worker declared under [`stream_writers`](../../workers.md#stream-writers) needs
|
||||||
[`outbound_federation_restricted_to`](#outbound_federation_restricted_to) needs a HTTP replication listener, and that
|
a HTTP replication listener, and that listener should be included in the `instance_map`.
|
||||||
listener should be included in the `instance_map`. The main process also needs an entry
|
The main process also needs an entry on the `instance_map`, and it should be listed under
|
||||||
on the `instance_map`, and it should be listed under `main` **if even one other worker
|
`main` **if even one other worker exists**. Ensure the port matches with what is declared
|
||||||
exists**. Ensure the port matches with what is declared inside the `listener` block for
|
inside the `listener` block for a `replication` listener.
|
||||||
a `replication` listener.
|
|
||||||
|
|
||||||
|
|
||||||
Example configuration:
|
Example configuration:
|
||||||
|
@ -3967,22 +3966,6 @@ stream_writers:
|
||||||
typing: worker1
|
typing: worker1
|
||||||
```
|
```
|
||||||
---
|
---
|
||||||
### `outbound_federation_restricted_to`
|
|
||||||
|
|
||||||
When using workers, you can restrict outbound federation traffic to only go through a
|
|
||||||
specific subset of workers. Any worker specified here must also be in the
|
|
||||||
[`instance_map`](#instance_map).
|
|
||||||
|
|
||||||
```yaml
|
|
||||||
outbound_federation_restricted_to:
|
|
||||||
- federation_sender1
|
|
||||||
- federation_sender2
|
|
||||||
```
|
|
||||||
|
|
||||||
Also see the [worker
|
|
||||||
documentation](../../workers.md#restrict-outbound-federation-traffic-to-a-specific-set-of-workers)
|
|
||||||
for more info.
|
|
||||||
---
|
|
||||||
### `run_background_tasks_on`
|
### `run_background_tasks_on`
|
||||||
|
|
||||||
The [worker](../../workers.md#background-tasks) that is used to run
|
The [worker](../../workers.md#background-tasks) that is used to run
|
||||||
|
|
|
@ -528,26 +528,6 @@ the stream writer for the `presence` stream:
|
||||||
|
|
||||||
^/_matrix/client/(api/v1|r0|v3|unstable)/presence/
|
^/_matrix/client/(api/v1|r0|v3|unstable)/presence/
|
||||||
|
|
||||||
#### Restrict outbound federation traffic to a specific set of workers
|
|
||||||
|
|
||||||
The `outbound_federation_restricted_to` configuration is useful to make sure outbound
|
|
||||||
federation traffic only goes through a specified subset of workers. This allows you to
|
|
||||||
set more strict access controls (like a firewall) for all workers and only allow the
|
|
||||||
`federation_sender`'s to contact the outside world.
|
|
||||||
|
|
||||||
```yaml
|
|
||||||
instance_map:
|
|
||||||
main:
|
|
||||||
host: localhost
|
|
||||||
port: 8030
|
|
||||||
federation_sender1:
|
|
||||||
host: localhost
|
|
||||||
port: 8034
|
|
||||||
|
|
||||||
outbound_federation_restricted_to:
|
|
||||||
- federation_sender1
|
|
||||||
```
|
|
||||||
|
|
||||||
#### Background tasks
|
#### Background tasks
|
||||||
|
|
||||||
There is also support for moving background tasks to a separate
|
There is also support for moving background tasks to a separate
|
||||||
|
|
|
@ -386,7 +386,6 @@ def listen_unix(
|
||||||
|
|
||||||
|
|
||||||
def listen_http(
|
def listen_http(
|
||||||
hs: "HomeServer",
|
|
||||||
listener_config: ListenerConfig,
|
listener_config: ListenerConfig,
|
||||||
root_resource: Resource,
|
root_resource: Resource,
|
||||||
version_string: str,
|
version_string: str,
|
||||||
|
@ -407,7 +406,6 @@ def listen_http(
|
||||||
version_string,
|
version_string,
|
||||||
max_request_body_size=max_request_body_size,
|
max_request_body_size=max_request_body_size,
|
||||||
reactor=reactor,
|
reactor=reactor,
|
||||||
federation_agent=hs.get_federation_http_client().agent,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
if isinstance(listener_config, TCPListenerConfig):
|
if isinstance(listener_config, TCPListenerConfig):
|
||||||
|
|
|
@ -221,7 +221,6 @@ class GenericWorkerServer(HomeServer):
|
||||||
root_resource = create_resource_tree(resources, OptionsResource())
|
root_resource = create_resource_tree(resources, OptionsResource())
|
||||||
|
|
||||||
_base.listen_http(
|
_base.listen_http(
|
||||||
self,
|
|
||||||
listener_config,
|
listener_config,
|
||||||
root_resource,
|
root_resource,
|
||||||
self.version_string,
|
self.version_string,
|
||||||
|
|
|
@ -139,7 +139,6 @@ class SynapseHomeServer(HomeServer):
|
||||||
root_resource = OptionsResource()
|
root_resource = OptionsResource()
|
||||||
|
|
||||||
ports = listen_http(
|
ports = listen_http(
|
||||||
self,
|
|
||||||
listener_config,
|
listener_config,
|
||||||
create_resource_tree(resources, root_resource),
|
create_resource_tree(resources, root_resource),
|
||||||
self.version_string,
|
self.version_string,
|
||||||
|
|
|
@ -15,7 +15,7 @@
|
||||||
|
|
||||||
import argparse
|
import argparse
|
||||||
import logging
|
import logging
|
||||||
from typing import Any, Dict, List, Optional, Union
|
from typing import Any, Dict, List, Union
|
||||||
|
|
||||||
import attr
|
import attr
|
||||||
from pydantic import BaseModel, Extra, StrictBool, StrictInt, StrictStr
|
from pydantic import BaseModel, Extra, StrictBool, StrictInt, StrictStr
|
||||||
|
@ -154,27 +154,6 @@ class WriterLocations:
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
@attr.s(auto_attribs=True)
|
|
||||||
class OutboundFederationRestrictedTo:
|
|
||||||
"""Whether we limit outbound federation to a certain set of instances.
|
|
||||||
|
|
||||||
Attributes:
|
|
||||||
instances: optional list of instances that can make outbound federation
|
|
||||||
requests. If None then all instances can make federation requests.
|
|
||||||
locations: list of instance locations to connect to proxy via.
|
|
||||||
"""
|
|
||||||
|
|
||||||
instances: Optional[List[str]]
|
|
||||||
locations: List[InstanceLocationConfig] = attr.Factory(list)
|
|
||||||
|
|
||||||
def __contains__(self, instance: str) -> bool:
|
|
||||||
# It feels a bit dirty to return `True` if `instances` is `None`, but it makes
|
|
||||||
# sense in downstream usage in the sense that if
|
|
||||||
# `outbound_federation_restricted_to` is not configured, then any instance can
|
|
||||||
# talk to federation (no restrictions so always return `True`).
|
|
||||||
return self.instances is None or instance in self.instances
|
|
||||||
|
|
||||||
|
|
||||||
class WorkerConfig(Config):
|
class WorkerConfig(Config):
|
||||||
"""The workers are processes run separately to the main synapse process.
|
"""The workers are processes run separately to the main synapse process.
|
||||||
They have their own pid_file and listener configuration. They use the
|
They have their own pid_file and listener configuration. They use the
|
||||||
|
@ -386,23 +365,6 @@ class WorkerConfig(Config):
|
||||||
new_option_name="update_user_directory_from_worker",
|
new_option_name="update_user_directory_from_worker",
|
||||||
)
|
)
|
||||||
|
|
||||||
outbound_federation_restricted_to = config.get(
|
|
||||||
"outbound_federation_restricted_to", None
|
|
||||||
)
|
|
||||||
self.outbound_federation_restricted_to = OutboundFederationRestrictedTo(
|
|
||||||
outbound_federation_restricted_to
|
|
||||||
)
|
|
||||||
if outbound_federation_restricted_to:
|
|
||||||
for instance in outbound_federation_restricted_to:
|
|
||||||
if instance not in self.instance_map:
|
|
||||||
raise ConfigError(
|
|
||||||
"Instance %r is configured in 'outbound_federation_restricted_to' but does not appear in `instance_map` config."
|
|
||||||
% (instance,)
|
|
||||||
)
|
|
||||||
self.outbound_federation_restricted_to.locations.append(
|
|
||||||
self.instance_map[instance]
|
|
||||||
)
|
|
||||||
|
|
||||||
def _should_this_worker_perform_duty(
|
def _should_this_worker_perform_duty(
|
||||||
self,
|
self,
|
||||||
config: Dict[str, Any],
|
config: Dict[str, Any],
|
||||||
|
|
|
@ -1037,12 +1037,7 @@ class _ReadBodyWithMaxSizeProtocol(protocol.Protocol):
|
||||||
if reason.check(ResponseDone):
|
if reason.check(ResponseDone):
|
||||||
self.deferred.callback(self.length)
|
self.deferred.callback(self.length)
|
||||||
elif reason.check(PotentialDataLoss):
|
elif reason.check(PotentialDataLoss):
|
||||||
# This applies to requests which don't set `Content-Length` or a
|
# stolen from https://github.com/twisted/treq/pull/49/files
|
||||||
# `Transfer-Encoding` in the response because in this case the end of the
|
|
||||||
# response is indicated by the connection being closed, an event which may
|
|
||||||
# also be due to a transient network problem or other error. But since this
|
|
||||||
# behavior is expected of some servers (like YouTube), let's ignore it.
|
|
||||||
# Stolen from https://github.com/twisted/treq/pull/49/files
|
|
||||||
# http://twistedmatrix.com/trac/ticket/4840
|
# http://twistedmatrix.com/trac/ticket/4840
|
||||||
self.deferred.callback(self.length)
|
self.deferred.callback(self.length)
|
||||||
else:
|
else:
|
||||||
|
|
|
@ -50,7 +50,7 @@ from twisted.internet.interfaces import IReactorTime
|
||||||
from twisted.internet.task import Cooperator
|
from twisted.internet.task import Cooperator
|
||||||
from twisted.web.client import ResponseFailed
|
from twisted.web.client import ResponseFailed
|
||||||
from twisted.web.http_headers import Headers
|
from twisted.web.http_headers import Headers
|
||||||
from twisted.web.iweb import IAgent, IBodyProducer, IResponse
|
from twisted.web.iweb import IBodyProducer, IResponse
|
||||||
|
|
||||||
import synapse.metrics
|
import synapse.metrics
|
||||||
import synapse.util.retryutils
|
import synapse.util.retryutils
|
||||||
|
@ -72,7 +72,6 @@ from synapse.http.client import (
|
||||||
read_body_with_max_size,
|
read_body_with_max_size,
|
||||||
)
|
)
|
||||||
from synapse.http.federation.matrix_federation_agent import MatrixFederationAgent
|
from synapse.http.federation.matrix_federation_agent import MatrixFederationAgent
|
||||||
from synapse.http.proxyagent import ProxyAgent
|
|
||||||
from synapse.http.types import QueryParams
|
from synapse.http.types import QueryParams
|
||||||
from synapse.logging import opentracing
|
from synapse.logging import opentracing
|
||||||
from synapse.logging.context import make_deferred_yieldable, run_in_background
|
from synapse.logging.context import make_deferred_yieldable, run_in_background
|
||||||
|
@ -394,32 +393,17 @@ class MatrixFederationHttpClient:
|
||||||
if hs.config.server.user_agent_suffix:
|
if hs.config.server.user_agent_suffix:
|
||||||
user_agent = "%s %s" % (user_agent, hs.config.server.user_agent_suffix)
|
user_agent = "%s %s" % (user_agent, hs.config.server.user_agent_suffix)
|
||||||
|
|
||||||
outbound_federation_restricted_to = (
|
federation_agent = MatrixFederationAgent(
|
||||||
hs.config.worker.outbound_federation_restricted_to
|
self.reactor,
|
||||||
|
tls_client_options_factory,
|
||||||
|
user_agent.encode("ascii"),
|
||||||
|
hs.config.server.federation_ip_range_allowlist,
|
||||||
|
hs.config.server.federation_ip_range_blocklist,
|
||||||
)
|
)
|
||||||
if hs.get_instance_name() in outbound_federation_restricted_to:
|
|
||||||
# Talk to federation directly
|
|
||||||
federation_agent: IAgent = MatrixFederationAgent(
|
|
||||||
self.reactor,
|
|
||||||
tls_client_options_factory,
|
|
||||||
user_agent.encode("ascii"),
|
|
||||||
hs.config.server.federation_ip_range_allowlist,
|
|
||||||
hs.config.server.federation_ip_range_blocklist,
|
|
||||||
)
|
|
||||||
else:
|
|
||||||
# We need to talk to federation via the proxy via one of the configured
|
|
||||||
# locations
|
|
||||||
federation_proxies = outbound_federation_restricted_to.locations
|
|
||||||
federation_agent = ProxyAgent(
|
|
||||||
self.reactor,
|
|
||||||
self.reactor,
|
|
||||||
tls_client_options_factory,
|
|
||||||
federation_proxies=federation_proxies,
|
|
||||||
)
|
|
||||||
|
|
||||||
# Use a BlocklistingAgentWrapper to prevent circumventing the IP
|
# Use a BlocklistingAgentWrapper to prevent circumventing the IP
|
||||||
# blocking via IP literals in server names
|
# blocking via IP literals in server names
|
||||||
self.agent: IAgent = BlocklistingAgentWrapper(
|
self.agent = BlocklistingAgentWrapper(
|
||||||
federation_agent,
|
federation_agent,
|
||||||
ip_blocklist=hs.config.server.federation_ip_range_blocklist,
|
ip_blocklist=hs.config.server.federation_ip_range_blocklist,
|
||||||
)
|
)
|
||||||
|
@ -428,6 +412,7 @@ class MatrixFederationHttpClient:
|
||||||
self._store = hs.get_datastores().main
|
self._store = hs.get_datastores().main
|
||||||
self.version_string_bytes = hs.version_string.encode("ascii")
|
self.version_string_bytes = hs.version_string.encode("ascii")
|
||||||
self.default_timeout_seconds = hs.config.federation.client_timeout_ms / 1000
|
self.default_timeout_seconds = hs.config.federation.client_timeout_ms / 1000
|
||||||
|
|
||||||
self.max_long_retry_delay_seconds = (
|
self.max_long_retry_delay_seconds = (
|
||||||
hs.config.federation.max_long_retry_delay_ms / 1000
|
hs.config.federation.max_long_retry_delay_ms / 1000
|
||||||
)
|
)
|
||||||
|
@ -1146,101 +1131,6 @@ class MatrixFederationHttpClient:
|
||||||
Succeeds when we get a 2xx HTTP response. The
|
Succeeds when we get a 2xx HTTP response. The
|
||||||
result will be the decoded JSON body.
|
result will be the decoded JSON body.
|
||||||
|
|
||||||
Raises:
|
|
||||||
HttpResponseException: If we get an HTTP response code >= 300
|
|
||||||
(except 429).
|
|
||||||
NotRetryingDestination: If we are not yet ready to retry this
|
|
||||||
server.
|
|
||||||
FederationDeniedError: If this destination is not on our
|
|
||||||
federation whitelist
|
|
||||||
RequestSendFailed: If there were problems connecting to the
|
|
||||||
remote, due to e.g. DNS failures, connection timeouts etc.
|
|
||||||
"""
|
|
||||||
json_dict, _ = await self.get_json_with_headers(
|
|
||||||
destination=destination,
|
|
||||||
path=path,
|
|
||||||
args=args,
|
|
||||||
retry_on_dns_fail=retry_on_dns_fail,
|
|
||||||
timeout=timeout,
|
|
||||||
ignore_backoff=ignore_backoff,
|
|
||||||
try_trailing_slash_on_400=try_trailing_slash_on_400,
|
|
||||||
parser=parser,
|
|
||||||
)
|
|
||||||
return json_dict
|
|
||||||
|
|
||||||
@overload
|
|
||||||
async def get_json_with_headers(
|
|
||||||
self,
|
|
||||||
destination: str,
|
|
||||||
path: str,
|
|
||||||
args: Optional[QueryParams] = None,
|
|
||||||
retry_on_dns_fail: bool = True,
|
|
||||||
timeout: Optional[int] = None,
|
|
||||||
ignore_backoff: bool = False,
|
|
||||||
try_trailing_slash_on_400: bool = False,
|
|
||||||
parser: Literal[None] = None,
|
|
||||||
) -> Tuple[JsonDict, Dict[bytes, List[bytes]]]:
|
|
||||||
...
|
|
||||||
|
|
||||||
@overload
|
|
||||||
async def get_json_with_headers(
|
|
||||||
self,
|
|
||||||
destination: str,
|
|
||||||
path: str,
|
|
||||||
args: Optional[QueryParams] = ...,
|
|
||||||
retry_on_dns_fail: bool = ...,
|
|
||||||
timeout: Optional[int] = ...,
|
|
||||||
ignore_backoff: bool = ...,
|
|
||||||
try_trailing_slash_on_400: bool = ...,
|
|
||||||
parser: ByteParser[T] = ...,
|
|
||||||
) -> Tuple[T, Dict[bytes, List[bytes]]]:
|
|
||||||
...
|
|
||||||
|
|
||||||
async def get_json_with_headers(
|
|
||||||
self,
|
|
||||||
destination: str,
|
|
||||||
path: str,
|
|
||||||
args: Optional[QueryParams] = None,
|
|
||||||
retry_on_dns_fail: bool = True,
|
|
||||||
timeout: Optional[int] = None,
|
|
||||||
ignore_backoff: bool = False,
|
|
||||||
try_trailing_slash_on_400: bool = False,
|
|
||||||
parser: Optional[ByteParser[T]] = None,
|
|
||||||
) -> Tuple[Union[JsonDict, T], Dict[bytes, List[bytes]]]:
|
|
||||||
"""GETs some json from the given host homeserver and path
|
|
||||||
|
|
||||||
Args:
|
|
||||||
destination: The remote server to send the HTTP request to.
|
|
||||||
|
|
||||||
path: The HTTP path.
|
|
||||||
|
|
||||||
args: A dictionary used to create query strings, defaults to
|
|
||||||
None.
|
|
||||||
|
|
||||||
retry_on_dns_fail: true if the request should be retried on DNS failures
|
|
||||||
|
|
||||||
timeout: number of milliseconds to wait for the response.
|
|
||||||
self._default_timeout (60s) by default.
|
|
||||||
|
|
||||||
Note that we may make several attempts to send the request; this
|
|
||||||
timeout applies to the time spent waiting for response headers for
|
|
||||||
*each* attempt (including connection time) as well as the time spent
|
|
||||||
reading the response body after a 200 response.
|
|
||||||
|
|
||||||
ignore_backoff: true to ignore the historical backoff data
|
|
||||||
and try the request anyway.
|
|
||||||
|
|
||||||
try_trailing_slash_on_400: True if on a 400 M_UNRECOGNIZED
|
|
||||||
response we should try appending a trailing slash to the end of
|
|
||||||
the request. Workaround for #3622 in Synapse <= v0.99.3.
|
|
||||||
|
|
||||||
parser: The parser to use to decode the response. Defaults to
|
|
||||||
parsing as JSON.
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
Succeeds when we get a 2xx HTTP response. The result will be a tuple of the
|
|
||||||
decoded JSON body and a dict of the response headers.
|
|
||||||
|
|
||||||
Raises:
|
Raises:
|
||||||
HttpResponseException: If we get an HTTP response code >= 300
|
HttpResponseException: If we get an HTTP response code >= 300
|
||||||
(except 429).
|
(except 429).
|
||||||
|
@ -1266,8 +1156,6 @@ class MatrixFederationHttpClient:
|
||||||
timeout=timeout,
|
timeout=timeout,
|
||||||
)
|
)
|
||||||
|
|
||||||
headers = dict(response.headers.getAllRawHeaders())
|
|
||||||
|
|
||||||
if timeout is not None:
|
if timeout is not None:
|
||||||
_sec_timeout = timeout / 1000
|
_sec_timeout = timeout / 1000
|
||||||
else:
|
else:
|
||||||
|
@ -1285,7 +1173,7 @@ class MatrixFederationHttpClient:
|
||||||
parser=parser,
|
parser=parser,
|
||||||
)
|
)
|
||||||
|
|
||||||
return body, headers
|
return body
|
||||||
|
|
||||||
async def delete_json(
|
async def delete_json(
|
||||||
self,
|
self,
|
||||||
|
|
|
@ -1,249 +0,0 @@
|
||||||
# Copyright 2023 The Matrix.org Foundation C.I.C.
|
|
||||||
#
|
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
# you may not use this file except in compliance with the License.
|
|
||||||
# You may obtain a copy of the License at
|
|
||||||
#
|
|
||||||
# http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
#
|
|
||||||
# Unless required by applicable law or agreed to in writing, software
|
|
||||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
# See the License for the specific language governing permissions and
|
|
||||||
# limitations under the License.
|
|
||||||
#
|
|
||||||
|
|
||||||
import json
|
|
||||||
import logging
|
|
||||||
import urllib.parse
|
|
||||||
from typing import TYPE_CHECKING, Any, Optional, Set, Tuple, cast
|
|
||||||
|
|
||||||
from twisted.internet import protocol
|
|
||||||
from twisted.internet.interfaces import ITCPTransport
|
|
||||||
from twisted.internet.protocol import connectionDone
|
|
||||||
from twisted.python import failure
|
|
||||||
from twisted.python.failure import Failure
|
|
||||||
from twisted.web.client import ResponseDone
|
|
||||||
from twisted.web.http_headers import Headers
|
|
||||||
from twisted.web.iweb import IAgent, IResponse
|
|
||||||
from twisted.web.resource import IResource
|
|
||||||
from twisted.web.server import Site
|
|
||||||
|
|
||||||
from synapse.api.errors import Codes
|
|
||||||
from synapse.http import QuieterFileBodyProducer
|
|
||||||
from synapse.http.server import _AsyncResource
|
|
||||||
from synapse.logging.context import make_deferred_yieldable, run_in_background
|
|
||||||
from synapse.types import ISynapseReactor
|
|
||||||
from synapse.util.async_helpers import timeout_deferred
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from synapse.http.site import SynapseRequest
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
# "Hop-by-hop" headers (as opposed to "end-to-end" headers) as defined by RFC2616
|
|
||||||
# section 13.5.1 and referenced in RFC9110 section 7.6.1. These are meant to only be
|
|
||||||
# consumed by the immediate recipient and not be forwarded on.
|
|
||||||
HOP_BY_HOP_HEADERS = {
|
|
||||||
"Connection",
|
|
||||||
"Keep-Alive",
|
|
||||||
"Proxy-Authenticate",
|
|
||||||
"Proxy-Authorization",
|
|
||||||
"TE",
|
|
||||||
"Trailers",
|
|
||||||
"Transfer-Encoding",
|
|
||||||
"Upgrade",
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
def parse_connection_header_value(
|
|
||||||
connection_header_value: Optional[bytes],
|
|
||||||
) -> Set[str]:
|
|
||||||
"""
|
|
||||||
Parse the `Connection` header to determine which headers we should not be copied
|
|
||||||
over from the remote response.
|
|
||||||
|
|
||||||
As defined by RFC2616 section 14.10 and RFC9110 section 7.6.1
|
|
||||||
|
|
||||||
Example: `Connection: close, X-Foo, X-Bar` will return `{"Close", "X-Foo", "X-Bar"}`
|
|
||||||
|
|
||||||
Even though "close" is a special directive, let's just treat it as just another
|
|
||||||
header for simplicity. If people want to check for this directive, they can simply
|
|
||||||
check for `"Close" in headers`.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
connection_header_value: The value of the `Connection` header.
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
The set of header names that should not be copied over from the remote response.
|
|
||||||
The keys are capitalized in canonical capitalization.
|
|
||||||
"""
|
|
||||||
headers = Headers()
|
|
||||||
extra_headers_to_remove: Set[str] = set()
|
|
||||||
if connection_header_value:
|
|
||||||
extra_headers_to_remove = {
|
|
||||||
headers._canonicalNameCaps(connection_option.strip()).decode("ascii")
|
|
||||||
for connection_option in connection_header_value.split(b",")
|
|
||||||
}
|
|
||||||
|
|
||||||
return extra_headers_to_remove
|
|
||||||
|
|
||||||
|
|
||||||
class ProxyResource(_AsyncResource):
|
|
||||||
"""
|
|
||||||
A stub resource that proxies any requests with a `matrix-federation://` scheme
|
|
||||||
through the given `federation_agent` to the remote homeserver and ferries back the
|
|
||||||
info.
|
|
||||||
"""
|
|
||||||
|
|
||||||
isLeaf = True
|
|
||||||
|
|
||||||
def __init__(self, reactor: ISynapseReactor, federation_agent: IAgent):
|
|
||||||
super().__init__(True)
|
|
||||||
|
|
||||||
self.reactor = reactor
|
|
||||||
self.agent = federation_agent
|
|
||||||
|
|
||||||
async def _async_render(self, request: "SynapseRequest") -> Tuple[int, Any]:
|
|
||||||
uri = urllib.parse.urlparse(request.uri)
|
|
||||||
assert uri.scheme == b"matrix-federation"
|
|
||||||
|
|
||||||
headers = Headers()
|
|
||||||
for header_name in (b"User-Agent", b"Authorization", b"Content-Type"):
|
|
||||||
header_value = request.getHeader(header_name)
|
|
||||||
if header_value:
|
|
||||||
headers.addRawHeader(header_name, header_value)
|
|
||||||
|
|
||||||
request_deferred = run_in_background(
|
|
||||||
self.agent.request,
|
|
||||||
request.method,
|
|
||||||
request.uri,
|
|
||||||
headers=headers,
|
|
||||||
bodyProducer=QuieterFileBodyProducer(request.content),
|
|
||||||
)
|
|
||||||
request_deferred = timeout_deferred(
|
|
||||||
request_deferred,
|
|
||||||
# This should be set longer than the timeout in `MatrixFederationHttpClient`
|
|
||||||
# so that it has enough time to complete and pass us the data before we give
|
|
||||||
# up.
|
|
||||||
timeout=90,
|
|
||||||
reactor=self.reactor,
|
|
||||||
)
|
|
||||||
|
|
||||||
response = await make_deferred_yieldable(request_deferred)
|
|
||||||
|
|
||||||
return response.code, response
|
|
||||||
|
|
||||||
def _send_response(
|
|
||||||
self,
|
|
||||||
request: "SynapseRequest",
|
|
||||||
code: int,
|
|
||||||
response_object: Any,
|
|
||||||
) -> None:
|
|
||||||
response = cast(IResponse, response_object)
|
|
||||||
response_headers = cast(Headers, response.headers)
|
|
||||||
|
|
||||||
request.setResponseCode(code)
|
|
||||||
|
|
||||||
# The `Connection` header also defines which headers should not be copied over.
|
|
||||||
connection_header = response_headers.getRawHeaders(b"connection")
|
|
||||||
extra_headers_to_remove = parse_connection_header_value(
|
|
||||||
connection_header[0] if connection_header else None
|
|
||||||
)
|
|
||||||
|
|
||||||
# Copy headers.
|
|
||||||
for k, v in response_headers.getAllRawHeaders():
|
|
||||||
# Do not copy over any hop-by-hop headers. These are meant to only be
|
|
||||||
# consumed by the immediate recipient and not be forwarded on.
|
|
||||||
header_key = k.decode("ascii")
|
|
||||||
if (
|
|
||||||
header_key in HOP_BY_HOP_HEADERS
|
|
||||||
or header_key in extra_headers_to_remove
|
|
||||||
):
|
|
||||||
continue
|
|
||||||
|
|
||||||
request.responseHeaders.setRawHeaders(k, v)
|
|
||||||
|
|
||||||
response.deliverBody(_ProxyResponseBody(request))
|
|
||||||
|
|
||||||
def _send_error_response(
|
|
||||||
self,
|
|
||||||
f: failure.Failure,
|
|
||||||
request: "SynapseRequest",
|
|
||||||
) -> None:
|
|
||||||
request.setResponseCode(502)
|
|
||||||
request.setHeader(b"Content-Type", b"application/json")
|
|
||||||
request.write(
|
|
||||||
(
|
|
||||||
json.dumps(
|
|
||||||
{
|
|
||||||
"errcode": Codes.UNKNOWN,
|
|
||||||
"err": "ProxyResource: Error when proxying request: %s %s -> %s"
|
|
||||||
% (
|
|
||||||
request.method.decode("ascii"),
|
|
||||||
request.uri.decode("ascii"),
|
|
||||||
f,
|
|
||||||
),
|
|
||||||
}
|
|
||||||
)
|
|
||||||
).encode()
|
|
||||||
)
|
|
||||||
request.finish()
|
|
||||||
|
|
||||||
|
|
||||||
class _ProxyResponseBody(protocol.Protocol):
|
|
||||||
"""
|
|
||||||
A protocol that proxies the given remote response data back out to the given local
|
|
||||||
request.
|
|
||||||
"""
|
|
||||||
|
|
||||||
transport: Optional[ITCPTransport] = None
|
|
||||||
|
|
||||||
def __init__(self, request: "SynapseRequest") -> None:
|
|
||||||
self._request = request
|
|
||||||
|
|
||||||
def dataReceived(self, data: bytes) -> None:
|
|
||||||
# Avoid sending response data to the local request that already disconnected
|
|
||||||
if self._request._disconnected and self.transport is not None:
|
|
||||||
# Close the connection (forcefully) since all the data will get
|
|
||||||
# discarded anyway.
|
|
||||||
self.transport.abortConnection()
|
|
||||||
return
|
|
||||||
|
|
||||||
self._request.write(data)
|
|
||||||
|
|
||||||
def connectionLost(self, reason: Failure = connectionDone) -> None:
|
|
||||||
# If the local request is already finished (successfully or failed), don't
|
|
||||||
# worry about sending anything back.
|
|
||||||
if self._request.finished:
|
|
||||||
return
|
|
||||||
|
|
||||||
if reason.check(ResponseDone):
|
|
||||||
self._request.finish()
|
|
||||||
else:
|
|
||||||
# Abort the underlying request since our remote request also failed.
|
|
||||||
self._request.transport.abortConnection()
|
|
||||||
|
|
||||||
|
|
||||||
class ProxySite(Site):
|
|
||||||
"""
|
|
||||||
Proxies any requests with a `matrix-federation://` scheme through the given
|
|
||||||
`federation_agent`. Otherwise, behaves like a normal `Site`.
|
|
||||||
"""
|
|
||||||
|
|
||||||
def __init__(
|
|
||||||
self,
|
|
||||||
resource: IResource,
|
|
||||||
reactor: ISynapseReactor,
|
|
||||||
federation_agent: IAgent,
|
|
||||||
):
|
|
||||||
super().__init__(resource, reactor=reactor)
|
|
||||||
|
|
||||||
self._proxy_resource = ProxyResource(reactor, federation_agent)
|
|
||||||
|
|
||||||
def getResourceFor(self, request: "SynapseRequest") -> IResource:
|
|
||||||
uri = urllib.parse.urlparse(request.uri)
|
|
||||||
if uri.scheme == b"matrix-federation":
|
|
||||||
return self._proxy_resource
|
|
||||||
|
|
||||||
return super().getResourceFor(request)
|
|
|
@ -12,9 +12,8 @@
|
||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
import logging
|
import logging
|
||||||
import random
|
|
||||||
import re
|
import re
|
||||||
from typing import Any, Collection, Dict, List, Optional, Sequence, Tuple
|
from typing import Any, Dict, Optional, Tuple
|
||||||
from urllib.parse import urlparse
|
from urllib.parse import urlparse
|
||||||
from urllib.request import ( # type: ignore[attr-defined]
|
from urllib.request import ( # type: ignore[attr-defined]
|
||||||
getproxies_environment,
|
getproxies_environment,
|
||||||
|
@ -25,12 +24,7 @@ from zope.interface import implementer
|
||||||
|
|
||||||
from twisted.internet import defer
|
from twisted.internet import defer
|
||||||
from twisted.internet.endpoints import HostnameEndpoint, wrapClientTLS
|
from twisted.internet.endpoints import HostnameEndpoint, wrapClientTLS
|
||||||
from twisted.internet.interfaces import (
|
from twisted.internet.interfaces import IReactorCore, IStreamClientEndpoint
|
||||||
IProtocol,
|
|
||||||
IProtocolFactory,
|
|
||||||
IReactorCore,
|
|
||||||
IStreamClientEndpoint,
|
|
||||||
)
|
|
||||||
from twisted.python.failure import Failure
|
from twisted.python.failure import Failure
|
||||||
from twisted.web.client import (
|
from twisted.web.client import (
|
||||||
URI,
|
URI,
|
||||||
|
@ -42,10 +36,8 @@ from twisted.web.error import SchemeNotSupported
|
||||||
from twisted.web.http_headers import Headers
|
from twisted.web.http_headers import Headers
|
||||||
from twisted.web.iweb import IAgent, IBodyProducer, IPolicyForHTTPS, IResponse
|
from twisted.web.iweb import IAgent, IBodyProducer, IPolicyForHTTPS, IResponse
|
||||||
|
|
||||||
from synapse.config.workers import InstanceLocationConfig
|
|
||||||
from synapse.http import redact_uri
|
from synapse.http import redact_uri
|
||||||
from synapse.http.connectproxyclient import HTTPConnectProxyEndpoint, ProxyCredentials
|
from synapse.http.connectproxyclient import HTTPConnectProxyEndpoint, ProxyCredentials
|
||||||
from synapse.logging.context import run_in_background
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
@ -82,10 +74,6 @@ class ProxyAgent(_AgentBase):
|
||||||
use_proxy: Whether proxy settings should be discovered and used
|
use_proxy: Whether proxy settings should be discovered and used
|
||||||
from conventional environment variables.
|
from conventional environment variables.
|
||||||
|
|
||||||
federation_proxies: An optional list of locations to proxy outbound federation
|
|
||||||
traffic through (only requests that use the `matrix-federation://` scheme
|
|
||||||
will be proxied).
|
|
||||||
|
|
||||||
Raises:
|
Raises:
|
||||||
ValueError if use_proxy is set and the environment variables
|
ValueError if use_proxy is set and the environment variables
|
||||||
contain an invalid proxy specification.
|
contain an invalid proxy specification.
|
||||||
|
@ -101,7 +89,6 @@ class ProxyAgent(_AgentBase):
|
||||||
bindAddress: Optional[bytes] = None,
|
bindAddress: Optional[bytes] = None,
|
||||||
pool: Optional[HTTPConnectionPool] = None,
|
pool: Optional[HTTPConnectionPool] = None,
|
||||||
use_proxy: bool = False,
|
use_proxy: bool = False,
|
||||||
federation_proxies: Collection[InstanceLocationConfig] = (),
|
|
||||||
):
|
):
|
||||||
contextFactory = contextFactory or BrowserLikePolicyForHTTPS()
|
contextFactory = contextFactory or BrowserLikePolicyForHTTPS()
|
||||||
|
|
||||||
|
@ -140,27 +127,6 @@ class ProxyAgent(_AgentBase):
|
||||||
self._policy_for_https = contextFactory
|
self._policy_for_https = contextFactory
|
||||||
self._reactor = reactor
|
self._reactor = reactor
|
||||||
|
|
||||||
self._federation_proxy_endpoint: Optional[IStreamClientEndpoint] = None
|
|
||||||
if federation_proxies:
|
|
||||||
endpoints = []
|
|
||||||
for federation_proxy in federation_proxies:
|
|
||||||
endpoint = HostnameEndpoint(
|
|
||||||
self.proxy_reactor,
|
|
||||||
federation_proxy.host,
|
|
||||||
federation_proxy.port,
|
|
||||||
)
|
|
||||||
|
|
||||||
if federation_proxy.tls:
|
|
||||||
tls_connection_creator = self._policy_for_https.creatorForNetloc(
|
|
||||||
federation_proxy.host,
|
|
||||||
federation_proxy.port,
|
|
||||||
)
|
|
||||||
endpoint = wrapClientTLS(tls_connection_creator, endpoint)
|
|
||||||
|
|
||||||
endpoints.append(endpoint)
|
|
||||||
|
|
||||||
self._federation_proxy_endpoint = _ProxyEndpoints(endpoints)
|
|
||||||
|
|
||||||
def request(
|
def request(
|
||||||
self,
|
self,
|
||||||
method: bytes,
|
method: bytes,
|
||||||
|
@ -248,14 +214,6 @@ class ProxyAgent(_AgentBase):
|
||||||
parsed_uri.port,
|
parsed_uri.port,
|
||||||
self.https_proxy_creds,
|
self.https_proxy_creds,
|
||||||
)
|
)
|
||||||
elif (
|
|
||||||
parsed_uri.scheme == b"matrix-federation"
|
|
||||||
and self._federation_proxy_endpoint
|
|
||||||
):
|
|
||||||
# Cache *all* connections under the same key, since we are only
|
|
||||||
# connecting to a single destination, the proxy:
|
|
||||||
endpoint = self._federation_proxy_endpoint
|
|
||||||
request_path = uri
|
|
||||||
else:
|
else:
|
||||||
# not using a proxy
|
# not using a proxy
|
||||||
endpoint = HostnameEndpoint(
|
endpoint = HostnameEndpoint(
|
||||||
|
@ -275,11 +233,6 @@ class ProxyAgent(_AgentBase):
|
||||||
endpoint = wrapClientTLS(tls_connection_creator, endpoint)
|
endpoint = wrapClientTLS(tls_connection_creator, endpoint)
|
||||||
elif parsed_uri.scheme == b"http":
|
elif parsed_uri.scheme == b"http":
|
||||||
pass
|
pass
|
||||||
elif (
|
|
||||||
parsed_uri.scheme == b"matrix-federation"
|
|
||||||
and self._federation_proxy_endpoint
|
|
||||||
):
|
|
||||||
pass
|
|
||||||
else:
|
else:
|
||||||
return defer.fail(
|
return defer.fail(
|
||||||
Failure(
|
Failure(
|
||||||
|
@ -384,31 +337,3 @@ def parse_proxy(
|
||||||
credentials = ProxyCredentials(b"".join([url.username, b":", url.password]))
|
credentials = ProxyCredentials(b"".join([url.username, b":", url.password]))
|
||||||
|
|
||||||
return url.scheme, url.hostname, url.port or default_port, credentials
|
return url.scheme, url.hostname, url.port or default_port, credentials
|
||||||
|
|
||||||
|
|
||||||
@implementer(IStreamClientEndpoint)
|
|
||||||
class _ProxyEndpoints:
|
|
||||||
"""An endpoint that randomly iterates through a given list of endpoints at
|
|
||||||
each connection attempt.
|
|
||||||
"""
|
|
||||||
|
|
||||||
def __init__(self, endpoints: Sequence[IStreamClientEndpoint]) -> None:
|
|
||||||
assert endpoints
|
|
||||||
self._endpoints = endpoints
|
|
||||||
|
|
||||||
def connect(
|
|
||||||
self, protocol_factory: IProtocolFactory
|
|
||||||
) -> "defer.Deferred[IProtocol]":
|
|
||||||
"""Implements IStreamClientEndpoint interface"""
|
|
||||||
|
|
||||||
return run_in_background(self._do_connect, protocol_factory)
|
|
||||||
|
|
||||||
async def _do_connect(self, protocol_factory: IProtocolFactory) -> IProtocol:
|
|
||||||
failures: List[Failure] = []
|
|
||||||
for endpoint in random.sample(self._endpoints, k=len(self._endpoints)):
|
|
||||||
try:
|
|
||||||
return await endpoint.connect(protocol_factory)
|
|
||||||
except Exception:
|
|
||||||
failures.append(Failure())
|
|
||||||
|
|
||||||
failures.pop().raiseException()
|
|
||||||
|
|
|
@ -18,7 +18,6 @@ import html
|
||||||
import logging
|
import logging
|
||||||
import types
|
import types
|
||||||
import urllib
|
import urllib
|
||||||
import urllib.parse
|
|
||||||
from http import HTTPStatus
|
from http import HTTPStatus
|
||||||
from http.client import FOUND
|
from http.client import FOUND
|
||||||
from inspect import isawaitable
|
from inspect import isawaitable
|
||||||
|
@ -66,6 +65,7 @@ from synapse.api.errors import (
|
||||||
UnrecognizedRequestError,
|
UnrecognizedRequestError,
|
||||||
)
|
)
|
||||||
from synapse.config.homeserver import HomeServerConfig
|
from synapse.config.homeserver import HomeServerConfig
|
||||||
|
from synapse.http.site import SynapseRequest
|
||||||
from synapse.logging.context import defer_to_thread, preserve_fn, run_in_background
|
from synapse.logging.context import defer_to_thread, preserve_fn, run_in_background
|
||||||
from synapse.logging.opentracing import active_span, start_active_span, trace_servlet
|
from synapse.logging.opentracing import active_span, start_active_span, trace_servlet
|
||||||
from synapse.util import json_encoder
|
from synapse.util import json_encoder
|
||||||
|
@ -76,7 +76,6 @@ from synapse.util.iterutils import chunk_seq
|
||||||
if TYPE_CHECKING:
|
if TYPE_CHECKING:
|
||||||
import opentracing
|
import opentracing
|
||||||
|
|
||||||
from synapse.http.site import SynapseRequest
|
|
||||||
from synapse.server import HomeServer
|
from synapse.server import HomeServer
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
@ -103,7 +102,7 @@ HTTP_STATUS_REQUEST_CANCELLED = 499
|
||||||
|
|
||||||
|
|
||||||
def return_json_error(
|
def return_json_error(
|
||||||
f: failure.Failure, request: "SynapseRequest", config: Optional[HomeServerConfig]
|
f: failure.Failure, request: SynapseRequest, config: Optional[HomeServerConfig]
|
||||||
) -> None:
|
) -> None:
|
||||||
"""Sends a JSON error response to clients."""
|
"""Sends a JSON error response to clients."""
|
||||||
|
|
||||||
|
@ -221,8 +220,8 @@ def return_html_error(
|
||||||
|
|
||||||
|
|
||||||
def wrap_async_request_handler(
|
def wrap_async_request_handler(
|
||||||
h: Callable[["_AsyncResource", "SynapseRequest"], Awaitable[None]]
|
h: Callable[["_AsyncResource", SynapseRequest], Awaitable[None]]
|
||||||
) -> Callable[["_AsyncResource", "SynapseRequest"], "defer.Deferred[None]"]:
|
) -> Callable[["_AsyncResource", SynapseRequest], "defer.Deferred[None]"]:
|
||||||
"""Wraps an async request handler so that it calls request.processing.
|
"""Wraps an async request handler so that it calls request.processing.
|
||||||
|
|
||||||
This helps ensure that work done by the request handler after the request is completed
|
This helps ensure that work done by the request handler after the request is completed
|
||||||
|
@ -236,7 +235,7 @@ def wrap_async_request_handler(
|
||||||
"""
|
"""
|
||||||
|
|
||||||
async def wrapped_async_request_handler(
|
async def wrapped_async_request_handler(
|
||||||
self: "_AsyncResource", request: "SynapseRequest"
|
self: "_AsyncResource", request: SynapseRequest
|
||||||
) -> None:
|
) -> None:
|
||||||
with request.processing():
|
with request.processing():
|
||||||
await h(self, request)
|
await h(self, request)
|
||||||
|
@ -301,7 +300,7 @@ class _AsyncResource(resource.Resource, metaclass=abc.ABCMeta):
|
||||||
|
|
||||||
self._extract_context = extract_context
|
self._extract_context = extract_context
|
||||||
|
|
||||||
def render(self, request: "SynapseRequest") -> int:
|
def render(self, request: SynapseRequest) -> int:
|
||||||
"""This gets called by twisted every time someone sends us a request."""
|
"""This gets called by twisted every time someone sends us a request."""
|
||||||
request.render_deferred = defer.ensureDeferred(
|
request.render_deferred = defer.ensureDeferred(
|
||||||
self._async_render_wrapper(request)
|
self._async_render_wrapper(request)
|
||||||
|
@ -309,7 +308,7 @@ class _AsyncResource(resource.Resource, metaclass=abc.ABCMeta):
|
||||||
return NOT_DONE_YET
|
return NOT_DONE_YET
|
||||||
|
|
||||||
@wrap_async_request_handler
|
@wrap_async_request_handler
|
||||||
async def _async_render_wrapper(self, request: "SynapseRequest") -> None:
|
async def _async_render_wrapper(self, request: SynapseRequest) -> None:
|
||||||
"""This is a wrapper that delegates to `_async_render` and handles
|
"""This is a wrapper that delegates to `_async_render` and handles
|
||||||
exceptions, return values, metrics, etc.
|
exceptions, return values, metrics, etc.
|
||||||
"""
|
"""
|
||||||
|
@ -327,15 +326,9 @@ class _AsyncResource(resource.Resource, metaclass=abc.ABCMeta):
|
||||||
# of our stack, and thus gives us a sensible stack
|
# of our stack, and thus gives us a sensible stack
|
||||||
# trace.
|
# trace.
|
||||||
f = failure.Failure()
|
f = failure.Failure()
|
||||||
logger.exception(
|
|
||||||
"Error handling request",
|
|
||||||
exc_info=(f.type, f.value, f.getTracebackObject()),
|
|
||||||
)
|
|
||||||
self._send_error_response(f, request)
|
self._send_error_response(f, request)
|
||||||
|
|
||||||
async def _async_render(
|
async def _async_render(self, request: SynapseRequest) -> Optional[Tuple[int, Any]]:
|
||||||
self, request: "SynapseRequest"
|
|
||||||
) -> Optional[Tuple[int, Any]]:
|
|
||||||
"""Delegates to `_async_render_<METHOD>` methods, or returns a 400 if
|
"""Delegates to `_async_render_<METHOD>` methods, or returns a 400 if
|
||||||
no appropriate method exists. Can be overridden in sub classes for
|
no appropriate method exists. Can be overridden in sub classes for
|
||||||
different routing.
|
different routing.
|
||||||
|
@ -365,7 +358,7 @@ class _AsyncResource(resource.Resource, metaclass=abc.ABCMeta):
|
||||||
@abc.abstractmethod
|
@abc.abstractmethod
|
||||||
def _send_response(
|
def _send_response(
|
||||||
self,
|
self,
|
||||||
request: "SynapseRequest",
|
request: SynapseRequest,
|
||||||
code: int,
|
code: int,
|
||||||
response_object: Any,
|
response_object: Any,
|
||||||
) -> None:
|
) -> None:
|
||||||
|
@ -375,7 +368,7 @@ class _AsyncResource(resource.Resource, metaclass=abc.ABCMeta):
|
||||||
def _send_error_response(
|
def _send_error_response(
|
||||||
self,
|
self,
|
||||||
f: failure.Failure,
|
f: failure.Failure,
|
||||||
request: "SynapseRequest",
|
request: SynapseRequest,
|
||||||
) -> None:
|
) -> None:
|
||||||
raise NotImplementedError()
|
raise NotImplementedError()
|
||||||
|
|
||||||
|
@ -391,7 +384,7 @@ class DirectServeJsonResource(_AsyncResource):
|
||||||
|
|
||||||
def _send_response(
|
def _send_response(
|
||||||
self,
|
self,
|
||||||
request: "SynapseRequest",
|
request: SynapseRequest,
|
||||||
code: int,
|
code: int,
|
||||||
response_object: Any,
|
response_object: Any,
|
||||||
) -> None:
|
) -> None:
|
||||||
|
@ -408,7 +401,7 @@ class DirectServeJsonResource(_AsyncResource):
|
||||||
def _send_error_response(
|
def _send_error_response(
|
||||||
self,
|
self,
|
||||||
f: failure.Failure,
|
f: failure.Failure,
|
||||||
request: "SynapseRequest",
|
request: SynapseRequest,
|
||||||
) -> None:
|
) -> None:
|
||||||
"""Implements _AsyncResource._send_error_response"""
|
"""Implements _AsyncResource._send_error_response"""
|
||||||
return_json_error(f, request, None)
|
return_json_error(f, request, None)
|
||||||
|
@ -480,7 +473,7 @@ class JsonResource(DirectServeJsonResource):
|
||||||
)
|
)
|
||||||
|
|
||||||
def _get_handler_for_request(
|
def _get_handler_for_request(
|
||||||
self, request: "SynapseRequest"
|
self, request: SynapseRequest
|
||||||
) -> Tuple[ServletCallback, str, Dict[str, str]]:
|
) -> Tuple[ServletCallback, str, Dict[str, str]]:
|
||||||
"""Finds a callback method to handle the given request.
|
"""Finds a callback method to handle the given request.
|
||||||
|
|
||||||
|
@ -510,7 +503,7 @@ class JsonResource(DirectServeJsonResource):
|
||||||
# Huh. No one wanted to handle that? Fiiiiiine.
|
# Huh. No one wanted to handle that? Fiiiiiine.
|
||||||
raise UnrecognizedRequestError(code=404)
|
raise UnrecognizedRequestError(code=404)
|
||||||
|
|
||||||
async def _async_render(self, request: "SynapseRequest") -> Tuple[int, Any]:
|
async def _async_render(self, request: SynapseRequest) -> Tuple[int, Any]:
|
||||||
callback, servlet_classname, group_dict = self._get_handler_for_request(request)
|
callback, servlet_classname, group_dict = self._get_handler_for_request(request)
|
||||||
|
|
||||||
request.is_render_cancellable = is_function_cancellable(callback)
|
request.is_render_cancellable = is_function_cancellable(callback)
|
||||||
|
@ -542,7 +535,7 @@ class JsonResource(DirectServeJsonResource):
|
||||||
def _send_error_response(
|
def _send_error_response(
|
||||||
self,
|
self,
|
||||||
f: failure.Failure,
|
f: failure.Failure,
|
||||||
request: "SynapseRequest",
|
request: SynapseRequest,
|
||||||
) -> None:
|
) -> None:
|
||||||
"""Implements _AsyncResource._send_error_response"""
|
"""Implements _AsyncResource._send_error_response"""
|
||||||
return_json_error(f, request, self.hs.config)
|
return_json_error(f, request, self.hs.config)
|
||||||
|
@ -558,7 +551,7 @@ class DirectServeHtmlResource(_AsyncResource):
|
||||||
|
|
||||||
def _send_response(
|
def _send_response(
|
||||||
self,
|
self,
|
||||||
request: "SynapseRequest",
|
request: SynapseRequest,
|
||||||
code: int,
|
code: int,
|
||||||
response_object: Any,
|
response_object: Any,
|
||||||
) -> None:
|
) -> None:
|
||||||
|
@ -572,7 +565,7 @@ class DirectServeHtmlResource(_AsyncResource):
|
||||||
def _send_error_response(
|
def _send_error_response(
|
||||||
self,
|
self,
|
||||||
f: failure.Failure,
|
f: failure.Failure,
|
||||||
request: "SynapseRequest",
|
request: SynapseRequest,
|
||||||
) -> None:
|
) -> None:
|
||||||
"""Implements _AsyncResource._send_error_response"""
|
"""Implements _AsyncResource._send_error_response"""
|
||||||
return_html_error(f, request, self.ERROR_TEMPLATE)
|
return_html_error(f, request, self.ERROR_TEMPLATE)
|
||||||
|
@ -599,7 +592,7 @@ class UnrecognizedRequestResource(resource.Resource):
|
||||||
errcode of M_UNRECOGNIZED.
|
errcode of M_UNRECOGNIZED.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def render(self, request: "SynapseRequest") -> int:
|
def render(self, request: SynapseRequest) -> int:
|
||||||
f = failure.Failure(UnrecognizedRequestError(code=404))
|
f = failure.Failure(UnrecognizedRequestError(code=404))
|
||||||
return_json_error(f, request, None)
|
return_json_error(f, request, None)
|
||||||
# A response has already been sent but Twisted requires either NOT_DONE_YET
|
# A response has already been sent but Twisted requires either NOT_DONE_YET
|
||||||
|
@ -629,7 +622,7 @@ class RootRedirect(resource.Resource):
|
||||||
class OptionsResource(resource.Resource):
|
class OptionsResource(resource.Resource):
|
||||||
"""Responds to OPTION requests for itself and all children."""
|
"""Responds to OPTION requests for itself and all children."""
|
||||||
|
|
||||||
def render_OPTIONS(self, request: "SynapseRequest") -> bytes:
|
def render_OPTIONS(self, request: SynapseRequest) -> bytes:
|
||||||
request.setResponseCode(204)
|
request.setResponseCode(204)
|
||||||
request.setHeader(b"Content-Length", b"0")
|
request.setHeader(b"Content-Length", b"0")
|
||||||
|
|
||||||
|
@ -744,7 +737,7 @@ def _encode_json_bytes(json_object: object) -> bytes:
|
||||||
|
|
||||||
|
|
||||||
def respond_with_json(
|
def respond_with_json(
|
||||||
request: "SynapseRequest",
|
request: SynapseRequest,
|
||||||
code: int,
|
code: int,
|
||||||
json_object: Any,
|
json_object: Any,
|
||||||
send_cors: bool = False,
|
send_cors: bool = False,
|
||||||
|
@ -794,7 +787,7 @@ def respond_with_json(
|
||||||
|
|
||||||
|
|
||||||
def respond_with_json_bytes(
|
def respond_with_json_bytes(
|
||||||
request: "SynapseRequest",
|
request: SynapseRequest,
|
||||||
code: int,
|
code: int,
|
||||||
json_bytes: bytes,
|
json_bytes: bytes,
|
||||||
send_cors: bool = False,
|
send_cors: bool = False,
|
||||||
|
@ -832,7 +825,7 @@ def respond_with_json_bytes(
|
||||||
|
|
||||||
|
|
||||||
async def _async_write_json_to_request_in_thread(
|
async def _async_write_json_to_request_in_thread(
|
||||||
request: "SynapseRequest",
|
request: SynapseRequest,
|
||||||
json_encoder: Callable[[Any], bytes],
|
json_encoder: Callable[[Any], bytes],
|
||||||
json_object: Any,
|
json_object: Any,
|
||||||
) -> None:
|
) -> None:
|
||||||
|
@ -890,7 +883,7 @@ def _write_bytes_to_request(request: Request, bytes_to_write: bytes) -> None:
|
||||||
_ByteProducer(request, bytes_generator)
|
_ByteProducer(request, bytes_generator)
|
||||||
|
|
||||||
|
|
||||||
def set_cors_headers(request: "SynapseRequest") -> None:
|
def set_cors_headers(request: SynapseRequest) -> None:
|
||||||
"""Set the CORS headers so that javascript running in a web browsers can
|
"""Set the CORS headers so that javascript running in a web browsers can
|
||||||
use this API
|
use this API
|
||||||
|
|
||||||
|
@ -988,7 +981,7 @@ def set_clickjacking_protection_headers(request: Request) -> None:
|
||||||
|
|
||||||
|
|
||||||
def respond_with_redirect(
|
def respond_with_redirect(
|
||||||
request: "SynapseRequest", url: bytes, statusCode: int = FOUND, cors: bool = False
|
request: SynapseRequest, url: bytes, statusCode: int = FOUND, cors: bool = False
|
||||||
) -> None:
|
) -> None:
|
||||||
"""
|
"""
|
||||||
Write a 302 (or other specified status code) response to the request, if it is still alive.
|
Write a 302 (or other specified status code) response to the request, if it is still alive.
|
||||||
|
|
|
@ -21,28 +21,25 @@ from zope.interface import implementer
|
||||||
|
|
||||||
from twisted.internet.address import UNIXAddress
|
from twisted.internet.address import UNIXAddress
|
||||||
from twisted.internet.defer import Deferred
|
from twisted.internet.defer import Deferred
|
||||||
from twisted.internet.interfaces import IAddress
|
from twisted.internet.interfaces import IAddress, IReactorTime
|
||||||
from twisted.python.failure import Failure
|
from twisted.python.failure import Failure
|
||||||
from twisted.web.http import HTTPChannel
|
from twisted.web.http import HTTPChannel
|
||||||
from twisted.web.iweb import IAgent
|
|
||||||
from twisted.web.resource import IResource, Resource
|
from twisted.web.resource import IResource, Resource
|
||||||
from twisted.web.server import Request
|
from twisted.web.server import Request, Site
|
||||||
|
|
||||||
from synapse.config.server import ListenerConfig
|
from synapse.config.server import ListenerConfig
|
||||||
from synapse.http import get_request_user_agent, redact_uri
|
from synapse.http import get_request_user_agent, redact_uri
|
||||||
from synapse.http.proxy import ProxySite
|
|
||||||
from synapse.http.request_metrics import RequestMetrics, requests_counter
|
from synapse.http.request_metrics import RequestMetrics, requests_counter
|
||||||
from synapse.logging.context import (
|
from synapse.logging.context import (
|
||||||
ContextRequest,
|
ContextRequest,
|
||||||
LoggingContext,
|
LoggingContext,
|
||||||
PreserveLoggingContext,
|
PreserveLoggingContext,
|
||||||
)
|
)
|
||||||
from synapse.types import ISynapseReactor, Requester
|
from synapse.types import Requester
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
if TYPE_CHECKING:
|
||||||
import opentracing
|
import opentracing
|
||||||
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
_next_request_seq = 0
|
_next_request_seq = 0
|
||||||
|
@ -105,7 +102,7 @@ class SynapseRequest(Request):
|
||||||
# A boolean indicating whether `render_deferred` should be cancelled if the
|
# A boolean indicating whether `render_deferred` should be cancelled if the
|
||||||
# client disconnects early. Expected to be set by the coroutine started by
|
# client disconnects early. Expected to be set by the coroutine started by
|
||||||
# `Resource.render`, if rendering is asynchronous.
|
# `Resource.render`, if rendering is asynchronous.
|
||||||
self.is_render_cancellable: bool = False
|
self.is_render_cancellable = False
|
||||||
|
|
||||||
global _next_request_seq
|
global _next_request_seq
|
||||||
self.request_seq = _next_request_seq
|
self.request_seq = _next_request_seq
|
||||||
|
@ -604,7 +601,7 @@ class _XForwardedForAddress:
|
||||||
host: str
|
host: str
|
||||||
|
|
||||||
|
|
||||||
class SynapseSite(ProxySite):
|
class SynapseSite(Site):
|
||||||
"""
|
"""
|
||||||
Synapse-specific twisted http Site
|
Synapse-specific twisted http Site
|
||||||
|
|
||||||
|
@ -626,8 +623,7 @@ class SynapseSite(ProxySite):
|
||||||
resource: IResource,
|
resource: IResource,
|
||||||
server_version_string: str,
|
server_version_string: str,
|
||||||
max_request_body_size: int,
|
max_request_body_size: int,
|
||||||
reactor: ISynapseReactor,
|
reactor: IReactorTime,
|
||||||
federation_agent: IAgent,
|
|
||||||
):
|
):
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
@ -642,11 +638,7 @@ class SynapseSite(ProxySite):
|
||||||
dropping the connection
|
dropping the connection
|
||||||
reactor: reactor to be used to manage connection timeouts
|
reactor: reactor to be used to manage connection timeouts
|
||||||
"""
|
"""
|
||||||
super().__init__(
|
Site.__init__(self, resource, reactor=reactor)
|
||||||
resource=resource,
|
|
||||||
reactor=reactor,
|
|
||||||
federation_agent=federation_agent,
|
|
||||||
)
|
|
||||||
|
|
||||||
self.site_tag = site_tag
|
self.site_tag = site_tag
|
||||||
self.reactor = reactor
|
self.reactor = reactor
|
||||||
|
@ -657,9 +649,7 @@ class SynapseSite(ProxySite):
|
||||||
|
|
||||||
request_id_header = config.http_options.request_id_header
|
request_id_header = config.http_options.request_id_header
|
||||||
|
|
||||||
self.experimental_cors_msc3886: bool = (
|
self.experimental_cors_msc3886 = config.http_options.experimental_cors_msc3886
|
||||||
config.http_options.experimental_cors_msc3886
|
|
||||||
)
|
|
||||||
|
|
||||||
def request_factory(channel: HTTPChannel, queued: bool) -> Request:
|
def request_factory(channel: HTTPChannel, queued: bool) -> Request:
|
||||||
return request_class(
|
return request_class(
|
||||||
|
|
|
@ -31,7 +31,9 @@ from tests.unittest import HomeserverTestCase
|
||||||
|
|
||||||
class FederationReaderOpenIDListenerTests(HomeserverTestCase):
|
class FederationReaderOpenIDListenerTests(HomeserverTestCase):
|
||||||
def make_homeserver(self, reactor: MemoryReactor, clock: Clock) -> HomeServer:
|
def make_homeserver(self, reactor: MemoryReactor, clock: Clock) -> HomeServer:
|
||||||
hs = self.setup_test_homeserver(homeserver_to_use=GenericWorkerServer)
|
hs = self.setup_test_homeserver(
|
||||||
|
federation_http_client=None, homeserver_to_use=GenericWorkerServer
|
||||||
|
)
|
||||||
return hs
|
return hs
|
||||||
|
|
||||||
def default_config(self) -> JsonDict:
|
def default_config(self) -> JsonDict:
|
||||||
|
@ -89,7 +91,9 @@ class FederationReaderOpenIDListenerTests(HomeserverTestCase):
|
||||||
@patch("synapse.app.homeserver.KeyResource", new=Mock())
|
@patch("synapse.app.homeserver.KeyResource", new=Mock())
|
||||||
class SynapseHomeserverOpenIDListenerTests(HomeserverTestCase):
|
class SynapseHomeserverOpenIDListenerTests(HomeserverTestCase):
|
||||||
def make_homeserver(self, reactor: MemoryReactor, clock: Clock) -> HomeServer:
|
def make_homeserver(self, reactor: MemoryReactor, clock: Clock) -> HomeServer:
|
||||||
hs = self.setup_test_homeserver(homeserver_to_use=SynapseHomeServer)
|
hs = self.setup_test_homeserver(
|
||||||
|
federation_http_client=None, homeserver_to_use=SynapseHomeServer
|
||||||
|
)
|
||||||
return hs
|
return hs
|
||||||
|
|
||||||
@parameterized.expand(
|
@parameterized.expand(
|
||||||
|
|
|
@ -41,6 +41,7 @@ class DeviceTestCase(unittest.HomeserverTestCase):
|
||||||
self.appservice_api = mock.Mock()
|
self.appservice_api = mock.Mock()
|
||||||
hs = self.setup_test_homeserver(
|
hs = self.setup_test_homeserver(
|
||||||
"server",
|
"server",
|
||||||
|
federation_http_client=None,
|
||||||
application_service_api=self.appservice_api,
|
application_service_api=self.appservice_api,
|
||||||
)
|
)
|
||||||
handler = hs.get_device_handler()
|
handler = hs.get_device_handler()
|
||||||
|
@ -400,7 +401,7 @@ class DeviceTestCase(unittest.HomeserverTestCase):
|
||||||
|
|
||||||
class DehydrationTestCase(unittest.HomeserverTestCase):
|
class DehydrationTestCase(unittest.HomeserverTestCase):
|
||||||
def make_homeserver(self, reactor: MemoryReactor, clock: Clock) -> HomeServer:
|
def make_homeserver(self, reactor: MemoryReactor, clock: Clock) -> HomeServer:
|
||||||
hs = self.setup_test_homeserver("server")
|
hs = self.setup_test_homeserver("server", federation_http_client=None)
|
||||||
handler = hs.get_device_handler()
|
handler = hs.get_device_handler()
|
||||||
assert isinstance(handler, DeviceHandler)
|
assert isinstance(handler, DeviceHandler)
|
||||||
self.handler = handler
|
self.handler = handler
|
||||||
|
|
|
@ -57,7 +57,7 @@ class FederationTestCase(unittest.FederatingHomeserverTestCase):
|
||||||
]
|
]
|
||||||
|
|
||||||
def make_homeserver(self, reactor: MemoryReactor, clock: Clock) -> HomeServer:
|
def make_homeserver(self, reactor: MemoryReactor, clock: Clock) -> HomeServer:
|
||||||
hs = self.setup_test_homeserver()
|
hs = self.setup_test_homeserver(federation_http_client=None)
|
||||||
self.handler = hs.get_federation_handler()
|
self.handler = hs.get_federation_handler()
|
||||||
self.store = hs.get_datastores().main
|
self.store = hs.get_datastores().main
|
||||||
return hs
|
return hs
|
||||||
|
|
|
@ -993,6 +993,7 @@ class PresenceJoinTestCase(unittest.HomeserverTestCase):
|
||||||
def make_homeserver(self, reactor: MemoryReactor, clock: Clock) -> HomeServer:
|
def make_homeserver(self, reactor: MemoryReactor, clock: Clock) -> HomeServer:
|
||||||
hs = self.setup_test_homeserver(
|
hs = self.setup_test_homeserver(
|
||||||
"server",
|
"server",
|
||||||
|
federation_http_client=None,
|
||||||
federation_sender=Mock(spec=FederationSender),
|
federation_sender=Mock(spec=FederationSender),
|
||||||
)
|
)
|
||||||
return hs
|
return hs
|
||||||
|
|
|
@ -17,8 +17,6 @@ import json
|
||||||
from typing import Dict, List, Set
|
from typing import Dict, List, Set
|
||||||
from unittest.mock import ANY, Mock, call
|
from unittest.mock import ANY, Mock, call
|
||||||
|
|
||||||
from netaddr import IPSet
|
|
||||||
|
|
||||||
from twisted.test.proto_helpers import MemoryReactor
|
from twisted.test.proto_helpers import MemoryReactor
|
||||||
from twisted.web.resource import Resource
|
from twisted.web.resource import Resource
|
||||||
|
|
||||||
|
@ -26,7 +24,6 @@ from synapse.api.constants import EduTypes
|
||||||
from synapse.api.errors import AuthError
|
from synapse.api.errors import AuthError
|
||||||
from synapse.federation.transport.server import TransportLayerServer
|
from synapse.federation.transport.server import TransportLayerServer
|
||||||
from synapse.handlers.typing import TypingWriterHandler
|
from synapse.handlers.typing import TypingWriterHandler
|
||||||
from synapse.http.federation.matrix_federation_agent import MatrixFederationAgent
|
|
||||||
from synapse.server import HomeServer
|
from synapse.server import HomeServer
|
||||||
from synapse.types import JsonDict, Requester, UserID, create_requester
|
from synapse.types import JsonDict, Requester, UserID, create_requester
|
||||||
from synapse.util import Clock
|
from synapse.util import Clock
|
||||||
|
@ -79,13 +76,6 @@ class TypingNotificationsTestCase(unittest.HomeserverTestCase):
|
||||||
# we mock out the federation client too
|
# we mock out the federation client too
|
||||||
self.mock_federation_client = Mock(spec=["put_json"])
|
self.mock_federation_client = Mock(spec=["put_json"])
|
||||||
self.mock_federation_client.put_json.return_value = make_awaitable((200, "OK"))
|
self.mock_federation_client.put_json.return_value = make_awaitable((200, "OK"))
|
||||||
self.mock_federation_client.agent = MatrixFederationAgent(
|
|
||||||
reactor,
|
|
||||||
tls_client_options_factory=None,
|
|
||||||
user_agent=b"SynapseInTrialTest/0.0.0",
|
|
||||||
ip_allowlist=None,
|
|
||||||
ip_blocklist=IPSet(),
|
|
||||||
)
|
|
||||||
|
|
||||||
# the tests assume that we are starting at unix time 1000
|
# the tests assume that we are starting at unix time 1000
|
||||||
reactor.pump((1000,))
|
reactor.pump((1000,))
|
||||||
|
|
|
@ -11,8 +11,8 @@
|
||||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
from typing import Any, Dict, Generator
|
from typing import Generator
|
||||||
from unittest.mock import ANY, Mock, create_autospec
|
from unittest.mock import Mock
|
||||||
|
|
||||||
from netaddr import IPSet
|
from netaddr import IPSet
|
||||||
from parameterized import parameterized
|
from parameterized import parameterized
|
||||||
|
@ -21,11 +21,10 @@ from twisted.internet import defer
|
||||||
from twisted.internet.defer import Deferred, TimeoutError
|
from twisted.internet.defer import Deferred, TimeoutError
|
||||||
from twisted.internet.error import ConnectingCancelledError, DNSLookupError
|
from twisted.internet.error import ConnectingCancelledError, DNSLookupError
|
||||||
from twisted.test.proto_helpers import MemoryReactor, StringTransport
|
from twisted.test.proto_helpers import MemoryReactor, StringTransport
|
||||||
from twisted.web.client import Agent, ResponseNeverReceived
|
from twisted.web.client import ResponseNeverReceived
|
||||||
from twisted.web.http import HTTPChannel
|
from twisted.web.http import HTTPChannel
|
||||||
from twisted.web.http_headers import Headers
|
|
||||||
|
|
||||||
from synapse.api.errors import HttpResponseException, RequestSendFailed
|
from synapse.api.errors import RequestSendFailed
|
||||||
from synapse.http.matrixfederationclient import (
|
from synapse.http.matrixfederationclient import (
|
||||||
ByteParser,
|
ByteParser,
|
||||||
MatrixFederationHttpClient,
|
MatrixFederationHttpClient,
|
||||||
|
@ -40,9 +39,7 @@ from synapse.logging.context import (
|
||||||
from synapse.server import HomeServer
|
from synapse.server import HomeServer
|
||||||
from synapse.util import Clock
|
from synapse.util import Clock
|
||||||
|
|
||||||
from tests.replication._base import BaseMultiWorkerStreamTestCase
|
|
||||||
from tests.server import FakeTransport
|
from tests.server import FakeTransport
|
||||||
from tests.test_utils import FakeResponse
|
|
||||||
from tests.unittest import HomeserverTestCase, override_config
|
from tests.unittest import HomeserverTestCase, override_config
|
||||||
|
|
||||||
|
|
||||||
|
@ -661,181 +658,3 @@ class FederationClientTests(HomeserverTestCase):
|
||||||
self.assertEqual(self.cl.max_short_retry_delay_seconds, 7)
|
self.assertEqual(self.cl.max_short_retry_delay_seconds, 7)
|
||||||
self.assertEqual(self.cl.max_long_retries, 20)
|
self.assertEqual(self.cl.max_long_retries, 20)
|
||||||
self.assertEqual(self.cl.max_short_retries, 5)
|
self.assertEqual(self.cl.max_short_retries, 5)
|
||||||
|
|
||||||
|
|
||||||
class FederationClientProxyTests(BaseMultiWorkerStreamTestCase):
|
|
||||||
def default_config(self) -> Dict[str, Any]:
|
|
||||||
conf = super().default_config()
|
|
||||||
conf["instance_map"] = {
|
|
||||||
"main": {"host": "testserv", "port": 8765},
|
|
||||||
"federation_sender": {"host": "testserv", "port": 1001},
|
|
||||||
}
|
|
||||||
return conf
|
|
||||||
|
|
||||||
@override_config({"outbound_federation_restricted_to": ["federation_sender"]})
|
|
||||||
def test_proxy_requests_through_federation_sender_worker(self) -> None:
|
|
||||||
"""
|
|
||||||
Test that all outbound federation requests go through the `federation_sender`
|
|
||||||
worker
|
|
||||||
"""
|
|
||||||
# Mock out the `MatrixFederationHttpClient` of the `federation_sender` instance
|
|
||||||
# so we can act like some remote server responding to requests
|
|
||||||
mock_client_on_federation_sender = Mock()
|
|
||||||
mock_agent_on_federation_sender = create_autospec(Agent, spec_set=True)
|
|
||||||
mock_client_on_federation_sender.agent = mock_agent_on_federation_sender
|
|
||||||
|
|
||||||
# Create the `federation_sender` worker
|
|
||||||
self.federation_sender = self.make_worker_hs(
|
|
||||||
"synapse.app.generic_worker",
|
|
||||||
{"worker_name": "federation_sender"},
|
|
||||||
federation_http_client=mock_client_on_federation_sender,
|
|
||||||
)
|
|
||||||
|
|
||||||
# Fake `remoteserv:8008` responding to requests
|
|
||||||
mock_agent_on_federation_sender.request.side_effect = (
|
|
||||||
lambda *args, **kwargs: defer.succeed(
|
|
||||||
FakeResponse.json(
|
|
||||||
payload={
|
|
||||||
"foo": "bar",
|
|
||||||
}
|
|
||||||
)
|
|
||||||
)
|
|
||||||
)
|
|
||||||
|
|
||||||
# This federation request from the main process should be proxied through the
|
|
||||||
# `federation_sender` worker off to the remote server
|
|
||||||
test_request_from_main_process_d = defer.ensureDeferred(
|
|
||||||
self.hs.get_federation_http_client().get_json("remoteserv:8008", "foo/bar")
|
|
||||||
)
|
|
||||||
|
|
||||||
# Pump the reactor so our deferred goes through the motions
|
|
||||||
self.pump()
|
|
||||||
|
|
||||||
# Make sure that the request was proxied through the `federation_sender` worker
|
|
||||||
mock_agent_on_federation_sender.request.assert_called_once_with(
|
|
||||||
b"GET",
|
|
||||||
b"matrix-federation://remoteserv:8008/foo/bar",
|
|
||||||
headers=ANY,
|
|
||||||
bodyProducer=ANY,
|
|
||||||
)
|
|
||||||
|
|
||||||
# Make sure the response is as expected back on the main worker
|
|
||||||
res = self.successResultOf(test_request_from_main_process_d)
|
|
||||||
self.assertEqual(res, {"foo": "bar"})
|
|
||||||
|
|
||||||
@override_config({"outbound_federation_restricted_to": ["federation_sender"]})
|
|
||||||
def test_proxy_request_with_network_error_through_federation_sender_worker(
|
|
||||||
self,
|
|
||||||
) -> None:
|
|
||||||
"""
|
|
||||||
Test that when the outbound federation request fails with a network related
|
|
||||||
error, a sensible error makes its way back to the main process.
|
|
||||||
"""
|
|
||||||
# Mock out the `MatrixFederationHttpClient` of the `federation_sender` instance
|
|
||||||
# so we can act like some remote server responding to requests
|
|
||||||
mock_client_on_federation_sender = Mock()
|
|
||||||
mock_agent_on_federation_sender = create_autospec(Agent, spec_set=True)
|
|
||||||
mock_client_on_federation_sender.agent = mock_agent_on_federation_sender
|
|
||||||
|
|
||||||
# Create the `federation_sender` worker
|
|
||||||
self.federation_sender = self.make_worker_hs(
|
|
||||||
"synapse.app.generic_worker",
|
|
||||||
{"worker_name": "federation_sender"},
|
|
||||||
federation_http_client=mock_client_on_federation_sender,
|
|
||||||
)
|
|
||||||
|
|
||||||
# Fake `remoteserv:8008` responding to requests
|
|
||||||
mock_agent_on_federation_sender.request.side_effect = (
|
|
||||||
lambda *args, **kwargs: defer.fail(ResponseNeverReceived("fake error"))
|
|
||||||
)
|
|
||||||
|
|
||||||
# This federation request from the main process should be proxied through the
|
|
||||||
# `federation_sender` worker off to the remote server
|
|
||||||
test_request_from_main_process_d = defer.ensureDeferred(
|
|
||||||
self.hs.get_federation_http_client().get_json("remoteserv:8008", "foo/bar")
|
|
||||||
)
|
|
||||||
|
|
||||||
# Pump the reactor so our deferred goes through the motions. We pump with 10
|
|
||||||
# seconds (0.1 * 100) so the `MatrixFederationHttpClient` runs out of retries
|
|
||||||
# and finally passes along the error response.
|
|
||||||
self.pump(0.1)
|
|
||||||
|
|
||||||
# Make sure that the request was proxied through the `federation_sender` worker
|
|
||||||
mock_agent_on_federation_sender.request.assert_called_with(
|
|
||||||
b"GET",
|
|
||||||
b"matrix-federation://remoteserv:8008/foo/bar",
|
|
||||||
headers=ANY,
|
|
||||||
bodyProducer=ANY,
|
|
||||||
)
|
|
||||||
|
|
||||||
# Make sure we get some sort of error back on the main worker
|
|
||||||
failure_res = self.failureResultOf(test_request_from_main_process_d)
|
|
||||||
self.assertIsInstance(failure_res.value, RequestSendFailed)
|
|
||||||
self.assertIsInstance(failure_res.value.inner_exception, HttpResponseException)
|
|
||||||
|
|
||||||
@override_config({"outbound_federation_restricted_to": ["federation_sender"]})
|
|
||||||
def test_proxy_requests_and_discards_hop_by_hop_headers(self) -> None:
|
|
||||||
"""
|
|
||||||
Test to make sure hop-by-hop headers and addional headers defined in the
|
|
||||||
`Connection` header are discarded when proxying requests
|
|
||||||
"""
|
|
||||||
# Mock out the `MatrixFederationHttpClient` of the `federation_sender` instance
|
|
||||||
# so we can act like some remote server responding to requests
|
|
||||||
mock_client_on_federation_sender = Mock()
|
|
||||||
mock_agent_on_federation_sender = create_autospec(Agent, spec_set=True)
|
|
||||||
mock_client_on_federation_sender.agent = mock_agent_on_federation_sender
|
|
||||||
|
|
||||||
# Create the `federation_sender` worker
|
|
||||||
self.federation_sender = self.make_worker_hs(
|
|
||||||
"synapse.app.generic_worker",
|
|
||||||
{"worker_name": "federation_sender"},
|
|
||||||
federation_http_client=mock_client_on_federation_sender,
|
|
||||||
)
|
|
||||||
|
|
||||||
# Fake `remoteserv:8008` responding to requests
|
|
||||||
mock_agent_on_federation_sender.request.side_effect = lambda *args, **kwargs: defer.succeed(
|
|
||||||
FakeResponse(
|
|
||||||
code=200,
|
|
||||||
body=b'{"foo": "bar"}',
|
|
||||||
headers=Headers(
|
|
||||||
{
|
|
||||||
"Content-Type": ["application/json"],
|
|
||||||
"Connection": ["close, X-Foo, X-Bar"],
|
|
||||||
# Should be removed because it's defined in the `Connection` header
|
|
||||||
"X-Foo": ["foo"],
|
|
||||||
"X-Bar": ["bar"],
|
|
||||||
# Should be removed because it's a hop-by-hop header
|
|
||||||
"Proxy-Authorization": "abcdef",
|
|
||||||
}
|
|
||||||
),
|
|
||||||
)
|
|
||||||
)
|
|
||||||
|
|
||||||
# This federation request from the main process should be proxied through the
|
|
||||||
# `federation_sender` worker off to the remote server
|
|
||||||
test_request_from_main_process_d = defer.ensureDeferred(
|
|
||||||
self.hs.get_federation_http_client().get_json_with_headers(
|
|
||||||
"remoteserv:8008", "foo/bar"
|
|
||||||
)
|
|
||||||
)
|
|
||||||
|
|
||||||
# Pump the reactor so our deferred goes through the motions
|
|
||||||
self.pump()
|
|
||||||
|
|
||||||
# Make sure that the request was proxied through the `federation_sender` worker
|
|
||||||
mock_agent_on_federation_sender.request.assert_called_once_with(
|
|
||||||
b"GET",
|
|
||||||
b"matrix-federation://remoteserv:8008/foo/bar",
|
|
||||||
headers=ANY,
|
|
||||||
bodyProducer=ANY,
|
|
||||||
)
|
|
||||||
|
|
||||||
res, headers = self.successResultOf(test_request_from_main_process_d)
|
|
||||||
header_names = set(headers.keys())
|
|
||||||
|
|
||||||
# Make sure the response does not include the hop-by-hop headers
|
|
||||||
self.assertNotIn(b"X-Foo", header_names)
|
|
||||||
self.assertNotIn(b"X-Bar", header_names)
|
|
||||||
self.assertNotIn(b"Proxy-Authorization", header_names)
|
|
||||||
# Make sure the response is as expected back on the main worker
|
|
||||||
self.assertEqual(res, {"foo": "bar"})
|
|
||||||
|
|
|
@ -1,53 +0,0 @@
|
||||||
# Copyright 2023 The Matrix.org Foundation C.I.C.
|
|
||||||
#
|
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
# you may not use this file except in compliance with the License.
|
|
||||||
# You may obtain a copy of the License at
|
|
||||||
#
|
|
||||||
# http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
#
|
|
||||||
# Unless required by applicable law or agreed to in writing, software
|
|
||||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
# See the License for the specific language governing permissions and
|
|
||||||
# limitations under the License.
|
|
||||||
from typing import Set
|
|
||||||
|
|
||||||
from parameterized import parameterized
|
|
||||||
|
|
||||||
from synapse.http.proxy import parse_connection_header_value
|
|
||||||
|
|
||||||
from tests.unittest import TestCase
|
|
||||||
|
|
||||||
|
|
||||||
class ProxyTests(TestCase):
|
|
||||||
@parameterized.expand(
|
|
||||||
[
|
|
||||||
[b"close, X-Foo, X-Bar", {"Close", "X-Foo", "X-Bar"}],
|
|
||||||
# No whitespace
|
|
||||||
[b"close,X-Foo,X-Bar", {"Close", "X-Foo", "X-Bar"}],
|
|
||||||
# More whitespace
|
|
||||||
[b"close, X-Foo, X-Bar", {"Close", "X-Foo", "X-Bar"}],
|
|
||||||
# "close" directive in not the first position
|
|
||||||
[b"X-Foo, X-Bar, close", {"X-Foo", "X-Bar", "Close"}],
|
|
||||||
# Normalizes header capitalization
|
|
||||||
[b"keep-alive, x-fOo, x-bAr", {"Keep-Alive", "X-Foo", "X-Bar"}],
|
|
||||||
# Handles header names with whitespace
|
|
||||||
[
|
|
||||||
b"keep-alive, x foo, x bar",
|
|
||||||
{"Keep-Alive", "X foo", "X bar"},
|
|
||||||
],
|
|
||||||
]
|
|
||||||
)
|
|
||||||
def test_parse_connection_header_value(
|
|
||||||
self,
|
|
||||||
connection_header_value: bytes,
|
|
||||||
expected_extra_headers_to_remove: Set[str],
|
|
||||||
) -> None:
|
|
||||||
"""
|
|
||||||
Tests that the connection header value is parsed correctly
|
|
||||||
"""
|
|
||||||
self.assertEqual(
|
|
||||||
expected_extra_headers_to_remove,
|
|
||||||
parse_connection_header_value(connection_header_value),
|
|
||||||
)
|
|
|
@ -69,10 +69,10 @@ class BaseStreamTestCase(unittest.HomeserverTestCase):
|
||||||
# Make a new HomeServer object for the worker
|
# Make a new HomeServer object for the worker
|
||||||
self.reactor.lookups["testserv"] = "1.2.3.4"
|
self.reactor.lookups["testserv"] = "1.2.3.4"
|
||||||
self.worker_hs = self.setup_test_homeserver(
|
self.worker_hs = self.setup_test_homeserver(
|
||||||
|
federation_http_client=None,
|
||||||
homeserver_to_use=GenericWorkerServer,
|
homeserver_to_use=GenericWorkerServer,
|
||||||
config=self._get_worker_hs_config(),
|
config=self._get_worker_hs_config(),
|
||||||
reactor=self.reactor,
|
reactor=self.reactor,
|
||||||
federation_http_client=None,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
# Since we use sqlite in memory databases we need to make sure the
|
# Since we use sqlite in memory databases we need to make sure the
|
||||||
|
@ -380,7 +380,6 @@ class BaseMultiWorkerStreamTestCase(unittest.HomeserverTestCase):
|
||||||
server_version_string="1",
|
server_version_string="1",
|
||||||
max_request_body_size=8192,
|
max_request_body_size=8192,
|
||||||
reactor=self.reactor,
|
reactor=self.reactor,
|
||||||
federation_agent=worker_hs.get_federation_http_client().agent,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
worker_hs.get_replication_command_handler().start_replication(worker_hs)
|
worker_hs.get_replication_command_handler().start_replication(worker_hs)
|
||||||
|
|
|
@ -14,18 +14,14 @@
|
||||||
import logging
|
import logging
|
||||||
from unittest.mock import Mock
|
from unittest.mock import Mock
|
||||||
|
|
||||||
from netaddr import IPSet
|
|
||||||
|
|
||||||
from synapse.api.constants import EventTypes, Membership
|
from synapse.api.constants import EventTypes, Membership
|
||||||
from synapse.events.builder import EventBuilderFactory
|
from synapse.events.builder import EventBuilderFactory
|
||||||
from synapse.handlers.typing import TypingWriterHandler
|
from synapse.handlers.typing import TypingWriterHandler
|
||||||
from synapse.http.federation.matrix_federation_agent import MatrixFederationAgent
|
|
||||||
from synapse.rest.admin import register_servlets_for_client_rest_resource
|
from synapse.rest.admin import register_servlets_for_client_rest_resource
|
||||||
from synapse.rest.client import login, room
|
from synapse.rest.client import login, room
|
||||||
from synapse.types import UserID, create_requester
|
from synapse.types import UserID, create_requester
|
||||||
|
|
||||||
from tests.replication._base import BaseMultiWorkerStreamTestCase
|
from tests.replication._base import BaseMultiWorkerStreamTestCase
|
||||||
from tests.server import get_clock
|
|
||||||
from tests.test_utils import make_awaitable
|
from tests.test_utils import make_awaitable
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
@ -45,25 +41,13 @@ class FederationSenderTestCase(BaseMultiWorkerStreamTestCase):
|
||||||
room.register_servlets,
|
room.register_servlets,
|
||||||
]
|
]
|
||||||
|
|
||||||
def setUp(self) -> None:
|
|
||||||
super().setUp()
|
|
||||||
|
|
||||||
reactor, _ = get_clock()
|
|
||||||
self.matrix_federation_agent = MatrixFederationAgent(
|
|
||||||
reactor,
|
|
||||||
tls_client_options_factory=None,
|
|
||||||
user_agent=b"SynapseInTrialTest/0.0.0",
|
|
||||||
ip_allowlist=None,
|
|
||||||
ip_blocklist=IPSet(),
|
|
||||||
)
|
|
||||||
|
|
||||||
def test_send_event_single_sender(self) -> None:
|
def test_send_event_single_sender(self) -> None:
|
||||||
"""Test that using a single federation sender worker correctly sends a
|
"""Test that using a single federation sender worker correctly sends a
|
||||||
new event.
|
new event.
|
||||||
"""
|
"""
|
||||||
mock_client = Mock(spec=["put_json"])
|
mock_client = Mock(spec=["put_json"])
|
||||||
mock_client.put_json.return_value = make_awaitable({})
|
mock_client.put_json.return_value = make_awaitable({})
|
||||||
mock_client.agent = self.matrix_federation_agent
|
|
||||||
self.make_worker_hs(
|
self.make_worker_hs(
|
||||||
"synapse.app.generic_worker",
|
"synapse.app.generic_worker",
|
||||||
{
|
{
|
||||||
|
@ -94,7 +78,6 @@ class FederationSenderTestCase(BaseMultiWorkerStreamTestCase):
|
||||||
"""
|
"""
|
||||||
mock_client1 = Mock(spec=["put_json"])
|
mock_client1 = Mock(spec=["put_json"])
|
||||||
mock_client1.put_json.return_value = make_awaitable({})
|
mock_client1.put_json.return_value = make_awaitable({})
|
||||||
mock_client1.agent = self.matrix_federation_agent
|
|
||||||
self.make_worker_hs(
|
self.make_worker_hs(
|
||||||
"synapse.app.generic_worker",
|
"synapse.app.generic_worker",
|
||||||
{
|
{
|
||||||
|
@ -109,7 +92,6 @@ class FederationSenderTestCase(BaseMultiWorkerStreamTestCase):
|
||||||
|
|
||||||
mock_client2 = Mock(spec=["put_json"])
|
mock_client2 = Mock(spec=["put_json"])
|
||||||
mock_client2.put_json.return_value = make_awaitable({})
|
mock_client2.put_json.return_value = make_awaitable({})
|
||||||
mock_client2.agent = self.matrix_federation_agent
|
|
||||||
self.make_worker_hs(
|
self.make_worker_hs(
|
||||||
"synapse.app.generic_worker",
|
"synapse.app.generic_worker",
|
||||||
{
|
{
|
||||||
|
@ -163,7 +145,6 @@ class FederationSenderTestCase(BaseMultiWorkerStreamTestCase):
|
||||||
"""
|
"""
|
||||||
mock_client1 = Mock(spec=["put_json"])
|
mock_client1 = Mock(spec=["put_json"])
|
||||||
mock_client1.put_json.return_value = make_awaitable({})
|
mock_client1.put_json.return_value = make_awaitable({})
|
||||||
mock_client1.agent = self.matrix_federation_agent
|
|
||||||
self.make_worker_hs(
|
self.make_worker_hs(
|
||||||
"synapse.app.generic_worker",
|
"synapse.app.generic_worker",
|
||||||
{
|
{
|
||||||
|
@ -178,7 +159,6 @@ class FederationSenderTestCase(BaseMultiWorkerStreamTestCase):
|
||||||
|
|
||||||
mock_client2 = Mock(spec=["put_json"])
|
mock_client2 = Mock(spec=["put_json"])
|
||||||
mock_client2.put_json.return_value = make_awaitable({})
|
mock_client2.put_json.return_value = make_awaitable({})
|
||||||
mock_client2.agent = self.matrix_federation_agent
|
|
||||||
self.make_worker_hs(
|
self.make_worker_hs(
|
||||||
"synapse.app.generic_worker",
|
"synapse.app.generic_worker",
|
||||||
{
|
{
|
||||||
|
|
|
@ -40,6 +40,7 @@ class PresenceTestCase(unittest.HomeserverTestCase):
|
||||||
|
|
||||||
hs = self.setup_test_homeserver(
|
hs = self.setup_test_homeserver(
|
||||||
"red",
|
"red",
|
||||||
|
federation_http_client=None,
|
||||||
federation_client=Mock(),
|
federation_client=Mock(),
|
||||||
presence_handler=self.presence_handler,
|
presence_handler=self.presence_handler,
|
||||||
)
|
)
|
||||||
|
|
|
@ -67,6 +67,8 @@ class RoomBase(unittest.HomeserverTestCase):
|
||||||
def make_homeserver(self, reactor: MemoryReactor, clock: Clock) -> HomeServer:
|
def make_homeserver(self, reactor: MemoryReactor, clock: Clock) -> HomeServer:
|
||||||
self.hs = self.setup_test_homeserver(
|
self.hs = self.setup_test_homeserver(
|
||||||
"red",
|
"red",
|
||||||
|
federation_http_client=None,
|
||||||
|
federation_client=Mock(),
|
||||||
)
|
)
|
||||||
|
|
||||||
self.hs.get_federation_handler = Mock() # type: ignore[assignment]
|
self.hs.get_federation_handler = Mock() # type: ignore[assignment]
|
||||||
|
|
|
@ -31,7 +31,7 @@ room_key: RoomKey = {
|
||||||
|
|
||||||
class E2eRoomKeysHandlerTestCase(unittest.HomeserverTestCase):
|
class E2eRoomKeysHandlerTestCase(unittest.HomeserverTestCase):
|
||||||
def make_homeserver(self, reactor: MemoryReactor, clock: Clock) -> HomeServer:
|
def make_homeserver(self, reactor: MemoryReactor, clock: Clock) -> HomeServer:
|
||||||
hs = self.setup_test_homeserver("server")
|
hs = self.setup_test_homeserver("server", federation_http_client=None)
|
||||||
self.store = hs.get_datastores().main
|
self.store = hs.get_datastores().main
|
||||||
return hs
|
return hs
|
||||||
|
|
||||||
|
|
|
@ -27,7 +27,7 @@ class PurgeTests(HomeserverTestCase):
|
||||||
servlets = [room.register_servlets]
|
servlets = [room.register_servlets]
|
||||||
|
|
||||||
def make_homeserver(self, reactor: MemoryReactor, clock: Clock) -> HomeServer:
|
def make_homeserver(self, reactor: MemoryReactor, clock: Clock) -> HomeServer:
|
||||||
hs = self.setup_test_homeserver("server")
|
hs = self.setup_test_homeserver("server", federation_http_client=None)
|
||||||
return hs
|
return hs
|
||||||
|
|
||||||
def prepare(self, reactor: MemoryReactor, clock: Clock, hs: HomeServer) -> None:
|
def prepare(self, reactor: MemoryReactor, clock: Clock, hs: HomeServer) -> None:
|
||||||
|
|
|
@ -45,7 +45,9 @@ def fake_listdir(filepath: str) -> List[str]:
|
||||||
|
|
||||||
class WorkerSchemaTests(HomeserverTestCase):
|
class WorkerSchemaTests(HomeserverTestCase):
|
||||||
def make_homeserver(self, reactor: MemoryReactor, clock: Clock) -> HomeServer:
|
def make_homeserver(self, reactor: MemoryReactor, clock: Clock) -> HomeServer:
|
||||||
hs = self.setup_test_homeserver(homeserver_to_use=GenericWorkerServer)
|
hs = self.setup_test_homeserver(
|
||||||
|
federation_http_client=None, homeserver_to_use=GenericWorkerServer
|
||||||
|
)
|
||||||
return hs
|
return hs
|
||||||
|
|
||||||
def default_config(self) -> JsonDict:
|
def default_config(self) -> JsonDict:
|
||||||
|
|
|
@ -38,7 +38,7 @@ from tests.http.server._base import test_disconnect
|
||||||
from tests.server import (
|
from tests.server import (
|
||||||
FakeChannel,
|
FakeChannel,
|
||||||
FakeSite,
|
FakeSite,
|
||||||
get_clock,
|
ThreadedMemoryReactorClock,
|
||||||
make_request,
|
make_request,
|
||||||
setup_test_homeserver,
|
setup_test_homeserver,
|
||||||
)
|
)
|
||||||
|
@ -46,11 +46,12 @@ from tests.server import (
|
||||||
|
|
||||||
class JsonResourceTests(unittest.TestCase):
|
class JsonResourceTests(unittest.TestCase):
|
||||||
def setUp(self) -> None:
|
def setUp(self) -> None:
|
||||||
reactor, clock = get_clock()
|
self.reactor = ThreadedMemoryReactorClock()
|
||||||
self.reactor = reactor
|
self.hs_clock = Clock(self.reactor)
|
||||||
self.homeserver = setup_test_homeserver(
|
self.homeserver = setup_test_homeserver(
|
||||||
self.addCleanup,
|
self.addCleanup,
|
||||||
clock=clock,
|
federation_http_client=None,
|
||||||
|
clock=self.hs_clock,
|
||||||
reactor=self.reactor,
|
reactor=self.reactor,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -208,13 +209,7 @@ class JsonResourceTests(unittest.TestCase):
|
||||||
|
|
||||||
class OptionsResourceTests(unittest.TestCase):
|
class OptionsResourceTests(unittest.TestCase):
|
||||||
def setUp(self) -> None:
|
def setUp(self) -> None:
|
||||||
reactor, clock = get_clock()
|
self.reactor = ThreadedMemoryReactorClock()
|
||||||
self.reactor = reactor
|
|
||||||
self.homeserver = setup_test_homeserver(
|
|
||||||
self.addCleanup,
|
|
||||||
clock=clock,
|
|
||||||
reactor=self.reactor,
|
|
||||||
)
|
|
||||||
|
|
||||||
class DummyResource(Resource):
|
class DummyResource(Resource):
|
||||||
isLeaf = True
|
isLeaf = True
|
||||||
|
@ -247,7 +242,6 @@ class OptionsResourceTests(unittest.TestCase):
|
||||||
"1.0",
|
"1.0",
|
||||||
max_request_body_size=4096,
|
max_request_body_size=4096,
|
||||||
reactor=self.reactor,
|
reactor=self.reactor,
|
||||||
federation_agent=self.homeserver.get_federation_http_client().agent,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
# render the request and return the channel
|
# render the request and return the channel
|
||||||
|
@ -350,8 +344,7 @@ class WrapHtmlRequestHandlerTests(unittest.TestCase):
|
||||||
await self.callback(request)
|
await self.callback(request)
|
||||||
|
|
||||||
def setUp(self) -> None:
|
def setUp(self) -> None:
|
||||||
reactor, _ = get_clock()
|
self.reactor = ThreadedMemoryReactorClock()
|
||||||
self.reactor = reactor
|
|
||||||
|
|
||||||
def test_good_response(self) -> None:
|
def test_good_response(self) -> None:
|
||||||
async def callback(request: SynapseRequest) -> None:
|
async def callback(request: SynapseRequest) -> None:
|
||||||
|
@ -469,9 +462,9 @@ class DirectServeJsonResourceCancellationTests(unittest.TestCase):
|
||||||
"""Tests for `DirectServeJsonResource` cancellation."""
|
"""Tests for `DirectServeJsonResource` cancellation."""
|
||||||
|
|
||||||
def setUp(self) -> None:
|
def setUp(self) -> None:
|
||||||
reactor, clock = get_clock()
|
self.reactor = ThreadedMemoryReactorClock()
|
||||||
self.reactor = reactor
|
self.clock = Clock(self.reactor)
|
||||||
self.resource = CancellableDirectServeJsonResource(clock)
|
self.resource = CancellableDirectServeJsonResource(self.clock)
|
||||||
self.site = FakeSite(self.resource, self.reactor)
|
self.site = FakeSite(self.resource, self.reactor)
|
||||||
|
|
||||||
def test_cancellable_disconnect(self) -> None:
|
def test_cancellable_disconnect(self) -> None:
|
||||||
|
@ -503,9 +496,9 @@ class DirectServeHtmlResourceCancellationTests(unittest.TestCase):
|
||||||
"""Tests for `DirectServeHtmlResource` cancellation."""
|
"""Tests for `DirectServeHtmlResource` cancellation."""
|
||||||
|
|
||||||
def setUp(self) -> None:
|
def setUp(self) -> None:
|
||||||
reactor, clock = get_clock()
|
self.reactor = ThreadedMemoryReactorClock()
|
||||||
self.reactor = reactor
|
self.clock = Clock(self.reactor)
|
||||||
self.resource = CancellableDirectServeHtmlResource(clock)
|
self.resource = CancellableDirectServeHtmlResource(self.clock)
|
||||||
self.site = FakeSite(self.resource, self.reactor)
|
self.site = FakeSite(self.resource, self.reactor)
|
||||||
|
|
||||||
def test_cancellable_disconnect(self) -> None:
|
def test_cancellable_disconnect(self) -> None:
|
||||||
|
|
|
@ -358,7 +358,6 @@ class HomeserverTestCase(TestCase):
|
||||||
server_version_string="1",
|
server_version_string="1",
|
||||||
max_request_body_size=4096,
|
max_request_body_size=4096,
|
||||||
reactor=self.reactor,
|
reactor=self.reactor,
|
||||||
federation_agent=self.hs.get_federation_http_client().agent,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
from tests.rest.client.utils import RestHelper
|
from tests.rest.client.utils import RestHelper
|
||||||
|
|
Loading…
Reference in a new issue