2018-07-09 08:09:20 +02:00
|
|
|
import json
|
2019-01-22 21:28:48 +01:00
|
|
|
import logging
|
Fix threadsafety in ThreadedMemoryReactorClock (#8497)
This could, very occasionally, cause:
```
tests.test_visibility.FilterEventsForServerTestCase.test_large_room
===============================================================================
[ERROR]
Traceback (most recent call last):
File "/src/tests/rest/media/v1/test_media_storage.py", line 86, in test_ensure_media_is_in_local_cache
self.wait_on_thread(x)
File "/src/tests/unittest.py", line 296, in wait_on_thread
self.reactor.advance(0.01)
File "/src/.tox/py35/lib/python3.5/site-packages/twisted/internet/task.py", line 826, in advance
self._sortCalls()
File "/src/.tox/py35/lib/python3.5/site-packages/twisted/internet/task.py", line 787, in _sortCalls
self.calls.sort(key=lambda a: a.getTime())
builtins.ValueError: list modified during sort
tests.rest.media.v1.test_media_storage.MediaStorageTests.test_ensure_media_is_in_local_cache
```
2020-10-09 18:22:25 +02:00
|
|
|
from collections import deque
|
2020-09-10 12:45:12 +02:00
|
|
|
from io import SEEK_END, BytesIO
|
2020-11-16 15:45:22 +01:00
|
|
|
from typing import Callable, Iterable, Optional, Tuple, Union
|
2018-06-27 11:37:24 +02:00
|
|
|
|
2018-07-09 08:09:20 +02:00
|
|
|
import attr
|
Fix threadsafety in ThreadedMemoryReactorClock (#8497)
This could, very occasionally, cause:
```
tests.test_visibility.FilterEventsForServerTestCase.test_large_room
===============================================================================
[ERROR]
Traceback (most recent call last):
File "/src/tests/rest/media/v1/test_media_storage.py", line 86, in test_ensure_media_is_in_local_cache
self.wait_on_thread(x)
File "/src/tests/unittest.py", line 296, in wait_on_thread
self.reactor.advance(0.01)
File "/src/.tox/py35/lib/python3.5/site-packages/twisted/internet/task.py", line 826, in advance
self._sortCalls()
File "/src/.tox/py35/lib/python3.5/site-packages/twisted/internet/task.py", line 787, in _sortCalls
self.calls.sort(key=lambda a: a.getTime())
builtins.ValueError: list modified during sort
tests.rest.media.v1.test_media_storage.MediaStorageTests.test_ensure_media_is_in_local_cache
```
2020-10-09 18:22:25 +02:00
|
|
|
from typing_extensions import Deque
|
2018-09-13 16:15:51 +02:00
|
|
|
from zope.interface import implementer
|
2018-07-09 08:09:20 +02:00
|
|
|
|
2018-09-13 16:15:51 +02:00
|
|
|
from twisted.internet import address, threads, udp
|
2019-01-29 10:38:29 +01:00
|
|
|
from twisted.internet._resolver import SimpleResolverComplexifier
|
|
|
|
from twisted.internet.defer import Deferred, fail, succeed
|
2018-09-13 16:15:51 +02:00
|
|
|
from twisted.internet.error import DNSLookupError
|
2019-08-28 13:18:53 +02:00
|
|
|
from twisted.internet.interfaces import (
|
|
|
|
IReactorPluggableNameResolver,
|
|
|
|
IReactorTCP,
|
|
|
|
IResolverSimple,
|
|
|
|
)
|
2018-07-09 08:09:20 +02:00
|
|
|
from twisted.python.failure import Failure
|
2019-08-28 13:18:53 +02:00
|
|
|
from twisted.test.proto_helpers import AccumulatingProtocol, MemoryReactorClock
|
2018-11-15 22:55:58 +01:00
|
|
|
from twisted.web.http import unquote
|
|
|
|
from twisted.web.http_headers import Headers
|
2020-01-03 15:19:48 +01:00
|
|
|
from twisted.web.server import Site
|
2018-06-27 11:37:24 +02:00
|
|
|
|
|
|
|
from synapse.http.site import SynapseRequest
|
2018-08-09 04:22:01 +02:00
|
|
|
from synapse.util import Clock
|
2018-07-09 08:09:20 +02:00
|
|
|
|
2018-06-27 11:37:24 +02:00
|
|
|
from tests.utils import setup_test_homeserver as _sth
|
|
|
|
|
2019-01-22 21:28:48 +01:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
2018-06-27 11:37:24 +02:00
|
|
|
|
2018-11-02 14:19:23 +01:00
|
|
|
class TimedOutException(Exception):
|
|
|
|
"""
|
|
|
|
A web query timed out.
|
|
|
|
"""
|
|
|
|
|
|
|
|
|
2018-06-27 11:37:24 +02:00
|
|
|
@attr.s
|
2020-09-04 12:54:56 +02:00
|
|
|
class FakeChannel:
|
2018-06-27 11:37:24 +02:00
|
|
|
"""
|
|
|
|
A fake Twisted Web Channel (the part that interfaces with the
|
|
|
|
wire).
|
|
|
|
"""
|
|
|
|
|
2020-01-03 15:19:48 +01:00
|
|
|
site = attr.ib(type=Site)
|
2018-11-06 17:00:00 +01:00
|
|
|
_reactor = attr.ib()
|
2020-10-30 11:55:24 +01:00
|
|
|
result = attr.ib(type=dict, default=attr.Factory(dict))
|
2018-08-15 15:43:41 +02:00
|
|
|
_producer = None
|
2018-06-27 11:37:24 +02:00
|
|
|
|
|
|
|
@property
|
|
|
|
def json_body(self):
|
|
|
|
if not self.result:
|
|
|
|
raise Exception("No result yet.")
|
2019-06-20 11:32:02 +02:00
|
|
|
return json.loads(self.result["body"].decode("utf8"))
|
2018-08-09 04:22:01 +02:00
|
|
|
|
|
|
|
@property
|
|
|
|
def code(self):
|
|
|
|
if not self.result:
|
|
|
|
raise Exception("No result yet.")
|
|
|
|
return int(self.result["code"])
|
2018-06-27 11:37:24 +02:00
|
|
|
|
2018-11-15 22:55:58 +01:00
|
|
|
@property
|
|
|
|
def headers(self):
|
|
|
|
if not self.result:
|
|
|
|
raise Exception("No result yet.")
|
|
|
|
h = Headers()
|
|
|
|
for i in self.result["headers"]:
|
|
|
|
h.addRawHeader(*i)
|
|
|
|
return h
|
|
|
|
|
2018-06-27 11:37:24 +02:00
|
|
|
def writeHeaders(self, version, code, reason, headers):
|
|
|
|
self.result["version"] = version
|
|
|
|
self.result["code"] = code
|
|
|
|
self.result["reason"] = reason
|
|
|
|
self.result["headers"] = headers
|
|
|
|
|
|
|
|
def write(self, content):
|
2018-11-07 15:37:43 +01:00
|
|
|
assert isinstance(content, bytes), "Should be bytes! " + repr(content)
|
|
|
|
|
2018-06-27 11:37:24 +02:00
|
|
|
if "body" not in self.result:
|
|
|
|
self.result["body"] = b""
|
|
|
|
|
|
|
|
self.result["body"] += content
|
|
|
|
|
2018-08-15 15:43:41 +02:00
|
|
|
def registerProducer(self, producer, streaming):
|
|
|
|
self._producer = producer
|
2018-11-06 17:00:00 +01:00
|
|
|
self.producerStreaming = streaming
|
|
|
|
|
|
|
|
def _produce():
|
|
|
|
if self._producer:
|
|
|
|
self._producer.resumeProducing()
|
|
|
|
self._reactor.callLater(0.1, _produce)
|
|
|
|
|
|
|
|
if not streaming:
|
|
|
|
self._reactor.callLater(0.0, _produce)
|
2018-08-15 15:43:41 +02:00
|
|
|
|
|
|
|
def unregisterProducer(self):
|
|
|
|
if self._producer is None:
|
|
|
|
return
|
|
|
|
|
|
|
|
self._producer = None
|
|
|
|
|
2018-06-27 11:37:24 +02:00
|
|
|
def requestDone(self, _self):
|
|
|
|
self.result["done"] = True
|
|
|
|
|
|
|
|
def getPeer(self):
|
2018-08-23 19:33:04 +02:00
|
|
|
# We give an address so that getClientIP returns a non null entry,
|
|
|
|
# causing us to record the MAU
|
2018-09-06 18:58:18 +02:00
|
|
|
return address.IPv4Address("TCP", "127.0.0.1", 3423)
|
2018-06-27 11:37:24 +02:00
|
|
|
|
|
|
|
def getHost(self):
|
|
|
|
return None
|
|
|
|
|
|
|
|
@property
|
|
|
|
def transport(self):
|
|
|
|
return self
|
|
|
|
|
|
|
|
|
|
|
|
class FakeSite:
|
|
|
|
"""
|
|
|
|
A fake Twisted Web Site, with mocks of the extra things that
|
|
|
|
Synapse adds.
|
|
|
|
"""
|
|
|
|
|
|
|
|
server_version_string = b"1"
|
|
|
|
site_tag = "test"
|
2019-03-20 19:00:02 +01:00
|
|
|
access_logger = logging.getLogger("synapse.access.http.fake")
|
2018-06-27 11:37:24 +02:00
|
|
|
|
|
|
|
|
2018-11-05 19:53:44 +01:00
|
|
|
def make_request(
|
2018-11-06 17:00:00 +01:00
|
|
|
reactor,
|
|
|
|
method,
|
|
|
|
path,
|
|
|
|
content=b"",
|
|
|
|
access_token=None,
|
|
|
|
request=SynapseRequest,
|
|
|
|
shorthand=True,
|
2019-03-04 11:05:39 +01:00
|
|
|
federation_auth_origin=None,
|
2020-09-10 12:45:12 +02:00
|
|
|
content_is_form=False,
|
2020-11-16 15:45:22 +01:00
|
|
|
custom_headers: Optional[
|
|
|
|
Iterable[Tuple[Union[bytes, str], Union[bytes, str]]]
|
|
|
|
] = None,
|
2018-11-05 19:53:44 +01:00
|
|
|
):
|
2018-06-27 11:37:24 +02:00
|
|
|
"""
|
|
|
|
Make a web request using the given method and path, feed it the
|
|
|
|
content, and return the Request and the Channel underneath.
|
2018-11-05 19:53:44 +01:00
|
|
|
|
|
|
|
Args:
|
|
|
|
method (bytes/unicode): The HTTP request method ("verb").
|
|
|
|
path (bytes/unicode): The HTTP path, suitably URL encoded (e.g.
|
|
|
|
escaped UTF-8 & spaces and such).
|
|
|
|
content (bytes or dict): The body of the request. JSON-encoded, if
|
|
|
|
a dict.
|
|
|
|
shorthand: Whether to try and be helpful and prefix the given URL
|
|
|
|
with the usual REST API path, if it doesn't contain it.
|
2019-03-04 11:05:39 +01:00
|
|
|
federation_auth_origin (bytes|None): if set to not-None, we will add a fake
|
|
|
|
Authorization header pretenting to be the given server name.
|
2020-09-10 12:45:12 +02:00
|
|
|
content_is_form: Whether the content is URL encoded form data. Adds the
|
|
|
|
'Content-Type': 'application/x-www-form-urlencoded' header.
|
2018-11-05 19:53:44 +01:00
|
|
|
|
2020-11-16 15:45:22 +01:00
|
|
|
custom_headers: (name, value) pairs to add as request headers
|
|
|
|
|
2018-11-05 19:53:44 +01:00
|
|
|
Returns:
|
2019-03-04 11:05:39 +01:00
|
|
|
Tuple[synapse.http.site.SynapseRequest, channel]
|
2018-06-27 11:37:24 +02:00
|
|
|
"""
|
2018-08-09 04:22:01 +02:00
|
|
|
if not isinstance(method, bytes):
|
2019-06-20 11:32:02 +02:00
|
|
|
method = method.encode("ascii")
|
2018-08-09 04:22:01 +02:00
|
|
|
|
|
|
|
if not isinstance(path, bytes):
|
2019-06-20 11:32:02 +02:00
|
|
|
path = path.encode("ascii")
|
2018-06-27 11:37:24 +02:00
|
|
|
|
2018-11-05 19:53:44 +01:00
|
|
|
# Decorate it to be the full path, if we're using shorthand
|
2019-10-31 12:30:25 +01:00
|
|
|
if (
|
|
|
|
shorthand
|
|
|
|
and not path.startswith(b"/_matrix")
|
|
|
|
and not path.startswith(b"/_synapse")
|
|
|
|
):
|
2018-07-17 12:43:18 +02:00
|
|
|
path = b"/_matrix/client/r0/" + path
|
2018-08-09 04:22:01 +02:00
|
|
|
path = path.replace(b"//", b"/")
|
2018-07-17 12:43:18 +02:00
|
|
|
|
2018-11-15 22:55:58 +01:00
|
|
|
if not path.startswith(b"/"):
|
|
|
|
path = b"/" + path
|
|
|
|
|
2020-06-16 14:51:47 +02:00
|
|
|
if isinstance(content, str):
|
2019-06-20 11:32:02 +02:00
|
|
|
content = content.encode("utf8")
|
2018-06-27 11:37:24 +02:00
|
|
|
|
|
|
|
site = FakeSite()
|
2020-01-03 15:19:48 +01:00
|
|
|
channel = FakeChannel(site, reactor)
|
2018-06-27 11:37:24 +02:00
|
|
|
|
2020-01-03 15:19:48 +01:00
|
|
|
req = request(channel)
|
2018-06-27 11:37:24 +02:00
|
|
|
req.process = lambda: b""
|
|
|
|
req.content = BytesIO(content)
|
2020-09-10 12:45:12 +02:00
|
|
|
# Twisted expects to be at the end of the content when parsing the request.
|
|
|
|
req.content.seek(SEEK_END)
|
2019-06-20 11:32:02 +02:00
|
|
|
req.postpath = list(map(unquote, path[1:].split(b"/")))
|
2018-08-23 19:33:04 +02:00
|
|
|
|
|
|
|
if access_token:
|
2018-10-30 13:55:43 +01:00
|
|
|
req.requestHeaders.addRawHeader(
|
2019-06-20 11:32:02 +02:00
|
|
|
b"Authorization", b"Bearer " + access_token.encode("ascii")
|
2018-10-30 13:55:43 +01:00
|
|
|
)
|
2018-08-23 19:33:04 +02:00
|
|
|
|
2019-03-04 11:05:39 +01:00
|
|
|
if federation_auth_origin is not None:
|
|
|
|
req.requestHeaders.addRawHeader(
|
2019-05-10 07:12:11 +02:00
|
|
|
b"Authorization",
|
|
|
|
b"X-Matrix origin=%s,key=,sig=" % (federation_auth_origin,),
|
2019-03-04 11:05:39 +01:00
|
|
|
)
|
|
|
|
|
2018-09-20 12:14:34 +02:00
|
|
|
if content:
|
2020-09-10 12:45:12 +02:00
|
|
|
if content_is_form:
|
|
|
|
req.requestHeaders.addRawHeader(
|
|
|
|
b"Content-Type", b"application/x-www-form-urlencoded"
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
# Assume the body is JSON
|
|
|
|
req.requestHeaders.addRawHeader(b"Content-Type", b"application/json")
|
2018-09-20 12:14:34 +02:00
|
|
|
|
2020-11-16 15:45:22 +01:00
|
|
|
if custom_headers:
|
|
|
|
for k, v in custom_headers:
|
|
|
|
req.requestHeaders.addRawHeader(k, v)
|
|
|
|
|
2018-06-27 11:37:24 +02:00
|
|
|
req.requestReceived(method, path, b"1.1")
|
|
|
|
|
|
|
|
return req, channel
|
|
|
|
|
|
|
|
|
2018-08-15 15:43:41 +02:00
|
|
|
def wait_until_result(clock, request, timeout=100):
|
2018-06-27 11:37:24 +02:00
|
|
|
"""
|
2018-08-15 15:43:41 +02:00
|
|
|
Wait until the request is finished.
|
2018-06-27 11:37:24 +02:00
|
|
|
"""
|
|
|
|
clock.run()
|
|
|
|
x = 0
|
|
|
|
|
2018-08-15 15:43:41 +02:00
|
|
|
while not request.finished:
|
|
|
|
|
|
|
|
# If there's a producer, tell it to resume producing so we get content
|
|
|
|
if request._channel._producer:
|
|
|
|
request._channel._producer.resumeProducing()
|
|
|
|
|
2018-06-27 11:37:24 +02:00
|
|
|
x += 1
|
|
|
|
|
|
|
|
if x > timeout:
|
2018-11-02 14:19:23 +01:00
|
|
|
raise TimedOutException("Timed out waiting for request to finish.")
|
2018-06-27 11:37:24 +02:00
|
|
|
|
|
|
|
clock.advance(0.1)
|
|
|
|
|
|
|
|
|
2018-07-17 12:43:18 +02:00
|
|
|
def render(request, resource, clock):
|
|
|
|
request.render(resource)
|
2018-08-15 15:43:41 +02:00
|
|
|
wait_until_result(clock, request)
|
2018-07-17 12:43:18 +02:00
|
|
|
|
|
|
|
|
2018-09-13 16:15:51 +02:00
|
|
|
@implementer(IReactorPluggableNameResolver)
|
2018-06-27 11:37:24 +02:00
|
|
|
class ThreadedMemoryReactorClock(MemoryReactorClock):
|
|
|
|
"""
|
|
|
|
A MemoryReactorClock that supports callFromThread.
|
|
|
|
"""
|
2018-08-10 15:54:09 +02:00
|
|
|
|
2018-09-13 16:15:51 +02:00
|
|
|
def __init__(self):
|
2019-05-13 22:01:14 +02:00
|
|
|
self.threadpool = ThreadPool(self)
|
|
|
|
|
2020-07-15 16:27:35 +02:00
|
|
|
self._tcp_callbacks = {}
|
2018-09-13 16:15:51 +02:00
|
|
|
self._udp = []
|
2019-01-29 10:38:29 +01:00
|
|
|
lookups = self.lookups = {}
|
Fix threadsafety in ThreadedMemoryReactorClock (#8497)
This could, very occasionally, cause:
```
tests.test_visibility.FilterEventsForServerTestCase.test_large_room
===============================================================================
[ERROR]
Traceback (most recent call last):
File "/src/tests/rest/media/v1/test_media_storage.py", line 86, in test_ensure_media_is_in_local_cache
self.wait_on_thread(x)
File "/src/tests/unittest.py", line 296, in wait_on_thread
self.reactor.advance(0.01)
File "/src/.tox/py35/lib/python3.5/site-packages/twisted/internet/task.py", line 826, in advance
self._sortCalls()
File "/src/.tox/py35/lib/python3.5/site-packages/twisted/internet/task.py", line 787, in _sortCalls
self.calls.sort(key=lambda a: a.getTime())
builtins.ValueError: list modified during sort
tests.rest.media.v1.test_media_storage.MediaStorageTests.test_ensure_media_is_in_local_cache
```
2020-10-09 18:22:25 +02:00
|
|
|
self._thread_callbacks = deque() # type: Deque[Callable[[], None]]()
|
2019-01-29 10:38:29 +01:00
|
|
|
|
|
|
|
@implementer(IResolverSimple)
|
2020-09-04 12:54:56 +02:00
|
|
|
class FakeResolver:
|
2019-01-29 10:38:29 +01:00
|
|
|
def getHostByName(self, name, timeout=None):
|
|
|
|
if name not in lookups:
|
2019-05-10 07:12:11 +02:00
|
|
|
return fail(DNSLookupError("OH NO: unknown %s" % (name,)))
|
2019-01-29 10:38:29 +01:00
|
|
|
return succeed(lookups[name])
|
|
|
|
|
|
|
|
self.nameResolver = SimpleResolverComplexifier(FakeResolver())
|
2020-09-18 15:56:44 +02:00
|
|
|
super().__init__()
|
2018-09-13 16:15:51 +02:00
|
|
|
|
2019-06-20 11:32:02 +02:00
|
|
|
def listenUDP(self, port, protocol, interface="", maxPacketSize=8196):
|
2018-09-13 16:15:51 +02:00
|
|
|
p = udp.Port(port, protocol, interface, maxPacketSize, self)
|
|
|
|
p.startListening()
|
|
|
|
self._udp.append(p)
|
|
|
|
return p
|
|
|
|
|
2018-06-27 11:37:24 +02:00
|
|
|
def callFromThread(self, callback, *args, **kwargs):
|
|
|
|
"""
|
|
|
|
Make the callback fire in the next reactor iteration.
|
|
|
|
"""
|
Fix threadsafety in ThreadedMemoryReactorClock (#8497)
This could, very occasionally, cause:
```
tests.test_visibility.FilterEventsForServerTestCase.test_large_room
===============================================================================
[ERROR]
Traceback (most recent call last):
File "/src/tests/rest/media/v1/test_media_storage.py", line 86, in test_ensure_media_is_in_local_cache
self.wait_on_thread(x)
File "/src/tests/unittest.py", line 296, in wait_on_thread
self.reactor.advance(0.01)
File "/src/.tox/py35/lib/python3.5/site-packages/twisted/internet/task.py", line 826, in advance
self._sortCalls()
File "/src/.tox/py35/lib/python3.5/site-packages/twisted/internet/task.py", line 787, in _sortCalls
self.calls.sort(key=lambda a: a.getTime())
builtins.ValueError: list modified during sort
tests.rest.media.v1.test_media_storage.MediaStorageTests.test_ensure_media_is_in_local_cache
```
2020-10-09 18:22:25 +02:00
|
|
|
cb = lambda: callback(*args, **kwargs)
|
|
|
|
# it's not safe to call callLater() here, so we append the callback to a
|
|
|
|
# separate queue.
|
|
|
|
self._thread_callbacks.append(cb)
|
2018-06-27 11:37:24 +02:00
|
|
|
|
2019-05-13 22:01:14 +02:00
|
|
|
def getThreadPool(self):
|
|
|
|
return self.threadpool
|
|
|
|
|
2020-07-15 16:27:35 +02:00
|
|
|
def add_tcp_client_callback(self, host, port, callback):
|
|
|
|
"""Add a callback that will be invoked when we receive a connection
|
|
|
|
attempt to the given IP/port using `connectTCP`.
|
|
|
|
|
|
|
|
Note that the callback gets run before we return the connection to the
|
|
|
|
client, which means callbacks cannot block while waiting for writes.
|
|
|
|
"""
|
|
|
|
self._tcp_callbacks[(host, port)] = callback
|
|
|
|
|
|
|
|
def connectTCP(self, host, port, factory, timeout=30, bindAddress=None):
|
|
|
|
"""Fake L{IReactorTCP.connectTCP}.
|
|
|
|
"""
|
|
|
|
|
|
|
|
conn = super().connectTCP(
|
|
|
|
host, port, factory, timeout=timeout, bindAddress=None
|
|
|
|
)
|
|
|
|
|
|
|
|
callback = self._tcp_callbacks.get((host, port))
|
|
|
|
if callback:
|
|
|
|
callback()
|
|
|
|
|
|
|
|
return conn
|
|
|
|
|
Fix threadsafety in ThreadedMemoryReactorClock (#8497)
This could, very occasionally, cause:
```
tests.test_visibility.FilterEventsForServerTestCase.test_large_room
===============================================================================
[ERROR]
Traceback (most recent call last):
File "/src/tests/rest/media/v1/test_media_storage.py", line 86, in test_ensure_media_is_in_local_cache
self.wait_on_thread(x)
File "/src/tests/unittest.py", line 296, in wait_on_thread
self.reactor.advance(0.01)
File "/src/.tox/py35/lib/python3.5/site-packages/twisted/internet/task.py", line 826, in advance
self._sortCalls()
File "/src/.tox/py35/lib/python3.5/site-packages/twisted/internet/task.py", line 787, in _sortCalls
self.calls.sort(key=lambda a: a.getTime())
builtins.ValueError: list modified during sort
tests.rest.media.v1.test_media_storage.MediaStorageTests.test_ensure_media_is_in_local_cache
```
2020-10-09 18:22:25 +02:00
|
|
|
def advance(self, amount):
|
|
|
|
# first advance our reactor's time, and run any "callLater" callbacks that
|
|
|
|
# makes ready
|
|
|
|
super().advance(amount)
|
|
|
|
|
|
|
|
# now run any "callFromThread" callbacks
|
|
|
|
while True:
|
|
|
|
try:
|
|
|
|
callback = self._thread_callbacks.popleft()
|
|
|
|
except IndexError:
|
|
|
|
break
|
|
|
|
callback()
|
|
|
|
|
|
|
|
# check for more "callLater" callbacks added by the thread callback
|
|
|
|
# This isn't required in a regular reactor, but it ends up meaning that
|
|
|
|
# our database queries can complete in a single call to `advance` [1] which
|
|
|
|
# simplifies tests.
|
|
|
|
#
|
|
|
|
# [1]: we replace the threadpool backing the db connection pool with a
|
|
|
|
# mock ThreadPool which doesn't really use threads; but we still use
|
|
|
|
# reactor.callFromThread to feed results back from the db functions to the
|
|
|
|
# main thread.
|
|
|
|
super().advance(0)
|
|
|
|
|
2019-05-13 22:01:14 +02:00
|
|
|
|
|
|
|
class ThreadPool:
|
|
|
|
"""
|
|
|
|
Threadless thread pool.
|
|
|
|
"""
|
|
|
|
|
|
|
|
def __init__(self, reactor):
|
|
|
|
self._reactor = reactor
|
|
|
|
|
|
|
|
def start(self):
|
|
|
|
pass
|
|
|
|
|
|
|
|
def stop(self):
|
|
|
|
pass
|
|
|
|
|
|
|
|
def callInThreadWithCallback(self, onResult, function, *args, **kwargs):
|
|
|
|
def _(res):
|
|
|
|
if isinstance(res, Failure):
|
|
|
|
onResult(False, res)
|
|
|
|
else:
|
|
|
|
onResult(True, res)
|
|
|
|
|
|
|
|
d = Deferred()
|
|
|
|
d.addCallback(lambda x: function(*args, **kwargs))
|
|
|
|
d.addBoth(_)
|
|
|
|
self._reactor.callLater(0, d.callback, True)
|
|
|
|
return d
|
|
|
|
|
2018-06-27 11:37:24 +02:00
|
|
|
|
2018-08-13 08:47:46 +02:00
|
|
|
def setup_test_homeserver(cleanup_func, *args, **kwargs):
|
2018-06-27 11:37:24 +02:00
|
|
|
"""
|
|
|
|
Set up a synchronous test server, driven by the reactor used by
|
|
|
|
the homeserver.
|
|
|
|
"""
|
2019-12-18 11:45:12 +01:00
|
|
|
server = _sth(cleanup_func, *args, **kwargs)
|
2018-08-13 08:47:46 +02:00
|
|
|
|
2018-06-27 11:37:24 +02:00
|
|
|
# Make the thread pool synchronous.
|
2019-12-18 11:45:12 +01:00
|
|
|
clock = server.get_clock()
|
|
|
|
|
|
|
|
for database in server.get_datastores().databases:
|
|
|
|
pool = database._db_pool
|
|
|
|
|
|
|
|
def runWithConnection(func, *args, **kwargs):
|
|
|
|
return threads.deferToThreadPool(
|
|
|
|
pool._reactor,
|
|
|
|
pool.threadpool,
|
|
|
|
pool._runWithConnection,
|
|
|
|
func,
|
|
|
|
*args,
|
2020-10-28 00:26:36 +01:00
|
|
|
**kwargs,
|
2019-12-18 11:45:12 +01:00
|
|
|
)
|
|
|
|
|
|
|
|
def runInteraction(interaction, *args, **kwargs):
|
|
|
|
return threads.deferToThreadPool(
|
|
|
|
pool._reactor,
|
|
|
|
pool.threadpool,
|
|
|
|
pool._runInteraction,
|
|
|
|
interaction,
|
|
|
|
*args,
|
2020-10-28 00:26:36 +01:00
|
|
|
**kwargs,
|
2019-12-18 11:45:12 +01:00
|
|
|
)
|
2018-06-27 11:37:24 +02:00
|
|
|
|
2019-03-04 11:05:39 +01:00
|
|
|
pool.runWithConnection = runWithConnection
|
|
|
|
pool.runInteraction = runInteraction
|
2019-05-13 22:01:14 +02:00
|
|
|
pool.threadpool = ThreadPool(clock._reactor)
|
2019-03-04 11:05:39 +01:00
|
|
|
pool.running = True
|
2019-12-18 11:45:12 +01:00
|
|
|
|
2020-10-02 16:09:31 +02:00
|
|
|
# We've just changed the Databases to run DB transactions on the same
|
|
|
|
# thread, so we need to disable the dedicated thread behaviour.
|
|
|
|
server.get_datastores().main.USE_DEDICATED_DB_THREADS_FOR_EVENT_FETCHING = False
|
|
|
|
|
2019-12-18 11:45:12 +01:00
|
|
|
return server
|
2018-08-09 04:22:01 +02:00
|
|
|
|
|
|
|
|
|
|
|
def get_clock():
|
|
|
|
clock = ThreadedMemoryReactorClock()
|
|
|
|
hs_clock = Clock(clock)
|
2019-08-30 17:28:26 +02:00
|
|
|
return clock, hs_clock
|
2018-09-18 19:17:15 +02:00
|
|
|
|
|
|
|
|
2019-01-23 12:25:36 +01:00
|
|
|
@attr.s(cmp=False)
|
2020-09-04 12:54:56 +02:00
|
|
|
class FakeTransport:
|
2018-09-18 19:17:15 +02:00
|
|
|
"""
|
|
|
|
A twisted.internet.interfaces.ITransport implementation which sends all its data
|
|
|
|
straight into an IProtocol object: it exists to connect two IProtocols together.
|
|
|
|
|
|
|
|
To use it, instantiate it with the receiving IProtocol, and then pass it to the
|
|
|
|
sending IProtocol's makeConnection method:
|
|
|
|
|
|
|
|
server = HTTPChannel()
|
|
|
|
client.makeConnection(FakeTransport(server, self.reactor))
|
|
|
|
|
|
|
|
If you want bidirectional communication, you'll need two instances.
|
|
|
|
"""
|
|
|
|
|
|
|
|
other = attr.ib()
|
|
|
|
"""The Protocol object which will receive any data written to this transport.
|
|
|
|
|
|
|
|
:type: twisted.internet.interfaces.IProtocol
|
|
|
|
"""
|
|
|
|
|
|
|
|
_reactor = attr.ib()
|
|
|
|
"""Test reactor
|
|
|
|
|
|
|
|
:type: twisted.internet.interfaces.IReactorTime
|
|
|
|
"""
|
|
|
|
|
2019-01-29 14:53:02 +01:00
|
|
|
_protocol = attr.ib(default=None)
|
|
|
|
"""The Protocol which is producing data for this transport. Optional, but if set
|
|
|
|
will get called back for connectionLost() notifications etc.
|
|
|
|
"""
|
|
|
|
|
2018-09-18 19:17:15 +02:00
|
|
|
disconnecting = False
|
2019-01-30 11:55:25 +01:00
|
|
|
disconnected = False
|
2019-11-25 17:45:50 +01:00
|
|
|
connected = True
|
2019-06-20 11:32:02 +02:00
|
|
|
buffer = attr.ib(default=b"")
|
2018-09-18 19:17:15 +02:00
|
|
|
producer = attr.ib(default=None)
|
2019-04-02 13:42:39 +02:00
|
|
|
autoflush = attr.ib(default=True)
|
2018-09-18 19:17:15 +02:00
|
|
|
|
|
|
|
def getPeer(self):
|
|
|
|
return None
|
|
|
|
|
|
|
|
def getHost(self):
|
|
|
|
return None
|
|
|
|
|
2019-01-29 14:53:02 +01:00
|
|
|
def loseConnection(self, reason=None):
|
|
|
|
if not self.disconnecting:
|
2019-01-30 11:55:25 +01:00
|
|
|
logger.info("FakeTransport: loseConnection(%s)", reason)
|
2019-01-29 14:53:02 +01:00
|
|
|
self.disconnecting = True
|
|
|
|
if self._protocol:
|
|
|
|
self._protocol.connectionLost(reason)
|
2019-11-01 15:07:44 +01:00
|
|
|
|
|
|
|
# if we still have data to write, delay until that is done
|
|
|
|
if self.buffer:
|
|
|
|
logger.info(
|
|
|
|
"FakeTransport: Delaying disconnect until buffer is flushed"
|
|
|
|
)
|
|
|
|
else:
|
2019-11-25 17:45:50 +01:00
|
|
|
self.connected = False
|
2019-11-01 15:07:44 +01:00
|
|
|
self.disconnected = True
|
2018-09-18 19:17:15 +02:00
|
|
|
|
|
|
|
def abortConnection(self):
|
2019-01-30 11:55:25 +01:00
|
|
|
logger.info("FakeTransport: abortConnection()")
|
2019-11-01 15:07:44 +01:00
|
|
|
|
|
|
|
if not self.disconnecting:
|
|
|
|
self.disconnecting = True
|
|
|
|
if self._protocol:
|
|
|
|
self._protocol.connectionLost(None)
|
|
|
|
|
|
|
|
self.disconnected = True
|
2018-09-18 19:17:15 +02:00
|
|
|
|
|
|
|
def pauseProducing(self):
|
2018-12-21 15:56:13 +01:00
|
|
|
if not self.producer:
|
|
|
|
return
|
|
|
|
|
2018-09-18 19:17:15 +02:00
|
|
|
self.producer.pauseProducing()
|
|
|
|
|
2018-12-21 15:56:13 +01:00
|
|
|
def resumeProducing(self):
|
|
|
|
if not self.producer:
|
|
|
|
return
|
|
|
|
self.producer.resumeProducing()
|
|
|
|
|
2018-09-18 19:17:15 +02:00
|
|
|
def unregisterProducer(self):
|
|
|
|
if not self.producer:
|
|
|
|
return
|
|
|
|
|
|
|
|
self.producer = None
|
|
|
|
|
|
|
|
def registerProducer(self, producer, streaming):
|
|
|
|
self.producer = producer
|
|
|
|
self.producerStreaming = streaming
|
|
|
|
|
|
|
|
def _produce():
|
|
|
|
d = self.producer.resumeProducing()
|
|
|
|
d.addCallback(lambda x: self._reactor.callLater(0.1, _produce))
|
|
|
|
|
|
|
|
if not streaming:
|
|
|
|
self._reactor.callLater(0.0, _produce)
|
|
|
|
|
|
|
|
def write(self, byt):
|
2019-11-01 15:07:44 +01:00
|
|
|
if self.disconnecting:
|
|
|
|
raise Exception("Writing to disconnecting FakeTransport")
|
|
|
|
|
2018-09-18 19:17:15 +02:00
|
|
|
self.buffer = self.buffer + byt
|
|
|
|
|
2019-01-22 21:28:48 +01:00
|
|
|
# always actually do the write asynchronously. Some protocols (notably the
|
|
|
|
# TLSMemoryBIOProtocol) get very confused if a read comes back while they are
|
|
|
|
# still doing a write. Doing a callLater here breaks the cycle.
|
2019-04-02 13:42:39 +02:00
|
|
|
if self.autoflush:
|
|
|
|
self._reactor.callLater(0.0, self.flush)
|
2018-09-18 19:17:15 +02:00
|
|
|
|
|
|
|
def writeSequence(self, seq):
|
|
|
|
for x in seq:
|
|
|
|
self.write(x)
|
2019-04-02 13:42:39 +02:00
|
|
|
|
|
|
|
def flush(self, maxbytes=None):
|
|
|
|
if not self.buffer:
|
|
|
|
# nothing to do. Don't write empty buffers: it upsets the
|
|
|
|
# TLSMemoryBIOProtocol
|
|
|
|
return
|
|
|
|
|
|
|
|
if self.disconnected:
|
|
|
|
return
|
|
|
|
|
|
|
|
if getattr(self.other, "transport") is None:
|
|
|
|
# the other has no transport yet; reschedule
|
|
|
|
if self.autoflush:
|
|
|
|
self._reactor.callLater(0.0, self.flush)
|
|
|
|
return
|
|
|
|
|
|
|
|
if maxbytes is not None:
|
|
|
|
to_write = self.buffer[:maxbytes]
|
|
|
|
else:
|
|
|
|
to_write = self.buffer
|
|
|
|
|
|
|
|
logger.info("%s->%s: %s", self._protocol, self.other, to_write)
|
|
|
|
|
|
|
|
try:
|
|
|
|
self.other.dataReceived(to_write)
|
|
|
|
except Exception as e:
|
2020-07-15 16:27:35 +02:00
|
|
|
logger.exception("Exception writing to protocol: %s", e)
|
2019-04-02 13:42:39 +02:00
|
|
|
return
|
|
|
|
|
2019-05-10 07:12:11 +02:00
|
|
|
self.buffer = self.buffer[len(to_write) :]
|
2019-04-02 13:42:39 +02:00
|
|
|
if self.buffer and self.autoflush:
|
|
|
|
self._reactor.callLater(0.0, self.flush)
|
2019-08-28 13:18:53 +02:00
|
|
|
|
2019-11-01 15:07:44 +01:00
|
|
|
if not self.buffer and self.disconnecting:
|
|
|
|
logger.info("FakeTransport: Buffer now empty, completing disconnect")
|
|
|
|
self.disconnected = True
|
|
|
|
|
2019-08-28 13:18:53 +02:00
|
|
|
|
|
|
|
def connect_client(reactor: IReactorTCP, client_id: int) -> AccumulatingProtocol:
|
|
|
|
"""
|
|
|
|
Connect a client to a fake TCP transport.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
reactor
|
|
|
|
factory: The connecting factory to build.
|
|
|
|
"""
|
2020-10-29 12:27:37 +01:00
|
|
|
factory = reactor.tcpClients.pop(client_id)[2]
|
2019-08-28 13:18:53 +02:00
|
|
|
client = factory.buildProtocol(None)
|
|
|
|
server = AccumulatingProtocol()
|
|
|
|
server.makeConnection(FakeTransport(client, reactor))
|
|
|
|
client.makeConnection(FakeTransport(server, reactor))
|
|
|
|
|
|
|
|
return client, server
|