mirror of
https://mau.dev/maunium/synapse.git
synced 2024-12-13 11:43:25 +01:00
Revert "Reduce the size of the HTTP connection pool for non-pushers" (#15530)
#15514 introduced a regression where Synapse would encounter `PartialDownloadError`s when fetching OpenID metadata for certain providers on startup. Due to #8088, this prevents Synapse from starting entirely. Revert the change while we decide what to do about the regression.
This commit is contained in:
parent
1c0e98717b
commit
3b837d856c
6 changed files with 16 additions and 31 deletions
|
@ -48,7 +48,6 @@ Internal Changes
|
||||||
- Bump packaging from 23.0 to 23.1. ([\#15510](https://github.com/matrix-org/synapse/issues/15510))
|
- Bump packaging from 23.0 to 23.1. ([\#15510](https://github.com/matrix-org/synapse/issues/15510))
|
||||||
- Bump types-requests from 2.28.11.16 to 2.29.0.0. ([\#15511](https://github.com/matrix-org/synapse/issues/15511))
|
- Bump types-requests from 2.28.11.16 to 2.29.0.0. ([\#15511](https://github.com/matrix-org/synapse/issues/15511))
|
||||||
- Bump setuptools-rust from 1.5.2 to 1.6.0. ([\#15512](https://github.com/matrix-org/synapse/issues/15512))
|
- Bump setuptools-rust from 1.5.2 to 1.6.0. ([\#15512](https://github.com/matrix-org/synapse/issues/15512))
|
||||||
- Reduce the size of the HTTP connection pool for non-pushers. ([\#15514](https://github.com/matrix-org/synapse/issues/15514))
|
|
||||||
- Update the check_schema_delta script to account for when the schema version has been bumped locally. ([\#15466](https://github.com/matrix-org/synapse/issues/15466))
|
- Update the check_schema_delta script to account for when the schema version has been bumped locally. ([\#15466](https://github.com/matrix-org/synapse/issues/15466))
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -768,7 +768,6 @@ class SimpleHttpClient(BaseHttpClient):
|
||||||
request if it were otherwise caught in a blacklist.
|
request if it were otherwise caught in a blacklist.
|
||||||
use_proxy: Whether proxy settings should be discovered and used
|
use_proxy: Whether proxy settings should be discovered and used
|
||||||
from conventional environment variables.
|
from conventional environment variables.
|
||||||
connection_pool: The connection pool to use for this client's agent.
|
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(
|
def __init__(
|
||||||
|
@ -778,7 +777,6 @@ class SimpleHttpClient(BaseHttpClient):
|
||||||
ip_whitelist: Optional[IPSet] = None,
|
ip_whitelist: Optional[IPSet] = None,
|
||||||
ip_blacklist: Optional[IPSet] = None,
|
ip_blacklist: Optional[IPSet] = None,
|
||||||
use_proxy: bool = False,
|
use_proxy: bool = False,
|
||||||
connection_pool: Optional[HTTPConnectionPool] = None,
|
|
||||||
):
|
):
|
||||||
super().__init__(hs, treq_args=treq_args)
|
super().__init__(hs, treq_args=treq_args)
|
||||||
self._ip_whitelist = ip_whitelist
|
self._ip_whitelist = ip_whitelist
|
||||||
|
@ -791,12 +789,22 @@ class SimpleHttpClient(BaseHttpClient):
|
||||||
self.reactor, self._ip_whitelist, self._ip_blacklist
|
self.reactor, self._ip_whitelist, self._ip_blacklist
|
||||||
)
|
)
|
||||||
|
|
||||||
|
# the pusher makes lots of concurrent SSL connections to Sygnal, and tends to
|
||||||
|
# do so in batches, so we need to allow the pool to keep lots of idle
|
||||||
|
# connections around.
|
||||||
|
pool = HTTPConnectionPool(self.reactor)
|
||||||
|
# XXX: The justification for using the cache factor here is that larger
|
||||||
|
# instances will need both more cache and more connections.
|
||||||
|
# Still, this should probably be a separate dial
|
||||||
|
pool.maxPersistentPerHost = max(int(100 * hs.config.caches.global_factor), 5)
|
||||||
|
pool.cachedConnectionTimeout = 2 * 60
|
||||||
|
|
||||||
self.agent: IAgent = ProxyAgent(
|
self.agent: IAgent = ProxyAgent(
|
||||||
self.reactor,
|
self.reactor,
|
||||||
hs.get_reactor(),
|
hs.get_reactor(),
|
||||||
connectTimeout=15,
|
connectTimeout=15,
|
||||||
contextFactory=self.hs.get_http_client_context_factory(),
|
contextFactory=self.hs.get_http_client_context_factory(),
|
||||||
pool=connection_pool,
|
pool=pool,
|
||||||
use_proxy=use_proxy,
|
use_proxy=use_proxy,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
|
@ -140,8 +140,7 @@ class HttpPusher(Pusher):
|
||||||
)
|
)
|
||||||
|
|
||||||
self.url = url
|
self.url = url
|
||||||
self.http_client = hs.get_pusher_http_client()
|
self.http_client = hs.get_proxied_blacklisted_http_client()
|
||||||
|
|
||||||
self.data_minus_url = {}
|
self.data_minus_url = {}
|
||||||
self.data_minus_url.update(self.data)
|
self.data_minus_url.update(self.data)
|
||||||
del self.data_minus_url["url"]
|
del self.data_minus_url["url"]
|
||||||
|
|
|
@ -27,7 +27,6 @@ from typing_extensions import TypeAlias
|
||||||
|
|
||||||
from twisted.internet.interfaces import IOpenSSLContextFactory
|
from twisted.internet.interfaces import IOpenSSLContextFactory
|
||||||
from twisted.internet.tcp import Port
|
from twisted.internet.tcp import Port
|
||||||
from twisted.web.client import HTTPConnectionPool
|
|
||||||
from twisted.web.iweb import IPolicyForHTTPS
|
from twisted.web.iweb import IPolicyForHTTPS
|
||||||
from twisted.web.resource import Resource
|
from twisted.web.resource import Resource
|
||||||
|
|
||||||
|
@ -454,26 +453,6 @@ class HomeServer(metaclass=abc.ABCMeta):
|
||||||
use_proxy=True,
|
use_proxy=True,
|
||||||
)
|
)
|
||||||
|
|
||||||
@cache_in_self
|
|
||||||
def get_pusher_http_client(self) -> SimpleHttpClient:
|
|
||||||
# the pusher makes lots of concurrent SSL connections to Sygnal, and tends to
|
|
||||||
# do so in batches, so we need to allow the pool to keep lots of idle
|
|
||||||
# connections around.
|
|
||||||
pool = HTTPConnectionPool(self.get_reactor())
|
|
||||||
# XXX: The justification for using the cache factor here is that larger
|
|
||||||
# instances will need both more cache and more connections.
|
|
||||||
# Still, this should probably be a separate dial
|
|
||||||
pool.maxPersistentPerHost = max(int(100 * self.config.caches.global_factor), 5)
|
|
||||||
pool.cachedConnectionTimeout = 2 * 60
|
|
||||||
|
|
||||||
return SimpleHttpClient(
|
|
||||||
self,
|
|
||||||
ip_whitelist=self.config.server.ip_range_whitelist,
|
|
||||||
ip_blacklist=self.config.server.ip_range_blacklist,
|
|
||||||
use_proxy=True,
|
|
||||||
connection_pool=pool,
|
|
||||||
)
|
|
||||||
|
|
||||||
@cache_in_self
|
@cache_in_self
|
||||||
def get_federation_http_client(self) -> MatrixFederationHttpClient:
|
def get_federation_http_client(self) -> MatrixFederationHttpClient:
|
||||||
"""
|
"""
|
||||||
|
|
|
@ -52,7 +52,7 @@ class HTTPPusherTests(HomeserverTestCase):
|
||||||
|
|
||||||
m.post_json_get_json = post_json_get_json
|
m.post_json_get_json = post_json_get_json
|
||||||
|
|
||||||
hs = self.setup_test_homeserver(pusher_http_client=m)
|
hs = self.setup_test_homeserver(proxied_blacklisted_http_client=m)
|
||||||
|
|
||||||
return hs
|
return hs
|
||||||
|
|
||||||
|
|
|
@ -93,7 +93,7 @@ class PusherShardTestCase(BaseMultiWorkerStreamTestCase):
|
||||||
self.make_worker_hs(
|
self.make_worker_hs(
|
||||||
"synapse.app.generic_worker",
|
"synapse.app.generic_worker",
|
||||||
{"worker_name": "pusher1", "pusher_instances": ["pusher1"]},
|
{"worker_name": "pusher1", "pusher_instances": ["pusher1"]},
|
||||||
pusher_http_client=http_client_mock,
|
proxied_blacklisted_http_client=http_client_mock,
|
||||||
)
|
)
|
||||||
|
|
||||||
event_id = self._create_pusher_and_send_msg("user")
|
event_id = self._create_pusher_and_send_msg("user")
|
||||||
|
@ -126,7 +126,7 @@ class PusherShardTestCase(BaseMultiWorkerStreamTestCase):
|
||||||
"worker_name": "pusher1",
|
"worker_name": "pusher1",
|
||||||
"pusher_instances": ["pusher1", "pusher2"],
|
"pusher_instances": ["pusher1", "pusher2"],
|
||||||
},
|
},
|
||||||
pusher_http_client=http_client_mock1,
|
proxied_blacklisted_http_client=http_client_mock1,
|
||||||
)
|
)
|
||||||
|
|
||||||
http_client_mock2 = Mock(spec_set=["post_json_get_json"])
|
http_client_mock2 = Mock(spec_set=["post_json_get_json"])
|
||||||
|
@ -140,7 +140,7 @@ class PusherShardTestCase(BaseMultiWorkerStreamTestCase):
|
||||||
"worker_name": "pusher2",
|
"worker_name": "pusher2",
|
||||||
"pusher_instances": ["pusher1", "pusher2"],
|
"pusher_instances": ["pusher1", "pusher2"],
|
||||||
},
|
},
|
||||||
pusher_http_client=http_client_mock2,
|
proxied_blacklisted_http_client=http_client_mock2,
|
||||||
)
|
)
|
||||||
|
|
||||||
# We choose a user name that we know should go to pusher1.
|
# We choose a user name that we know should go to pusher1.
|
||||||
|
|
Loading…
Reference in a new issue