0
0
Fork 1
mirror of https://mau.dev/maunium/synapse.git synced 2024-12-15 13:33:54 +01:00

Revert "Fix #8518 (sync requests being cached wrongly on timeout) (#9358)"

This reverts commit f5c93fc993.

This is being backed out due to a regression (#9507) and additional
review feedback being provided.
This commit is contained in:
Patrick Cloke 2021-03-02 09:43:34 -05:00
parent 7f5d753d06
commit aee10768d8
3 changed files with 3 additions and 35 deletions

View file

@ -1 +0,0 @@
Added a fix that invalidates cache for empty timed-out sync responses.

View file

@ -277,9 +277,8 @@ class SyncHandler:
user_id = sync_config.user.to_string() user_id = sync_config.user.to_string()
await self.auth.check_auth_blocking(requester=requester) await self.auth.check_auth_blocking(requester=requester)
res = await self.response_cache.wrap_conditional( res = await self.response_cache.wrap(
sync_config.request_key, sync_config.request_key,
lambda result: since_token != result.next_batch,
self._wait_for_sync_for_user, self._wait_for_sync_for_user,
sync_config, sync_config,
since_token, since_token,

View file

@ -13,7 +13,7 @@
# See the License for the specific language governing permissions and # See the License for the specific language governing permissions and
# limitations under the License. # limitations under the License.
import logging import logging
from typing import TYPE_CHECKING, Any, Callable, Dict, Generic, Optional, Set, TypeVar from typing import TYPE_CHECKING, Any, Callable, Dict, Generic, Optional, TypeVar
from twisted.internet import defer from twisted.internet import defer
@ -40,7 +40,6 @@ class ResponseCache(Generic[T]):
def __init__(self, hs: "HomeServer", name: str, timeout_ms: float = 0): def __init__(self, hs: "HomeServer", name: str, timeout_ms: float = 0):
# Requests that haven't finished yet. # Requests that haven't finished yet.
self.pending_result_cache = {} # type: Dict[T, ObservableDeferred] self.pending_result_cache = {} # type: Dict[T, ObservableDeferred]
self.pending_conditionals = {} # type: Dict[T, Set[Callable[[Any], bool]]]
self.clock = hs.get_clock() self.clock = hs.get_clock()
self.timeout_sec = timeout_ms / 1000.0 self.timeout_sec = timeout_ms / 1000.0
@ -102,11 +101,7 @@ class ResponseCache(Generic[T]):
self.pending_result_cache[key] = result self.pending_result_cache[key] = result
def remove(r): def remove(r):
should_cache = all( if self.timeout_sec:
func(r) for func in self.pending_conditionals.pop(key, [])
)
if self.timeout_sec and should_cache:
self.clock.call_later( self.clock.call_later(
self.timeout_sec, self.pending_result_cache.pop, key, None self.timeout_sec, self.pending_result_cache.pop, key, None
) )
@ -117,31 +112,6 @@ class ResponseCache(Generic[T]):
result.addBoth(remove) result.addBoth(remove)
return result.observe() return result.observe()
def add_conditional(self, key: T, conditional: Callable[[Any], bool]):
self.pending_conditionals.setdefault(key, set()).add(conditional)
def wrap_conditional(
self,
key: T,
should_cache: Callable[[Any], bool],
callback: "Callable[..., Any]",
*args: Any,
**kwargs: Any
) -> defer.Deferred:
"""The same as wrap(), but adds a conditional to the final execution.
When the final execution completes, *all* conditionals need to return True for it to properly cache,
else it'll not be cached in a timed fashion.
"""
# See if there's already a result on this key that hasn't yet completed. Due to the single-threaded nature of
# python, adding a key immediately in the same execution thread will not cause a race condition.
result = self.get(key)
if not result or isinstance(result, defer.Deferred) and not result.called:
self.add_conditional(key, should_cache)
return self.wrap(key, callback, *args, **kwargs)
def wrap( def wrap(
self, key: T, callback: "Callable[..., Any]", *args: Any, **kwargs: Any self, key: T, callback: "Callable[..., Any]", *args: Any, **kwargs: Any
) -> defer.Deferred: ) -> defer.Deferred: