mirror of
https://mau.dev/maunium/synapse.git
synced 2024-12-25 17:04:00 +01:00
Bug fixes.
This commit is contained in:
parent
b4886264a3
commit
5bf318e9a6
2 changed files with 57 additions and 27 deletions
|
@ -274,6 +274,8 @@ def setup():
|
||||||
|
|
||||||
hs.get_pusherpool().start()
|
hs.get_pusherpool().start()
|
||||||
|
|
||||||
|
hs.get_state_handler().start_caching()
|
||||||
|
|
||||||
if config.daemonize:
|
if config.daemonize:
|
||||||
print config.pid_file
|
print config.pid_file
|
||||||
daemon = Daemonize(
|
daemon = Daemonize(
|
||||||
|
|
|
@ -43,6 +43,10 @@ AuthEventTypes = (
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
SIZE_OF_CACHE = 1000
|
||||||
|
EVICTION_TIMEOUT_SECONDS = 20
|
||||||
|
|
||||||
|
|
||||||
class _StateCacheEntry(object):
|
class _StateCacheEntry(object):
|
||||||
def __init__(self, state, state_group, ts):
|
def __init__(self, state, state_group, ts):
|
||||||
self.state = state
|
self.state = state
|
||||||
|
@ -59,13 +63,22 @@ class StateHandler(object):
|
||||||
self.store = hs.get_datastore()
|
self.store = hs.get_datastore()
|
||||||
self.hs = hs
|
self.hs = hs
|
||||||
|
|
||||||
# set of event_ids -> _StateCacheEntry.
|
# dict of set of event_ids -> _StateCacheEntry.
|
||||||
|
self._state_cache = None
|
||||||
|
|
||||||
|
def start_caching(self):
|
||||||
|
logger.debug("start_caching")
|
||||||
|
|
||||||
self._state_cache = {}
|
self._state_cache = {}
|
||||||
|
|
||||||
def f():
|
def f():
|
||||||
|
logger.debug("Pruning")
|
||||||
|
try:
|
||||||
self._prune_cache()
|
self._prune_cache()
|
||||||
|
except:
|
||||||
|
logger.exception("Prune")
|
||||||
|
|
||||||
self.clock.looping_call(f, 10*1000)
|
self.clock.looping_call(f, 5*1000)
|
||||||
|
|
||||||
@defer.inlineCallbacks
|
@defer.inlineCallbacks
|
||||||
def get_current_state(self, room_id, event_type=None, state_key=""):
|
def get_current_state(self, room_id, event_type=None, state_key=""):
|
||||||
|
@ -86,7 +99,10 @@ class StateHandler(object):
|
||||||
for e_id, _, _ in events
|
for e_id, _, _ in events
|
||||||
]
|
]
|
||||||
|
|
||||||
|
cache = None
|
||||||
|
if self._state_cache is not None:
|
||||||
cache = self._state_cache.get(frozenset(event_ids), None)
|
cache = self._state_cache.get(frozenset(event_ids), None)
|
||||||
|
|
||||||
if cache:
|
if cache:
|
||||||
cache.ts = self.clock.time_msec()
|
cache.ts = self.clock.time_msec()
|
||||||
state = cache.state
|
state = cache.state
|
||||||
|
@ -199,6 +215,7 @@ class StateHandler(object):
|
||||||
"""
|
"""
|
||||||
logger.debug("resolve_state_groups event_ids %s", event_ids)
|
logger.debug("resolve_state_groups event_ids %s", event_ids)
|
||||||
|
|
||||||
|
if self._state_cache is not None:
|
||||||
cache = self._state_cache.get(frozenset(event_ids), None)
|
cache = self._state_cache.get(frozenset(event_ids), None)
|
||||||
if cache and cache.state_group:
|
if cache and cache.state_group:
|
||||||
cache.ts = self.clock.time_msec()
|
cache.ts = self.clock.time_msec()
|
||||||
|
@ -208,7 +225,9 @@ class StateHandler(object):
|
||||||
prev_states = [prev_state]
|
prev_states = [prev_state]
|
||||||
else:
|
else:
|
||||||
prev_states = []
|
prev_states = []
|
||||||
defer.returnValue((cache.state_group, cache.state, prev_states))
|
defer.returnValue(
|
||||||
|
(cache.state_group, cache.state, prev_states)
|
||||||
|
)
|
||||||
|
|
||||||
state_groups = yield self.store.get_state_groups(
|
state_groups = yield self.store.get_state_groups(
|
||||||
event_ids
|
event_ids
|
||||||
|
@ -233,6 +252,7 @@ class StateHandler(object):
|
||||||
else:
|
else:
|
||||||
prev_states = []
|
prev_states = []
|
||||||
|
|
||||||
|
if self._state_cache is not None:
|
||||||
cache = _StateCacheEntry(
|
cache = _StateCacheEntry(
|
||||||
state=state,
|
state=state,
|
||||||
state_group=name,
|
state_group=name,
|
||||||
|
@ -292,6 +312,7 @@ class StateHandler(object):
|
||||||
ts=self.clock.time_msec()
|
ts=self.clock.time_msec()
|
||||||
)
|
)
|
||||||
|
|
||||||
|
if self._state_cache is not None:
|
||||||
self._state_cache[frozenset(event_ids)] = cache
|
self._state_cache[frozenset(event_ids)] = cache
|
||||||
|
|
||||||
defer.returnValue((None, new_state, prev_states))
|
defer.returnValue((None, new_state, prev_states))
|
||||||
|
@ -379,26 +400,33 @@ class StateHandler(object):
|
||||||
return sorted(events, key=key_func)
|
return sorted(events, key=key_func)
|
||||||
|
|
||||||
def _prune_cache(self):
|
def _prune_cache(self):
|
||||||
logger.debug("_prune_cache. before len: ", len(self._state_cache))
|
logger.debug("_prune_cache")
|
||||||
|
logger.debug(
|
||||||
|
"_prune_cache. before len: %d",
|
||||||
|
len(self._state_cache.keys())
|
||||||
|
)
|
||||||
|
|
||||||
now = self.clock.time_msec()
|
now = self.clock.time_msec()
|
||||||
|
|
||||||
if len(self._state_cache) > 100:
|
if len(self._state_cache.keys()) > SIZE_OF_CACHE:
|
||||||
sorted_entries = sorted(
|
sorted_entries = sorted(
|
||||||
self._state_cache.items(),
|
self._state_cache.items(),
|
||||||
key=lambda k, v: v.ts,
|
key=lambda k, v: v.ts,
|
||||||
)
|
)
|
||||||
|
|
||||||
for k, _ in sorted_entries[100:]:
|
for k, _ in sorted_entries[SIZE_OF_CACHE:]:
|
||||||
self._state_cache.pop(k)
|
self._state_cache.pop(k)
|
||||||
|
|
||||||
keys_to_delete = set()
|
keys_to_delete = set()
|
||||||
|
|
||||||
for key, cache_entry in self._state_cache.items():
|
for key, cache_entry in self._state_cache.items():
|
||||||
if now - cache_entry.ts > 60*1000:
|
if now - cache_entry.ts > EVICTION_TIMEOUT_SECONDS*1000:
|
||||||
keys_to_delete.add(key)
|
keys_to_delete.add(key)
|
||||||
|
|
||||||
for k in keys_to_delete:
|
for k in keys_to_delete:
|
||||||
self._state_cache.pop(k)
|
self._state_cache.pop(k)
|
||||||
|
|
||||||
logger.debug("_prune_cache. after len: ", len(self._state_cache))
|
logger.debug(
|
||||||
|
"_prune_cache. after len: %d",
|
||||||
|
len(self._state_cache.keys())
|
||||||
|
)
|
||||||
|
|
Loading…
Reference in a new issue