mirror of
https://mau.dev/maunium/synapse.git
synced 2024-12-15 15:33:50 +01:00
Fix the background update
This commit is contained in:
parent
90b503216c
commit
940a161192
3 changed files with 19 additions and 17 deletions
|
@ -93,16 +93,19 @@ class BackgroundUpdateStore(SQLBaseStore):
|
||||||
|
|
||||||
sleep = defer.Deferred()
|
sleep = defer.Deferred()
|
||||||
self._background_update_timer = self._clock.call_later(
|
self._background_update_timer = self._clock.call_later(
|
||||||
self.BACKGROUND_UPDATE_INTERVAL_MS / 1000., sleep.callback
|
self.BACKGROUND_UPDATE_INTERVAL_MS / 1000., sleep.callback, None
|
||||||
)
|
)
|
||||||
try:
|
try:
|
||||||
yield sleep
|
yield sleep
|
||||||
finally:
|
finally:
|
||||||
self._background_update_timer = None
|
self._background_update_timer = None
|
||||||
|
|
||||||
result = yield self.do_background_update(
|
try:
|
||||||
self.BACKGROUND_UPDATE_DURATION_MS
|
result = yield self.do_background_update(
|
||||||
)
|
self.BACKGROUND_UPDATE_DURATION_MS
|
||||||
|
)
|
||||||
|
except:
|
||||||
|
logger.exception("Error doing update")
|
||||||
|
|
||||||
if result is None:
|
if result is None:
|
||||||
logger.info(
|
logger.info(
|
||||||
|
@ -169,7 +172,7 @@ class BackgroundUpdateStore(SQLBaseStore):
|
||||||
duration_ms = time_stop - time_start
|
duration_ms = time_stop - time_start
|
||||||
|
|
||||||
logger.info(
|
logger.info(
|
||||||
"Updating %. Updated %r items in %rms."
|
"Updating %r. Updated %r items in %rms."
|
||||||
" (total_rate=%r/ms, current_rate=%r/ms, total_updated=%r)",
|
" (total_rate=%r/ms, current_rate=%r/ms, total_updated=%r)",
|
||||||
update_name, items_updated, duration_ms,
|
update_name, items_updated, duration_ms,
|
||||||
performance.total_items_per_ms(),
|
performance.total_items_per_ms(),
|
||||||
|
|
|
@ -47,11 +47,10 @@ def run_upgrade(cur, database_engine, *args, **kwargs):
|
||||||
if isinstance(database_engine, PostgresEngine):
|
if isinstance(database_engine, PostgresEngine):
|
||||||
for statement in get_statements(POSTGRES_TABLE.splitlines()):
|
for statement in get_statements(POSTGRES_TABLE.splitlines()):
|
||||||
cur.execute(statement)
|
cur.execute(statement)
|
||||||
return
|
elif isinstance(database_engine, Sqlite3Engine):
|
||||||
|
|
||||||
if isinstance(database_engine, Sqlite3Engine):
|
|
||||||
cur.execute(SQLITE_TABLE)
|
cur.execute(SQLITE_TABLE)
|
||||||
return
|
else:
|
||||||
|
raise Exception("Unrecognized database engine")
|
||||||
|
|
||||||
cur.execute("SELECT MIN(stream_ordering) FROM events")
|
cur.execute("SELECT MIN(stream_ordering) FROM events")
|
||||||
rows = cur.fetchall()
|
rows = cur.fetchall()
|
||||||
|
|
|
@ -46,18 +46,18 @@ class SearchStore(BackgroundUpdateStore):
|
||||||
|
|
||||||
def reindex_search_txn(txn):
|
def reindex_search_txn(txn):
|
||||||
sql = (
|
sql = (
|
||||||
"SELECT stream_id, event_id FROM events"
|
"SELECT stream_ordering, event_id FROM events"
|
||||||
" WHERE ? <= stream_ordering AND stream_ordering < ?"
|
" WHERE ? <= stream_ordering AND stream_ordering < ?"
|
||||||
" AND (%s)"
|
" AND (%s)"
|
||||||
" ORDER BY stream_ordering DESC"
|
" ORDER BY stream_ordering DESC"
|
||||||
" LIMIT ?"
|
" LIMIT ?"
|
||||||
) % (" OR ".join("type = '%s'" % TYPES),)
|
) % (" OR ".join("type = '%s'" % (t,) for t in TYPES),)
|
||||||
|
|
||||||
txn.execute(sql, target_min_stream_id, max_stream_id, batch_size)
|
txn.execute(sql, (target_min_stream_id, max_stream_id, batch_size))
|
||||||
|
|
||||||
rows = txn.fetch_all()
|
rows = txn.fetchall()
|
||||||
if not rows:
|
if not rows:
|
||||||
return None
|
return 0
|
||||||
|
|
||||||
min_stream_id = rows[-1][0]
|
min_stream_id = rows[-1][0]
|
||||||
event_ids = [row[1] for row in rows]
|
event_ids = [row[1] for row in rows]
|
||||||
|
@ -102,7 +102,7 @@ class SearchStore(BackgroundUpdateStore):
|
||||||
|
|
||||||
for index in range(0, len(event_search_rows), INSERT_CLUMP_SIZE):
|
for index in range(0, len(event_search_rows), INSERT_CLUMP_SIZE):
|
||||||
clump = event_search_rows[index:index + INSERT_CLUMP_SIZE]
|
clump = event_search_rows[index:index + INSERT_CLUMP_SIZE]
|
||||||
txn.execute_many(sql, clump)
|
txn.executemany(sql, clump)
|
||||||
|
|
||||||
progress = {
|
progress = {
|
||||||
"target_min_stream_id_inclusive": target_min_stream_id,
|
"target_min_stream_id_inclusive": target_min_stream_id,
|
||||||
|
@ -116,11 +116,11 @@ class SearchStore(BackgroundUpdateStore):
|
||||||
|
|
||||||
return len(event_search_rows)
|
return len(event_search_rows)
|
||||||
|
|
||||||
result = yield self.runInteration(
|
result = yield self.runInteraction(
|
||||||
self.EVENT_SEARCH_UPDATE_NAME, reindex_search_txn
|
self.EVENT_SEARCH_UPDATE_NAME, reindex_search_txn
|
||||||
)
|
)
|
||||||
|
|
||||||
if result is None:
|
if not result:
|
||||||
yield self._end_background_update(self.EVENT_SEARCH_UPDATE_NAME)
|
yield self._end_background_update(self.EVENT_SEARCH_UPDATE_NAME)
|
||||||
|
|
||||||
defer.returnValue(result)
|
defer.returnValue(result)
|
||||||
|
|
Loading…
Reference in a new issue