mirror of
https://mau.dev/maunium/synapse.git
synced 2024-11-10 12:02:43 +01:00
Add more logging to debug slow startup (#8264)
I'm hoping this will provide some pointers for debugging https://github.com/matrix-org/synapse/issues/7968.
This commit is contained in:
parent
96312536f2
commit
0dae7d80bf
5 changed files with 33 additions and 7 deletions
1
changelog.d/8264.misc
Normal file
1
changelog.d/8264.misc
Normal file
|
@ -0,0 +1 @@
|
||||||
|
Add more logging to debug slow startup.
|
|
@ -47,9 +47,14 @@ class Databases:
|
||||||
engine = create_engine(database_config.config)
|
engine = create_engine(database_config.config)
|
||||||
|
|
||||||
with make_conn(database_config, engine) as db_conn:
|
with make_conn(database_config, engine) as db_conn:
|
||||||
logger.info("Preparing database %r...", db_name)
|
logger.info("[database config %r]: Checking database server", db_name)
|
||||||
|
|
||||||
engine.check_database(db_conn)
|
engine.check_database(db_conn)
|
||||||
|
|
||||||
|
logger.info(
|
||||||
|
"[database config %r]: Preparing for databases %r",
|
||||||
|
db_name,
|
||||||
|
database_config.databases,
|
||||||
|
)
|
||||||
prepare_database(
|
prepare_database(
|
||||||
db_conn, engine, hs.config, databases=database_config.databases,
|
db_conn, engine, hs.config, databases=database_config.databases,
|
||||||
)
|
)
|
||||||
|
@ -57,7 +62,9 @@ class Databases:
|
||||||
database = DatabasePool(hs, database_config, engine)
|
database = DatabasePool(hs, database_config, engine)
|
||||||
|
|
||||||
if "main" in database_config.databases:
|
if "main" in database_config.databases:
|
||||||
logger.info("Starting 'main' data store")
|
logger.info(
|
||||||
|
"[database config %r]: Starting 'main' database", db_name
|
||||||
|
)
|
||||||
|
|
||||||
# Sanity check we don't try and configure the main store on
|
# Sanity check we don't try and configure the main store on
|
||||||
# multiple databases.
|
# multiple databases.
|
||||||
|
@ -72,7 +79,9 @@ class Databases:
|
||||||
persist_events = PersistEventsStore(hs, database, main)
|
persist_events = PersistEventsStore(hs, database, main)
|
||||||
|
|
||||||
if "state" in database_config.databases:
|
if "state" in database_config.databases:
|
||||||
logger.info("Starting 'state' data store")
|
logger.info(
|
||||||
|
"[database config %r]: Starting 'state' database", db_name
|
||||||
|
)
|
||||||
|
|
||||||
# Sanity check we don't try and configure the state store on
|
# Sanity check we don't try and configure the state store on
|
||||||
# multiple databases.
|
# multiple databases.
|
||||||
|
@ -85,7 +94,7 @@ class Databases:
|
||||||
|
|
||||||
self.databases.append(database)
|
self.databases.append(database)
|
||||||
|
|
||||||
logger.info("Database %r prepared", db_name)
|
logger.info("[database config %r]: prepared", db_name)
|
||||||
|
|
||||||
# Closing the context manager doesn't close the connection.
|
# Closing the context manager doesn't close the connection.
|
||||||
# psycopg will close the connection when the object gets GCed, but *only*
|
# psycopg will close the connection when the object gets GCed, but *only*
|
||||||
|
@ -98,10 +107,10 @@ class Databases:
|
||||||
|
|
||||||
# Sanity check that we have actually configured all the required stores.
|
# Sanity check that we have actually configured all the required stores.
|
||||||
if not main:
|
if not main:
|
||||||
raise Exception("No 'main' data store configured")
|
raise Exception("No 'main' database configured")
|
||||||
|
|
||||||
if not state:
|
if not state:
|
||||||
raise Exception("No 'state' data store configured")
|
raise Exception("No 'state' database configured")
|
||||||
|
|
||||||
# We use local variables here to ensure that the databases do not have
|
# We use local variables here to ensure that the databases do not have
|
||||||
# optional types.
|
# optional types.
|
||||||
|
|
|
@ -591,6 +591,7 @@ def check_database_before_upgrade(cur, database_engine, config: HomeServerConfig
|
||||||
"""Called before upgrading an existing database to check that it is broadly sane
|
"""Called before upgrading an existing database to check that it is broadly sane
|
||||||
compared with the configuration.
|
compared with the configuration.
|
||||||
"""
|
"""
|
||||||
|
logger.info("Running sanity-checks on database...")
|
||||||
domain = config.server_name
|
domain = config.server_name
|
||||||
|
|
||||||
sql = database_engine.convert_param_style(
|
sql = database_engine.convert_param_style(
|
||||||
|
|
|
@ -82,10 +82,18 @@ def prepare_database(db_conn, database_engine, config, databases=["main", "state
|
||||||
|
|
||||||
try:
|
try:
|
||||||
cur = db_conn.cursor()
|
cur = db_conn.cursor()
|
||||||
|
|
||||||
|
logger.info("%r: Checking existing schema version", databases)
|
||||||
version_info = _get_or_create_schema_state(cur, database_engine)
|
version_info = _get_or_create_schema_state(cur, database_engine)
|
||||||
|
|
||||||
if version_info:
|
if version_info:
|
||||||
user_version, delta_files, upgraded = version_info
|
user_version, delta_files, upgraded = version_info
|
||||||
|
logger.info(
|
||||||
|
"%r: Existing schema is %i (+%i deltas)",
|
||||||
|
databases,
|
||||||
|
user_version,
|
||||||
|
len(delta_files),
|
||||||
|
)
|
||||||
|
|
||||||
# config should only be None when we are preparing an in-memory SQLite db,
|
# config should only be None when we are preparing an in-memory SQLite db,
|
||||||
# which should be empty.
|
# which should be empty.
|
||||||
|
@ -111,6 +119,8 @@ def prepare_database(db_conn, database_engine, config, databases=["main", "state
|
||||||
databases=databases,
|
databases=databases,
|
||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
|
logger.info("%r: Initialising new database", databases)
|
||||||
|
|
||||||
# if it's a worker app, refuse to upgrade the database, to avoid multiple
|
# if it's a worker app, refuse to upgrade the database, to avoid multiple
|
||||||
# workers doing it at once.
|
# workers doing it at once.
|
||||||
if config and config.worker_app is not None:
|
if config and config.worker_app is not None:
|
||||||
|
|
|
@ -15,6 +15,7 @@
|
||||||
|
|
||||||
import contextlib
|
import contextlib
|
||||||
import heapq
|
import heapq
|
||||||
|
import logging
|
||||||
import threading
|
import threading
|
||||||
from collections import deque
|
from collections import deque
|
||||||
from typing import Dict, List, Set
|
from typing import Dict, List, Set
|
||||||
|
@ -24,6 +25,8 @@ from typing_extensions import Deque
|
||||||
from synapse.storage.database import DatabasePool, LoggingTransaction
|
from synapse.storage.database import DatabasePool, LoggingTransaction
|
||||||
from synapse.storage.util.sequence import PostgresSequenceGenerator
|
from synapse.storage.util.sequence import PostgresSequenceGenerator
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
class IdGenerator:
|
class IdGenerator:
|
||||||
def __init__(self, db_conn, table, column):
|
def __init__(self, db_conn, table, column):
|
||||||
|
@ -48,6 +51,8 @@ def _load_current_id(db_conn, table, column, step=1):
|
||||||
Returns:
|
Returns:
|
||||||
int
|
int
|
||||||
"""
|
"""
|
||||||
|
# debug logging for https://github.com/matrix-org/synapse/issues/7968
|
||||||
|
logger.info("initialising stream generator for %s(%s)", table, column)
|
||||||
cur = db_conn.cursor()
|
cur = db_conn.cursor()
|
||||||
if step == 1:
|
if step == 1:
|
||||||
cur.execute("SELECT MAX(%s) FROM %s" % (column, table))
|
cur.execute("SELECT MAX(%s) FROM %s" % (column, table))
|
||||||
|
|
Loading…
Reference in a new issue