forked from MirrorHub/synapse
Add type hints to schema deltas (#15497)
Cleans-up the schema delta files: * Removes no-op functions. * Adds missing type hints to function parameters. * Fixes any issues with type hints. This also renames one (very old) schema delta to avoid a conflict that mypy complains about.
This commit is contained in:
parent
a346b43837
commit
6efa674004
30 changed files with 132 additions and 144 deletions
1
changelog.d/15497.misc
Normal file
1
changelog.d/15497.misc
Normal file
|
@ -0,0 +1 @@
|
||||||
|
Improve type hints.
|
10
mypy.ini
10
mypy.ini
|
@ -21,16 +21,6 @@ files =
|
||||||
tests/,
|
tests/,
|
||||||
build_rust.py
|
build_rust.py
|
||||||
|
|
||||||
# Note: Better exclusion syntax coming in mypy > 0.910
|
|
||||||
# https://github.com/python/mypy/pull/11329
|
|
||||||
#
|
|
||||||
# For now, set the (?x) flag enable "verbose" regexes
|
|
||||||
# https://docs.python.org/3/library/re.html#re.X
|
|
||||||
exclude = (?x)
|
|
||||||
^(
|
|
||||||
|synapse/storage/schema/
|
|
||||||
)$
|
|
||||||
|
|
||||||
[mypy-synapse.metrics._reactor_metrics]
|
[mypy-synapse.metrics._reactor_metrics]
|
||||||
# This module imports select.epoll. That exists on Linux, but doesn't on macOS.
|
# This module imports select.epoll. That exists on Linux, but doesn't on macOS.
|
||||||
# See https://github.com/matrix-org/synapse/pull/11771.
|
# See https://github.com/matrix-org/synapse/pull/11771.
|
||||||
|
|
|
@ -22,7 +22,7 @@ import attr
|
||||||
from typing_extensions import Counter as CounterType
|
from typing_extensions import Counter as CounterType
|
||||||
|
|
||||||
from synapse.config.homeserver import HomeServerConfig
|
from synapse.config.homeserver import HomeServerConfig
|
||||||
from synapse.storage.database import LoggingDatabaseConnection
|
from synapse.storage.database import LoggingDatabaseConnection, LoggingTransaction
|
||||||
from synapse.storage.engines import BaseDatabaseEngine, PostgresEngine, Sqlite3Engine
|
from synapse.storage.engines import BaseDatabaseEngine, PostgresEngine, Sqlite3Engine
|
||||||
from synapse.storage.schema import SCHEMA_COMPAT_VERSION, SCHEMA_VERSION
|
from synapse.storage.schema import SCHEMA_COMPAT_VERSION, SCHEMA_VERSION
|
||||||
from synapse.storage.types import Cursor
|
from synapse.storage.types import Cursor
|
||||||
|
@ -168,7 +168,9 @@ def prepare_database(
|
||||||
|
|
||||||
|
|
||||||
def _setup_new_database(
|
def _setup_new_database(
|
||||||
cur: Cursor, database_engine: BaseDatabaseEngine, databases: Collection[str]
|
cur: LoggingTransaction,
|
||||||
|
database_engine: BaseDatabaseEngine,
|
||||||
|
databases: Collection[str],
|
||||||
) -> None:
|
) -> None:
|
||||||
"""Sets up the physical database by finding a base set of "full schemas" and
|
"""Sets up the physical database by finding a base set of "full schemas" and
|
||||||
then applying any necessary deltas, including schemas from the given data
|
then applying any necessary deltas, including schemas from the given data
|
||||||
|
@ -289,7 +291,7 @@ def _setup_new_database(
|
||||||
|
|
||||||
|
|
||||||
def _upgrade_existing_database(
|
def _upgrade_existing_database(
|
||||||
cur: Cursor,
|
cur: LoggingTransaction,
|
||||||
current_schema_state: _SchemaState,
|
current_schema_state: _SchemaState,
|
||||||
database_engine: BaseDatabaseEngine,
|
database_engine: BaseDatabaseEngine,
|
||||||
config: Optional[HomeServerConfig],
|
config: Optional[HomeServerConfig],
|
||||||
|
|
|
@ -24,10 +24,13 @@ UTF-8 bytes, so we have to do it in Python.
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
|
from synapse.storage.database import LoggingTransaction
|
||||||
|
from synapse.storage.engines import BaseDatabaseEngine
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur, database_engine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
logger.info("Porting pushers table...")
|
logger.info("Porting pushers table...")
|
||||||
cur.execute(
|
cur.execute(
|
||||||
"""
|
"""
|
||||||
|
@ -61,8 +64,8 @@ def run_create(cur, database_engine, *args, **kwargs):
|
||||||
"""
|
"""
|
||||||
)
|
)
|
||||||
count = 0
|
count = 0
|
||||||
for row in cur.fetchall():
|
for tuple_row in cur.fetchall():
|
||||||
row = list(row)
|
row = list(tuple_row)
|
||||||
row[8] = bytes(row[8]).decode("utf-8")
|
row[8] = bytes(row[8]).decode("utf-8")
|
||||||
row[11] = bytes(row[11]).decode("utf-8")
|
row[11] = bytes(row[11]).decode("utf-8")
|
||||||
cur.execute(
|
cur.execute(
|
||||||
|
@ -81,7 +84,3 @@ def run_create(cur, database_engine, *args, **kwargs):
|
||||||
cur.execute("DROP TABLE pushers")
|
cur.execute("DROP TABLE pushers")
|
||||||
cur.execute("ALTER TABLE pushers2 RENAME TO pushers")
|
cur.execute("ALTER TABLE pushers2 RENAME TO pushers")
|
||||||
logger.info("Moved %d pushers to new table", count)
|
logger.info("Moved %d pushers to new table", count)
|
||||||
|
|
||||||
|
|
||||||
def run_upgrade(*args, **kwargs):
|
|
||||||
pass
|
|
||||||
|
|
|
@ -14,7 +14,8 @@
|
||||||
import json
|
import json
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
from synapse.storage.engines import PostgresEngine, Sqlite3Engine
|
from synapse.storage.database import LoggingTransaction
|
||||||
|
from synapse.storage.engines import BaseDatabaseEngine, PostgresEngine, Sqlite3Engine
|
||||||
from synapse.storage.prepare_database import get_statements
|
from synapse.storage.prepare_database import get_statements
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
@ -41,7 +42,7 @@ SQLITE_TABLE = (
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur, database_engine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
if isinstance(database_engine, PostgresEngine):
|
if isinstance(database_engine, PostgresEngine):
|
||||||
for statement in get_statements(POSTGRES_TABLE.splitlines()):
|
for statement in get_statements(POSTGRES_TABLE.splitlines()):
|
||||||
cur.execute(statement)
|
cur.execute(statement)
|
||||||
|
@ -72,7 +73,3 @@ def run_create(cur, database_engine, *args, **kwargs):
|
||||||
)
|
)
|
||||||
|
|
||||||
cur.execute(sql, ("event_search", progress_json))
|
cur.execute(sql, ("event_search", progress_json))
|
||||||
|
|
||||||
|
|
||||||
def run_upgrade(*args, **kwargs):
|
|
||||||
pass
|
|
||||||
|
|
|
@ -14,6 +14,8 @@
|
||||||
import json
|
import json
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
|
from synapse.storage.database import LoggingTransaction
|
||||||
|
from synapse.storage.engines import BaseDatabaseEngine
|
||||||
from synapse.storage.prepare_database import get_statements
|
from synapse.storage.prepare_database import get_statements
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
@ -25,7 +27,7 @@ ALTER_TABLE = (
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur, database_engine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
for statement in get_statements(ALTER_TABLE.splitlines()):
|
for statement in get_statements(ALTER_TABLE.splitlines()):
|
||||||
cur.execute(statement)
|
cur.execute(statement)
|
||||||
|
|
||||||
|
@ -51,7 +53,3 @@ def run_create(cur, database_engine, *args, **kwargs):
|
||||||
)
|
)
|
||||||
|
|
||||||
cur.execute(sql, ("event_origin_server_ts", progress_json))
|
cur.execute(sql, ("event_origin_server_ts", progress_json))
|
||||||
|
|
||||||
|
|
||||||
def run_upgrade(*args, **kwargs):
|
|
||||||
pass
|
|
||||||
|
|
|
@ -12,13 +12,17 @@
|
||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
import logging
|
import logging
|
||||||
|
from typing import Dict, Iterable, List, Tuple, cast
|
||||||
|
|
||||||
from synapse.config.appservice import load_appservices
|
from synapse.config.appservice import load_appservices
|
||||||
|
from synapse.config.homeserver import HomeServerConfig
|
||||||
|
from synapse.storage.database import LoggingTransaction
|
||||||
|
from synapse.storage.engines import BaseDatabaseEngine
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur, database_engine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
# NULL indicates user was not registered by an appservice.
|
# NULL indicates user was not registered by an appservice.
|
||||||
try:
|
try:
|
||||||
cur.execute("ALTER TABLE users ADD COLUMN appservice_id TEXT")
|
cur.execute("ALTER TABLE users ADD COLUMN appservice_id TEXT")
|
||||||
|
@ -27,9 +31,13 @@ def run_create(cur, database_engine, *args, **kwargs):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
def run_upgrade(cur, database_engine, config, *args, **kwargs):
|
def run_upgrade(
|
||||||
|
cur: LoggingTransaction,
|
||||||
|
database_engine: BaseDatabaseEngine,
|
||||||
|
config: HomeServerConfig,
|
||||||
|
) -> None:
|
||||||
cur.execute("SELECT name FROM users")
|
cur.execute("SELECT name FROM users")
|
||||||
rows = cur.fetchall()
|
rows = cast(Iterable[Tuple[str]], cur.fetchall())
|
||||||
|
|
||||||
config_files = []
|
config_files = []
|
||||||
try:
|
try:
|
||||||
|
@ -39,7 +47,7 @@ def run_upgrade(cur, database_engine, config, *args, **kwargs):
|
||||||
|
|
||||||
appservices = load_appservices(config.server.server_name, config_files)
|
appservices = load_appservices(config.server.server_name, config_files)
|
||||||
|
|
||||||
owned = {}
|
owned: Dict[str, List[str]] = {}
|
||||||
|
|
||||||
for row in rows:
|
for row in rows:
|
||||||
user_id = row[0]
|
user_id = row[0]
|
||||||
|
|
|
@ -20,14 +20,17 @@
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
|
from synapse.storage.database import LoggingTransaction
|
||||||
|
from synapse.storage.engines import BaseDatabaseEngine
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
def token_to_stream_ordering(token):
|
def token_to_stream_ordering(token: str) -> int:
|
||||||
return int(token[1:].split("_")[0])
|
return int(token[1:].split("_")[0])
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur, database_engine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
logger.info("Porting pushers table, delta 31...")
|
logger.info("Porting pushers table, delta 31...")
|
||||||
cur.execute(
|
cur.execute(
|
||||||
"""
|
"""
|
||||||
|
@ -61,8 +64,8 @@ def run_create(cur, database_engine, *args, **kwargs):
|
||||||
"""
|
"""
|
||||||
)
|
)
|
||||||
count = 0
|
count = 0
|
||||||
for row in cur.fetchall():
|
for tuple_row in cur.fetchall():
|
||||||
row = list(row)
|
row = list(tuple_row)
|
||||||
row[12] = token_to_stream_ordering(row[12])
|
row[12] = token_to_stream_ordering(row[12])
|
||||||
cur.execute(
|
cur.execute(
|
||||||
"""
|
"""
|
||||||
|
@ -80,7 +83,3 @@ def run_create(cur, database_engine, *args, **kwargs):
|
||||||
cur.execute("DROP TABLE pushers")
|
cur.execute("DROP TABLE pushers")
|
||||||
cur.execute("ALTER TABLE pushers2 RENAME TO pushers")
|
cur.execute("ALTER TABLE pushers2 RENAME TO pushers")
|
||||||
logger.info("Moved %d pushers to new table", count)
|
logger.info("Moved %d pushers to new table", count)
|
||||||
|
|
||||||
|
|
||||||
def run_upgrade(cur, database_engine, *args, **kwargs):
|
|
||||||
pass
|
|
|
@ -14,7 +14,8 @@
|
||||||
import json
|
import json
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
from synapse.storage.engines import PostgresEngine
|
from synapse.storage.database import LoggingTransaction
|
||||||
|
from synapse.storage.engines import BaseDatabaseEngine, PostgresEngine
|
||||||
from synapse.storage.prepare_database import get_statements
|
from synapse.storage.prepare_database import get_statements
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
@ -26,7 +27,7 @@ ALTER TABLE event_search ADD COLUMN stream_ordering BIGINT;
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur, database_engine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
if not isinstance(database_engine, PostgresEngine):
|
if not isinstance(database_engine, PostgresEngine):
|
||||||
return
|
return
|
||||||
|
|
||||||
|
@ -56,7 +57,3 @@ def run_create(cur, database_engine, *args, **kwargs):
|
||||||
)
|
)
|
||||||
|
|
||||||
cur.execute(sql, ("event_search_order", progress_json))
|
cur.execute(sql, ("event_search_order", progress_json))
|
||||||
|
|
||||||
|
|
||||||
def run_upgrade(cur, database_engine, *args, **kwargs):
|
|
||||||
pass
|
|
||||||
|
|
|
@ -14,6 +14,8 @@
|
||||||
import json
|
import json
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
|
from synapse.storage.database import LoggingTransaction
|
||||||
|
from synapse.storage.engines import BaseDatabaseEngine
|
||||||
from synapse.storage.prepare_database import get_statements
|
from synapse.storage.prepare_database import get_statements
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
@ -25,7 +27,7 @@ ALTER TABLE events ADD COLUMN contains_url BOOLEAN;
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur, database_engine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
for statement in get_statements(ALTER_TABLE.splitlines()):
|
for statement in get_statements(ALTER_TABLE.splitlines()):
|
||||||
cur.execute(statement)
|
cur.execute(statement)
|
||||||
|
|
||||||
|
@ -51,7 +53,3 @@ def run_create(cur, database_engine, *args, **kwargs):
|
||||||
)
|
)
|
||||||
|
|
||||||
cur.execute(sql, ("event_fields_sender_url", progress_json))
|
cur.execute(sql, ("event_fields_sender_url", progress_json))
|
||||||
|
|
||||||
|
|
||||||
def run_upgrade(cur, database_engine, *args, **kwargs):
|
|
||||||
pass
|
|
||||||
|
|
|
@ -14,14 +14,22 @@
|
||||||
|
|
||||||
import time
|
import time
|
||||||
|
|
||||||
|
from synapse.config.homeserver import HomeServerConfig
|
||||||
|
from synapse.storage.database import LoggingTransaction
|
||||||
|
from synapse.storage.engines import BaseDatabaseEngine
|
||||||
|
|
||||||
ALTER_TABLE = "ALTER TABLE remote_media_cache ADD COLUMN last_access_ts BIGINT"
|
ALTER_TABLE = "ALTER TABLE remote_media_cache ADD COLUMN last_access_ts BIGINT"
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur, database_engine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
cur.execute(ALTER_TABLE)
|
cur.execute(ALTER_TABLE)
|
||||||
|
|
||||||
|
|
||||||
def run_upgrade(cur, database_engine, *args, **kwargs):
|
def run_upgrade(
|
||||||
|
cur: LoggingTransaction,
|
||||||
|
database_engine: BaseDatabaseEngine,
|
||||||
|
config: HomeServerConfig,
|
||||||
|
) -> None:
|
||||||
cur.execute(
|
cur.execute(
|
||||||
"UPDATE remote_media_cache SET last_access_ts = ?",
|
"UPDATE remote_media_cache SET last_access_ts = ?",
|
||||||
(int(time.time() * 1000),),
|
(int(time.time() * 1000),),
|
||||||
|
|
|
@ -14,7 +14,8 @@
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
from synapse.storage.engines import PostgresEngine
|
from synapse.storage.database import LoggingTransaction
|
||||||
|
from synapse.storage.engines import BaseDatabaseEngine, PostgresEngine
|
||||||
from synapse.storage.prepare_database import get_statements
|
from synapse.storage.prepare_database import get_statements
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
@ -34,13 +35,9 @@ CREATE INDEX cache_invalidation_stream_id ON cache_invalidation_stream(stream_id
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur, database_engine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
if not isinstance(database_engine, PostgresEngine):
|
if not isinstance(database_engine, PostgresEngine):
|
||||||
return
|
return
|
||||||
|
|
||||||
for statement in get_statements(CREATE_TABLE.splitlines()):
|
for statement in get_statements(CREATE_TABLE.splitlines()):
|
||||||
cur.execute(statement)
|
cur.execute(statement)
|
||||||
|
|
||||||
|
|
||||||
def run_upgrade(cur, database_engine, *args, **kwargs):
|
|
||||||
pass
|
|
||||||
|
|
|
@ -14,19 +14,16 @@
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
from synapse.storage.engines import PostgresEngine
|
from synapse.storage.database import LoggingTransaction
|
||||||
|
from synapse.storage.engines import BaseDatabaseEngine, PostgresEngine
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur, database_engine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
if isinstance(database_engine, PostgresEngine):
|
if isinstance(database_engine, PostgresEngine):
|
||||||
cur.execute("TRUNCATE received_transactions")
|
cur.execute("TRUNCATE received_transactions")
|
||||||
else:
|
else:
|
||||||
cur.execute("DELETE FROM received_transactions")
|
cur.execute("DELETE FROM received_transactions")
|
||||||
|
|
||||||
cur.execute("CREATE INDEX received_transactions_ts ON received_transactions(ts)")
|
cur.execute("CREATE INDEX received_transactions_ts ON received_transactions(ts)")
|
||||||
|
|
||||||
|
|
||||||
def run_upgrade(cur, database_engine, *args, **kwargs):
|
|
||||||
pass
|
|
||||||
|
|
|
@ -14,7 +14,8 @@
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
from synapse.storage.engines import PostgresEngine
|
from synapse.storage.database import LoggingTransaction
|
||||||
|
from synapse.storage.engines import BaseDatabaseEngine, PostgresEngine
|
||||||
from synapse.storage.prepare_database import get_statements
|
from synapse.storage.prepare_database import get_statements
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
@ -68,7 +69,7 @@ CREATE INDEX evauth_edges_id ON event_auth(event_id);
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur, database_engine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
for statement in get_statements(DROP_INDICES.splitlines()):
|
for statement in get_statements(DROP_INDICES.splitlines()):
|
||||||
cur.execute(statement)
|
cur.execute(statement)
|
||||||
|
|
||||||
|
@ -79,7 +80,3 @@ def run_create(cur, database_engine, *args, **kwargs):
|
||||||
|
|
||||||
for statement in get_statements(drop_constraint.splitlines()):
|
for statement in get_statements(drop_constraint.splitlines()):
|
||||||
cur.execute(statement)
|
cur.execute(statement)
|
||||||
|
|
||||||
|
|
||||||
def run_upgrade(cur, database_engine, *args, **kwargs):
|
|
||||||
pass
|
|
||||||
|
|
|
@ -14,7 +14,8 @@
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
from synapse.storage.engines import PostgresEngine, Sqlite3Engine
|
from synapse.storage.database import LoggingTransaction
|
||||||
|
from synapse.storage.engines import BaseDatabaseEngine, PostgresEngine, Sqlite3Engine
|
||||||
from synapse.storage.prepare_database import get_statements
|
from synapse.storage.prepare_database import get_statements
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
@ -66,7 +67,7 @@ CREATE VIRTUAL TABLE user_directory_search
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur, database_engine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
for statement in get_statements(BOTH_TABLES.splitlines()):
|
for statement in get_statements(BOTH_TABLES.splitlines()):
|
||||||
cur.execute(statement)
|
cur.execute(statement)
|
||||||
|
|
||||||
|
@ -78,7 +79,3 @@ def run_create(cur, database_engine, *args, **kwargs):
|
||||||
cur.execute(statement)
|
cur.execute(statement)
|
||||||
else:
|
else:
|
||||||
raise Exception("Unrecognized database engine")
|
raise Exception("Unrecognized database engine")
|
||||||
|
|
||||||
|
|
||||||
def run_upgrade(*args, **kwargs):
|
|
||||||
pass
|
|
||||||
|
|
|
@ -12,7 +12,9 @@
|
||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
|
||||||
from synapse.storage.engines import PostgresEngine
|
|
||||||
|
from synapse.storage.database import LoggingTransaction
|
||||||
|
from synapse.storage.engines import BaseDatabaseEngine, PostgresEngine
|
||||||
from synapse.storage.prepare_database import get_statements
|
from synapse.storage.prepare_database import get_statements
|
||||||
|
|
||||||
FIX_INDEXES = """
|
FIX_INDEXES = """
|
||||||
|
@ -34,7 +36,7 @@ CREATE INDEX group_rooms_r_idx ON group_rooms(room_id);
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur, database_engine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
rowid = "ctid" if isinstance(database_engine, PostgresEngine) else "rowid"
|
rowid = "ctid" if isinstance(database_engine, PostgresEngine) else "rowid"
|
||||||
|
|
||||||
# remove duplicates from group_users & group_invites tables
|
# remove duplicates from group_users & group_invites tables
|
||||||
|
@ -57,7 +59,3 @@ def run_create(cur, database_engine, *args, **kwargs):
|
||||||
|
|
||||||
for statement in get_statements(FIX_INDEXES.splitlines()):
|
for statement in get_statements(FIX_INDEXES.splitlines()):
|
||||||
cur.execute(statement)
|
cur.execute(statement)
|
||||||
|
|
||||||
|
|
||||||
def run_upgrade(*args, **kwargs):
|
|
||||||
pass
|
|
||||||
|
|
|
@ -53,16 +53,13 @@ SQLite:
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
from synapse.storage.engines import PostgresEngine
|
from synapse.storage.database import LoggingTransaction
|
||||||
|
from synapse.storage.engines import BaseDatabaseEngine, PostgresEngine
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur, database_engine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
def run_upgrade(cur, database_engine, *args, **kwargs):
|
|
||||||
if isinstance(database_engine, PostgresEngine):
|
if isinstance(database_engine, PostgresEngine):
|
||||||
cur.execute(
|
cur.execute(
|
||||||
"""
|
"""
|
||||||
|
@ -76,7 +73,9 @@ def run_upgrade(cur, database_engine, *args, **kwargs):
|
||||||
cur.execute(
|
cur.execute(
|
||||||
"SELECT sql FROM sqlite_master WHERE tbl_name='events' AND type='table'"
|
"SELECT sql FROM sqlite_master WHERE tbl_name='events' AND type='table'"
|
||||||
)
|
)
|
||||||
(oldsql,) = cur.fetchone()
|
row = cur.fetchone()
|
||||||
|
assert row is not None
|
||||||
|
(oldsql,) = row
|
||||||
|
|
||||||
sql = oldsql.replace("content TEXT NOT NULL", "content TEXT")
|
sql = oldsql.replace("content TEXT NOT NULL", "content TEXT")
|
||||||
if sql == oldsql:
|
if sql == oldsql:
|
||||||
|
@ -85,7 +84,9 @@ def run_upgrade(cur, database_engine, *args, **kwargs):
|
||||||
logger.info("Replacing definition of 'events' with: %s", sql)
|
logger.info("Replacing definition of 'events' with: %s", sql)
|
||||||
|
|
||||||
cur.execute("PRAGMA schema_version")
|
cur.execute("PRAGMA schema_version")
|
||||||
(oldver,) = cur.fetchone()
|
row = cur.fetchone()
|
||||||
|
assert row is not None
|
||||||
|
(oldver,) = row
|
||||||
cur.execute("PRAGMA writable_schema=ON")
|
cur.execute("PRAGMA writable_schema=ON")
|
||||||
cur.execute(
|
cur.execute(
|
||||||
"UPDATE sqlite_master SET sql=? WHERE tbl_name='events' AND type='table'",
|
"UPDATE sqlite_master SET sql=? WHERE tbl_name='events' AND type='table'",
|
||||||
|
|
|
@ -1,7 +1,8 @@
|
||||||
import logging
|
import logging
|
||||||
from io import StringIO
|
from io import StringIO
|
||||||
|
|
||||||
from synapse.storage.engines import PostgresEngine
|
from synapse.storage.database import LoggingTransaction
|
||||||
|
from synapse.storage.engines import BaseDatabaseEngine, PostgresEngine
|
||||||
from synapse.storage.prepare_database import execute_statements_from_stream
|
from synapse.storage.prepare_database import execute_statements_from_stream
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
@ -16,11 +17,7 @@ This migration updates the user_filters table as follows:
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
|
||||||
def run_upgrade(cur, database_engine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur, database_engine, *args, **kwargs):
|
|
||||||
if isinstance(database_engine, PostgresEngine):
|
if isinstance(database_engine, PostgresEngine):
|
||||||
select_clause = """
|
select_clause = """
|
||||||
SELECT DISTINCT ON (user_id, filter_id) user_id, filter_id, filter_json
|
SELECT DISTINCT ON (user_id, filter_id) user_id, filter_id, filter_json
|
||||||
|
|
|
@ -27,7 +27,16 @@
|
||||||
# equivalent behaviour as if the server had remained in the room).
|
# equivalent behaviour as if the server had remained in the room).
|
||||||
|
|
||||||
|
|
||||||
def run_upgrade(cur, database_engine, config, *args, **kwargs):
|
from synapse.config.homeserver import HomeServerConfig
|
||||||
|
from synapse.storage.database import LoggingTransaction
|
||||||
|
from synapse.storage.engines import BaseDatabaseEngine
|
||||||
|
|
||||||
|
|
||||||
|
def run_upgrade(
|
||||||
|
cur: LoggingTransaction,
|
||||||
|
database_engine: BaseDatabaseEngine,
|
||||||
|
config: HomeServerConfig,
|
||||||
|
) -> None:
|
||||||
# We need to do the insert in `run_upgrade` section as we don't have access
|
# We need to do the insert in `run_upgrade` section as we don't have access
|
||||||
# to `config` in `run_create`.
|
# to `config` in `run_create`.
|
||||||
|
|
||||||
|
@ -77,7 +86,7 @@ def run_upgrade(cur, database_engine, config, *args, **kwargs):
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur, database_engine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
cur.execute(
|
cur.execute(
|
||||||
"""
|
"""
|
||||||
CREATE TABLE local_current_membership (
|
CREATE TABLE local_current_membership (
|
||||||
|
|
|
@ -20,18 +20,14 @@ entries, and with a UNIQUE index.
|
||||||
import logging
|
import logging
|
||||||
from io import StringIO
|
from io import StringIO
|
||||||
|
|
||||||
|
from synapse.storage.database import LoggingTransaction
|
||||||
from synapse.storage.engines import BaseDatabaseEngine, PostgresEngine
|
from synapse.storage.engines import BaseDatabaseEngine, PostgresEngine
|
||||||
from synapse.storage.prepare_database import execute_statements_from_stream
|
from synapse.storage.prepare_database import execute_statements_from_stream
|
||||||
from synapse.storage.types import Cursor
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
def run_upgrade(*args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur: Cursor, database_engine: BaseDatabaseEngine, *args, **kwargs):
|
|
||||||
# some instances might already have this index, in which case we can skip this
|
# some instances might already have this index, in which case we can skip this
|
||||||
if isinstance(database_engine, PostgresEngine):
|
if isinstance(database_engine, PostgresEngine):
|
||||||
cur.execute(
|
cur.execute(
|
||||||
|
|
|
@ -16,19 +16,16 @@
|
||||||
Adds a postgres SEQUENCE for generating guest user IDs.
|
Adds a postgres SEQUENCE for generating guest user IDs.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
from synapse.storage.database import LoggingTransaction
|
||||||
from synapse.storage.databases.main.registration import (
|
from synapse.storage.databases.main.registration import (
|
||||||
find_max_generated_user_id_localpart,
|
find_max_generated_user_id_localpart,
|
||||||
)
|
)
|
||||||
from synapse.storage.engines import PostgresEngine
|
from synapse.storage.engines import BaseDatabaseEngine, PostgresEngine
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur, database_engine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
if not isinstance(database_engine, PostgresEngine):
|
if not isinstance(database_engine, PostgresEngine):
|
||||||
return
|
return
|
||||||
|
|
||||||
next_id = find_max_generated_user_id_localpart(cur) + 1
|
next_id = find_max_generated_user_id_localpart(cur) + 1
|
||||||
cur.execute("CREATE SEQUENCE user_id_seq START WITH %s", (next_id,))
|
cur.execute("CREATE SEQUENCE user_id_seq START WITH %s", (next_id,))
|
||||||
|
|
||||||
|
|
||||||
def run_upgrade(*args, **kwargs):
|
|
||||||
pass
|
|
||||||
|
|
|
@ -20,18 +20,14 @@ import logging
|
||||||
from io import StringIO
|
from io import StringIO
|
||||||
|
|
||||||
from synapse.storage._base import db_to_json
|
from synapse.storage._base import db_to_json
|
||||||
|
from synapse.storage.database import LoggingTransaction
|
||||||
from synapse.storage.engines import BaseDatabaseEngine
|
from synapse.storage.engines import BaseDatabaseEngine
|
||||||
from synapse.storage.prepare_database import execute_statements_from_stream
|
from synapse.storage.prepare_database import execute_statements_from_stream
|
||||||
from synapse.storage.types import Cursor
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
def run_upgrade(cur: Cursor, database_engine: BaseDatabaseEngine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur: Cursor, database_engine: BaseDatabaseEngine, *args, **kwargs):
|
|
||||||
logger.info("Creating ignored_users table")
|
logger.info("Creating ignored_users table")
|
||||||
execute_statements_from_stream(cur, StringIO(_create_commands))
|
execute_statements_from_stream(cur, StringIO(_create_commands))
|
||||||
|
|
||||||
|
|
|
@ -16,11 +16,11 @@
|
||||||
This migration handles the process of changing the type of `room_depth.min_depth` to
|
This migration handles the process of changing the type of `room_depth.min_depth` to
|
||||||
a BIGINT.
|
a BIGINT.
|
||||||
"""
|
"""
|
||||||
|
from synapse.storage.database import LoggingTransaction
|
||||||
from synapse.storage.engines import BaseDatabaseEngine, PostgresEngine
|
from synapse.storage.engines import BaseDatabaseEngine, PostgresEngine
|
||||||
from synapse.storage.types import Cursor
|
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur: Cursor, database_engine: BaseDatabaseEngine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
if not isinstance(database_engine, PostgresEngine):
|
if not isinstance(database_engine, PostgresEngine):
|
||||||
# this only applies to postgres - sqlite does not distinguish between big and
|
# this only applies to postgres - sqlite does not distinguish between big and
|
||||||
# little ints.
|
# little ints.
|
||||||
|
@ -64,7 +64,3 @@ def run_create(cur: Cursor, database_engine: BaseDatabaseEngine, *args, **kwargs
|
||||||
(6103, 'replace_room_depth_min_depth', '{}', 'populate_room_depth2')
|
(6103, 'replace_room_depth_min_depth', '{}', 'populate_room_depth2')
|
||||||
"""
|
"""
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
def run_upgrade(cur: Cursor, database_engine: BaseDatabaseEngine, *args, **kwargs):
|
|
||||||
pass
|
|
||||||
|
|
|
@ -18,11 +18,11 @@ This migration adds triggers to the partial_state_events tables to enforce uniqu
|
||||||
|
|
||||||
Triggers cannot be expressed in .sql files, so we have to use a separate file.
|
Triggers cannot be expressed in .sql files, so we have to use a separate file.
|
||||||
"""
|
"""
|
||||||
|
from synapse.storage.database import LoggingTransaction
|
||||||
from synapse.storage.engines import BaseDatabaseEngine, PostgresEngine, Sqlite3Engine
|
from synapse.storage.engines import BaseDatabaseEngine, PostgresEngine, Sqlite3Engine
|
||||||
from synapse.storage.types import Cursor
|
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur: Cursor, database_engine: BaseDatabaseEngine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
# complain if the room_id in partial_state_events doesn't match
|
# complain if the room_id in partial_state_events doesn't match
|
||||||
# that in `events`. We already have a fk constraint which ensures that the event
|
# that in `events`. We already have a fk constraint which ensures that the event
|
||||||
# exists in `events`, so all we have to do is raise if there is a row with a
|
# exists in `events`, so all we have to do is raise if there is a row with a
|
||||||
|
|
|
@ -17,10 +17,11 @@
|
||||||
Adds a postgres SEQUENCE for generating application service transaction IDs.
|
Adds a postgres SEQUENCE for generating application service transaction IDs.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
from synapse.storage.engines import PostgresEngine
|
from synapse.storage.database import LoggingTransaction
|
||||||
|
from synapse.storage.engines import BaseDatabaseEngine, PostgresEngine
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur, database_engine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
if isinstance(database_engine, PostgresEngine):
|
if isinstance(database_engine, PostgresEngine):
|
||||||
# If we already have some AS TXNs we want to start from the current
|
# If we already have some AS TXNs we want to start from the current
|
||||||
# maximum value. There are two potential places this is stored - the
|
# maximum value. There are two potential places this is stored - the
|
||||||
|
@ -30,10 +31,12 @@ def run_create(cur, database_engine, *args, **kwargs):
|
||||||
|
|
||||||
cur.execute("SELECT COALESCE(max(txn_id), 0) FROM application_services_txns")
|
cur.execute("SELECT COALESCE(max(txn_id), 0) FROM application_services_txns")
|
||||||
row = cur.fetchone()
|
row = cur.fetchone()
|
||||||
|
assert row is not None
|
||||||
txn_max = row[0]
|
txn_max = row[0]
|
||||||
|
|
||||||
cur.execute("SELECT COALESCE(max(last_txn), 0) FROM application_services_state")
|
cur.execute("SELECT COALESCE(max(last_txn), 0) FROM application_services_state")
|
||||||
row = cur.fetchone()
|
row = cur.fetchone()
|
||||||
|
assert row is not None
|
||||||
last_txn_max = row[0]
|
last_txn_max = row[0]
|
||||||
|
|
||||||
start_val = max(last_txn_max, txn_max) + 1
|
start_val = max(last_txn_max, txn_max) + 1
|
||||||
|
|
|
@ -14,10 +14,11 @@
|
||||||
|
|
||||||
import json
|
import json
|
||||||
|
|
||||||
from synapse.storage.types import Cursor
|
from synapse.storage.database import LoggingTransaction
|
||||||
|
from synapse.storage.engines import BaseDatabaseEngine
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur: Cursor, database_engine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
"""Add a bg update to populate the `state_key` and `rejection_reason` columns of `events`"""
|
"""Add a bg update to populate the `state_key` and `rejection_reason` columns of `events`"""
|
||||||
|
|
||||||
# we know that any new events will have the columns populated (and that has been
|
# we know that any new events will have the columns populated (and that has been
|
||||||
|
@ -27,7 +28,9 @@ def run_create(cur: Cursor, database_engine, *args, **kwargs):
|
||||||
# current min and max stream orderings, since that is guaranteed to include all
|
# current min and max stream orderings, since that is guaranteed to include all
|
||||||
# the events that were stored before the new columns were added.
|
# the events that were stored before the new columns were added.
|
||||||
cur.execute("SELECT MIN(stream_ordering), MAX(stream_ordering) FROM events")
|
cur.execute("SELECT MIN(stream_ordering), MAX(stream_ordering) FROM events")
|
||||||
(min_stream_ordering, max_stream_ordering) = cur.fetchone()
|
row = cur.fetchone()
|
||||||
|
assert row is not None
|
||||||
|
(min_stream_ordering, max_stream_ordering) = row
|
||||||
|
|
||||||
if min_stream_ordering is None:
|
if min_stream_ordering is None:
|
||||||
# no rows, nothing to do.
|
# no rows, nothing to do.
|
||||||
|
|
|
@ -19,9 +19,16 @@ for its completion can be removed.
|
||||||
|
|
||||||
Note the background job must still remain defined in the database class.
|
Note the background job must still remain defined in the database class.
|
||||||
"""
|
"""
|
||||||
|
from synapse.config.homeserver import HomeServerConfig
|
||||||
|
from synapse.storage.database import LoggingTransaction
|
||||||
|
from synapse.storage.engines import BaseDatabaseEngine
|
||||||
|
|
||||||
|
|
||||||
def run_upgrade(cur, database_engine, *args, **kwargs):
|
def run_upgrade(
|
||||||
|
cur: LoggingTransaction,
|
||||||
|
database_engine: BaseDatabaseEngine,
|
||||||
|
config: HomeServerConfig,
|
||||||
|
) -> None:
|
||||||
cur.execute("SELECT update_name FROM background_updates")
|
cur.execute("SELECT update_name FROM background_updates")
|
||||||
rows = cur.fetchall()
|
rows = cur.fetchall()
|
||||||
for row in rows:
|
for row in rows:
|
||||||
|
|
|
@ -13,11 +13,11 @@
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
import json
|
import json
|
||||||
|
|
||||||
|
from synapse.storage.database import LoggingTransaction
|
||||||
from synapse.storage.engines import BaseDatabaseEngine, Sqlite3Engine
|
from synapse.storage.engines import BaseDatabaseEngine, Sqlite3Engine
|
||||||
from synapse.storage.types import Cursor
|
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur: Cursor, database_engine: BaseDatabaseEngine) -> None:
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
"""
|
"""
|
||||||
Upgrade the event_search table to use the porter tokenizer if it isn't already
|
Upgrade the event_search table to use the porter tokenizer if it isn't already
|
||||||
|
|
||||||
|
@ -38,6 +38,7 @@ def run_create(cur: Cursor, database_engine: BaseDatabaseEngine) -> None:
|
||||||
# Re-run the background job to re-populate the event_search table.
|
# Re-run the background job to re-populate the event_search table.
|
||||||
cur.execute("SELECT MIN(stream_ordering) FROM events")
|
cur.execute("SELECT MIN(stream_ordering) FROM events")
|
||||||
row = cur.fetchone()
|
row = cur.fetchone()
|
||||||
|
assert row is not None
|
||||||
min_stream_id = row[0]
|
min_stream_id = row[0]
|
||||||
|
|
||||||
# If there are not any events, nothing to do.
|
# If there are not any events, nothing to do.
|
||||||
|
@ -46,6 +47,7 @@ def run_create(cur: Cursor, database_engine: BaseDatabaseEngine) -> None:
|
||||||
|
|
||||||
cur.execute("SELECT MAX(stream_ordering) FROM events")
|
cur.execute("SELECT MAX(stream_ordering) FROM events")
|
||||||
row = cur.fetchone()
|
row = cur.fetchone()
|
||||||
|
assert row is not None
|
||||||
max_stream_id = row[0]
|
max_stream_id = row[0]
|
||||||
|
|
||||||
progress = {
|
progress = {
|
||||||
|
|
|
@ -17,11 +17,11 @@
|
||||||
This migration adds triggers to the room membership tables to enforce consistency.
|
This migration adds triggers to the room membership tables to enforce consistency.
|
||||||
Triggers cannot be expressed in .sql files, so we have to use a separate file.
|
Triggers cannot be expressed in .sql files, so we have to use a separate file.
|
||||||
"""
|
"""
|
||||||
|
from synapse.storage.database import LoggingTransaction
|
||||||
from synapse.storage.engines import BaseDatabaseEngine, PostgresEngine, Sqlite3Engine
|
from synapse.storage.engines import BaseDatabaseEngine, PostgresEngine, Sqlite3Engine
|
||||||
from synapse.storage.types import Cursor
|
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur: Cursor, database_engine: BaseDatabaseEngine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
# Complain if the `event_stream_ordering` in membership tables doesn't match
|
# Complain if the `event_stream_ordering` in membership tables doesn't match
|
||||||
# the `stream_ordering` row with the same `event_id` in `events`.
|
# the `stream_ordering` row with the same `event_id` in `events`.
|
||||||
if isinstance(database_engine, Sqlite3Engine):
|
if isinstance(database_engine, Sqlite3Engine):
|
||||||
|
|
|
@ -12,15 +12,17 @@
|
||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
|
||||||
from synapse.storage.engines import PostgresEngine
|
from synapse.storage.database import LoggingTransaction
|
||||||
|
from synapse.storage.engines import BaseDatabaseEngine, PostgresEngine
|
||||||
|
|
||||||
|
|
||||||
def run_create(cur, database_engine, *args, **kwargs):
|
def run_create(cur: LoggingTransaction, database_engine: BaseDatabaseEngine) -> None:
|
||||||
if isinstance(database_engine, PostgresEngine):
|
if isinstance(database_engine, PostgresEngine):
|
||||||
# if we already have some state groups, we want to start making new
|
# if we already have some state groups, we want to start making new
|
||||||
# ones with a higher id.
|
# ones with a higher id.
|
||||||
cur.execute("SELECT max(id) FROM state_groups")
|
cur.execute("SELECT max(id) FROM state_groups")
|
||||||
row = cur.fetchone()
|
row = cur.fetchone()
|
||||||
|
assert row is not None
|
||||||
|
|
||||||
if row[0] is None:
|
if row[0] is None:
|
||||||
start_val = 1
|
start_val = 1
|
||||||
|
@ -28,7 +30,3 @@ def run_create(cur, database_engine, *args, **kwargs):
|
||||||
start_val = row[0] + 1
|
start_val = row[0] + 1
|
||||||
|
|
||||||
cur.execute("CREATE SEQUENCE state_group_id_seq START WITH %s", (start_val,))
|
cur.execute("CREATE SEQUENCE state_group_id_seq START WITH %s", (start_val,))
|
||||||
|
|
||||||
|
|
||||||
def run_upgrade(*args, **kwargs):
|
|
||||||
pass
|
|
||||||
|
|
Loading…
Reference in a new issue