mirror of
https://mau.dev/maunium/synapse.git
synced 2024-11-17 07:21:37 +01:00
Merge pull request #5597 from matrix-org/erikj/admin_api_cmd
Create basic admin command app
This commit is contained in:
commit
c831c5b2bb
8 changed files with 357 additions and 13 deletions
1
changelog.d/5597.feature
Normal file
1
changelog.d/5597.feature
Normal file
|
@ -0,0 +1 @@
|
||||||
|
Add a basic admin command app to allow server operators to run Synapse admin commands separately from the main production instance.
|
|
@ -48,7 +48,7 @@ def register_sighup(func):
|
||||||
_sighup_callbacks.append(func)
|
_sighup_callbacks.append(func)
|
||||||
|
|
||||||
|
|
||||||
def start_worker_reactor(appname, config):
|
def start_worker_reactor(appname, config, run_command=reactor.run):
|
||||||
""" Run the reactor in the main process
|
""" Run the reactor in the main process
|
||||||
|
|
||||||
Daemonizes if necessary, and then configures some resources, before starting
|
Daemonizes if necessary, and then configures some resources, before starting
|
||||||
|
@ -57,6 +57,7 @@ def start_worker_reactor(appname, config):
|
||||||
Args:
|
Args:
|
||||||
appname (str): application name which will be sent to syslog
|
appname (str): application name which will be sent to syslog
|
||||||
config (synapse.config.Config): config object
|
config (synapse.config.Config): config object
|
||||||
|
run_command (Callable[]): callable that actually runs the reactor
|
||||||
"""
|
"""
|
||||||
|
|
||||||
logger = logging.getLogger(config.worker_app)
|
logger = logging.getLogger(config.worker_app)
|
||||||
|
@ -69,11 +70,19 @@ def start_worker_reactor(appname, config):
|
||||||
daemonize=config.worker_daemonize,
|
daemonize=config.worker_daemonize,
|
||||||
print_pidfile=config.print_pidfile,
|
print_pidfile=config.print_pidfile,
|
||||||
logger=logger,
|
logger=logger,
|
||||||
|
run_command=run_command,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
def start_reactor(
|
def start_reactor(
|
||||||
appname, soft_file_limit, gc_thresholds, pid_file, daemonize, print_pidfile, logger
|
appname,
|
||||||
|
soft_file_limit,
|
||||||
|
gc_thresholds,
|
||||||
|
pid_file,
|
||||||
|
daemonize,
|
||||||
|
print_pidfile,
|
||||||
|
logger,
|
||||||
|
run_command=reactor.run,
|
||||||
):
|
):
|
||||||
""" Run the reactor in the main process
|
""" Run the reactor in the main process
|
||||||
|
|
||||||
|
@ -88,6 +97,7 @@ def start_reactor(
|
||||||
daemonize (bool): true to run the reactor in a background process
|
daemonize (bool): true to run the reactor in a background process
|
||||||
print_pidfile (bool): whether to print the pid file, if daemonize is True
|
print_pidfile (bool): whether to print the pid file, if daemonize is True
|
||||||
logger (logging.Logger): logger instance to pass to Daemonize
|
logger (logging.Logger): logger instance to pass to Daemonize
|
||||||
|
run_command (Callable[]): callable that actually runs the reactor
|
||||||
"""
|
"""
|
||||||
|
|
||||||
install_dns_limiter(reactor)
|
install_dns_limiter(reactor)
|
||||||
|
@ -97,7 +107,7 @@ def start_reactor(
|
||||||
change_resource_limit(soft_file_limit)
|
change_resource_limit(soft_file_limit)
|
||||||
if gc_thresholds:
|
if gc_thresholds:
|
||||||
gc.set_threshold(*gc_thresholds)
|
gc.set_threshold(*gc_thresholds)
|
||||||
reactor.run()
|
run_command()
|
||||||
|
|
||||||
# make sure that we run the reactor with the sentinel log context,
|
# make sure that we run the reactor with the sentinel log context,
|
||||||
# otherwise other PreserveLoggingContext instances will get confused
|
# otherwise other PreserveLoggingContext instances will get confused
|
||||||
|
|
264
synapse/app/admin_cmd.py
Normal file
264
synapse/app/admin_cmd.py
Normal file
|
@ -0,0 +1,264 @@
|
||||||
|
#!/usr/bin/env python
|
||||||
|
# -*- coding: utf-8 -*-
|
||||||
|
# Copyright 2019 Matrix.org Foundation C.I.C.
|
||||||
|
#
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
# you may not use this file except in compliance with the License.
|
||||||
|
# You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
# See the License for the specific language governing permissions and
|
||||||
|
# limitations under the License.
|
||||||
|
import argparse
|
||||||
|
import logging
|
||||||
|
import os
|
||||||
|
import sys
|
||||||
|
import tempfile
|
||||||
|
|
||||||
|
from canonicaljson import json
|
||||||
|
|
||||||
|
from twisted.internet import defer, task
|
||||||
|
|
||||||
|
import synapse
|
||||||
|
from synapse.app import _base
|
||||||
|
from synapse.config._base import ConfigError
|
||||||
|
from synapse.config.homeserver import HomeServerConfig
|
||||||
|
from synapse.config.logger import setup_logging
|
||||||
|
from synapse.handlers.admin import ExfiltrationWriter
|
||||||
|
from synapse.replication.slave.storage._base import BaseSlavedStore
|
||||||
|
from synapse.replication.slave.storage.account_data import SlavedAccountDataStore
|
||||||
|
from synapse.replication.slave.storage.appservice import SlavedApplicationServiceStore
|
||||||
|
from synapse.replication.slave.storage.client_ips import SlavedClientIpStore
|
||||||
|
from synapse.replication.slave.storage.deviceinbox import SlavedDeviceInboxStore
|
||||||
|
from synapse.replication.slave.storage.devices import SlavedDeviceStore
|
||||||
|
from synapse.replication.slave.storage.events import SlavedEventStore
|
||||||
|
from synapse.replication.slave.storage.filtering import SlavedFilteringStore
|
||||||
|
from synapse.replication.slave.storage.groups import SlavedGroupServerStore
|
||||||
|
from synapse.replication.slave.storage.presence import SlavedPresenceStore
|
||||||
|
from synapse.replication.slave.storage.push_rule import SlavedPushRuleStore
|
||||||
|
from synapse.replication.slave.storage.receipts import SlavedReceiptsStore
|
||||||
|
from synapse.replication.slave.storage.registration import SlavedRegistrationStore
|
||||||
|
from synapse.replication.slave.storage.room import RoomStore
|
||||||
|
from synapse.replication.tcp.client import ReplicationClientHandler
|
||||||
|
from synapse.server import HomeServer
|
||||||
|
from synapse.storage.engines import create_engine
|
||||||
|
from synapse.util.logcontext import LoggingContext
|
||||||
|
from synapse.util.versionstring import get_version_string
|
||||||
|
|
||||||
|
logger = logging.getLogger("synapse.app.admin_cmd")
|
||||||
|
|
||||||
|
|
||||||
|
class AdminCmdSlavedStore(
|
||||||
|
SlavedReceiptsStore,
|
||||||
|
SlavedAccountDataStore,
|
||||||
|
SlavedApplicationServiceStore,
|
||||||
|
SlavedRegistrationStore,
|
||||||
|
SlavedFilteringStore,
|
||||||
|
SlavedPresenceStore,
|
||||||
|
SlavedGroupServerStore,
|
||||||
|
SlavedDeviceInboxStore,
|
||||||
|
SlavedDeviceStore,
|
||||||
|
SlavedPushRuleStore,
|
||||||
|
SlavedEventStore,
|
||||||
|
SlavedClientIpStore,
|
||||||
|
RoomStore,
|
||||||
|
BaseSlavedStore,
|
||||||
|
):
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
class AdminCmdServer(HomeServer):
|
||||||
|
DATASTORE_CLASS = AdminCmdSlavedStore
|
||||||
|
|
||||||
|
def _listen_http(self, listener_config):
|
||||||
|
pass
|
||||||
|
|
||||||
|
def start_listening(self, listeners):
|
||||||
|
pass
|
||||||
|
|
||||||
|
def build_tcp_replication(self):
|
||||||
|
return AdminCmdReplicationHandler(self)
|
||||||
|
|
||||||
|
|
||||||
|
class AdminCmdReplicationHandler(ReplicationClientHandler):
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def on_rdata(self, stream_name, token, rows):
|
||||||
|
pass
|
||||||
|
|
||||||
|
def get_streams_to_replicate(self):
|
||||||
|
return {}
|
||||||
|
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def export_data_command(hs, args):
|
||||||
|
"""Export data for a user.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
hs (HomeServer)
|
||||||
|
args (argparse.Namespace)
|
||||||
|
"""
|
||||||
|
|
||||||
|
user_id = args.user_id
|
||||||
|
directory = args.output_directory
|
||||||
|
|
||||||
|
res = yield hs.get_handlers().admin_handler.export_user_data(
|
||||||
|
user_id, FileExfiltrationWriter(user_id, directory=directory)
|
||||||
|
)
|
||||||
|
print(res)
|
||||||
|
|
||||||
|
|
||||||
|
class FileExfiltrationWriter(ExfiltrationWriter):
|
||||||
|
"""An ExfiltrationWriter that writes the users data to a directory.
|
||||||
|
Returns the directory location on completion.
|
||||||
|
|
||||||
|
Note: This writes to disk on the main reactor thread.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
user_id (str): The user whose data is being exfiltrated.
|
||||||
|
directory (str|None): The directory to write the data to, if None then
|
||||||
|
will write to a temporary directory.
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(self, user_id, directory=None):
|
||||||
|
self.user_id = user_id
|
||||||
|
|
||||||
|
if directory:
|
||||||
|
self.base_directory = directory
|
||||||
|
else:
|
||||||
|
self.base_directory = tempfile.mkdtemp(
|
||||||
|
prefix="synapse-exfiltrate__%s__" % (user_id,)
|
||||||
|
)
|
||||||
|
|
||||||
|
os.makedirs(self.base_directory, exist_ok=True)
|
||||||
|
if list(os.listdir(self.base_directory)):
|
||||||
|
raise Exception("Directory must be empty")
|
||||||
|
|
||||||
|
def write_events(self, room_id, events):
|
||||||
|
room_directory = os.path.join(self.base_directory, "rooms", room_id)
|
||||||
|
os.makedirs(room_directory, exist_ok=True)
|
||||||
|
events_file = os.path.join(room_directory, "events")
|
||||||
|
|
||||||
|
with open(events_file, "a") as f:
|
||||||
|
for event in events:
|
||||||
|
print(json.dumps(event.get_pdu_json()), file=f)
|
||||||
|
|
||||||
|
def write_state(self, room_id, event_id, state):
|
||||||
|
room_directory = os.path.join(self.base_directory, "rooms", room_id)
|
||||||
|
state_directory = os.path.join(room_directory, "state")
|
||||||
|
os.makedirs(state_directory, exist_ok=True)
|
||||||
|
|
||||||
|
event_file = os.path.join(state_directory, event_id)
|
||||||
|
|
||||||
|
with open(event_file, "a") as f:
|
||||||
|
for event in state.values():
|
||||||
|
print(json.dumps(event.get_pdu_json()), file=f)
|
||||||
|
|
||||||
|
def write_invite(self, room_id, event, state):
|
||||||
|
self.write_events(room_id, [event])
|
||||||
|
|
||||||
|
# We write the invite state somewhere else as they aren't full events
|
||||||
|
# and are only a subset of the state at the event.
|
||||||
|
room_directory = os.path.join(self.base_directory, "rooms", room_id)
|
||||||
|
os.makedirs(room_directory, exist_ok=True)
|
||||||
|
|
||||||
|
invite_state = os.path.join(room_directory, "invite_state")
|
||||||
|
|
||||||
|
with open(invite_state, "a") as f:
|
||||||
|
for event in state.values():
|
||||||
|
print(json.dumps(event), file=f)
|
||||||
|
|
||||||
|
def finished(self):
|
||||||
|
return self.base_directory
|
||||||
|
|
||||||
|
|
||||||
|
def start(config_options):
|
||||||
|
parser = argparse.ArgumentParser(description="Synapse Admin Command")
|
||||||
|
HomeServerConfig.add_arguments_to_parser(parser)
|
||||||
|
|
||||||
|
subparser = parser.add_subparsers(
|
||||||
|
title="Admin Commands",
|
||||||
|
required=True,
|
||||||
|
dest="command",
|
||||||
|
metavar="<admin_command>",
|
||||||
|
help="The admin command to perform.",
|
||||||
|
)
|
||||||
|
export_data_parser = subparser.add_parser(
|
||||||
|
"export-data", help="Export all data for a user"
|
||||||
|
)
|
||||||
|
export_data_parser.add_argument("user_id", help="User to extra data from")
|
||||||
|
export_data_parser.add_argument(
|
||||||
|
"--output-directory",
|
||||||
|
action="store",
|
||||||
|
metavar="DIRECTORY",
|
||||||
|
required=False,
|
||||||
|
help="The directory to store the exported data in. Must be empty. Defaults"
|
||||||
|
" to creating a temp directory.",
|
||||||
|
)
|
||||||
|
export_data_parser.set_defaults(func=export_data_command)
|
||||||
|
|
||||||
|
try:
|
||||||
|
config, args = HomeServerConfig.load_config_with_parser(parser, config_options)
|
||||||
|
except ConfigError as e:
|
||||||
|
sys.stderr.write("\n" + str(e) + "\n")
|
||||||
|
sys.exit(1)
|
||||||
|
|
||||||
|
if config.worker_app is not None:
|
||||||
|
assert config.worker_app == "synapse.app.admin_cmd"
|
||||||
|
|
||||||
|
# Update the config with some basic overrides so that don't have to specify
|
||||||
|
# a full worker config.
|
||||||
|
config.worker_app = "synapse.app.admin_cmd"
|
||||||
|
|
||||||
|
if (
|
||||||
|
not config.worker_daemonize
|
||||||
|
and not config.worker_log_file
|
||||||
|
and not config.worker_log_config
|
||||||
|
):
|
||||||
|
# Since we're meant to be run as a "command" let's not redirect stdio
|
||||||
|
# unless we've actually set log config.
|
||||||
|
config.no_redirect_stdio = True
|
||||||
|
|
||||||
|
# Explicitly disable background processes
|
||||||
|
config.update_user_directory = False
|
||||||
|
config.start_pushers = False
|
||||||
|
config.send_federation = False
|
||||||
|
|
||||||
|
setup_logging(config, use_worker_options=True)
|
||||||
|
|
||||||
|
synapse.events.USE_FROZEN_DICTS = config.use_frozen_dicts
|
||||||
|
|
||||||
|
database_engine = create_engine(config.database_config)
|
||||||
|
|
||||||
|
ss = AdminCmdServer(
|
||||||
|
config.server_name,
|
||||||
|
db_config=config.database_config,
|
||||||
|
config=config,
|
||||||
|
version_string="Synapse/" + get_version_string(synapse),
|
||||||
|
database_engine=database_engine,
|
||||||
|
)
|
||||||
|
|
||||||
|
ss.setup()
|
||||||
|
|
||||||
|
# We use task.react as the basic run command as it correctly handles tearing
|
||||||
|
# down the reactor when the deferreds resolve and setting the return value.
|
||||||
|
# We also make sure that `_base.start` gets run before we actually run the
|
||||||
|
# command.
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def run(_reactor):
|
||||||
|
with LoggingContext("command"):
|
||||||
|
yield _base.start(ss, [])
|
||||||
|
yield args.func(ss, args)
|
||||||
|
|
||||||
|
_base.start_worker_reactor(
|
||||||
|
"synapse-admin-cmd", config, run_command=lambda: task.react(run)
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
with LoggingContext("main"):
|
||||||
|
start(sys.argv[1:])
|
|
@ -137,12 +137,42 @@ class Config(object):
|
||||||
return file_stream.read()
|
return file_stream.read()
|
||||||
|
|
||||||
def invoke_all(self, name, *args, **kargs):
|
def invoke_all(self, name, *args, **kargs):
|
||||||
|
"""Invoke all instance methods with the given name and arguments in the
|
||||||
|
class's MRO.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
name (str): Name of function to invoke
|
||||||
|
*args
|
||||||
|
**kwargs
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
list: The list of the return values from each method called
|
||||||
|
"""
|
||||||
results = []
|
results = []
|
||||||
for cls in type(self).mro():
|
for cls in type(self).mro():
|
||||||
if name in cls.__dict__:
|
if name in cls.__dict__:
|
||||||
results.append(getattr(cls, name)(self, *args, **kargs))
|
results.append(getattr(cls, name)(self, *args, **kargs))
|
||||||
return results
|
return results
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def invoke_all_static(cls, name, *args, **kargs):
|
||||||
|
"""Invoke all static methods with the given name and arguments in the
|
||||||
|
class's MRO.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
name (str): Name of function to invoke
|
||||||
|
*args
|
||||||
|
**kwargs
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
list: The list of the return values from each method called
|
||||||
|
"""
|
||||||
|
results = []
|
||||||
|
for c in cls.mro():
|
||||||
|
if name in c.__dict__:
|
||||||
|
results.append(getattr(c, name)(*args, **kargs))
|
||||||
|
return results
|
||||||
|
|
||||||
def generate_config(
|
def generate_config(
|
||||||
self,
|
self,
|
||||||
config_dir_path,
|
config_dir_path,
|
||||||
|
@ -202,6 +232,23 @@ class Config(object):
|
||||||
Returns: Config object.
|
Returns: Config object.
|
||||||
"""
|
"""
|
||||||
config_parser = argparse.ArgumentParser(description=description)
|
config_parser = argparse.ArgumentParser(description=description)
|
||||||
|
cls.add_arguments_to_parser(config_parser)
|
||||||
|
obj, _ = cls.load_config_with_parser(config_parser, argv)
|
||||||
|
|
||||||
|
return obj
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def add_arguments_to_parser(cls, config_parser):
|
||||||
|
"""Adds all the config flags to an ArgumentParser.
|
||||||
|
|
||||||
|
Doesn't support config-file-generation: used by the worker apps.
|
||||||
|
|
||||||
|
Used for workers where we want to add extra flags/subcommands.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
config_parser (ArgumentParser): App description
|
||||||
|
"""
|
||||||
|
|
||||||
config_parser.add_argument(
|
config_parser.add_argument(
|
||||||
"-c",
|
"-c",
|
||||||
"--config-path",
|
"--config-path",
|
||||||
|
@ -219,16 +266,34 @@ class Config(object):
|
||||||
" Defaults to the directory containing the last config file",
|
" Defaults to the directory containing the last config file",
|
||||||
)
|
)
|
||||||
|
|
||||||
|
cls.invoke_all_static("add_arguments", config_parser)
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def load_config_with_parser(cls, parser, argv):
|
||||||
|
"""Parse the commandline and config files with the given parser
|
||||||
|
|
||||||
|
Doesn't support config-file-generation: used by the worker apps.
|
||||||
|
|
||||||
|
Used for workers where we want to add extra flags/subcommands.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
parser (ArgumentParser)
|
||||||
|
argv (list[str])
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
tuple[HomeServerConfig, argparse.Namespace]: Returns the parsed
|
||||||
|
config object and the parsed argparse.Namespace object from
|
||||||
|
`parser.parse_args(..)`
|
||||||
|
"""
|
||||||
|
|
||||||
obj = cls()
|
obj = cls()
|
||||||
|
|
||||||
obj.invoke_all("add_arguments", config_parser)
|
config_args = parser.parse_args(argv)
|
||||||
|
|
||||||
config_args = config_parser.parse_args(argv)
|
|
||||||
|
|
||||||
config_files = find_config_files(search_paths=config_args.config_path)
|
config_files = find_config_files(search_paths=config_args.config_path)
|
||||||
|
|
||||||
if not config_files:
|
if not config_files:
|
||||||
config_parser.error("Must supply a config file.")
|
parser.error("Must supply a config file.")
|
||||||
|
|
||||||
if config_args.keys_directory:
|
if config_args.keys_directory:
|
||||||
config_dir_path = config_args.keys_directory
|
config_dir_path = config_args.keys_directory
|
||||||
|
@ -244,7 +309,7 @@ class Config(object):
|
||||||
|
|
||||||
obj.invoke_all("read_arguments", config_args)
|
obj.invoke_all("read_arguments", config_args)
|
||||||
|
|
||||||
return obj
|
return obj, config_args
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def load_or_generate_config(cls, description, argv):
|
def load_or_generate_config(cls, description, argv):
|
||||||
|
@ -401,7 +466,7 @@ class Config(object):
|
||||||
formatter_class=argparse.RawDescriptionHelpFormatter,
|
formatter_class=argparse.RawDescriptionHelpFormatter,
|
||||||
)
|
)
|
||||||
|
|
||||||
obj.invoke_all("add_arguments", parser)
|
obj.invoke_all_static("add_arguments", parser)
|
||||||
args = parser.parse_args(remaining_args)
|
args = parser.parse_args(remaining_args)
|
||||||
|
|
||||||
config_dict = read_config_files(config_files)
|
config_dict = read_config_files(config_files)
|
||||||
|
|
|
@ -69,7 +69,8 @@ class DatabaseConfig(Config):
|
||||||
if database_path is not None:
|
if database_path is not None:
|
||||||
self.database_config["args"]["database"] = database_path
|
self.database_config["args"]["database"] = database_path
|
||||||
|
|
||||||
def add_arguments(self, parser):
|
@staticmethod
|
||||||
|
def add_arguments(parser):
|
||||||
db_group = parser.add_argument_group("database")
|
db_group = parser.add_argument_group("database")
|
||||||
db_group.add_argument(
|
db_group.add_argument(
|
||||||
"-d",
|
"-d",
|
||||||
|
|
|
@ -103,7 +103,8 @@ class LoggingConfig(Config):
|
||||||
if args.log_file is not None:
|
if args.log_file is not None:
|
||||||
self.log_file = args.log_file
|
self.log_file = args.log_file
|
||||||
|
|
||||||
def add_arguments(cls, parser):
|
@staticmethod
|
||||||
|
def add_arguments(parser):
|
||||||
logging_group = parser.add_argument_group("logging")
|
logging_group = parser.add_argument_group("logging")
|
||||||
logging_group.add_argument(
|
logging_group.add_argument(
|
||||||
"-v",
|
"-v",
|
||||||
|
|
|
@ -237,7 +237,8 @@ class RegistrationConfig(Config):
|
||||||
% locals()
|
% locals()
|
||||||
)
|
)
|
||||||
|
|
||||||
def add_arguments(self, parser):
|
@staticmethod
|
||||||
|
def add_arguments(parser):
|
||||||
reg_group = parser.add_argument_group("registration")
|
reg_group = parser.add_argument_group("registration")
|
||||||
reg_group.add_argument(
|
reg_group.add_argument(
|
||||||
"--enable-registration",
|
"--enable-registration",
|
||||||
|
|
|
@ -639,7 +639,8 @@ class ServerConfig(Config):
|
||||||
if args.print_pidfile is not None:
|
if args.print_pidfile is not None:
|
||||||
self.print_pidfile = args.print_pidfile
|
self.print_pidfile = args.print_pidfile
|
||||||
|
|
||||||
def add_arguments(self, parser):
|
@staticmethod
|
||||||
|
def add_arguments(parser):
|
||||||
server_group = parser.add_argument_group("server")
|
server_group = parser.add_argument_group("server")
|
||||||
server_group.add_argument(
|
server_group.add_argument(
|
||||||
"-D",
|
"-D",
|
||||||
|
|
Loading…
Reference in a new issue