0
0
Fork 1
mirror of https://mau.dev/maunium/synapse.git synced 2024-12-13 21:33:20 +01:00

Various clean-ups to the logging context code (#8935)

This commit is contained in:
Patrick Cloke 2020-12-14 14:19:47 -05:00 committed by GitHub
parent 895e04319b
commit 1619802228
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
11 changed files with 20 additions and 39 deletions

View file

@ -1 +1 @@
Improve structured logging tests. Various clean-ups to the structured logging and logging context code.

1
changelog.d/8935.misc Normal file
View file

@ -0,0 +1 @@
Various clean-ups to the structured logging and logging context code.

View file

@ -206,7 +206,7 @@ def _setup_stdlib_logging(config, log_config_path, logBeginner: LogBeginner) ->
# filter options, but care must when using e.g. MemoryHandler to buffer # filter options, but care must when using e.g. MemoryHandler to buffer
# writes. # writes.
log_context_filter = LoggingContextFilter(request="") log_context_filter = LoggingContextFilter()
log_metadata_filter = MetadataFilter({"server_name": config.server_name}) log_metadata_filter = MetadataFilter({"server_name": config.server_name})
old_factory = logging.getLogRecordFactory() old_factory = logging.getLogRecordFactory()

View file

@ -128,8 +128,7 @@ class SynapseRequest(Request):
# create a LogContext for this request # create a LogContext for this request
request_id = self.get_request_id() request_id = self.get_request_id()
logcontext = self.logcontext = LoggingContext(request_id) self.logcontext = LoggingContext(request_id, request=request_id)
logcontext.request = request_id
# override the Server header which is set by twisted # override the Server header which is set by twisted
self.setHeader("Server", self.site.server_version_string) self.setHeader("Server", self.site.server_version_string)

View file

@ -203,10 +203,6 @@ class _Sentinel:
def copy_to(self, record): def copy_to(self, record):
pass pass
def copy_to_twisted_log_entry(self, record):
record["request"] = None
record["scope"] = None
def start(self, rusage: "Optional[resource._RUsage]"): def start(self, rusage: "Optional[resource._RUsage]"):
pass pass
@ -372,13 +368,6 @@ class LoggingContext:
# we also track the current scope: # we also track the current scope:
record.scope = self.scope record.scope = self.scope
def copy_to_twisted_log_entry(self, record) -> None:
"""
Copy logging fields from this context to a Twisted log record.
"""
record["request"] = self.request
record["scope"] = self.scope
def start(self, rusage: "Optional[resource._RUsage]") -> None: def start(self, rusage: "Optional[resource._RUsage]") -> None:
""" """
Record that this logcontext is currently running. Record that this logcontext is currently running.
@ -542,13 +531,10 @@ class LoggingContext:
class LoggingContextFilter(logging.Filter): class LoggingContextFilter(logging.Filter):
"""Logging filter that adds values from the current logging context to each """Logging filter that adds values from the current logging context to each
record. record.
Args:
**defaults: Default values to avoid formatters complaining about
missing fields
""" """
def __init__(self, **defaults) -> None: def __init__(self, request: str = ""):
self.defaults = defaults self._default_request = request
def filter(self, record) -> Literal[True]: def filter(self, record) -> Literal[True]:
"""Add each fields from the logging contexts to the record. """Add each fields from the logging contexts to the record.
@ -556,14 +542,14 @@ class LoggingContextFilter(logging.Filter):
True to include the record in the log output. True to include the record in the log output.
""" """
context = current_context() context = current_context()
for key, value in self.defaults.items(): record.request = self._default_request
setattr(record, key, value)
# context should never be None, but if it somehow ends up being, then # context should never be None, but if it somehow ends up being, then
# we end up in a death spiral of infinite loops, so let's check, for # we end up in a death spiral of infinite loops, so let's check, for
# robustness' sake. # robustness' sake.
if context is not None: if context is not None:
context.copy_to(record) # Logging is interested in the request.
record.request = context.request
return True return True

View file

@ -199,8 +199,7 @@ def run_as_background_process(desc: str, func, *args, bg_start_span=True, **kwar
_background_process_start_count.labels(desc).inc() _background_process_start_count.labels(desc).inc()
_background_process_in_flight_count.labels(desc).inc() _background_process_in_flight_count.labels(desc).inc()
with BackgroundProcessLoggingContext(desc) as context: with BackgroundProcessLoggingContext(desc, "%s-%i" % (desc, count)) as context:
context.request = "%s-%i" % (desc, count)
try: try:
ctx = noop_context_manager() ctx = noop_context_manager()
if bg_start_span: if bg_start_span:
@ -244,8 +243,8 @@ class BackgroundProcessLoggingContext(LoggingContext):
__slots__ = ["_proc"] __slots__ = ["_proc"]
def __init__(self, name: str): def __init__(self, name: str, request: Optional[str] = None):
super().__init__(name) super().__init__(name, request=request)
self._proc = _BackgroundProcess(name, self) self._proc = _BackgroundProcess(name, self)

View file

@ -172,8 +172,7 @@ class BaseReplicationStreamProtocol(LineOnlyReceiver):
# a logcontext which we use for processing incoming commands. We declare it as a # a logcontext which we use for processing incoming commands. We declare it as a
# background process so that the CPU stats get reported to prometheus. # background process so that the CPU stats get reported to prometheus.
ctx_name = "replication-conn-%s" % self.conn_id ctx_name = "replication-conn-%s" % self.conn_id
self._logging_context = BackgroundProcessLoggingContext(ctx_name) self._logging_context = BackgroundProcessLoggingContext(ctx_name, ctx_name)
self._logging_context.request = ctx_name
def connectionMade(self): def connectionMade(self):
logger.info("[%s] Connection established", self.id()) logger.info("[%s] Connection established", self.id())

View file

@ -126,7 +126,7 @@ class FederationTestCase(unittest.HomeserverTestCase):
room_version, room_version,
) )
with LoggingContext(request="send_rejected"): with LoggingContext("send_rejected"):
d = run_in_background(self.handler.on_receive_pdu, OTHER_SERVER, ev) d = run_in_background(self.handler.on_receive_pdu, OTHER_SERVER, ev)
self.get_success(d) self.get_success(d)
@ -178,7 +178,7 @@ class FederationTestCase(unittest.HomeserverTestCase):
room_version, room_version,
) )
with LoggingContext(request="send_rejected"): with LoggingContext("send_rejected"):
d = run_in_background(self.handler.on_receive_pdu, OTHER_SERVER, ev) d = run_in_background(self.handler.on_receive_pdu, OTHER_SERVER, ev)
self.get_success(d) self.get_success(d)
@ -198,7 +198,7 @@ class FederationTestCase(unittest.HomeserverTestCase):
# the auth code requires that a signature exists, but doesn't check that # the auth code requires that a signature exists, but doesn't check that
# signature... go figure. # signature... go figure.
join_event.signatures[other_server] = {"x": "y"} join_event.signatures[other_server] = {"x": "y"}
with LoggingContext(request="send_join"): with LoggingContext("send_join"):
d = run_in_background( d = run_in_background(
self.handler.on_send_join_request, other_server, join_event self.handler.on_send_join_request, other_server, join_event
) )

View file

@ -117,11 +117,10 @@ class TerseJsonTestCase(LoggerCleanupMixin, TestCase):
""" """
handler = logging.StreamHandler(self.output) handler = logging.StreamHandler(self.output)
handler.setFormatter(JsonFormatter()) handler.setFormatter(JsonFormatter())
handler.addFilter(LoggingContextFilter(request="")) handler.addFilter(LoggingContextFilter())
logger = self.get_logger(handler) logger = self.get_logger(handler)
with LoggingContext() as context_one: with LoggingContext(request="test"):
context_one.request = "test"
logger.info("Hello there, %s!", "wally") logger.info("Hello there, %s!", "wally")
log = self.get_log_line() log = self.get_log_line()
@ -132,9 +131,7 @@ class TerseJsonTestCase(LoggerCleanupMixin, TestCase):
"level", "level",
"namespace", "namespace",
"request", "request",
"scope",
] ]
self.assertCountEqual(log.keys(), expected_log_keys) self.assertCountEqual(log.keys(), expected_log_keys)
self.assertEqual(log["log"], "Hello there, wally!") self.assertEqual(log["log"], "Hello there, wally!")
self.assertEqual(log["request"], "test") self.assertEqual(log["request"], "test")
self.assertIsNone(log["scope"])

View file

@ -134,7 +134,7 @@ class MessageAcceptTests(unittest.HomeserverTestCase):
} }
) )
with LoggingContext(request="lying_event"): with LoggingContext():
failure = self.get_failure( failure = self.get_failure(
self.handler.on_receive_pdu( self.handler.on_receive_pdu(
"test.serv", lying_event, sent_to_us_directly=True "test.serv", lying_event, sent_to_us_directly=True

View file

@ -48,7 +48,7 @@ def setup_logging():
handler = ToTwistedHandler() handler = ToTwistedHandler()
formatter = logging.Formatter(log_format) formatter = logging.Formatter(log_format)
handler.setFormatter(formatter) handler.setFormatter(formatter)
handler.addFilter(LoggingContextFilter(request="")) handler.addFilter(LoggingContextFilter())
root_logger.addHandler(handler) root_logger.addHandler(handler)
log_level = os.environ.get("SYNAPSE_TEST_LOG_LEVEL", "ERROR") log_level = os.environ.get("SYNAPSE_TEST_LOG_LEVEL", "ERROR")