synapse/synapse/http/site.py

386 lines
14 KiB
Python
Raw Normal View History

2016-04-22 17:09:55 +02:00
# Copyright 2016 OpenMarket Ltd
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import contextlib
import logging
import time
from typing import Optional, Union
2016-04-22 17:09:55 +02:00
from twisted.python.failure import Failure
2018-07-09 08:09:20 +02:00
from twisted.web.server import Request, Site
from synapse.config.server import ListenerConfig
from synapse.http import get_request_user_agent, redact_uri
from synapse.http.request_metrics import RequestMetrics, requests_counter
from synapse.logging.context import LoggingContext, PreserveLoggingContext
from synapse.types import Requester
logger = logging.getLogger(__name__)
_next_request_seq = 0
2016-04-22 17:09:55 +02:00
class SynapseRequest(Request):
"""Class which encapsulates an HTTP request to synapse.
All of the requests processed in synapse are of this type.
It extends twisted's twisted.web.server.Request, and adds:
* Unique request ID
* A log context associated with the request
* Redaction of access_token query-params in __repr__
* Logging at start and end
* Metrics to record CPU, wallclock and DB time by endpoint.
It also provides a method `processing`, which returns a context manager. If this
method is called, the request won't be logged until the context manager is closed;
this is useful for asynchronous request handlers which may go on processing the
request even after the client has disconnected.
Attributes:
2020-05-22 11:12:17 +02:00
logcontext: the log context for this request
"""
2019-06-20 11:32:02 +02:00
def __init__(self, channel, *args, **kw):
Request.__init__(self, channel, *args, **kw)
self.site = channel.site
2019-06-20 11:32:02 +02:00
self._channel = channel # this is used by the tests
2020-05-22 11:12:17 +02:00
self.start_time = 0.0
2016-04-22 17:09:55 +02:00
# The requester, if authenticated. For federation requests this is the
# server name, for client requests this is the Requester object.
self.requester = None # type: Optional[Union[Requester, str]]
# we can't yet create the logcontext, as we don't know the method.
2020-05-22 11:12:17 +02:00
self.logcontext = None # type: Optional[LoggingContext]
global _next_request_seq
self.request_seq = _next_request_seq
_next_request_seq += 1
# whether an asynchronous request handler has called processing()
self._is_processing = False
# the time when the asynchronous request handler completed its processing
self._processing_finished_time = None
# what time we finished sending the response to the client (or the connection
# dropped)
self.finish_time = None
2016-04-22 17:09:55 +02:00
def __repr__(self):
# We overwrite this so that we don't log ``access_token``
2019-06-20 11:32:02 +02:00
return "<%s at 0x%x method=%r uri=%r clientproto=%r site=%r>" % (
2016-04-22 17:09:55 +02:00
self.__class__.__name__,
id(self),
self.get_method(),
2016-04-22 17:09:55 +02:00
self.get_redacted_uri(),
2019-06-20 11:32:02 +02:00
self.clientproto.decode("ascii", errors="replace"),
2016-04-22 17:09:55 +02:00
self.site.site_tag,
)
def get_request_id(self):
return "%s-%i" % (self.get_method(), self.request_seq)
2016-04-22 17:09:55 +02:00
def get_redacted_uri(self):
2018-09-13 19:57:02 +02:00
uri = self.uri
if isinstance(uri, bytes):
uri = self.uri.decode("ascii", errors="replace")
2018-09-13 19:57:02 +02:00
return redact_uri(uri)
2016-04-22 17:09:55 +02:00
def get_method(self):
"""Gets the method associated with the request (or placeholder if not
method has yet been received).
Note: This is necessary as the placeholder value in twisted is str
rather than bytes, so we need to sanitise `self.method`.
Returns:
str
"""
method = self.method
if isinstance(method, bytes):
2019-06-20 11:32:02 +02:00
method = self.method.decode("ascii")
return method
def render(self, resrc):
# this is called once a Resource has been found to serve the request; in our
# case the Resource in question will normally be a JsonResource.
# create a LogContext for this request
request_id = self.get_request_id()
self.logcontext = LoggingContext(request_id, request=request_id)
# override the Server header which is set by twisted
self.setHeader("Server", self.site.server_version_string)
with PreserveLoggingContext(self.logcontext):
# we start the request metrics timer here with an initial stab
# at the servlet name. For most requests that name will be
# JsonResource (or a subclass), and JsonResource._async_render
# will update it once it picks a servlet.
servlet_name = resrc.__class__.__name__
self._started_processing(servlet_name)
Request.render(self, resrc)
# record the arrival of the request *after*
# dispatching to the handler, so that the handler
# can update the servlet name in the request
# metrics
2019-06-20 11:32:02 +02:00
requests_counter.labels(self.get_method(), self.request_metrics.name).inc()
@contextlib.contextmanager
def processing(self):
"""Record the fact that we are processing this request.
Returns a context manager; the correct way to use this is:
async def handle_request(request):
with request.processing("FooServlet"):
await really_handle_the_request()
Once the context manager is closed, the completion of the request will be logged,
and the various metrics will be updated.
"""
if self._is_processing:
raise RuntimeError("Request is already processing")
self._is_processing = True
try:
yield
except Exception:
# this should already have been caught, and sent back to the client as a 500.
2020-10-23 18:38:40 +02:00
logger.exception(
"Asynchronous message handler raised an uncaught exception"
)
finally:
# the request handler has finished its work and either sent the whole response
# back, or handed over responsibility to a Producer.
self._processing_finished_time = time.time()
self._is_processing = False
# if we've already sent the response, log it now; otherwise, we wait for the
# response to be sent.
if self.finish_time is not None:
self._finished_processing()
def finish(self):
"""Called when all response data has been written to this Request.
Overrides twisted.web.server.Request.finish to record the finish time and do
logging.
"""
self.finish_time = time.time()
Request.finish(self)
if not self._is_processing:
2020-05-22 11:12:17 +02:00
assert self.logcontext is not None
with PreserveLoggingContext(self.logcontext):
self._finished_processing()
def connectionLost(self, reason):
"""Called when the client connection is closed before the response is written.
Overrides twisted.web.server.Request.connectionLost to record the finish time and
do logging.
"""
# There is a bug in Twisted where reason is not wrapped in a Failure object
# Detect this and wrap it manually as a workaround
# More information: https://github.com/matrix-org/synapse/issues/7441
if not isinstance(reason, Failure):
reason = Failure(reason)
self.finish_time = time.time()
Request.connectionLost(self, reason)
if self.logcontext is None:
logger.info(
"Connection from %s lost before request headers were read", self.client
)
return
# we only get here if the connection to the client drops before we send
# the response.
#
# It's useful to log it here so that we can get an idea of when
# the client disconnects.
with PreserveLoggingContext(self.logcontext):
logger.info("Connection from client lost before response was sent")
if not self._is_processing:
self._finished_processing()
def _started_processing(self, servlet_name):
"""Record the fact that we are processing this request.
This will log the request's arrival. Once the request completes,
be sure to call finished_processing.
Args:
servlet_name (str): the name of the servlet which will be
processing this request. This is used in the metrics.
It is possible to update this afterwards by updating
self.request_metrics.name.
"""
2018-05-28 11:10:27 +02:00
self.start_time = time.time()
self.request_metrics = RequestMetrics()
self.request_metrics.start(
2019-06-20 11:32:02 +02:00
self.start_time, name=servlet_name, method=self.get_method()
)
self.site.access_logger.debug(
2016-04-22 17:09:55 +02:00
"%s - %s - Received request: %s %s",
self.getClientIP(),
self.site.site_tag,
self.get_method(),
2019-06-20 11:32:02 +02:00
self.get_redacted_uri(),
2016-04-22 17:09:55 +02:00
)
def _finished_processing(self):
"""Log the completion of this request and update the metrics"""
2020-05-22 11:12:17 +02:00
assert self.logcontext is not None
usage = self.logcontext.get_resource_usage()
if self._processing_finished_time is None:
# we completed the request without anything calling processing()
self._processing_finished_time = time.time()
2016-04-22 17:09:55 +02:00
# the time between receiving the request and the request handler finishing
processing_time = self._processing_finished_time - self.start_time
# the time between the request handler finishing and the response being sent
# to the client (nb may be negative)
response_send_time = self.finish_time - self._processing_finished_time
# Convert the requester into a string that we can log
authenticated_entity = None
if isinstance(self.requester, str):
authenticated_entity = self.requester
elif isinstance(self.requester, Requester):
authenticated_entity = self.requester.authenticated_entity
# If this is a request where the target user doesn't match the user who
# authenticated (e.g. and admin is puppetting a user) then we log both.
if self.requester.user.to_string() != authenticated_entity:
authenticated_entity = "{},{}".format(
authenticated_entity,
self.requester.user.to_string(),
)
elif self.requester is not None:
# This shouldn't happen, but we log it so we don't lose information
# and can see that we're doing something wrong.
authenticated_entity = repr(self.requester) # type: ignore[unreachable]
2018-07-01 12:56:33 +02:00
user_agent = get_request_user_agent(self, "-")
code = str(self.code)
if not self.finished:
# we didn't send the full response before we gave up (presumably because
# the connection dropped)
code += "!"
2018-07-01 12:56:33 +02:00
2020-08-07 15:21:24 +02:00
log_level = logging.INFO if self._should_log_request() else logging.DEBUG
self.site.access_logger.log(
log_level,
2016-04-22 17:09:55 +02:00
"%s - %s - {%s}"
" Processed request: %.3fsec/%.3fsec (%.3fsec, %.3fsec) (%.3fsec/%.3fsec/%d)"
2019-06-20 11:32:02 +02:00
' %sB %s "%s %s %s" "%s" [%d dbevts]',
2016-04-22 17:09:55 +02:00
self.getClientIP(),
self.site.site_tag,
2018-07-01 12:56:33 +02:00
authenticated_entity,
processing_time,
response_send_time,
usage.ru_utime,
usage.ru_stime,
usage.db_sched_duration_sec,
usage.db_txn_duration_sec,
int(usage.db_txn_count),
2016-04-22 17:09:55 +02:00
self.sentLength,
code,
self.get_method(),
2016-04-22 17:09:55 +02:00
self.get_redacted_uri(),
2019-06-20 11:32:02 +02:00
self.clientproto.decode("ascii", errors="replace"),
2018-07-01 12:56:33 +02:00
user_agent,
usage.evt_db_fetch_count,
2016-04-22 17:09:55 +02:00
)
try:
2018-09-17 18:16:50 +02:00
self.request_metrics.stop(self.finish_time, self.code, self.sentLength)
except Exception as e:
logger.warning("Failed to stop metrics: %r", e)
2020-08-07 15:21:24 +02:00
def _should_log_request(self) -> bool:
"""Whether we should log at INFO that we processed the request."""
if self.path == b"/health":
return False
if self.method == b"OPTIONS":
return False
return True
2020-08-07 15:21:24 +02:00
2016-04-22 17:09:55 +02:00
class XForwardedForRequest(SynapseRequest):
def __init__(self, *args, **kw):
SynapseRequest.__init__(self, *args, **kw)
"""
Add a layer on top of another request that only uses the value of an
X-Forwarded-For header as the result of C{getClientIP}.
"""
2019-06-20 11:32:02 +02:00
2016-04-22 17:09:55 +02:00
def getClientIP(self):
"""
@return: The client address (the first address) in the value of the
I{X-Forwarded-For header}. If the header is not present, return
C{b"-"}.
"""
2019-06-20 11:32:02 +02:00
return (
self.requestHeaders.getRawHeaders(b"x-forwarded-for", [b"-"])[0]
.split(b",")[0]
.strip()
.decode("ascii")
)
2016-04-22 17:09:55 +02:00
class SynapseSite(Site):
"""
Subclass of a twisted http Site that does access logging with python's
standard logging
"""
2019-06-20 11:32:02 +02:00
def __init__(
self,
logger_name,
site_tag,
config: ListenerConfig,
2019-06-20 11:32:02 +02:00
resource,
server_version_string,
*args,
**kwargs
):
2016-04-22 17:09:55 +02:00
Site.__init__(self, resource, *args, **kwargs)
self.site_tag = site_tag
assert config.http_options is not None
proxied = config.http_options.x_forwarded
self.requestFactory = XForwardedForRequest if proxied else SynapseRequest
2016-04-22 17:09:55 +02:00
self.access_logger = logging.getLogger(logger_name)
2019-06-20 11:32:02 +02:00
self.server_version_string = server_version_string.encode("ascii")
2016-04-22 17:09:55 +02:00
def log(self, request):
pass