mirror of
https://mau.dev/maunium/synapse.git
synced 2024-11-17 07:21:37 +01:00
Use monotonic clock where possible for metrics
Fixes intermittent errors observed on Apple hardware which were caused by time.clock() appearing to go backwards when called from different threads. Also fixes a bug where database activity times were logged as 1/1000 of their correct ratio due to confusion between milliseconds and seconds.
This commit is contained in:
parent
4f68188d0b
commit
15bf32e062
2 changed files with 15 additions and 6 deletions
1
changelog.d/5498.bugfix
Normal file
1
changelog.d/5498.bugfix
Normal file
|
@ -0,0 +1 @@
|
||||||
|
Fix intermittent exceptions on Apple hardware. Also fix bug that caused database activity times to be under-reported in log lines.
|
|
@ -38,6 +38,14 @@ from synapse.util.caches.descriptors import Cache
|
||||||
from synapse.util.logcontext import LoggingContext, PreserveLoggingContext
|
from synapse.util.logcontext import LoggingContext, PreserveLoggingContext
|
||||||
from synapse.util.stringutils import exception_to_unicode
|
from synapse.util.stringutils import exception_to_unicode
|
||||||
|
|
||||||
|
# import a function which will return a monotonic time, in seconds
|
||||||
|
try:
|
||||||
|
# on python 3, use time.monotonic, since time.clock can go backwards
|
||||||
|
from time import monotonic as monotonic_time
|
||||||
|
except ImportError:
|
||||||
|
# ... but python 2 doesn't have it
|
||||||
|
from time import clock as monotonic_time
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
|
@ -352,14 +360,14 @@ class SQLBaseStore(object):
|
||||||
)
|
)
|
||||||
|
|
||||||
def start_profiling(self):
|
def start_profiling(self):
|
||||||
self._previous_loop_ts = self._clock.time_msec()
|
self._previous_loop_ts = monotonic_time()
|
||||||
|
|
||||||
def loop():
|
def loop():
|
||||||
curr = self._current_txn_total_time
|
curr = self._current_txn_total_time
|
||||||
prev = self._previous_txn_total_time
|
prev = self._previous_txn_total_time
|
||||||
self._previous_txn_total_time = curr
|
self._previous_txn_total_time = curr
|
||||||
|
|
||||||
time_now = self._clock.time_msec()
|
time_now = monotonic_time()
|
||||||
time_then = self._previous_loop_ts
|
time_then = self._previous_loop_ts
|
||||||
self._previous_loop_ts = time_now
|
self._previous_loop_ts = time_now
|
||||||
|
|
||||||
|
@ -385,7 +393,7 @@ class SQLBaseStore(object):
|
||||||
def _new_transaction(
|
def _new_transaction(
|
||||||
self, conn, desc, after_callbacks, exception_callbacks, func, *args, **kwargs
|
self, conn, desc, after_callbacks, exception_callbacks, func, *args, **kwargs
|
||||||
):
|
):
|
||||||
start = time.time()
|
start = monotonic_time()
|
||||||
txn_id = self._TXN_ID
|
txn_id = self._TXN_ID
|
||||||
|
|
||||||
# We don't really need these to be unique, so lets stop it from
|
# We don't really need these to be unique, so lets stop it from
|
||||||
|
@ -451,7 +459,7 @@ class SQLBaseStore(object):
|
||||||
logger.debug("[TXN FAIL] {%s} %s", name, e)
|
logger.debug("[TXN FAIL] {%s} %s", name, e)
|
||||||
raise
|
raise
|
||||||
finally:
|
finally:
|
||||||
end = time.time()
|
end = monotonic_time()
|
||||||
duration = end - start
|
duration = end - start
|
||||||
|
|
||||||
LoggingContext.current_context().add_database_transaction(duration)
|
LoggingContext.current_context().add_database_transaction(duration)
|
||||||
|
@ -525,11 +533,11 @@ class SQLBaseStore(object):
|
||||||
)
|
)
|
||||||
parent_context = None
|
parent_context = None
|
||||||
|
|
||||||
start_time = time.time()
|
start_time = monotonic_time()
|
||||||
|
|
||||||
def inner_func(conn, *args, **kwargs):
|
def inner_func(conn, *args, **kwargs):
|
||||||
with LoggingContext("runWithConnection", parent_context) as context:
|
with LoggingContext("runWithConnection", parent_context) as context:
|
||||||
sched_duration_sec = time.time() - start_time
|
sched_duration_sec = monotonic_time() - start_time
|
||||||
sql_scheduling_timer.observe(sched_duration_sec)
|
sql_scheduling_timer.observe(sched_duration_sec)
|
||||||
context.add_database_scheduled(sched_duration_sec)
|
context.add_database_scheduled(sched_duration_sec)
|
||||||
|
|
||||||
|
|
Loading…
Reference in a new issue