forked from MirrorHub/synapse
Code style.
This commit is contained in:
parent
71da2bed55
commit
b8d30899b1
3 changed files with 85 additions and 46 deletions
|
@ -790,8 +790,11 @@ class _TransactionQueue(object):
|
||||||
retry_timings.retry_last_ts, retry_timings.retry_interval
|
retry_timings.retry_last_ts, retry_timings.retry_interval
|
||||||
)
|
)
|
||||||
if retry_last_ts + retry_interval > int(self._clock.time_msec()):
|
if retry_last_ts + retry_interval > int(self._clock.time_msec()):
|
||||||
logger.info("TX [%s] not ready for retry yet - "
|
logger.info(
|
||||||
"dropping transaction for now", destination)
|
"TX [%s] not ready for retry yet - "
|
||||||
|
"dropping transaction for now",
|
||||||
|
destination,
|
||||||
|
)
|
||||||
return
|
return
|
||||||
else:
|
else:
|
||||||
logger.info("TX [%s] is ready for retry", destination)
|
logger.info("TX [%s] is ready for retry", destination)
|
||||||
|
@ -811,10 +814,14 @@ class _TransactionQueue(object):
|
||||||
if not pending_pdus and not pending_edus and not pending_failures:
|
if not pending_pdus and not pending_edus and not pending_failures:
|
||||||
return
|
return
|
||||||
|
|
||||||
logger.debug("TX [%s] Attempting new transaction "
|
logger.debug(
|
||||||
|
"TX [%s] Attempting new transaction "
|
||||||
"(pdus: %d, edus: %d, failures: %d)",
|
"(pdus: %d, edus: %d, failures: %d)",
|
||||||
destination,
|
destination,
|
||||||
len(pending_pdus), len(pending_edus), len(pending_failures))
|
len(pending_pdus),
|
||||||
|
len(pending_edus),
|
||||||
|
len(pending_failures)
|
||||||
|
)
|
||||||
|
|
||||||
# Sort based on the order field
|
# Sort based on the order field
|
||||||
pending_pdus.sort(key=lambda t: t[2])
|
pending_pdus.sort(key=lambda t: t[2])
|
||||||
|
@ -847,8 +854,11 @@ class _TransactionQueue(object):
|
||||||
yield self.transaction_actions.prepare_to_send(transaction)
|
yield self.transaction_actions.prepare_to_send(transaction)
|
||||||
|
|
||||||
logger.debug("TX [%s] Persisted transaction", destination)
|
logger.debug("TX [%s] Persisted transaction", destination)
|
||||||
logger.info("TX [%s] Sending transaction [%s]", destination,
|
logger.info(
|
||||||
transaction.transaction_id)
|
"TX [%s] Sending transaction [%s]",
|
||||||
|
destination,
|
||||||
|
transaction.transaction_id,
|
||||||
|
)
|
||||||
|
|
||||||
# Actually send the transaction
|
# Actually send the transaction
|
||||||
|
|
||||||
|
@ -905,8 +915,11 @@ class _TransactionQueue(object):
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
# We capture this here as there as nothing actually listens
|
# We capture this here as there as nothing actually listens
|
||||||
# for this finishing functions deferred.
|
# for this finishing functions deferred.
|
||||||
logger.warn("TX [%s] Problem in _attempt_transaction: %s",
|
logger.warn(
|
||||||
destination, e)
|
"TX [%s] Problem in _attempt_transaction: %s",
|
||||||
|
destination,
|
||||||
|
e,
|
||||||
|
)
|
||||||
|
|
||||||
self.set_retrying(destination, retry_interval)
|
self.set_retrying(destination, retry_interval)
|
||||||
|
|
||||||
|
@ -925,6 +938,7 @@ class _TransactionQueue(object):
|
||||||
def set_retrying(self, destination, retry_interval):
|
def set_retrying(self, destination, retry_interval):
|
||||||
# track that this destination is having problems and we should
|
# track that this destination is having problems and we should
|
||||||
# give it a chance to recover before trying it again
|
# give it a chance to recover before trying it again
|
||||||
|
|
||||||
if retry_interval:
|
if retry_interval:
|
||||||
retry_interval *= 2
|
retry_interval *= 2
|
||||||
# plateau at hourly retries for now
|
# plateau at hourly retries for now
|
||||||
|
@ -932,5 +946,9 @@ class _TransactionQueue(object):
|
||||||
retry_interval = 60 * 60 * 1000
|
retry_interval = 60 * 60 * 1000
|
||||||
else:
|
else:
|
||||||
retry_interval = 2000 # try again at first after 2 seconds
|
retry_interval = 2000 # try again at first after 2 seconds
|
||||||
yield self.store.set_destination_retry_timings(destination,
|
|
||||||
int(self._clock.time_msec()), retry_interval)
|
yield self.store.set_destination_retry_timings(
|
||||||
|
destination,
|
||||||
|
int(self._clock.time_msec()),
|
||||||
|
retry_interval
|
||||||
|
)
|
||||||
|
|
|
@ -130,12 +130,20 @@ class MatrixFederationHttpClient(object):
|
||||||
break
|
break
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
if not retry_on_dns_fail and isinstance(e, DNSLookupError):
|
if not retry_on_dns_fail and isinstance(e, DNSLookupError):
|
||||||
logger.warn("DNS Lookup failed to %s with %s", destination,
|
logger.warn(
|
||||||
e)
|
"DNS Lookup failed to %s with %s",
|
||||||
|
destination,
|
||||||
|
e
|
||||||
|
)
|
||||||
raise SynapseError(400, "Domain specified not found.")
|
raise SynapseError(400, "Domain specified not found.")
|
||||||
|
|
||||||
logger.warn("Sending request failed to %s: %s %s : %s",
|
logger.warn(
|
||||||
destination, method, url_bytes, e)
|
"Sending request failed to %s: %s %s : %s",
|
||||||
|
destination,
|
||||||
|
method,
|
||||||
|
url_bytes,
|
||||||
|
e
|
||||||
|
)
|
||||||
_print_ex(e)
|
_print_ex(e)
|
||||||
|
|
||||||
if retries_left:
|
if retries_left:
|
||||||
|
@ -144,9 +152,14 @@ class MatrixFederationHttpClient(object):
|
||||||
else:
|
else:
|
||||||
raise
|
raise
|
||||||
|
|
||||||
logger.info("Received response %d %s for %s: %s %s",
|
logger.info(
|
||||||
response.code, response.phrase,
|
"Received response %d %s for %s: %s %s",
|
||||||
destination, method, url_bytes)
|
response.code,
|
||||||
|
response.phrase,
|
||||||
|
destination,
|
||||||
|
method,
|
||||||
|
url_bytes
|
||||||
|
)
|
||||||
|
|
||||||
if 200 <= response.code < 300:
|
if 200 <= response.code < 300:
|
||||||
# We need to update the transactions table to say it was sent?
|
# We need to update the transactions table to say it was sent?
|
||||||
|
|
|
@ -251,16 +251,22 @@ class TransactionStore(SQLBaseStore):
|
||||||
"""
|
"""
|
||||||
|
|
||||||
self.destination_retry_cache[destination] = (
|
self.destination_retry_cache[destination] = (
|
||||||
DestinationsTable.EntryType(destination,
|
DestinationsTable.EntryType(
|
||||||
retry_last_ts, retry_interval)
|
destination,
|
||||||
|
retry_last_ts,
|
||||||
|
retry_interval
|
||||||
|
)
|
||||||
)
|
)
|
||||||
|
|
||||||
# XXX: we could chose to not bother persisting this if our cache thinks
|
# XXX: we could chose to not bother persisting this if our cache thinks
|
||||||
# this is a NOOP
|
# this is a NOOP
|
||||||
return self.runInteraction(
|
return self.runInteraction(
|
||||||
"set_destination_retry_timings",
|
"set_destination_retry_timings",
|
||||||
self._set_destination_retry_timings, destination,
|
self._set_destination_retry_timings,
|
||||||
retry_last_ts, retry_interval)
|
destination,
|
||||||
|
retry_last_ts,
|
||||||
|
retry_interval,
|
||||||
|
)
|
||||||
|
|
||||||
def _set_destination_retry_timings(cls, txn, destination,
|
def _set_destination_retry_timings(cls, txn, destination,
|
||||||
retry_last_ts, retry_interval):
|
retry_last_ts, retry_interval):
|
||||||
|
@ -291,6 +297,7 @@ class TransactionStore(SQLBaseStore):
|
||||||
txn.execute(query)
|
txn.execute(query)
|
||||||
return DestinationsTable.decode_results(txn.fetchall())
|
return DestinationsTable.decode_results(txn.fetchall())
|
||||||
|
|
||||||
|
|
||||||
class ReceivedTransactionsTable(Table):
|
class ReceivedTransactionsTable(Table):
|
||||||
table_name = "received_transactions"
|
table_name = "received_transactions"
|
||||||
|
|
||||||
|
@ -333,6 +340,7 @@ class TransactionsToPduTable(Table):
|
||||||
|
|
||||||
EntryType = namedtuple("TransactionsToPduEntry", fields)
|
EntryType = namedtuple("TransactionsToPduEntry", fields)
|
||||||
|
|
||||||
|
|
||||||
class DestinationsTable(Table):
|
class DestinationsTable(Table):
|
||||||
table_name = "destinations"
|
table_name = "destinations"
|
||||||
|
|
||||||
|
|
Loading…
Reference in a new issue