mirror of
https://mau.dev/maunium/synapse.git
synced 2024-12-30 09:24:40 +01:00
Add type hints for HTTP and email pushers. (#8880)
This commit is contained in:
parent
02e588856a
commit
92d87c6882
8 changed files with 168 additions and 101 deletions
1
changelog.d/8880.misc
Normal file
1
changelog.d/8880.misc
Normal file
|
@ -0,0 +1 @@
|
||||||
|
Add type hints to push module.
|
3
mypy.ini
3
mypy.ini
|
@ -55,7 +55,10 @@ files =
|
||||||
synapse/metrics,
|
synapse/metrics,
|
||||||
synapse/module_api,
|
synapse/module_api,
|
||||||
synapse/notifier.py,
|
synapse/notifier.py,
|
||||||
|
synapse/push/emailpusher.py,
|
||||||
|
synapse/push/httppusher.py,
|
||||||
synapse/push/mailer.py,
|
synapse/push/mailer.py,
|
||||||
|
synapse/push/pusher.py,
|
||||||
synapse/push/pusherpool.py,
|
synapse/push/pusherpool.py,
|
||||||
synapse/push/push_rule_evaluator.py,
|
synapse/push/push_rule_evaluator.py,
|
||||||
synapse/replication,
|
synapse/replication,
|
||||||
|
|
|
@ -13,6 +13,56 @@
|
||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
|
||||||
|
import abc
|
||||||
|
from typing import TYPE_CHECKING, Any, Dict, Optional
|
||||||
|
|
||||||
|
from synapse.types import RoomStreamToken
|
||||||
|
|
||||||
|
if TYPE_CHECKING:
|
||||||
|
from synapse.app.homeserver import HomeServer
|
||||||
|
|
||||||
|
|
||||||
|
class Pusher(metaclass=abc.ABCMeta):
|
||||||
|
def __init__(self, hs: "HomeServer", pusherdict: Dict[str, Any]):
|
||||||
|
self.hs = hs
|
||||||
|
self.store = self.hs.get_datastore()
|
||||||
|
self.clock = self.hs.get_clock()
|
||||||
|
|
||||||
|
self.pusher_id = pusherdict["id"]
|
||||||
|
self.user_id = pusherdict["user_name"]
|
||||||
|
self.app_id = pusherdict["app_id"]
|
||||||
|
self.pushkey = pusherdict["pushkey"]
|
||||||
|
|
||||||
|
# This is the highest stream ordering we know it's safe to process.
|
||||||
|
# When new events arrive, we'll be given a window of new events: we
|
||||||
|
# should honour this rather than just looking for anything higher
|
||||||
|
# because of potential out-of-order event serialisation. This starts
|
||||||
|
# off as None though as we don't know any better.
|
||||||
|
self.max_stream_ordering = None # type: Optional[int]
|
||||||
|
|
||||||
|
@abc.abstractmethod
|
||||||
|
def on_new_notifications(self, max_token: RoomStreamToken) -> None:
|
||||||
|
raise NotImplementedError()
|
||||||
|
|
||||||
|
@abc.abstractmethod
|
||||||
|
def on_new_receipts(self, min_stream_id: int, max_stream_id: int) -> None:
|
||||||
|
raise NotImplementedError()
|
||||||
|
|
||||||
|
@abc.abstractmethod
|
||||||
|
def on_started(self, have_notifs: bool) -> None:
|
||||||
|
"""Called when this pusher has been started.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
should_check_for_notifs: Whether we should immediately
|
||||||
|
check for push to send. Set to False only if it's known there
|
||||||
|
is nothing to send
|
||||||
|
"""
|
||||||
|
raise NotImplementedError()
|
||||||
|
|
||||||
|
@abc.abstractmethod
|
||||||
|
def on_stop(self) -> None:
|
||||||
|
raise NotImplementedError()
|
||||||
|
|
||||||
|
|
||||||
class PusherConfigException(Exception):
|
class PusherConfigException(Exception):
|
||||||
"""An error occurred when creating a pusher."""
|
"""An error occurred when creating a pusher."""
|
||||||
|
|
|
@ -14,12 +14,19 @@
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
|
from typing import TYPE_CHECKING, Any, Dict, List, Optional
|
||||||
|
|
||||||
|
from twisted.internet.base import DelayedCall
|
||||||
from twisted.internet.error import AlreadyCalled, AlreadyCancelled
|
from twisted.internet.error import AlreadyCalled, AlreadyCancelled
|
||||||
|
|
||||||
from synapse.metrics.background_process_metrics import run_as_background_process
|
from synapse.metrics.background_process_metrics import run_as_background_process
|
||||||
|
from synapse.push import Pusher
|
||||||
|
from synapse.push.mailer import Mailer
|
||||||
from synapse.types import RoomStreamToken
|
from synapse.types import RoomStreamToken
|
||||||
|
|
||||||
|
if TYPE_CHECKING:
|
||||||
|
from synapse.app.homeserver import HomeServer
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
# The amount of time we always wait before ever emailing about a notification
|
# The amount of time we always wait before ever emailing about a notification
|
||||||
|
@ -46,7 +53,7 @@ THROTTLE_RESET_AFTER_MS = 12 * 60 * 60 * 1000
|
||||||
INCLUDE_ALL_UNREAD_NOTIFS = False
|
INCLUDE_ALL_UNREAD_NOTIFS = False
|
||||||
|
|
||||||
|
|
||||||
class EmailPusher:
|
class EmailPusher(Pusher):
|
||||||
"""
|
"""
|
||||||
A pusher that sends email notifications about events (approximately)
|
A pusher that sends email notifications about events (approximately)
|
||||||
when they happen.
|
when they happen.
|
||||||
|
@ -54,37 +61,31 @@ class EmailPusher:
|
||||||
factor out the common parts
|
factor out the common parts
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, hs, pusherdict, mailer):
|
def __init__(self, hs: "HomeServer", pusherdict: Dict[str, Any], mailer: Mailer):
|
||||||
self.hs = hs
|
super().__init__(hs, pusherdict)
|
||||||
self.mailer = mailer
|
self.mailer = mailer
|
||||||
|
|
||||||
self.store = self.hs.get_datastore()
|
self.store = self.hs.get_datastore()
|
||||||
self.clock = self.hs.get_clock()
|
|
||||||
self.pusher_id = pusherdict["id"]
|
|
||||||
self.user_id = pusherdict["user_name"]
|
|
||||||
self.app_id = pusherdict["app_id"]
|
|
||||||
self.email = pusherdict["pushkey"]
|
self.email = pusherdict["pushkey"]
|
||||||
self.last_stream_ordering = pusherdict["last_stream_ordering"]
|
self.last_stream_ordering = pusherdict["last_stream_ordering"]
|
||||||
self.timed_call = None
|
self.timed_call = None # type: Optional[DelayedCall]
|
||||||
self.throttle_params = None
|
self.throttle_params = {} # type: Dict[str, Dict[str, int]]
|
||||||
|
self._inited = False
|
||||||
# See httppusher
|
|
||||||
self.max_stream_ordering = None
|
|
||||||
|
|
||||||
self._is_processing = False
|
self._is_processing = False
|
||||||
|
|
||||||
def on_started(self, should_check_for_notifs):
|
def on_started(self, should_check_for_notifs: bool) -> None:
|
||||||
"""Called when this pusher has been started.
|
"""Called when this pusher has been started.
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
should_check_for_notifs (bool): Whether we should immediately
|
should_check_for_notifs: Whether we should immediately
|
||||||
check for push to send. Set to False only if it's known there
|
check for push to send. Set to False only if it's known there
|
||||||
is nothing to send
|
is nothing to send
|
||||||
"""
|
"""
|
||||||
if should_check_for_notifs and self.mailer is not None:
|
if should_check_for_notifs and self.mailer is not None:
|
||||||
self._start_processing()
|
self._start_processing()
|
||||||
|
|
||||||
def on_stop(self):
|
def on_stop(self) -> None:
|
||||||
if self.timed_call:
|
if self.timed_call:
|
||||||
try:
|
try:
|
||||||
self.timed_call.cancel()
|
self.timed_call.cancel()
|
||||||
|
@ -92,7 +93,7 @@ class EmailPusher:
|
||||||
pass
|
pass
|
||||||
self.timed_call = None
|
self.timed_call = None
|
||||||
|
|
||||||
def on_new_notifications(self, max_token: RoomStreamToken):
|
def on_new_notifications(self, max_token: RoomStreamToken) -> None:
|
||||||
# We just use the minimum stream ordering and ignore the vector clock
|
# We just use the minimum stream ordering and ignore the vector clock
|
||||||
# component. This is safe to do as long as we *always* ignore the vector
|
# component. This is safe to do as long as we *always* ignore the vector
|
||||||
# clock components.
|
# clock components.
|
||||||
|
@ -106,23 +107,23 @@ class EmailPusher:
|
||||||
self.max_stream_ordering = max_stream_ordering
|
self.max_stream_ordering = max_stream_ordering
|
||||||
self._start_processing()
|
self._start_processing()
|
||||||
|
|
||||||
def on_new_receipts(self, min_stream_id, max_stream_id):
|
def on_new_receipts(self, min_stream_id: int, max_stream_id: int) -> None:
|
||||||
# We could wake up and cancel the timer but there tend to be quite a
|
# We could wake up and cancel the timer but there tend to be quite a
|
||||||
# lot of read receipts so it's probably less work to just let the
|
# lot of read receipts so it's probably less work to just let the
|
||||||
# timer fire
|
# timer fire
|
||||||
pass
|
pass
|
||||||
|
|
||||||
def on_timer(self):
|
def on_timer(self) -> None:
|
||||||
self.timed_call = None
|
self.timed_call = None
|
||||||
self._start_processing()
|
self._start_processing()
|
||||||
|
|
||||||
def _start_processing(self):
|
def _start_processing(self) -> None:
|
||||||
if self._is_processing:
|
if self._is_processing:
|
||||||
return
|
return
|
||||||
|
|
||||||
run_as_background_process("emailpush.process", self._process)
|
run_as_background_process("emailpush.process", self._process)
|
||||||
|
|
||||||
def _pause_processing(self):
|
def _pause_processing(self) -> None:
|
||||||
"""Used by tests to temporarily pause processing of events.
|
"""Used by tests to temporarily pause processing of events.
|
||||||
|
|
||||||
Asserts that its not currently processing.
|
Asserts that its not currently processing.
|
||||||
|
@ -130,25 +131,26 @@ class EmailPusher:
|
||||||
assert not self._is_processing
|
assert not self._is_processing
|
||||||
self._is_processing = True
|
self._is_processing = True
|
||||||
|
|
||||||
def _resume_processing(self):
|
def _resume_processing(self) -> None:
|
||||||
"""Used by tests to resume processing of events after pausing.
|
"""Used by tests to resume processing of events after pausing.
|
||||||
"""
|
"""
|
||||||
assert self._is_processing
|
assert self._is_processing
|
||||||
self._is_processing = False
|
self._is_processing = False
|
||||||
self._start_processing()
|
self._start_processing()
|
||||||
|
|
||||||
async def _process(self):
|
async def _process(self) -> None:
|
||||||
# we should never get here if we are already processing
|
# we should never get here if we are already processing
|
||||||
assert not self._is_processing
|
assert not self._is_processing
|
||||||
|
|
||||||
try:
|
try:
|
||||||
self._is_processing = True
|
self._is_processing = True
|
||||||
|
|
||||||
if self.throttle_params is None:
|
if not self._inited:
|
||||||
# this is our first loop: load up the throttle params
|
# this is our first loop: load up the throttle params
|
||||||
self.throttle_params = await self.store.get_throttle_params_by_room(
|
self.throttle_params = await self.store.get_throttle_params_by_room(
|
||||||
self.pusher_id
|
self.pusher_id
|
||||||
)
|
)
|
||||||
|
self._inited = True
|
||||||
|
|
||||||
# if the max ordering changes while we're running _unsafe_process,
|
# if the max ordering changes while we're running _unsafe_process,
|
||||||
# call it again, and so on until we've caught up.
|
# call it again, and so on until we've caught up.
|
||||||
|
@ -163,17 +165,19 @@ class EmailPusher:
|
||||||
finally:
|
finally:
|
||||||
self._is_processing = False
|
self._is_processing = False
|
||||||
|
|
||||||
async def _unsafe_process(self):
|
async def _unsafe_process(self) -> None:
|
||||||
"""
|
"""
|
||||||
Main logic of the push loop without the wrapper function that sets
|
Main logic of the push loop without the wrapper function that sets
|
||||||
up logging, measures and guards against multiple instances of it
|
up logging, measures and guards against multiple instances of it
|
||||||
being run.
|
being run.
|
||||||
"""
|
"""
|
||||||
start = 0 if INCLUDE_ALL_UNREAD_NOTIFS else self.last_stream_ordering
|
start = 0 if INCLUDE_ALL_UNREAD_NOTIFS else self.last_stream_ordering
|
||||||
fn = self.store.get_unread_push_actions_for_user_in_range_for_email
|
assert self.max_stream_ordering is not None
|
||||||
unprocessed = await fn(self.user_id, start, self.max_stream_ordering)
|
unprocessed = await self.store.get_unread_push_actions_for_user_in_range_for_email(
|
||||||
|
self.user_id, start, self.max_stream_ordering
|
||||||
|
)
|
||||||
|
|
||||||
soonest_due_at = None
|
soonest_due_at = None # type: Optional[int]
|
||||||
|
|
||||||
if not unprocessed:
|
if not unprocessed:
|
||||||
await self.save_last_stream_ordering_and_success(self.max_stream_ordering)
|
await self.save_last_stream_ordering_and_success(self.max_stream_ordering)
|
||||||
|
@ -230,7 +234,9 @@ class EmailPusher:
|
||||||
self.seconds_until(soonest_due_at), self.on_timer
|
self.seconds_until(soonest_due_at), self.on_timer
|
||||||
)
|
)
|
||||||
|
|
||||||
async def save_last_stream_ordering_and_success(self, last_stream_ordering):
|
async def save_last_stream_ordering_and_success(
|
||||||
|
self, last_stream_ordering: Optional[int]
|
||||||
|
) -> None:
|
||||||
if last_stream_ordering is None:
|
if last_stream_ordering is None:
|
||||||
# This happens if we haven't yet processed anything
|
# This happens if we haven't yet processed anything
|
||||||
return
|
return
|
||||||
|
@ -248,27 +254,29 @@ class EmailPusher:
|
||||||
# lets just stop and return.
|
# lets just stop and return.
|
||||||
self.on_stop()
|
self.on_stop()
|
||||||
|
|
||||||
def seconds_until(self, ts_msec):
|
def seconds_until(self, ts_msec: int) -> float:
|
||||||
secs = (ts_msec - self.clock.time_msec()) / 1000
|
secs = (ts_msec - self.clock.time_msec()) / 1000
|
||||||
return max(secs, 0)
|
return max(secs, 0)
|
||||||
|
|
||||||
def get_room_throttle_ms(self, room_id):
|
def get_room_throttle_ms(self, room_id: str) -> int:
|
||||||
if room_id in self.throttle_params:
|
if room_id in self.throttle_params:
|
||||||
return self.throttle_params[room_id]["throttle_ms"]
|
return self.throttle_params[room_id]["throttle_ms"]
|
||||||
else:
|
else:
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
def get_room_last_sent_ts(self, room_id):
|
def get_room_last_sent_ts(self, room_id: str) -> int:
|
||||||
if room_id in self.throttle_params:
|
if room_id in self.throttle_params:
|
||||||
return self.throttle_params[room_id]["last_sent_ts"]
|
return self.throttle_params[room_id]["last_sent_ts"]
|
||||||
else:
|
else:
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
def room_ready_to_notify_at(self, room_id):
|
def room_ready_to_notify_at(self, room_id: str) -> int:
|
||||||
"""
|
"""
|
||||||
Determines whether throttling should prevent us from sending an email
|
Determines whether throttling should prevent us from sending an email
|
||||||
for the given room
|
for the given room
|
||||||
Returns: The timestamp when we are next allowed to send an email notif
|
|
||||||
|
Returns:
|
||||||
|
The timestamp when we are next allowed to send an email notif
|
||||||
for this room
|
for this room
|
||||||
"""
|
"""
|
||||||
last_sent_ts = self.get_room_last_sent_ts(room_id)
|
last_sent_ts = self.get_room_last_sent_ts(room_id)
|
||||||
|
@ -277,7 +285,9 @@ class EmailPusher:
|
||||||
may_send_at = last_sent_ts + throttle_ms
|
may_send_at = last_sent_ts + throttle_ms
|
||||||
return may_send_at
|
return may_send_at
|
||||||
|
|
||||||
async def sent_notif_update_throttle(self, room_id, notified_push_action):
|
async def sent_notif_update_throttle(
|
||||||
|
self, room_id: str, notified_push_action: dict
|
||||||
|
) -> None:
|
||||||
# We have sent a notification, so update the throttle accordingly.
|
# We have sent a notification, so update the throttle accordingly.
|
||||||
# If the event that triggered the notif happened more than
|
# If the event that triggered the notif happened more than
|
||||||
# THROTTLE_RESET_AFTER_MS after the previous one that triggered a
|
# THROTTLE_RESET_AFTER_MS after the previous one that triggered a
|
||||||
|
@ -315,7 +325,7 @@ class EmailPusher:
|
||||||
self.pusher_id, room_id, self.throttle_params[room_id]
|
self.pusher_id, room_id, self.throttle_params[room_id]
|
||||||
)
|
)
|
||||||
|
|
||||||
async def send_notification(self, push_actions, reason):
|
async def send_notification(self, push_actions: List[dict], reason: dict) -> None:
|
||||||
logger.info("Sending notif email for user %r", self.user_id)
|
logger.info("Sending notif email for user %r", self.user_id)
|
||||||
|
|
||||||
await self.mailer.send_notification_mail(
|
await self.mailer.send_notification_mail(
|
||||||
|
|
|
@ -15,19 +15,24 @@
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
import logging
|
import logging
|
||||||
import urllib.parse
|
import urllib.parse
|
||||||
|
from typing import TYPE_CHECKING, Any, Dict, Iterable, Union
|
||||||
|
|
||||||
from prometheus_client import Counter
|
from prometheus_client import Counter
|
||||||
|
|
||||||
from twisted.internet.error import AlreadyCalled, AlreadyCancelled
|
from twisted.internet.error import AlreadyCalled, AlreadyCancelled
|
||||||
|
|
||||||
from synapse.api.constants import EventTypes
|
from synapse.api.constants import EventTypes
|
||||||
|
from synapse.events import EventBase
|
||||||
from synapse.logging import opentracing
|
from synapse.logging import opentracing
|
||||||
from synapse.metrics.background_process_metrics import run_as_background_process
|
from synapse.metrics.background_process_metrics import run_as_background_process
|
||||||
from synapse.push import PusherConfigException
|
from synapse.push import Pusher, PusherConfigException
|
||||||
from synapse.types import RoomStreamToken
|
from synapse.types import RoomStreamToken
|
||||||
|
|
||||||
from . import push_rule_evaluator, push_tools
|
from . import push_rule_evaluator, push_tools
|
||||||
|
|
||||||
|
if TYPE_CHECKING:
|
||||||
|
from synapse.app.homeserver import HomeServer
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
http_push_processed_counter = Counter(
|
http_push_processed_counter = Counter(
|
||||||
|
@ -51,24 +56,18 @@ http_badges_failed_counter = Counter(
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
class HttpPusher:
|
class HttpPusher(Pusher):
|
||||||
INITIAL_BACKOFF_SEC = 1 # in seconds because that's what Twisted takes
|
INITIAL_BACKOFF_SEC = 1 # in seconds because that's what Twisted takes
|
||||||
MAX_BACKOFF_SEC = 60 * 60
|
MAX_BACKOFF_SEC = 60 * 60
|
||||||
|
|
||||||
# This one's in ms because we compare it against the clock
|
# This one's in ms because we compare it against the clock
|
||||||
GIVE_UP_AFTER_MS = 24 * 60 * 60 * 1000
|
GIVE_UP_AFTER_MS = 24 * 60 * 60 * 1000
|
||||||
|
|
||||||
def __init__(self, hs, pusherdict):
|
def __init__(self, hs: "HomeServer", pusherdict: Dict[str, Any]):
|
||||||
self.hs = hs
|
super().__init__(hs, pusherdict)
|
||||||
self.store = self.hs.get_datastore()
|
|
||||||
self.storage = self.hs.get_storage()
|
self.storage = self.hs.get_storage()
|
||||||
self.clock = self.hs.get_clock()
|
|
||||||
self.state_handler = self.hs.get_state_handler()
|
|
||||||
self.user_id = pusherdict["user_name"]
|
|
||||||
self.app_id = pusherdict["app_id"]
|
|
||||||
self.app_display_name = pusherdict["app_display_name"]
|
self.app_display_name = pusherdict["app_display_name"]
|
||||||
self.device_display_name = pusherdict["device_display_name"]
|
self.device_display_name = pusherdict["device_display_name"]
|
||||||
self.pushkey = pusherdict["pushkey"]
|
|
||||||
self.pushkey_ts = pusherdict["ts"]
|
self.pushkey_ts = pusherdict["ts"]
|
||||||
self.data = pusherdict["data"]
|
self.data = pusherdict["data"]
|
||||||
self.last_stream_ordering = pusherdict["last_stream_ordering"]
|
self.last_stream_ordering = pusherdict["last_stream_ordering"]
|
||||||
|
@ -78,13 +77,6 @@ class HttpPusher:
|
||||||
self._is_processing = False
|
self._is_processing = False
|
||||||
self._group_unread_count_by_room = hs.config.push_group_unread_count_by_room
|
self._group_unread_count_by_room = hs.config.push_group_unread_count_by_room
|
||||||
|
|
||||||
# This is the highest stream ordering we know it's safe to process.
|
|
||||||
# When new events arrive, we'll be given a window of new events: we
|
|
||||||
# should honour this rather than just looking for anything higher
|
|
||||||
# because of potential out-of-order event serialisation. This starts
|
|
||||||
# off as None though as we don't know any better.
|
|
||||||
self.max_stream_ordering = None
|
|
||||||
|
|
||||||
if "data" not in pusherdict:
|
if "data" not in pusherdict:
|
||||||
raise PusherConfigException("No 'data' key for HTTP pusher")
|
raise PusherConfigException("No 'data' key for HTTP pusher")
|
||||||
self.data = pusherdict["data"]
|
self.data = pusherdict["data"]
|
||||||
|
@ -119,18 +111,18 @@ class HttpPusher:
|
||||||
self.data_minus_url.update(self.data)
|
self.data_minus_url.update(self.data)
|
||||||
del self.data_minus_url["url"]
|
del self.data_minus_url["url"]
|
||||||
|
|
||||||
def on_started(self, should_check_for_notifs):
|
def on_started(self, should_check_for_notifs: bool) -> None:
|
||||||
"""Called when this pusher has been started.
|
"""Called when this pusher has been started.
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
should_check_for_notifs (bool): Whether we should immediately
|
should_check_for_notifs: Whether we should immediately
|
||||||
check for push to send. Set to False only if it's known there
|
check for push to send. Set to False only if it's known there
|
||||||
is nothing to send
|
is nothing to send
|
||||||
"""
|
"""
|
||||||
if should_check_for_notifs:
|
if should_check_for_notifs:
|
||||||
self._start_processing()
|
self._start_processing()
|
||||||
|
|
||||||
def on_new_notifications(self, max_token: RoomStreamToken):
|
def on_new_notifications(self, max_token: RoomStreamToken) -> None:
|
||||||
# We just use the minimum stream ordering and ignore the vector clock
|
# We just use the minimum stream ordering and ignore the vector clock
|
||||||
# component. This is safe to do as long as we *always* ignore the vector
|
# component. This is safe to do as long as we *always* ignore the vector
|
||||||
# clock components.
|
# clock components.
|
||||||
|
@ -141,14 +133,14 @@ class HttpPusher:
|
||||||
)
|
)
|
||||||
self._start_processing()
|
self._start_processing()
|
||||||
|
|
||||||
def on_new_receipts(self, min_stream_id, max_stream_id):
|
def on_new_receipts(self, min_stream_id: int, max_stream_id: int) -> None:
|
||||||
# Note that the min here shouldn't be relied upon to be accurate.
|
# Note that the min here shouldn't be relied upon to be accurate.
|
||||||
|
|
||||||
# We could check the receipts are actually m.read receipts here,
|
# We could check the receipts are actually m.read receipts here,
|
||||||
# but currently that's the only type of receipt anyway...
|
# but currently that's the only type of receipt anyway...
|
||||||
run_as_background_process("http_pusher.on_new_receipts", self._update_badge)
|
run_as_background_process("http_pusher.on_new_receipts", self._update_badge)
|
||||||
|
|
||||||
async def _update_badge(self):
|
async def _update_badge(self) -> None:
|
||||||
# XXX as per https://github.com/matrix-org/matrix-doc/issues/2627, this seems
|
# XXX as per https://github.com/matrix-org/matrix-doc/issues/2627, this seems
|
||||||
# to be largely redundant. perhaps we can remove it.
|
# to be largely redundant. perhaps we can remove it.
|
||||||
badge = await push_tools.get_badge_count(
|
badge = await push_tools.get_badge_count(
|
||||||
|
@ -158,10 +150,10 @@ class HttpPusher:
|
||||||
)
|
)
|
||||||
await self._send_badge(badge)
|
await self._send_badge(badge)
|
||||||
|
|
||||||
def on_timer(self):
|
def on_timer(self) -> None:
|
||||||
self._start_processing()
|
self._start_processing()
|
||||||
|
|
||||||
def on_stop(self):
|
def on_stop(self) -> None:
|
||||||
if self.timed_call:
|
if self.timed_call:
|
||||||
try:
|
try:
|
||||||
self.timed_call.cancel()
|
self.timed_call.cancel()
|
||||||
|
@ -169,13 +161,13 @@ class HttpPusher:
|
||||||
pass
|
pass
|
||||||
self.timed_call = None
|
self.timed_call = None
|
||||||
|
|
||||||
def _start_processing(self):
|
def _start_processing(self) -> None:
|
||||||
if self._is_processing:
|
if self._is_processing:
|
||||||
return
|
return
|
||||||
|
|
||||||
run_as_background_process("httppush.process", self._process)
|
run_as_background_process("httppush.process", self._process)
|
||||||
|
|
||||||
async def _process(self):
|
async def _process(self) -> None:
|
||||||
# we should never get here if we are already processing
|
# we should never get here if we are already processing
|
||||||
assert not self._is_processing
|
assert not self._is_processing
|
||||||
|
|
||||||
|
@ -194,7 +186,7 @@ class HttpPusher:
|
||||||
finally:
|
finally:
|
||||||
self._is_processing = False
|
self._is_processing = False
|
||||||
|
|
||||||
async def _unsafe_process(self):
|
async def _unsafe_process(self) -> None:
|
||||||
"""
|
"""
|
||||||
Looks for unset notifications and dispatch them, in order
|
Looks for unset notifications and dispatch them, in order
|
||||||
Never call this directly: use _process which will only allow this to
|
Never call this directly: use _process which will only allow this to
|
||||||
|
@ -202,6 +194,7 @@ class HttpPusher:
|
||||||
"""
|
"""
|
||||||
|
|
||||||
fn = self.store.get_unread_push_actions_for_user_in_range_for_http
|
fn = self.store.get_unread_push_actions_for_user_in_range_for_http
|
||||||
|
assert self.max_stream_ordering is not None
|
||||||
unprocessed = await fn(
|
unprocessed = await fn(
|
||||||
self.user_id, self.last_stream_ordering, self.max_stream_ordering
|
self.user_id, self.last_stream_ordering, self.max_stream_ordering
|
||||||
)
|
)
|
||||||
|
@ -271,17 +264,12 @@ class HttpPusher:
|
||||||
)
|
)
|
||||||
self.backoff_delay = HttpPusher.INITIAL_BACKOFF_SEC
|
self.backoff_delay = HttpPusher.INITIAL_BACKOFF_SEC
|
||||||
self.last_stream_ordering = push_action["stream_ordering"]
|
self.last_stream_ordering = push_action["stream_ordering"]
|
||||||
pusher_still_exists = await self.store.update_pusher_last_stream_ordering(
|
await self.store.update_pusher_last_stream_ordering(
|
||||||
self.app_id,
|
self.app_id,
|
||||||
self.pushkey,
|
self.pushkey,
|
||||||
self.user_id,
|
self.user_id,
|
||||||
self.last_stream_ordering,
|
self.last_stream_ordering,
|
||||||
)
|
)
|
||||||
if not pusher_still_exists:
|
|
||||||
# The pusher has been deleted while we were processing, so
|
|
||||||
# lets just stop and return.
|
|
||||||
self.on_stop()
|
|
||||||
return
|
|
||||||
|
|
||||||
self.failing_since = None
|
self.failing_since = None
|
||||||
await self.store.update_pusher_failing_since(
|
await self.store.update_pusher_failing_since(
|
||||||
|
@ -297,7 +285,7 @@ class HttpPusher:
|
||||||
)
|
)
|
||||||
break
|
break
|
||||||
|
|
||||||
async def _process_one(self, push_action):
|
async def _process_one(self, push_action: dict) -> bool:
|
||||||
if "notify" not in push_action["actions"]:
|
if "notify" not in push_action["actions"]:
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
@ -328,7 +316,9 @@ class HttpPusher:
|
||||||
await self.hs.remove_pusher(self.app_id, pk, self.user_id)
|
await self.hs.remove_pusher(self.app_id, pk, self.user_id)
|
||||||
return True
|
return True
|
||||||
|
|
||||||
async def _build_notification_dict(self, event, tweaks, badge):
|
async def _build_notification_dict(
|
||||||
|
self, event: EventBase, tweaks: Dict[str, bool], badge: int
|
||||||
|
) -> Dict[str, Any]:
|
||||||
priority = "low"
|
priority = "low"
|
||||||
if (
|
if (
|
||||||
event.type == EventTypes.Encrypted
|
event.type == EventTypes.Encrypted
|
||||||
|
@ -358,9 +348,7 @@ class HttpPusher:
|
||||||
}
|
}
|
||||||
return d
|
return d
|
||||||
|
|
||||||
ctx = await push_tools.get_context_for_event(
|
ctx = await push_tools.get_context_for_event(self.storage, event, self.user_id)
|
||||||
self.storage, self.state_handler, event, self.user_id
|
|
||||||
)
|
|
||||||
|
|
||||||
d = {
|
d = {
|
||||||
"notification": {
|
"notification": {
|
||||||
|
@ -400,7 +388,9 @@ class HttpPusher:
|
||||||
|
|
||||||
return d
|
return d
|
||||||
|
|
||||||
async def dispatch_push(self, event, tweaks, badge):
|
async def dispatch_push(
|
||||||
|
self, event: EventBase, tweaks: Dict[str, bool], badge: int
|
||||||
|
) -> Union[bool, Iterable[str]]:
|
||||||
notification_dict = await self._build_notification_dict(event, tweaks, badge)
|
notification_dict = await self._build_notification_dict(event, tweaks, badge)
|
||||||
if not notification_dict:
|
if not notification_dict:
|
||||||
return []
|
return []
|
||||||
|
|
|
@ -12,6 +12,9 @@
|
||||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
# See the License for the specific language governing permissions and
|
# See the License for the specific language governing permissions and
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
from typing import Dict
|
||||||
|
|
||||||
|
from synapse.events import EventBase
|
||||||
from synapse.push.presentable_names import calculate_room_name, name_from_member_event
|
from synapse.push.presentable_names import calculate_room_name, name_from_member_event
|
||||||
from synapse.storage import Storage
|
from synapse.storage import Storage
|
||||||
from synapse.storage.databases.main import DataStore
|
from synapse.storage.databases.main import DataStore
|
||||||
|
@ -46,7 +49,9 @@ async def get_badge_count(store: DataStore, user_id: str, group_by_room: bool) -
|
||||||
return badge
|
return badge
|
||||||
|
|
||||||
|
|
||||||
async def get_context_for_event(storage: Storage, state_handler, ev, user_id):
|
async def get_context_for_event(
|
||||||
|
storage: Storage, ev: EventBase, user_id: str
|
||||||
|
) -> Dict[str, str]:
|
||||||
ctx = {}
|
ctx = {}
|
||||||
|
|
||||||
room_state_ids = await storage.state.get_state_ids_for_event(ev.event_id)
|
room_state_ids = await storage.state.get_state_ids_for_event(ev.event_id)
|
||||||
|
|
|
@ -14,25 +14,31 @@
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
|
from typing import TYPE_CHECKING, Any, Callable, Dict, Optional
|
||||||
|
|
||||||
|
from synapse.push import Pusher
|
||||||
from synapse.push.emailpusher import EmailPusher
|
from synapse.push.emailpusher import EmailPusher
|
||||||
|
from synapse.push.httppusher import HttpPusher
|
||||||
from synapse.push.mailer import Mailer
|
from synapse.push.mailer import Mailer
|
||||||
|
|
||||||
from .httppusher import HttpPusher
|
if TYPE_CHECKING:
|
||||||
|
from synapse.app.homeserver import HomeServer
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
class PusherFactory:
|
class PusherFactory:
|
||||||
def __init__(self, hs):
|
def __init__(self, hs: "HomeServer"):
|
||||||
self.hs = hs
|
self.hs = hs
|
||||||
self.config = hs.config
|
self.config = hs.config
|
||||||
|
|
||||||
self.pusher_types = {"http": HttpPusher}
|
self.pusher_types = {
|
||||||
|
"http": HttpPusher
|
||||||
|
} # type: Dict[str, Callable[[HomeServer, dict], Pusher]]
|
||||||
|
|
||||||
logger.info("email enable notifs: %r", hs.config.email_enable_notifs)
|
logger.info("email enable notifs: %r", hs.config.email_enable_notifs)
|
||||||
if hs.config.email_enable_notifs:
|
if hs.config.email_enable_notifs:
|
||||||
self.mailers = {} # app_name -> Mailer
|
self.mailers = {} # type: Dict[str, Mailer]
|
||||||
|
|
||||||
self._notif_template_html = hs.config.email_notif_template_html
|
self._notif_template_html = hs.config.email_notif_template_html
|
||||||
self._notif_template_text = hs.config.email_notif_template_text
|
self._notif_template_text = hs.config.email_notif_template_text
|
||||||
|
@ -41,7 +47,7 @@ class PusherFactory:
|
||||||
|
|
||||||
logger.info("defined email pusher type")
|
logger.info("defined email pusher type")
|
||||||
|
|
||||||
def create_pusher(self, pusherdict):
|
def create_pusher(self, pusherdict: Dict[str, Any]) -> Optional[Pusher]:
|
||||||
kind = pusherdict["kind"]
|
kind = pusherdict["kind"]
|
||||||
f = self.pusher_types.get(kind, None)
|
f = self.pusher_types.get(kind, None)
|
||||||
if not f:
|
if not f:
|
||||||
|
@ -49,7 +55,9 @@ class PusherFactory:
|
||||||
logger.debug("creating %s pusher for %r", kind, pusherdict)
|
logger.debug("creating %s pusher for %r", kind, pusherdict)
|
||||||
return f(self.hs, pusherdict)
|
return f(self.hs, pusherdict)
|
||||||
|
|
||||||
def _create_email_pusher(self, _hs, pusherdict):
|
def _create_email_pusher(
|
||||||
|
self, _hs: "HomeServer", pusherdict: Dict[str, Any]
|
||||||
|
) -> EmailPusher:
|
||||||
app_name = self._app_name_from_pusherdict(pusherdict)
|
app_name = self._app_name_from_pusherdict(pusherdict)
|
||||||
mailer = self.mailers.get(app_name)
|
mailer = self.mailers.get(app_name)
|
||||||
if not mailer:
|
if not mailer:
|
||||||
|
@ -62,7 +70,7 @@ class PusherFactory:
|
||||||
self.mailers[app_name] = mailer
|
self.mailers[app_name] = mailer
|
||||||
return EmailPusher(self.hs, pusherdict, mailer)
|
return EmailPusher(self.hs, pusherdict, mailer)
|
||||||
|
|
||||||
def _app_name_from_pusherdict(self, pusherdict):
|
def _app_name_from_pusherdict(self, pusherdict: Dict[str, Any]) -> str:
|
||||||
data = pusherdict["data"]
|
data = pusherdict["data"]
|
||||||
|
|
||||||
if isinstance(data, dict):
|
if isinstance(data, dict):
|
||||||
|
|
|
@ -15,7 +15,7 @@
|
||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
from typing import TYPE_CHECKING, Dict, Union
|
from typing import TYPE_CHECKING, Any, Dict, Optional
|
||||||
|
|
||||||
from prometheus_client import Gauge
|
from prometheus_client import Gauge
|
||||||
|
|
||||||
|
@ -23,9 +23,7 @@ from synapse.metrics.background_process_metrics import (
|
||||||
run_as_background_process,
|
run_as_background_process,
|
||||||
wrap_as_background_process,
|
wrap_as_background_process,
|
||||||
)
|
)
|
||||||
from synapse.push import PusherConfigException
|
from synapse.push import Pusher, PusherConfigException
|
||||||
from synapse.push.emailpusher import EmailPusher
|
|
||||||
from synapse.push.httppusher import HttpPusher
|
|
||||||
from synapse.push.pusher import PusherFactory
|
from synapse.push.pusher import PusherFactory
|
||||||
from synapse.types import RoomStreamToken
|
from synapse.types import RoomStreamToken
|
||||||
from synapse.util.async_helpers import concurrently_execute
|
from synapse.util.async_helpers import concurrently_execute
|
||||||
|
@ -77,7 +75,7 @@ class PusherPool:
|
||||||
self._last_room_stream_id_seen = self.store.get_room_max_stream_ordering()
|
self._last_room_stream_id_seen = self.store.get_room_max_stream_ordering()
|
||||||
|
|
||||||
# map from user id to app_id:pushkey to pusher
|
# map from user id to app_id:pushkey to pusher
|
||||||
self.pushers = {} # type: Dict[str, Dict[str, Union[HttpPusher, EmailPusher]]]
|
self.pushers = {} # type: Dict[str, Dict[str, Pusher]]
|
||||||
|
|
||||||
def start(self):
|
def start(self):
|
||||||
"""Starts the pushers off in a background process.
|
"""Starts the pushers off in a background process.
|
||||||
|
@ -99,11 +97,11 @@ class PusherPool:
|
||||||
lang,
|
lang,
|
||||||
data,
|
data,
|
||||||
profile_tag="",
|
profile_tag="",
|
||||||
):
|
) -> Optional[Pusher]:
|
||||||
"""Creates a new pusher and adds it to the pool
|
"""Creates a new pusher and adds it to the pool
|
||||||
|
|
||||||
Returns:
|
Returns:
|
||||||
EmailPusher|HttpPusher
|
The newly created pusher.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
time_now_msec = self.clock.time_msec()
|
time_now_msec = self.clock.time_msec()
|
||||||
|
@ -267,17 +265,19 @@ class PusherPool:
|
||||||
except Exception:
|
except Exception:
|
||||||
logger.exception("Exception in pusher on_new_receipts")
|
logger.exception("Exception in pusher on_new_receipts")
|
||||||
|
|
||||||
async def start_pusher_by_id(self, app_id, pushkey, user_id):
|
async def start_pusher_by_id(
|
||||||
|
self, app_id: str, pushkey: str, user_id: str
|
||||||
|
) -> Optional[Pusher]:
|
||||||
"""Look up the details for the given pusher, and start it
|
"""Look up the details for the given pusher, and start it
|
||||||
|
|
||||||
Returns:
|
Returns:
|
||||||
EmailPusher|HttpPusher|None: The pusher started, if any
|
The pusher started, if any
|
||||||
"""
|
"""
|
||||||
if not self._should_start_pushers:
|
if not self._should_start_pushers:
|
||||||
return
|
return None
|
||||||
|
|
||||||
if not self._pusher_shard_config.should_handle(self._instance_name, user_id):
|
if not self._pusher_shard_config.should_handle(self._instance_name, user_id):
|
||||||
return
|
return None
|
||||||
|
|
||||||
resultlist = await self.store.get_pushers_by_app_id_and_pushkey(app_id, pushkey)
|
resultlist = await self.store.get_pushers_by_app_id_and_pushkey(app_id, pushkey)
|
||||||
|
|
||||||
|
@ -303,19 +303,19 @@ class PusherPool:
|
||||||
|
|
||||||
logger.info("Started pushers")
|
logger.info("Started pushers")
|
||||||
|
|
||||||
async def _start_pusher(self, pusherdict):
|
async def _start_pusher(self, pusherdict: Dict[str, Any]) -> Optional[Pusher]:
|
||||||
"""Start the given pusher
|
"""Start the given pusher
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
pusherdict (dict): dict with the values pulled from the db table
|
pusherdict: dict with the values pulled from the db table
|
||||||
|
|
||||||
Returns:
|
Returns:
|
||||||
EmailPusher|HttpPusher
|
The newly created pusher or None.
|
||||||
"""
|
"""
|
||||||
if not self._pusher_shard_config.should_handle(
|
if not self._pusher_shard_config.should_handle(
|
||||||
self._instance_name, pusherdict["user_name"]
|
self._instance_name, pusherdict["user_name"]
|
||||||
):
|
):
|
||||||
return
|
return None
|
||||||
|
|
||||||
try:
|
try:
|
||||||
p = self.pusher_factory.create_pusher(pusherdict)
|
p = self.pusher_factory.create_pusher(pusherdict)
|
||||||
|
@ -328,15 +328,15 @@ class PusherPool:
|
||||||
pusherdict.get("pushkey"),
|
pusherdict.get("pushkey"),
|
||||||
e,
|
e,
|
||||||
)
|
)
|
||||||
return
|
return None
|
||||||
except Exception:
|
except Exception:
|
||||||
logger.exception(
|
logger.exception(
|
||||||
"Couldn't start pusher id %i: caught Exception", pusherdict["id"],
|
"Couldn't start pusher id %i: caught Exception", pusherdict["id"],
|
||||||
)
|
)
|
||||||
return
|
return None
|
||||||
|
|
||||||
if not p:
|
if not p:
|
||||||
return
|
return None
|
||||||
|
|
||||||
appid_pushkey = "%s:%s" % (pusherdict["app_id"], pusherdict["pushkey"])
|
appid_pushkey = "%s:%s" % (pusherdict["app_id"], pusherdict["pushkey"])
|
||||||
|
|
||||||
|
|
Loading…
Reference in a new issue