mirror of
https://mau.dev/maunium/synapse.git
synced 2024-11-18 16:02:15 +01:00
290 lines
11 KiB
Python
290 lines
11 KiB
Python
# -*- coding: utf-8 -*-
|
|
# Copyright 2015, 2016 OpenMarket Ltd
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
|
|
from twisted.internet import defer
|
|
|
|
from synapse.streams.config import PaginationConfig
|
|
from synapse.types import StreamToken
|
|
|
|
import synapse.util.async
|
|
import push_rule_evaluator as push_rule_evaluator
|
|
|
|
import logging
|
|
import random
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
# Pushers could now be moved to pull out of the event_push_actions table instead
|
|
# of listening on the event stream: this would avoid them having to run the
|
|
# rules again.
|
|
class Pusher(object):
|
|
INITIAL_BACKOFF = 1000
|
|
MAX_BACKOFF = 60 * 60 * 1000
|
|
GIVE_UP_AFTER = 24 * 60 * 60 * 1000
|
|
|
|
def __init__(self, _hs, profile_tag, user_id, app_id,
|
|
app_display_name, device_display_name, pushkey, pushkey_ts,
|
|
data, last_token, last_success, failing_since):
|
|
self.hs = _hs
|
|
self.evStreamHandler = self.hs.get_handlers().event_stream_handler
|
|
self.store = self.hs.get_datastore()
|
|
self.clock = self.hs.get_clock()
|
|
self.profile_tag = profile_tag
|
|
self.user_id = user_id
|
|
self.app_id = app_id
|
|
self.app_display_name = app_display_name
|
|
self.device_display_name = device_display_name
|
|
self.pushkey = pushkey
|
|
self.pushkey_ts = pushkey_ts
|
|
self.data = data
|
|
self.last_token = last_token
|
|
self.last_success = last_success # not actually used
|
|
self.backoff_delay = Pusher.INITIAL_BACKOFF
|
|
self.failing_since = failing_since
|
|
self.alive = True
|
|
|
|
# The last value of last_active_time that we saw
|
|
self.last_last_active_time = 0
|
|
self.has_unread = True
|
|
|
|
@defer.inlineCallbacks
|
|
def get_context_for_event(self, ev):
|
|
name_aliases = yield self.store.get_room_name_and_aliases(
|
|
ev['room_id']
|
|
)
|
|
|
|
ctx = {'aliases': name_aliases[1]}
|
|
if name_aliases[0] is not None:
|
|
ctx['name'] = name_aliases[0]
|
|
|
|
their_member_events_for_room = yield self.store.get_current_state(
|
|
room_id=ev['room_id'],
|
|
event_type='m.room.member',
|
|
state_key=ev['user_id']
|
|
)
|
|
for mev in their_member_events_for_room:
|
|
if mev.content['membership'] == 'join' and 'displayname' in mev.content:
|
|
dn = mev.content['displayname']
|
|
if dn is not None:
|
|
ctx['sender_display_name'] = dn
|
|
|
|
defer.returnValue(ctx)
|
|
|
|
@defer.inlineCallbacks
|
|
def start(self):
|
|
if not self.last_token:
|
|
# First-time setup: get a token to start from (we can't
|
|
# just start from no token, ie. 'now'
|
|
# because we need the result to be reproduceable in case
|
|
# we fail to dispatch the push)
|
|
config = PaginationConfig(from_token=None, limit='1')
|
|
chunk = yield self.evStreamHandler.get_stream(
|
|
self.user_id, config, timeout=0, affect_presence=False,
|
|
only_room_events=True
|
|
)
|
|
self.last_token = chunk['end']
|
|
self.store.update_pusher_last_token(
|
|
self.app_id, self.pushkey, self.user_id, self.last_token
|
|
)
|
|
logger.info("Pusher %s for user %s starting from token %s",
|
|
self.pushkey, self.user_id, self.last_token)
|
|
|
|
wait = 0
|
|
while self.alive:
|
|
try:
|
|
if wait > 0:
|
|
yield synapse.util.async.sleep(wait)
|
|
yield self.get_and_dispatch()
|
|
wait = 0
|
|
except:
|
|
if wait == 0:
|
|
wait = 1
|
|
else:
|
|
wait = min(wait * 2, 1800)
|
|
logger.exception(
|
|
"Exception in pusher loop for pushkey %s. Pausing for %ds",
|
|
self.pushkey, wait
|
|
)
|
|
|
|
@defer.inlineCallbacks
|
|
def get_and_dispatch(self):
|
|
from_tok = StreamToken.from_string(self.last_token)
|
|
config = PaginationConfig(from_token=from_tok, limit='1')
|
|
timeout = (300 + random.randint(-60, 60)) * 1000
|
|
chunk = yield self.evStreamHandler.get_stream(
|
|
self.user_id, config, timeout=timeout, affect_presence=False,
|
|
only_room_events=True
|
|
)
|
|
|
|
# limiting to 1 may get 1 event plus 1 presence event, so
|
|
# pick out the actual event
|
|
single_event = None
|
|
for c in chunk['chunk']:
|
|
if 'event_id' in c: # Hmmm...
|
|
single_event = c
|
|
break
|
|
if not single_event:
|
|
self.last_token = chunk['end']
|
|
logger.debug("Event stream timeout for pushkey %s", self.pushkey)
|
|
yield self.store.update_pusher_last_token(
|
|
self.app_id,
|
|
self.pushkey,
|
|
self.user_id,
|
|
self.last_token
|
|
)
|
|
return
|
|
|
|
if not self.alive:
|
|
return
|
|
|
|
processed = False
|
|
|
|
rule_evaluator = yield \
|
|
push_rule_evaluator.evaluator_for_user_id_and_profile_tag(
|
|
self.user_id, self.profile_tag, single_event['room_id'], self.store
|
|
)
|
|
|
|
actions = yield rule_evaluator.actions_for_event(single_event)
|
|
tweaks = rule_evaluator.tweaks_for_actions(actions)
|
|
|
|
if 'notify' in actions:
|
|
rejected = yield self.dispatch_push(single_event, tweaks)
|
|
self.has_unread = True
|
|
if isinstance(rejected, list) or isinstance(rejected, tuple):
|
|
processed = True
|
|
for pk in rejected:
|
|
if pk != self.pushkey:
|
|
# for sanity, we only remove the pushkey if it
|
|
# was the one we actually sent...
|
|
logger.warn(
|
|
("Ignoring rejected pushkey %s because we"
|
|
" didn't send it"), pk
|
|
)
|
|
else:
|
|
logger.info(
|
|
"Pushkey %s was rejected: removing",
|
|
pk
|
|
)
|
|
yield self.hs.get_pusherpool().remove_pusher(
|
|
self.app_id, pk, self.user_id
|
|
)
|
|
else:
|
|
processed = True
|
|
|
|
if not self.alive:
|
|
return
|
|
|
|
if processed:
|
|
self.backoff_delay = Pusher.INITIAL_BACKOFF
|
|
self.last_token = chunk['end']
|
|
yield self.store.update_pusher_last_token_and_success(
|
|
self.app_id,
|
|
self.pushkey,
|
|
self.user_id,
|
|
self.last_token,
|
|
self.clock.time_msec()
|
|
)
|
|
if self.failing_since:
|
|
self.failing_since = None
|
|
yield self.store.update_pusher_failing_since(
|
|
self.app_id,
|
|
self.pushkey,
|
|
self.user_id,
|
|
self.failing_since)
|
|
else:
|
|
if not self.failing_since:
|
|
self.failing_since = self.clock.time_msec()
|
|
yield self.store.update_pusher_failing_since(
|
|
self.app_id,
|
|
self.pushkey,
|
|
self.user_id,
|
|
self.failing_since
|
|
)
|
|
|
|
if (self.failing_since and
|
|
self.failing_since <
|
|
self.clock.time_msec() - Pusher.GIVE_UP_AFTER):
|
|
# we really only give up so that if the URL gets
|
|
# fixed, we don't suddenly deliver a load
|
|
# of old notifications.
|
|
logger.warn("Giving up on a notification to user %s, "
|
|
"pushkey %s",
|
|
self.user_id, self.pushkey)
|
|
self.backoff_delay = Pusher.INITIAL_BACKOFF
|
|
self.last_token = chunk['end']
|
|
yield self.store.update_pusher_last_token(
|
|
self.app_id,
|
|
self.pushkey,
|
|
self.user_id,
|
|
self.last_token
|
|
)
|
|
|
|
self.failing_since = None
|
|
yield self.store.update_pusher_failing_since(
|
|
self.app_id,
|
|
self.pushkey,
|
|
self.user_id,
|
|
self.failing_since
|
|
)
|
|
else:
|
|
logger.warn("Failed to dispatch push for user %s "
|
|
"(failing for %dms)."
|
|
"Trying again in %dms",
|
|
self.user_id,
|
|
self.clock.time_msec() - self.failing_since,
|
|
self.backoff_delay)
|
|
yield synapse.util.async.sleep(self.backoff_delay / 1000.0)
|
|
self.backoff_delay *= 2
|
|
if self.backoff_delay > Pusher.MAX_BACKOFF:
|
|
self.backoff_delay = Pusher.MAX_BACKOFF
|
|
|
|
def stop(self):
|
|
self.alive = False
|
|
|
|
def dispatch_push(self, p, tweaks):
|
|
"""
|
|
Overridden by implementing classes to actually deliver the notification
|
|
Args:
|
|
p: The event to notify for as a single event from the event stream
|
|
Returns: If the notification was delivered, an array containing any
|
|
pushkeys that were rejected by the push gateway.
|
|
False if the notification could not be delivered (ie.
|
|
should be retried).
|
|
"""
|
|
pass
|
|
|
|
def reset_badge_count(self):
|
|
pass
|
|
|
|
def presence_changed(self, state):
|
|
"""
|
|
We clear badge counts whenever a user's last_active time is bumped
|
|
This is by no means perfect but I think it's the best we can do
|
|
without read receipts.
|
|
"""
|
|
if 'last_active' in state.state:
|
|
last_active = state.state['last_active']
|
|
if last_active > self.last_last_active_time:
|
|
self.last_last_active_time = last_active
|
|
if self.has_unread:
|
|
logger.info("Resetting badge count for %s", self.user_id)
|
|
self.reset_badge_count()
|
|
self.has_unread = False
|
|
|
|
|
|
class PusherConfigException(Exception):
|
|
def __init__(self, msg):
|
|
super(PusherConfigException, self).__init__(msg)
|