2013-04-23 18:51:17 +02:00
|
|
|
from __future__ import absolute_import
|
|
|
|
|
2013-01-23 23:24:44 +01:00
|
|
|
from django.conf import settings
|
2013-09-13 23:33:11 +02:00
|
|
|
from django.utils.timezone import now
|
|
|
|
|
2013-11-05 17:02:34 +01:00
|
|
|
from zerver.models import Message, UserProfile, \
|
|
|
|
Recipient, get_user_profile_by_id
|
2013-01-23 22:25:22 +01:00
|
|
|
|
2013-07-29 23:03:31 +02:00
|
|
|
from zerver.decorator import JsonableError
|
2013-09-13 22:35:27 +02:00
|
|
|
from zerver.lib.cache import cache_get_many, message_cache_key, \
|
2013-09-17 20:29:43 +02:00
|
|
|
user_profile_by_id_cache_key, cache_save_user_profile
|
2014-01-09 22:04:58 +01:00
|
|
|
from zerver.lib.cache_helpers import cache_with_key
|
2013-07-29 23:03:31 +02:00
|
|
|
from zerver.lib.queue import queue_json_publish
|
2013-10-17 23:51:25 +02:00
|
|
|
from zerver.lib.event_queue import get_client_descriptors_for_user,\
|
|
|
|
get_client_descriptors_for_realm_all_streams
|
2013-09-13 23:33:11 +02:00
|
|
|
from zerver.lib.timestamp import timestamp_to_datetime
|
2013-01-23 22:25:22 +01:00
|
|
|
|
2013-05-07 23:19:52 +02:00
|
|
|
import time
|
2013-01-23 22:25:22 +01:00
|
|
|
import logging
|
2013-01-23 23:24:44 +01:00
|
|
|
import requests
|
2013-06-18 23:55:55 +02:00
|
|
|
import ujson
|
2013-09-13 23:33:11 +02:00
|
|
|
import datetime
|
2013-01-23 22:25:22 +01:00
|
|
|
|
2013-09-13 23:33:11 +02:00
|
|
|
# Send email notifications to idle users
|
|
|
|
# after they are idle for 1 hour
|
|
|
|
NOTIFY_AFTER_IDLE_HOURS = 1
|
|
|
|
|
2013-05-22 23:49:02 +02:00
|
|
|
def build_offline_notification_event(user_profile_id, message_id):
|
|
|
|
return {"user_profile_id": user_profile_id,
|
|
|
|
"message_id": message_id,
|
|
|
|
"timestamp": time.time()}
|
|
|
|
|
|
|
|
def missedmessage_hook(user_profile_id, queue, last_for_client):
|
|
|
|
# Only process missedmessage hook when the last queue for a
|
|
|
|
# client has been garbage collected
|
|
|
|
if not last_for_client:
|
|
|
|
return
|
|
|
|
|
2014-01-21 19:26:26 +01:00
|
|
|
message_ids_to_notify = []
|
2013-05-22 23:49:02 +02:00
|
|
|
for event in queue.event_queue.contents():
|
|
|
|
if not event['type'] == 'message' or not event['flags']:
|
|
|
|
continue
|
|
|
|
|
|
|
|
if 'mentioned' in event['flags'] and not 'read' in event['flags']:
|
2014-01-21 19:26:26 +01:00
|
|
|
notify_info = dict(message_id=event['message']['id'])
|
2013-05-22 23:49:02 +02:00
|
|
|
|
2014-01-21 19:26:26 +01:00
|
|
|
if not event.get('push_notified', False):
|
|
|
|
notify_info['send_push'] = True
|
|
|
|
if not event.get('email_notified', False):
|
|
|
|
notify_info['send_email'] = True
|
|
|
|
message_ids_to_notify.append(notify_info)
|
|
|
|
|
|
|
|
for notify_info in message_ids_to_notify:
|
|
|
|
msg_id = notify_info['message_id']
|
2013-05-22 23:49:02 +02:00
|
|
|
event = build_offline_notification_event(user_profile_id, msg_id)
|
2014-01-21 19:26:26 +01:00
|
|
|
if notify_info.get('send_push', False):
|
|
|
|
queue_json_publish("missedmessage_mobile_notifications", event, lambda event: None)
|
|
|
|
if notify_info.get('send_email', False):
|
|
|
|
queue_json_publish("missedmessage_}emails", event, lambda event: None)
|
2013-05-22 23:49:02 +02:00
|
|
|
|
2014-01-27 19:35:14 +01:00
|
|
|
@cache_with_key(message_cache_key, timeout=3600*24)
|
|
|
|
def get_message_by_id_dbwarn(message_id):
|
|
|
|
if not settings.TEST_SUITE:
|
|
|
|
logging.warning("Tornado failed to load message from memcached when delivering!")
|
|
|
|
return Message.objects.select_related().get(id=message_id)
|
|
|
|
|
2014-01-09 21:55:24 +01:00
|
|
|
def receiver_is_idle(user_profile_id, realm_presences):
|
2013-09-13 22:35:27 +02:00
|
|
|
# If a user has no message-receiving event queues, they've got no open zulip
|
|
|
|
# session so we notify them
|
2014-01-09 21:55:24 +01:00
|
|
|
all_client_descriptors = get_client_descriptors_for_user(user_profile_id)
|
2013-12-10 16:35:16 +01:00
|
|
|
message_event_queues = [client for client in all_client_descriptors if client.accepts_messages()]
|
2013-09-13 22:35:27 +02:00
|
|
|
off_zulip = len(message_event_queues) == 0
|
|
|
|
|
2013-09-13 23:33:11 +02:00
|
|
|
# It's possible a recipient is not in the realm of a sender. We don't have
|
|
|
|
# presence information in this case (and it's hard to get without an additional
|
|
|
|
# db query) so we simply don't try to guess if this cross-realm recipient
|
|
|
|
# has been idle for too long
|
2014-01-09 21:55:24 +01:00
|
|
|
if realm_presences is None or not user_profile_id in realm_presences:
|
2013-09-13 23:33:11 +02:00
|
|
|
return off_zulip
|
|
|
|
|
|
|
|
# If the most recent online status from a user is >1hr in the past, we notify
|
|
|
|
# them regardless of whether or not they have an open window
|
2014-01-09 21:55:24 +01:00
|
|
|
user_presence = realm_presences[user_profile_id]
|
2013-09-13 23:33:11 +02:00
|
|
|
idle_too_long = False
|
|
|
|
newest = None
|
|
|
|
for client, status in user_presence.iteritems():
|
|
|
|
if newest is None or status['timestamp'] > newest['timestamp']:
|
|
|
|
newest = status
|
|
|
|
|
|
|
|
update_time = timestamp_to_datetime(newest['timestamp'])
|
|
|
|
if now() - update_time > datetime.timedelta(hours=NOTIFY_AFTER_IDLE_HOURS):
|
|
|
|
idle_too_long = True
|
|
|
|
|
|
|
|
return off_zulip or idle_too_long
|
2013-09-13 22:35:27 +02:00
|
|
|
|
2014-01-27 20:00:31 +01:00
|
|
|
def process_new_message(event_template):
|
|
|
|
realm_presences = event_template['presences']
|
|
|
|
sender_queue_id = event_template.get('sender_queue_id', None)
|
|
|
|
if "message_dict_markdown" in event_template:
|
|
|
|
message_dict_markdown = event_template['message_dict_markdown']
|
|
|
|
message_dict_no_markdown = event_template['message_dict_no_markdown']
|
2014-01-27 19:35:14 +01:00
|
|
|
else:
|
|
|
|
# We can delete this and get_message_by_id_dbwarn after the
|
|
|
|
# next prod deploy
|
2014-01-27 20:00:31 +01:00
|
|
|
message = get_message_by_id_dbwarn(event_template['message'])
|
2014-01-27 19:35:14 +01:00
|
|
|
message_dict_markdown = message.to_dict(True)
|
|
|
|
message_dict_no_markdown = message.to_dict(False)
|
2014-01-23 22:26:41 +01:00
|
|
|
sender_id = message_dict_markdown['sender_id']
|
|
|
|
message_id = message_dict_markdown['id']
|
|
|
|
message_type = message_dict_markdown['type']
|
|
|
|
sending_client = message_dict_markdown['client']
|
2013-05-24 16:55:05 +02:00
|
|
|
|
2014-01-07 23:14:13 +01:00
|
|
|
# To remove duplicate clients: Maps queue ID to {'client': Client, 'flags': flags}
|
2013-10-17 23:51:25 +02:00
|
|
|
send_to_clients = dict()
|
|
|
|
|
2014-01-21 19:26:26 +01:00
|
|
|
# Extra user-specific data to include
|
|
|
|
extra_user_data = {}
|
|
|
|
|
2014-01-27 20:00:31 +01:00
|
|
|
if 'stream_name' in event_template and not event_template.get("invite_only"):
|
|
|
|
for client in get_client_descriptors_for_realm_all_streams(event_template['realm_id']):
|
2014-01-07 23:14:13 +01:00
|
|
|
send_to_clients[client.event_queue.id] = {'client': client, 'flags': None}
|
|
|
|
if sender_queue_id is not None and client.event_queue.id == sender_queue_id:
|
|
|
|
send_to_clients[client.event_queue.id]['is_sender'] = True
|
2013-10-17 23:51:25 +02:00
|
|
|
|
2014-01-27 20:00:31 +01:00
|
|
|
for user_data in event_template['users']:
|
2013-06-18 18:01:41 +02:00
|
|
|
user_profile_id = user_data['id']
|
|
|
|
flags = user_data.get('flags', [])
|
|
|
|
|
2013-03-27 22:19:24 +01:00
|
|
|
for client in get_client_descriptors_for_user(user_profile_id):
|
2014-01-07 23:14:13 +01:00
|
|
|
send_to_clients[client.event_queue.id] = {'client': client, 'flags': flags}
|
|
|
|
if sender_queue_id is not None and client.event_queue.id == sender_queue_id:
|
|
|
|
send_to_clients[client.event_queue.id]['is_sender'] = True
|
2013-03-26 18:06:00 +01:00
|
|
|
|
2013-05-22 23:52:17 +02:00
|
|
|
# If the recipient was offline and the message was a single or group PM to him
|
|
|
|
# or she was @-notified potentially notify more immediately
|
2014-01-23 22:26:41 +01:00
|
|
|
received_pm = message_type == "private" and user_profile_id != sender_id
|
2013-05-22 23:52:17 +02:00
|
|
|
mentioned = 'mentioned' in flags
|
2014-01-17 23:38:18 +01:00
|
|
|
idle = receiver_is_idle(user_profile_id, realm_presences)
|
|
|
|
always_push_notify = user_data.get('always_push_notify', False)
|
|
|
|
if (received_pm or mentioned) and (idle or always_push_notify):
|
2014-01-27 20:00:31 +01:00
|
|
|
notice = build_offline_notification_event(user_profile_id, message_id)
|
|
|
|
queue_json_publish("missedmessage_mobile_notifications", notice, lambda notice: None)
|
2014-01-21 19:26:26 +01:00
|
|
|
notified = dict(push_notified=True)
|
2014-01-17 23:38:18 +01:00
|
|
|
# Don't send missed message emails if always_push_notify is True
|
|
|
|
if idle:
|
|
|
|
# We require RabbitMQ to do this, as we can't call the email handler
|
|
|
|
# from the Tornado process. So if there's no rabbitmq support do nothing
|
2014-01-27 20:00:31 +01:00
|
|
|
queue_json_publish("missedmessage_emails", notice, lambda notice: None)
|
2014-01-21 19:26:26 +01:00
|
|
|
notified['email_notified'] = True
|
|
|
|
|
|
|
|
extra_user_data[user_profile_id] = notified
|
2013-05-07 23:19:52 +02:00
|
|
|
|
2014-01-07 23:14:13 +01:00
|
|
|
for client_data in send_to_clients.itervalues():
|
|
|
|
client = client_data['client']
|
|
|
|
flags = client_data['flags']
|
|
|
|
is_sender = client_data.get('is_sender', False)
|
2014-01-21 19:26:26 +01:00
|
|
|
extra_data = extra_user_data.get(client.user_profile_id, None)
|
2014-01-07 23:14:13 +01:00
|
|
|
|
2013-12-10 16:35:16 +01:00
|
|
|
if not client.accepts_messages():
|
|
|
|
# The actual check is the accepts_event() check below;
|
|
|
|
# this line is just an optimization to avoid copying
|
|
|
|
# message data unnecessarily
|
2013-10-17 23:51:25 +02:00
|
|
|
continue
|
|
|
|
|
|
|
|
if client.apply_markdown:
|
|
|
|
message_dict = message_dict_markdown
|
|
|
|
else:
|
|
|
|
message_dict = message_dict_no_markdown
|
|
|
|
|
|
|
|
# Make sure Zephyr mirroring bots know whether stream is invite-only
|
2014-01-27 20:00:31 +01:00
|
|
|
if "mirror" in client.client_type.name and event_template.get("invite_only"):
|
2013-10-17 23:51:25 +02:00
|
|
|
message_dict = message_dict.copy()
|
|
|
|
message_dict["invite_only_stream"] = True
|
|
|
|
|
2014-01-27 20:00:31 +01:00
|
|
|
user_event = dict(type='message', message=message_dict, flags=flags)
|
2014-01-21 19:26:26 +01:00
|
|
|
if extra_data is not None:
|
2014-01-27 20:00:31 +01:00
|
|
|
user_event.update(extra_data)
|
2014-01-07 23:14:13 +01:00
|
|
|
|
|
|
|
if is_sender:
|
2014-01-27 20:00:31 +01:00
|
|
|
local_message_id = event_template.get('local_id', None)
|
2014-01-07 23:14:13 +01:00
|
|
|
if local_message_id is not None:
|
2014-01-27 20:00:31 +01:00
|
|
|
user_event["local_message_id"] = local_message_id
|
2014-01-07 23:14:13 +01:00
|
|
|
|
2014-01-27 20:00:31 +01:00
|
|
|
if not client.accepts_event(user_event):
|
2013-12-10 16:35:16 +01:00
|
|
|
continue
|
|
|
|
|
|
|
|
# The below prevents (Zephyr) mirroring loops.
|
2014-01-23 22:26:41 +01:00
|
|
|
if ('mirror' in sending_client and
|
|
|
|
sending_client.lower() == client.client_type.name.lower()):
|
2013-12-10 16:35:16 +01:00
|
|
|
continue
|
2014-01-27 20:00:31 +01:00
|
|
|
client.add_event(user_event)
|
2013-10-17 23:51:25 +02:00
|
|
|
|
2014-01-25 00:06:40 +01:00
|
|
|
def process_event(event, users):
|
|
|
|
for user_profile_id in users:
|
2013-03-29 15:35:37 +01:00
|
|
|
for client in get_client_descriptors_for_user(user_profile_id):
|
2013-12-10 16:35:16 +01:00
|
|
|
if client.accepts_event(event):
|
2013-03-29 15:35:37 +01:00
|
|
|
client.add_event(event.copy())
|
|
|
|
|
2014-01-27 20:00:31 +01:00
|
|
|
def process_userdata_event(event_template, users):
|
2014-01-25 00:06:40 +01:00
|
|
|
for user_data in users:
|
2014-01-24 23:58:13 +01:00
|
|
|
user_profile_id = user_data['id']
|
2014-01-27 20:00:31 +01:00
|
|
|
user_event = event_template.copy() # shallow, but deep enough for our needs
|
2014-01-24 23:58:13 +01:00
|
|
|
for key in user_data.keys():
|
|
|
|
if key != "id":
|
|
|
|
user_event[key] = user_data[key]
|
2014-01-08 19:37:15 +01:00
|
|
|
|
|
|
|
for client in get_client_descriptors_for_user(user_profile_id):
|
|
|
|
if client.accepts_event(user_event):
|
|
|
|
client.add_event(user_event)
|
|
|
|
|
2013-01-23 23:24:44 +01:00
|
|
|
def process_notification(data):
|
2014-01-25 00:06:40 +01:00
|
|
|
event = data['event']
|
|
|
|
users = data['users']
|
|
|
|
if event['type'] in ["update_message"]:
|
|
|
|
process_userdata_event(event, users)
|
|
|
|
elif event['type'] == "message":
|
2013-01-23 23:24:44 +01:00
|
|
|
process_new_message(data)
|
|
|
|
else:
|
2014-01-25 00:06:40 +01:00
|
|
|
process_event(event, users)
|
2013-01-23 23:24:44 +01:00
|
|
|
|
|
|
|
# Runs in the Django process to send a notification to Tornado.
|
|
|
|
#
|
|
|
|
# We use JSON rather than bare form parameters, so that we can represent
|
|
|
|
# different types and for compatibility with non-HTTP transports.
|
2013-03-22 20:55:40 +01:00
|
|
|
|
|
|
|
def send_notification_http(data):
|
2013-10-11 21:31:20 +02:00
|
|
|
if settings.TORNADO_SERVER and not settings.RUNNING_INSIDE_TORNADO:
|
2013-03-29 21:04:49 +01:00
|
|
|
requests.post(settings.TORNADO_SERVER + '/notify_tornado', data=dict(
|
2013-06-18 23:55:55 +02:00
|
|
|
data = ujson.dumps(data),
|
2013-03-29 21:04:49 +01:00
|
|
|
secret = settings.SHARED_SECRET))
|
2013-07-02 00:11:59 +02:00
|
|
|
else:
|
|
|
|
process_notification(data)
|
2013-03-22 20:55:40 +01:00
|
|
|
|
2013-03-25 20:37:00 +01:00
|
|
|
def send_notification(data):
|
|
|
|
return queue_json_publish("notify_tornado", data, send_notification_http)
|
2014-01-24 23:24:44 +01:00
|
|
|
|
|
|
|
def send_event(event, users):
|
|
|
|
return send_notification(dict(event=event, users=users))
|