zulip/zerver/tests/test_digest.py

619 lines
25 KiB
Python

import datetime
import time
from typing import List
from unittest import mock
from django.test import override_settings
from django.utils.timezone import now as timezone_now
from confirmation.models import one_click_unsubscribe_link
from zerver.actions.create_user import do_create_user
from zerver.actions.realm_settings import do_set_realm_property
from zerver.actions.users import do_deactivate_user
from zerver.lib.digest import (
DigestTopic,
_enqueue_emails_for_realm,
bulk_handle_digest_email,
bulk_write_realm_audit_logs,
enqueue_emails,
gather_new_streams,
get_hot_topics,
get_modified_streams,
get_recent_streams,
)
from zerver.lib.message import get_last_message_id
from zerver.lib.streams import create_stream_if_needed
from zerver.lib.test_classes import ZulipTestCase
from zerver.lib.test_helpers import cache_tries_captured
from zerver.models import (
Client,
Message,
Realm,
RealmAuditLog,
Stream,
UserActivityInterval,
UserProfile,
flush_per_request_caches,
get_client,
get_realm,
get_stream,
)
class TestDigestEmailMessages(ZulipTestCase):
@mock.patch("zerver.lib.digest.enough_traffic")
@mock.patch("zerver.lib.digest.send_future_email")
def test_multiple_stream_senders(
self, mock_send_future_email: mock.MagicMock, mock_enough_traffic: mock.MagicMock
) -> None:
othello = self.example_user("othello")
self.subscribe(othello, "Verona")
one_day_ago = timezone_now() - datetime.timedelta(days=1)
Message.objects.all().update(date_sent=one_day_ago)
one_hour_ago = timezone_now() - datetime.timedelta(seconds=3600)
cutoff = time.mktime(one_hour_ago.timetuple())
senders = ["hamlet", "cordelia", "iago", "prospero", "ZOE"]
self.simulate_stream_conversation("Verona", senders)
# Remove RealmAuditLog rows, so we don't exclude polonius.
RealmAuditLog.objects.all().delete()
flush_per_request_caches()
# When this test is run in isolation, one additional query is run which
# is equivalent to
# ContentType.objects.get(app_label='zerver', model='userprofile')
# This code is run when we call `confirmation.models.create_confirmation_link`.
# To trigger this, we call the one_click_unsubscribe_link function below.
one_click_unsubscribe_link(othello, "digest")
with self.assert_database_query_count(9):
bulk_handle_digest_email([othello.id], cutoff)
self.assertEqual(mock_send_future_email.call_count, 1)
kwargs = mock_send_future_email.call_args[1]
self.assertEqual(kwargs["to_user_ids"], [othello.id])
hot_convo = kwargs["context"]["hot_conversations"][0]
expected_participants = {self.example_user(sender).full_name for sender in senders}
self.assertEqual(set(hot_convo["participants"]), expected_participants)
self.assertEqual(hot_convo["count"], 5 - 2) # 5 messages, but 2 shown
teaser_messages = hot_convo["first_few_messages"][0]["senders"]
self.assertIn("some content", teaser_messages[0]["content"][0]["plain"])
self.assertIn(teaser_messages[0]["sender"], expected_participants)
def test_bulk_handle_digest_email_skips_deactivated_users(self) -> None:
"""
A user id may be added to the queue before the user is deactivated. In such a case,
the function responsible for sending the email should correctly skip them.
"""
realm = get_realm("zulip")
hamlet = self.example_user("hamlet")
user_ids = list(
UserProfile.objects.filter(is_bot=False, realm=realm).values_list("id", flat=True)
)
do_deactivate_user(hamlet, acting_user=None)
with mock.patch("zerver.lib.digest.enough_traffic", return_value=True), mock.patch(
"zerver.lib.digest.send_future_email"
) as mock_send_email:
bulk_handle_digest_email(user_ids, 1)
emailed_user_ids = [
call_args[1]["to_user_ids"][0] for call_args in mock_send_email.call_args_list
]
self.assertEqual(
set(emailed_user_ids), {user_id for user_id in user_ids if user_id != hamlet.id}
)
@mock.patch("zerver.lib.digest.enough_traffic")
@mock.patch("zerver.lib.digest.send_future_email")
def test_guest_user_multiple_stream_sender(
self, mock_send_future_email: mock.MagicMock, mock_enough_traffic: mock.MagicMock
) -> None:
othello = self.example_user("othello")
hamlet = self.example_user("hamlet")
cordelia = self.example_user("cordelia")
polonius = self.example_user("polonius")
create_stream_if_needed(cordelia.realm, "web_public_stream", is_web_public=True)
self.subscribe(othello, "web_public_stream")
self.subscribe(hamlet, "web_public_stream")
self.subscribe(cordelia, "web_public_stream")
self.subscribe(polonius, "web_public_stream")
one_day_ago = timezone_now() - datetime.timedelta(days=1)
Message.objects.all().update(date_sent=one_day_ago)
one_hour_ago = timezone_now() - datetime.timedelta(seconds=3600)
cutoff = time.mktime(one_hour_ago.timetuple())
senders = ["hamlet", "cordelia", "othello", "desdemona"]
self.simulate_stream_conversation("web_public_stream", senders)
# Remove RealmAuditoLog rows, so we don't exclude polonius.
RealmAuditLog.objects.all().delete()
flush_per_request_caches()
# When this test is run in isolation, one additional query is run which
# is equivalent to
# ContentType.objects.get(app_label='zerver', model='userprofile')
# This code is run when we call `confirmation.models.create_confirmation_link`.
# To trigger this, we call the one_click_unsubscribe_link function below.
one_click_unsubscribe_link(polonius, "digest")
with self.assert_database_query_count(9):
bulk_handle_digest_email([polonius.id], cutoff)
self.assertEqual(mock_send_future_email.call_count, 1)
kwargs = mock_send_future_email.call_args[1]
self.assertEqual(kwargs["to_user_ids"], [polonius.id])
new_stream_names = kwargs["context"]["new_streams"]["plain"]
self.assertTrue("web_public_stream" in new_stream_names)
def test_no_logging(self) -> None:
hamlet = self.example_user("hamlet")
startlen = len(RealmAuditLog.objects.all())
bulk_write_realm_audit_logs([])
self.assert_length(RealmAuditLog.objects.all(), startlen)
bulk_write_realm_audit_logs([hamlet])
self.assert_length(RealmAuditLog.objects.all(), startlen + 1)
def test_soft_deactivated_user_multiple_stream_senders(self) -> None:
one_day_ago = timezone_now() - datetime.timedelta(days=1)
Message.objects.all().update(date_sent=one_day_ago)
digest_users = [
self.example_user("othello"),
self.example_user("aaron"),
self.example_user("desdemona"),
self.example_user("polonius"),
]
digest_users.sort(key=lambda user: user.id)
for digest_user in digest_users:
for stream in ["Verona", "Scotland", "Denmark"]:
self.subscribe(digest_user, stream)
RealmAuditLog.objects.all().delete()
# Send messages to a stream and unsubscribe - subscribe from that stream
senders = ["hamlet", "cordelia", "iago", "prospero", "ZOE"]
self.simulate_stream_conversation("Verona", senders)
for digest_user in digest_users:
self.unsubscribe(digest_user, "Verona")
self.subscribe(digest_user, "Verona")
# Send messages to other streams
self.simulate_stream_conversation("Scotland", senders)
self.simulate_stream_conversation("Denmark", senders)
one_hour_ago = timezone_now() - datetime.timedelta(seconds=3600)
cutoff = time.mktime(one_hour_ago.timetuple())
flush_per_request_caches()
# When this test is run in isolation, one additional query is run which
# is equivalent to
# ContentType.objects.get(app_label='zerver', model='userprofile')
# This code is run when we call `confirmation.models.create_confirmation_link`.
# To trigger this, we call the one_click_unsubscribe_link function below.
one_click_unsubscribe_link(digest_users[0], "digest")
with mock.patch("zerver.lib.digest.send_future_email") as mock_send_future_email:
digest_user_ids = [user.id for user in digest_users]
with self.assert_database_query_count(12):
with cache_tries_captured() as cache_tries:
bulk_handle_digest_email(digest_user_ids, cutoff)
self.assert_length(cache_tries, 0)
self.assert_length(digest_users, mock_send_future_email.call_count)
for i, digest_user in enumerate(digest_users):
kwargs = mock_send_future_email.call_args_list[i][1]
self.assertEqual(kwargs["to_user_ids"], [digest_user.id])
hot_conversations = kwargs["context"]["hot_conversations"]
self.assertEqual(2, len(hot_conversations), [digest_user.id])
hot_convo = hot_conversations[0]
expected_participants = {self.example_user(sender).full_name for sender in senders}
self.assertEqual(set(hot_convo["participants"]), expected_participants)
self.assertEqual(hot_convo["count"], 5 - 2) # 5 messages, but 2 shown
teaser_messages = hot_convo["first_few_messages"][0]["senders"]
self.assertIn("some content", teaser_messages[0]["content"][0]["plain"])
self.assertIn(teaser_messages[0]["sender"], expected_participants)
last_message_id = get_last_message_id()
for digest_user in digest_users:
log_rows = RealmAuditLog.objects.filter(
modified_user_id=digest_user.id,
event_type=RealmAuditLog.USER_DIGEST_EMAIL_CREATED,
)
(log,) = log_rows
self.assertEqual(log.event_last_message_id, last_message_id)
def test_streams_recently_modified_for_user(self) -> None:
othello = self.example_user("othello")
cordelia = self.example_user("cordelia")
for stream in ["Verona", "Scotland", "Denmark"]:
self.subscribe(othello, stream)
self.subscribe(cordelia, stream)
realm = othello.realm
denmark = get_stream("Denmark", realm)
verona = get_stream("Verona", realm)
two_hours_ago = timezone_now() - datetime.timedelta(hours=2)
one_hour_ago = timezone_now() - datetime.timedelta(hours=1)
# Delete all RealmAuditLogs to start with a clean slate.
RealmAuditLog.objects.all().delete()
# Unsubscribe and subscribe Othello from a stream
self.unsubscribe(othello, "Denmark")
self.subscribe(othello, "Denmark")
recent_streams = get_modified_streams([othello.id], one_hour_ago)
self.assertEqual(recent_streams[othello.id], {denmark.id})
# Backdate all our logs (so that Denmark will no longer
# appear like a recently modified stream for Othello).
RealmAuditLog.objects.all().update(event_time=two_hours_ago)
# Now Denmark no longer appears recent to Othello.
recent_streams = get_modified_streams([othello.id], one_hour_ago)
self.assertEqual(recent_streams[othello.id], set())
# Unsubscribe and subscribe from a stream
self.unsubscribe(othello, "Verona")
self.subscribe(othello, "Verona")
# Now, Verona, but not Denmark, appears recent.
recent_streams = get_modified_streams([othello.id], one_hour_ago)
self.assertEqual(recent_streams[othello.id], {verona.id})
# make sure we don't mix up Othello and Cordelia
self.unsubscribe(cordelia, "Denmark")
self.subscribe(cordelia, "Denmark")
recent_streams = get_modified_streams([othello.id, cordelia.id], one_hour_ago)
self.assertEqual(recent_streams[cordelia.id], {denmark.id})
def active_human_users(self, realm: Realm) -> List[UserProfile]:
users = list(
UserProfile.objects.filter(
realm=realm,
is_active=True,
is_bot=False,
enable_digest_emails=True,
)
)
assert len(users) >= 5
return users
def test_twelve_hour_exemption(self) -> None:
RealmAuditLog.objects.all().delete()
realm = get_realm("zulip")
cutoff = timezone_now() - datetime.timedelta(days=5)
with mock.patch("zerver.lib.digest.queue_digest_user_ids") as queue_mock:
_enqueue_emails_for_realm(realm, cutoff)
users = self.active_human_users(realm)
num_queued_users = len(queue_mock.call_args[0][0])
self.assert_length(users, num_queued_users)
# Simulate that we have sent digests for all our users.
bulk_write_realm_audit_logs(users)
# Now if we run again, we won't get any users, since they will have
# recent RealmAuditLog rows.
with mock.patch("zerver.lib.digest.queue_digest_user_ids") as queue_mock:
_enqueue_emails_for_realm(realm, cutoff)
self.assertEqual(queue_mock.call_count, 0)
@override_settings(SEND_DIGEST_EMAILS=True)
@override_settings(SYSTEM_ONLY_REALMS=["zulipinternal"])
def test_enqueue_emails(self) -> None:
# code coverage - digest.should_process_digest()
def call_enqueue_emails(realm: Realm) -> int:
do_set_realm_property(realm, "digest_emails_enabled", True, acting_user=None)
do_set_realm_property(
realm, "digest_weekday", timezone_now().weekday(), acting_user=None
)
cutoff = timezone_now() - datetime.timedelta(days=0)
with mock.patch(
"zerver.worker.queue_processors.bulk_handle_digest_email"
) as queue_mock:
enqueue_emails(cutoff)
return 0 if queue_mock.call_args is None else len(queue_mock.call_args[0][0])
num_queued_users = call_enqueue_emails(get_realm("zulipinternal"))
self.assertEqual(num_queued_users, 0)
num_queued_users = call_enqueue_emails(get_realm("zulip"))
self.assertEqual(num_queued_users, 10)
@override_settings(SEND_DIGEST_EMAILS=True)
def test_inactive_users_queued_for_digest(self) -> None:
UserActivityInterval.objects.all().delete()
RealmAuditLog.objects.all().delete()
# Turn on realm digest emails for all realms
Realm.objects.update(digest_emails_enabled=True)
cutoff = timezone_now() - datetime.timedelta(days=5)
realm = get_realm("zulip")
users = self.active_human_users(realm)
# Check that all users without a UserActivityInterval entry are considered
# inactive users and get enqueued.
with mock.patch("zerver.worker.queue_processors.bulk_handle_digest_email") as queue_mock:
_enqueue_emails_for_realm(realm, cutoff)
num_queued_users = len(queue_mock.call_args[0][0])
self.assert_length(users, num_queued_users)
for user in users:
last_visit = timezone_now() - datetime.timedelta(days=1)
UserActivityInterval.objects.create(
start=last_visit,
end=last_visit,
user_profile=user,
)
# Now we expect no users, due to recent activity.
with mock.patch("zerver.worker.queue_processors.bulk_handle_digest_email") as queue_mock:
_enqueue_emails_for_realm(realm, cutoff)
self.assertEqual(queue_mock.call_count, 0)
# Now, backdate all our users activity.
last_visit = timezone_now() - datetime.timedelta(days=7)
UserActivityInterval.objects.all().update(start=last_visit, end=last_visit)
with mock.patch("zerver.worker.queue_processors.bulk_handle_digest_email") as queue_mock:
_enqueue_emails_for_realm(realm, cutoff)
num_queued_users = len(queue_mock.call_args[0][0])
self.assert_length(users, num_queued_users)
def tuesday(self) -> datetime.datetime:
return datetime.datetime(year=2016, month=1, day=5, tzinfo=datetime.timezone.utc)
@override_settings(SEND_DIGEST_EMAILS=False)
def test_disabled(self) -> None:
RealmAuditLog.objects.all().delete()
tuesday = self.tuesday()
cutoff = tuesday - datetime.timedelta(days=5)
with mock.patch("zerver.lib.digest.timezone_now", return_value=tuesday):
with mock.patch("zerver.lib.digest.queue_digest_user_ids") as queue_mock:
enqueue_emails(cutoff)
queue_mock.assert_not_called()
@override_settings(SEND_DIGEST_EMAILS=True)
def test_only_enqueue_on_valid_day(self) -> None:
RealmAuditLog.objects.all().delete()
not_tuesday = datetime.datetime(year=2016, month=1, day=6, tzinfo=datetime.timezone.utc)
cutoff = not_tuesday - datetime.timedelta(days=5)
with mock.patch("zerver.lib.digest.timezone_now", return_value=not_tuesday):
with mock.patch("zerver.lib.digest.queue_digest_user_ids") as queue_mock:
enqueue_emails(cutoff)
queue_mock.assert_not_called()
@override_settings(SEND_DIGEST_EMAILS=True)
def test_no_email_digest_for_bots(self) -> None:
RealmAuditLog.objects.all().delete()
cutoff = timezone_now() - datetime.timedelta(days=5)
realm = get_realm("zulip")
realm.digest_emails_enabled = True
realm.save()
bot = do_create_user(
"some_bot@example.com",
"password",
realm,
"some_bot",
bot_type=UserProfile.DEFAULT_BOT,
acting_user=None,
)
# Check that bots are not sent emails
with mock.patch("zerver.lib.digest.queue_digest_user_ids") as queue_mock:
_enqueue_emails_for_realm(realm, cutoff)
num_queued_users = len(queue_mock.call_args[0][0])
assert num_queued_users >= 5
for arg in queue_mock.call_args_list:
user_ids = arg[0][0]
for user_id in user_ids:
self.assertNotEqual(user_id, bot.id)
@override_settings(SEND_DIGEST_EMAILS=True)
def test_new_stream_link(self) -> None:
Stream.objects.all().delete()
cutoff = timezone_now() - datetime.timedelta(days=5)
cordelia = self.example_user("cordelia")
stream = create_stream_if_needed(cordelia.realm, "New stream")[0]
stream.date_created = timezone_now()
stream.save()
realm = cordelia.realm
recent_streams = get_recent_streams(realm, cutoff)
stream_count, stream_info = gather_new_streams(
realm, recent_streams, can_access_public=True
)
self.assertEqual(stream_count, 1)
expected_html = f"<a href='http://zulip.testserver/#narrow/stream/{stream.id}-New-stream'>New stream</a>"
self.assertEqual(stream_info["html"][0], expected_html)
# guests don't see our stream
stream_count, stream_info = gather_new_streams(
realm, recent_streams, can_access_public=False
)
self.assertEqual(stream_count, 0)
self.assertEqual(stream_info["html"], [])
# but they do if we make it web-public
stream.is_web_public = True
stream.save()
recent_streams = get_recent_streams(realm, cutoff)
stream_count, stream_info = gather_new_streams(
realm, recent_streams, can_access_public=True
)
self.assertEqual(stream_count, 1)
# Make the stream appear to be older.
stream.date_created = timezone_now() - datetime.timedelta(days=7)
stream.save()
recent_streams = get_recent_streams(realm, cutoff)
stream_count, stream_info = gather_new_streams(
realm, recent_streams, can_access_public=True
)
self.assertEqual(stream_count, 0)
self.assertEqual(stream_info["html"], [])
def simulate_stream_conversation(self, stream: str, senders: List[str]) -> List[int]:
client = "website" # this makes `sent_by_human` return True
sending_client = get_client(client)
message_ids = [] # List[int]
for sender_name in senders:
sender = self.example_user(sender_name)
self.subscribe(sender, stream)
content = f"some content for {stream} from {sender_name}"
message_id = self.send_stream_message(sender, stream, content)
message_ids.append(message_id)
Message.objects.filter(id__in=message_ids).update(sending_client=sending_client)
return message_ids
class TestDigestContentInBrowser(ZulipTestCase):
def test_get_digest_content_in_browser(self) -> None:
self.login("hamlet")
result = self.client_get("/digest/")
self.assert_in_success_response(["Click here to log in to Zulip and catch up."], result)
class TestDigestTopics(ZulipTestCase):
def populate_topic(
self,
topic: DigestTopic,
humans: int,
human_messages: int,
bots: int,
bot_messages: int,
realm: Realm,
) -> None:
def send_messages(client: Client, users: int, messages: int) -> None:
messages_sent = 0
while messages_sent < messages:
for index, username in enumerate(self.example_user_map, start=1):
topic.add_message(
Message(
sender=self.example_user(username), sending_client=client, realm=realm
)
)
messages_sent += 1
if messages_sent == messages:
break
if index == users:
break
send_messages(Client(name="zulipmobile"), humans, human_messages)
send_messages(Client(name="bot"), bots, bot_messages)
def test_get_hot_topics(self) -> None:
realm = get_realm("zulip")
denmark = get_stream("Denmark", realm)
verona = get_stream("Verona", realm)
diverse_topic_a = DigestTopic((denmark.id, "5 humans talking"))
self.populate_topic(
diverse_topic_a, humans=5, human_messages=10, bots=0, bot_messages=0, realm=realm
)
diverse_topic_b = DigestTopic((denmark.id, "4 humans talking"))
self.populate_topic(
diverse_topic_b, humans=4, human_messages=15, bots=0, bot_messages=0, realm=realm
)
diverse_topic_c = DigestTopic((verona.id, "5 humans talking in another stream"))
self.populate_topic(
diverse_topic_c, humans=5, human_messages=15, bots=0, bot_messages=0, realm=realm
)
diverse_topic_d = DigestTopic((denmark.id, "3 humans and 2 bots talking"))
self.populate_topic(
diverse_topic_d, humans=3, human_messages=15, bots=2, bot_messages=10, realm=realm
)
diverse_topic_e = DigestTopic((denmark.id, "3 humans talking"))
self.populate_topic(
diverse_topic_a, humans=3, human_messages=20, bots=0, bot_messages=0, realm=realm
)
lengthy_topic_a = DigestTopic((denmark.id, "2 humans talking a lot"))
self.populate_topic(
lengthy_topic_a, humans=2, human_messages=40, bots=0, bot_messages=0, realm=realm
)
lengthy_topic_b = DigestTopic((denmark.id, "2 humans talking"))
self.populate_topic(
lengthy_topic_b, humans=2, human_messages=30, bots=0, bot_messages=0, realm=realm
)
lengthy_topic_c = DigestTopic((denmark.id, "a human and bot talking"))
self.populate_topic(
lengthy_topic_c, humans=1, human_messages=20, bots=1, bot_messages=20, realm=realm
)
lengthy_topic_d = DigestTopic((verona.id, "2 humans talking in another stream"))
self.populate_topic(
lengthy_topic_d, humans=2, human_messages=35, bots=0, bot_messages=0, realm=realm
)
topics = [
diverse_topic_a,
diverse_topic_b,
diverse_topic_c,
diverse_topic_d,
diverse_topic_e,
lengthy_topic_a,
lengthy_topic_b,
lengthy_topic_c,
lengthy_topic_d,
]
self.assertEqual(
get_hot_topics(topics, {denmark.id, 0}),
[diverse_topic_a, diverse_topic_b, lengthy_topic_a, lengthy_topic_b],
)
self.assertEqual(
get_hot_topics(topics, {denmark.id, verona.id}),
[diverse_topic_a, diverse_topic_c, lengthy_topic_a, lengthy_topic_d],
)
self.assertEqual(get_hot_topics(topics, {verona.id}), [diverse_topic_c, lengthy_topic_d])
self.assertEqual(get_hot_topics(topics, set()), [])