2016-10-25 19:51:31 +02:00
|
|
|
# -*- coding: utf-8 -*-
|
|
|
|
import types
|
|
|
|
from datetime import datetime, timedelta
|
|
|
|
|
2017-04-15 04:03:56 +02:00
|
|
|
from django.utils.timezone import now as timezone_now
|
2016-11-10 19:30:09 +01:00
|
|
|
from zerver.lib.test_classes import ZulipTestCase
|
2017-05-14 21:14:26 +02:00
|
|
|
from zerver.lib.upload import create_attachment
|
|
|
|
from zerver.models import Message, Realm, Recipient, UserProfile, UserMessage, ArchivedUserMessage, \
|
|
|
|
ArchivedMessage, Attachment, ArchivedAttachment
|
2018-09-13 16:29:12 +02:00
|
|
|
from zerver.lib.retention import get_expired_messages, move_messages_to_archive
|
2016-10-25 19:51:31 +02:00
|
|
|
|
2018-09-13 16:29:12 +02:00
|
|
|
from typing import Any, List, Tuple
|
2016-10-25 19:51:31 +02:00
|
|
|
|
|
|
|
|
|
|
|
class TestRetentionLib(ZulipTestCase):
|
|
|
|
"""
|
|
|
|
Test receiving expired messages retention tool.
|
|
|
|
"""
|
|
|
|
|
2017-11-19 04:02:03 +01:00
|
|
|
def setUp(self) -> None:
|
2017-10-27 08:28:23 +02:00
|
|
|
super().setUp()
|
2017-01-08 20:24:05 +01:00
|
|
|
self.zulip_realm = self._set_realm_message_retention_value('zulip', 30)
|
2017-03-04 09:19:37 +01:00
|
|
|
self.mit_realm = self._set_realm_message_retention_value('zephyr', 100)
|
2018-08-08 23:02:44 +02:00
|
|
|
Message.objects.all().update(pub_date=timezone_now())
|
2016-10-25 19:51:31 +02:00
|
|
|
|
|
|
|
@staticmethod
|
2017-11-19 04:02:03 +01:00
|
|
|
def _set_realm_message_retention_value(realm_str: str, retention_period: int) -> Realm:
|
2017-01-08 20:24:05 +01:00
|
|
|
realm = Realm.objects.get(string_id=realm_str)
|
2016-10-25 19:51:31 +02:00
|
|
|
realm.message_retention_days = retention_period
|
|
|
|
realm.save()
|
|
|
|
return realm
|
|
|
|
|
|
|
|
@staticmethod
|
2017-11-19 04:02:03 +01:00
|
|
|
def _change_messages_pub_date(msgs_ids: List[int], pub_date: datetime) -> Any:
|
2016-10-25 19:51:31 +02:00
|
|
|
messages = Message.objects.filter(id__in=msgs_ids).order_by('id')
|
|
|
|
messages.update(pub_date=pub_date)
|
|
|
|
return messages
|
|
|
|
|
2017-11-19 04:02:03 +01:00
|
|
|
def _make_mit_messages(self, message_quantity: int, pub_date: datetime) -> Any:
|
2016-10-25 19:51:31 +02:00
|
|
|
# send messages from mit.edu realm and change messages pub date
|
2017-05-23 03:28:23 +02:00
|
|
|
sender = self.mit_user('espuser')
|
2017-05-23 02:33:53 +02:00
|
|
|
recipient = self.mit_user('starnine')
|
2017-11-26 01:45:15 +01:00
|
|
|
msgs_ids = [self.send_personal_message(sender.email, recipient.email,
|
|
|
|
sender_realm='zephyr') for i in
|
2016-10-25 19:51:31 +02:00
|
|
|
range(message_quantity)]
|
|
|
|
mit_messages = self._change_messages_pub_date(msgs_ids, pub_date)
|
|
|
|
return mit_messages
|
|
|
|
|
2017-11-19 04:02:03 +01:00
|
|
|
def test_expired_messages_result_type(self) -> None:
|
2016-10-25 19:51:31 +02:00
|
|
|
# Check return type of get_expired_message method.
|
|
|
|
result = get_expired_messages()
|
|
|
|
self.assertIsInstance(result, types.GeneratorType)
|
|
|
|
|
2017-11-19 04:02:03 +01:00
|
|
|
def test_no_expired_messages(self) -> None:
|
2016-10-25 19:51:31 +02:00
|
|
|
result = list(get_expired_messages())
|
|
|
|
self.assertFalse(result)
|
|
|
|
|
2017-11-19 04:02:03 +01:00
|
|
|
def test_expired_messages_in_each_realm(self) -> None:
|
2016-10-25 19:51:31 +02:00
|
|
|
# Check result realm messages order and result content
|
|
|
|
# when all realm has expired messages.
|
2017-04-15 04:03:56 +02:00
|
|
|
expired_mit_messages = self._make_mit_messages(3, timezone_now() - timedelta(days=101))
|
|
|
|
self._make_mit_messages(4, timezone_now() - timedelta(days=50))
|
2016-10-25 19:51:31 +02:00
|
|
|
zulip_messages_ids = Message.objects.order_by('id').filter(
|
|
|
|
sender__realm=self.zulip_realm).values_list('id', flat=True)[3:10]
|
|
|
|
expired_zulip_messages = self._change_messages_pub_date(zulip_messages_ids,
|
2017-04-15 04:03:56 +02:00
|
|
|
timezone_now() - timedelta(days=31))
|
2016-10-25 19:51:31 +02:00
|
|
|
# Iterate by result
|
|
|
|
expired_messages_result = [messages_list for messages_list in get_expired_messages()]
|
|
|
|
self.assertEqual(len(expired_messages_result), 2)
|
|
|
|
# Check mit.edu realm expired messages.
|
|
|
|
self.assertEqual(len(expired_messages_result[0]['expired_messages']), 3)
|
|
|
|
self.assertEqual(expired_messages_result[0]['realm_id'], self.mit_realm.id)
|
|
|
|
# Check zulip.com realm expired messages.
|
|
|
|
self.assertEqual(len(expired_messages_result[1]['expired_messages']), 7)
|
|
|
|
self.assertEqual(expired_messages_result[1]['realm_id'], self.zulip_realm.id)
|
|
|
|
# Compare expected messages ids with result messages ids.
|
|
|
|
self.assertEqual(
|
|
|
|
sorted([message.id for message in expired_mit_messages]),
|
|
|
|
[message.id for message in expired_messages_result[0]['expired_messages']]
|
|
|
|
)
|
|
|
|
self.assertEqual(
|
|
|
|
sorted([message.id for message in expired_zulip_messages]),
|
|
|
|
[message.id for message in expired_messages_result[1]['expired_messages']]
|
|
|
|
)
|
|
|
|
|
2017-11-19 04:02:03 +01:00
|
|
|
def test_expired_messages_in_one_realm(self) -> None:
|
2016-10-25 19:51:31 +02:00
|
|
|
# Check realm with expired messages and messages
|
|
|
|
# with one day to expiration data.
|
2017-04-15 04:03:56 +02:00
|
|
|
expired_mit_messages = self._make_mit_messages(5, timezone_now() - timedelta(days=101))
|
|
|
|
actual_mit_messages = self._make_mit_messages(3, timezone_now() - timedelta(days=99))
|
2016-10-25 19:51:31 +02:00
|
|
|
expired_messages_result = list(get_expired_messages())
|
|
|
|
expired_mit_messages_ids = [message.id for message in expired_mit_messages]
|
|
|
|
expired_mit_messages_result_ids = [message.id for message in
|
|
|
|
expired_messages_result[0]['expired_messages']]
|
|
|
|
actual_mit_messages_ids = [message.id for message in actual_mit_messages]
|
|
|
|
self.assertEqual(len(expired_messages_result), 1)
|
|
|
|
self.assertEqual(len(expired_messages_result[0]['expired_messages']), 5)
|
|
|
|
self.assertEqual(expired_messages_result[0]['realm_id'], self.mit_realm.id)
|
|
|
|
# Compare expected messages ids with result messages ids.
|
|
|
|
self.assertEqual(
|
|
|
|
sorted(expired_mit_messages_ids),
|
|
|
|
expired_mit_messages_result_ids
|
|
|
|
)
|
|
|
|
# Check actual mit.edu messages are not contained in expired messages list
|
|
|
|
self.assertEqual(
|
|
|
|
set(actual_mit_messages_ids) - set(expired_mit_messages_ids),
|
|
|
|
set(actual_mit_messages_ids)
|
|
|
|
)
|
2017-05-14 21:14:26 +02:00
|
|
|
|
|
|
|
|
|
|
|
class TestMoveMessageToArchive(ZulipTestCase):
|
|
|
|
|
2017-11-19 04:02:03 +01:00
|
|
|
def setUp(self) -> None:
|
2017-10-27 08:28:23 +02:00
|
|
|
super().setUp()
|
2017-05-14 21:14:26 +02:00
|
|
|
self.sender = 'hamlet@zulip.com'
|
|
|
|
self.recipient = 'cordelia@zulip.com'
|
|
|
|
|
2017-11-19 04:02:03 +01:00
|
|
|
def _create_attachments(self) -> None:
|
2017-05-14 21:14:26 +02:00
|
|
|
sample_size = 10
|
|
|
|
dummy_files = [
|
|
|
|
('zulip.txt', '1/31/4CBjtTLYZhk66pZrF8hnYGwc/zulip.txt', sample_size),
|
|
|
|
('temp_file.py', '1/31/4CBjtTLYZhk66pZrF8hnYGwc/temp_file.py', sample_size),
|
2018-09-13 16:29:12 +02:00
|
|
|
('abc.py', '1/31/4CBjtTLYZhk66pZrF8hnYGwc/abc.py', sample_size),
|
|
|
|
('hello.txt', '1/31/4CBjtTLYZhk66pZrF8hnYGwc/hello.txt', sample_size),
|
|
|
|
('new.py', '1/31/4CBjtTLYZhk66pZrF8hnYGwc/new.py', sample_size)
|
2017-05-14 21:14:26 +02:00
|
|
|
]
|
|
|
|
user_profile = self.example_user('hamlet')
|
|
|
|
for file_name, path_id, size in dummy_files:
|
|
|
|
create_attachment(file_name, path_id, user_profile, size)
|
|
|
|
|
2018-09-13 16:29:12 +02:00
|
|
|
def _check_messages_before_archiving(self, msg_ids: List[int]) -> Tuple[List[int], List[int]]:
|
|
|
|
user_msgs_ids_before = list(UserMessage.objects.filter(
|
|
|
|
message_id__in=msg_ids).order_by('id').values_list('id', flat=True))
|
|
|
|
all_msgs_ids_before = list(Message.objects.filter().order_by('id').values_list('id', flat=True))
|
2017-05-14 21:14:26 +02:00
|
|
|
self.assertEqual(ArchivedUserMessage.objects.count(), 0)
|
|
|
|
self.assertEqual(ArchivedMessage.objects.count(), 0)
|
2018-09-13 16:29:12 +02:00
|
|
|
return (user_msgs_ids_before, all_msgs_ids_before)
|
|
|
|
|
|
|
|
def _check_messages_after_archiving(self, msg_ids: List[int], user_msgs_ids_before: List[int],
|
|
|
|
all_msgs_ids_before: List[int]) -> None:
|
|
|
|
self.assertEqual(ArchivedMessage.objects.all().count(), len(msg_ids))
|
|
|
|
self.assertEqual(Message.objects.filter().count(), len(all_msgs_ids_before) - len(msg_ids))
|
|
|
|
self.assertEqual(UserMessage.objects.filter(message_id__in=msg_ids).count(), 0)
|
|
|
|
arc_user_messages_ids_after = list(ArchivedUserMessage.objects.filter().order_by('id').values_list('id', flat=True))
|
2017-05-14 21:14:26 +02:00
|
|
|
self.assertEqual(arc_user_messages_ids_after, user_msgs_ids_before)
|
|
|
|
|
2018-09-13 16:29:12 +02:00
|
|
|
def test_personal_messages_archiving(self) -> None:
|
|
|
|
msg_ids = []
|
|
|
|
for i in range(0, 3):
|
|
|
|
msg_ids.append(self.send_personal_message(self.sender, self.recipient))
|
|
|
|
(user_msgs_ids_before, all_msgs_ids_before) = self._check_messages_before_archiving(msg_ids)
|
|
|
|
move_messages_to_archive(message_ids=msg_ids)
|
|
|
|
self._check_messages_after_archiving(msg_ids, user_msgs_ids_before, all_msgs_ids_before)
|
|
|
|
|
|
|
|
def test_stream_messages_archiving(self) -> None:
|
|
|
|
msg_ids = []
|
|
|
|
for i in range(0, 3):
|
|
|
|
msg_ids.append(self.send_stream_message(self.sender, "Verona"))
|
|
|
|
(user_msgs_ids_before, all_msgs_ids_before) = self._check_messages_before_archiving(msg_ids)
|
|
|
|
move_messages_to_archive(message_ids=msg_ids)
|
|
|
|
self._check_messages_after_archiving(msg_ids, user_msgs_ids_before, all_msgs_ids_before)
|
|
|
|
|
|
|
|
def test_archiving_messages_second_time(self) -> None:
|
|
|
|
msg_ids = []
|
|
|
|
for i in range(0, 3):
|
|
|
|
msg_ids.append(self.send_stream_message(self.sender, "Verona"))
|
|
|
|
(user_msgs_ids_before, all_msgs_ids_before) = self._check_messages_before_archiving(msg_ids)
|
|
|
|
move_messages_to_archive(message_ids=msg_ids)
|
|
|
|
self._check_messages_after_archiving(msg_ids, user_msgs_ids_before, all_msgs_ids_before)
|
2017-05-14 21:14:26 +02:00
|
|
|
with self.assertRaises(Message.DoesNotExist):
|
2018-09-13 16:29:12 +02:00
|
|
|
move_messages_to_archive(message_ids=msg_ids)
|
2017-05-14 21:14:26 +02:00
|
|
|
|
2018-09-13 16:29:12 +02:00
|
|
|
def test_archiving_messages_with_attachment(self) -> None:
|
2017-05-14 21:14:26 +02:00
|
|
|
self._create_attachments()
|
2018-09-13 16:29:12 +02:00
|
|
|
body1 = """Some files here ...[zulip.txt](
|
2017-05-14 21:14:26 +02:00
|
|
|
http://localhost:9991/user_uploads/1/31/4CBjtTLYZhk66pZrF8hnYGwc/zulip.txt)
|
|
|
|
http://localhost:9991/user_uploads/1/31/4CBjtTLYZhk66pZrF8hnYGwc/temp_file.py ....
|
|
|
|
Some more.... http://localhost:9991/user_uploads/1/31/4CBjtTLYZhk66pZrF8hnYGwc/abc.py
|
|
|
|
"""
|
2018-09-13 16:29:12 +02:00
|
|
|
body2 = """Some files here
|
|
|
|
http://localhost:9991/user_uploads/1/31/4CBjtTLYZhk66pZrF8hnYGwc/zulip.txt ...
|
|
|
|
http://localhost:9991/user_uploads/1/31/4CBjtTLYZhk66pZrF8hnYGwc/hello.txt ....
|
|
|
|
http://localhost:9991/user_uploads/1/31/4CBjtTLYZhk66pZrF8hnYGwc/new.py ....
|
|
|
|
"""
|
|
|
|
msg_ids = []
|
|
|
|
msg_ids.append(self.send_personal_message(self.sender, self.recipient, body1))
|
|
|
|
msg_ids.append(self.send_personal_message(self.sender, self.recipient, body2))
|
|
|
|
|
|
|
|
attachment_id_to_message_ids = {}
|
|
|
|
attachments = Attachment.objects.filter(messages__id__in=msg_ids)
|
|
|
|
for attachment in attachments:
|
2018-10-25 20:44:48 +02:00
|
|
|
attachment_id_to_message_ids[attachment.id] = {message.id for message in attachment.messages.all()}
|
2018-09-13 16:29:12 +02:00
|
|
|
|
|
|
|
(user_msgs_ids_before, all_msgs_ids_before) = self._check_messages_before_archiving(msg_ids)
|
|
|
|
attachments_ids_before = list(attachments.order_by("id").values_list("id", flat=True))
|
2017-05-14 21:14:26 +02:00
|
|
|
self.assertEqual(ArchivedAttachment.objects.count(), 0)
|
2018-09-13 16:29:12 +02:00
|
|
|
move_messages_to_archive(message_ids=msg_ids)
|
|
|
|
self._check_messages_after_archiving(msg_ids, user_msgs_ids_before, all_msgs_ids_before)
|
2017-05-14 21:14:26 +02:00
|
|
|
self.assertEqual(Attachment.objects.count(), 0)
|
2018-09-13 16:29:12 +02:00
|
|
|
archived_attachments = ArchivedAttachment.objects.filter(messages__id__in=msg_ids)
|
|
|
|
arc_attachments_ids_after = list(archived_attachments.order_by("id").values_list("id", flat=True))
|
2017-05-14 21:14:26 +02:00
|
|
|
self.assertEqual(attachments_ids_before, arc_attachments_ids_after)
|
2018-09-13 16:29:12 +02:00
|
|
|
for attachment in archived_attachments:
|
|
|
|
self.assertEqual(attachment_id_to_message_ids[attachment.id],
|
2018-10-25 20:44:48 +02:00
|
|
|
{message.id for message in attachment.messages.all()})
|
2017-05-14 21:14:26 +02:00
|
|
|
|
2017-11-19 04:02:03 +01:00
|
|
|
def test_archiving_message_with_shared_attachment(self) -> None:
|
2017-05-14 21:14:26 +02:00
|
|
|
# Check do not removing attachments which is used in other messages.
|
|
|
|
self._create_attachments()
|
|
|
|
body = """Some files here ...[zulip.txt](
|
|
|
|
http://localhost:9991/user_uploads/1/31/4CBjtTLYZhk66pZrF8hnYGwc/zulip.txt)
|
|
|
|
http://localhost:9991/user_uploads/1/31/4CBjtTLYZhk66pZrF8hnYGwc/temp_file.py ....
|
2018-09-13 16:29:12 +02:00
|
|
|
Some more.... http://localhost:9991/user_uploads/1/31/4CBjtTLYZhk66pZrF8hnYGwc/abc.py ...
|
|
|
|
http://localhost:9991/user_uploads/1/31/4CBjtTLYZhk66pZrF8hnYGwc/new.py ....
|
|
|
|
http://localhost:9991/user_uploads/1/31/4CBjtTLYZhk66pZrF8hnYGwc/hello.txt ....
|
2017-05-14 21:14:26 +02:00
|
|
|
"""
|
2017-10-28 18:00:16 +02:00
|
|
|
msg_id = self.send_personal_message(self.sender, self.recipient, body)
|
|
|
|
|
|
|
|
# Simulate a reply with the same contents.
|
|
|
|
msg_id_shared_attachments = self.send_personal_message(
|
|
|
|
from_email=self.recipient,
|
|
|
|
to_email=self.sender,
|
|
|
|
content=body,
|
|
|
|
)
|
|
|
|
|
2018-09-13 16:29:12 +02:00
|
|
|
(user_msgs_ids_before, all_msgs_ids_before) = self._check_messages_before_archiving([msg_id])
|
2017-05-14 21:14:26 +02:00
|
|
|
attachments_ids_before = list(Attachment.objects.filter(
|
|
|
|
messages__id=msg_id).order_by("id").values_list("id", flat=True))
|
|
|
|
self.assertEqual(ArchivedAttachment.objects.count(), 0)
|
2018-09-13 16:29:12 +02:00
|
|
|
move_messages_to_archive(message_ids=[msg_id])
|
|
|
|
self._check_messages_after_archiving([msg_id], user_msgs_ids_before, all_msgs_ids_before)
|
|
|
|
self.assertEqual(Attachment.objects.count(), 5)
|
2017-05-14 21:14:26 +02:00
|
|
|
arc_attachments_ids_after = list(ArchivedAttachment.objects.filter(
|
|
|
|
messages__id=msg_id).order_by("id").values_list("id", flat=True))
|
|
|
|
self.assertEqual(attachments_ids_before, arc_attachments_ids_after)
|
2018-09-13 16:29:12 +02:00
|
|
|
move_messages_to_archive(message_ids=[msg_id_shared_attachments])
|
2017-05-14 21:14:26 +02:00
|
|
|
self.assertEqual(Attachment.objects.count(), 0)
|