from datetime import datetime, timedelta from typing import Any, Dict, List, Optional, Tuple from unittest import mock from django.conf import settings from django.utils.timezone import now as timezone_now from zerver.lib.actions import ( do_add_submessage, do_create_realm, do_delete_messages, do_set_realm_property, internal_send_private_message, ) from zerver.lib.retention import ( archive_messages, clean_archived_data, get_realms_and_streams_for_archiving, move_messages_to_archive, restore_all_data_from_archive, restore_retention_policy_deletions_for_stream, ) from zerver.lib.test_classes import ZulipTestCase from zerver.lib.test_helpers import queries_captured, zulip_reaction_info from zerver.lib.upload import create_attachment from zerver.models import ( ArchivedAttachment, ArchivedMessage, ArchivedReaction, ArchivedSubMessage, ArchivedUserMessage, ArchiveTransaction, Attachment, Message, Reaction, Realm, Stream, SubMessage, UserMessage, get_realm, get_stream, get_system_bot, ) # Class with helper functions useful for testing archiving of reactions: from zerver.tornado.django_api import send_event ZULIP_REALM_DAYS = 30 MIT_REALM_DAYS = 100 class RetentionTestingBase(ZulipTestCase): def _get_usermessage_ids(self, message_ids: List[int]) -> List[int]: return list( UserMessage.objects.filter(message_id__in=message_ids).values_list("id", flat=True) ) def _verify_archive_data( self, expected_message_ids: List[int], expected_usermessage_ids: List[int] ) -> None: self.assertEqual( set(ArchivedMessage.objects.values_list("id", flat=True)), set(expected_message_ids), ) self.assertEqual( set(ArchivedUserMessage.objects.values_list("id", flat=True)), set(expected_usermessage_ids), ) # Archived Messages and UserMessages should have been removed from the normal tables: self.assertEqual(Message.objects.filter(id__in=expected_message_ids).count(), 0) self.assertEqual(UserMessage.objects.filter(id__in=expected_usermessage_ids).count(), 0) def _verify_restored_data( self, expected_message_ids: List[int], expected_usermessage_ids: List[int] ) -> None: # Check that the data was restored: self.assertEqual( set(Message.objects.filter(id__in=expected_message_ids).values_list("id", flat=True)), set(expected_message_ids), ) self.assertEqual( set( UserMessage.objects.filter(id__in=expected_usermessage_ids).values_list( "id", flat=True ) ), set(expected_usermessage_ids), ) # The Messages and UserMessages should still be in the archive - we don't delete them. self.assertEqual( set(ArchivedMessage.objects.values_list("id", flat=True)), set(expected_message_ids), ) self.assertEqual( set(ArchivedUserMessage.objects.values_list("id", flat=True)), set(expected_usermessage_ids), ) class ArchiveMessagesTestingBase(RetentionTestingBase): def setUp(self) -> None: super().setUp() self.zulip_realm = get_realm("zulip") self.mit_realm = get_realm("zephyr") self._set_realm_message_retention_value(self.zulip_realm, ZULIP_REALM_DAYS) self._set_realm_message_retention_value(self.mit_realm, MIT_REALM_DAYS) # Set publication date of all existing messages to "now", so that we have full # control over what's expired and what isn't. Message.objects.all().update(date_sent=timezone_now()) def _set_realm_message_retention_value(self, realm: Realm, retention_period: int) -> None: realm.message_retention_days = retention_period realm.save() def _set_stream_message_retention_value( self, stream: Stream, retention_period: Optional[int] ) -> None: stream.message_retention_days = retention_period stream.save() def _change_messages_date_sent(self, msgs_ids: List[int], date_sent: datetime) -> None: Message.objects.filter(id__in=msgs_ids).update(date_sent=date_sent) def _make_mit_messages(self, message_quantity: int, date_sent: datetime) -> Any: # send messages from mit.edu realm and change messages pub date sender = self.mit_user("espuser") recipient = self.mit_user("starnine") msg_ids = [self.send_personal_message(sender, recipient) for i in range(message_quantity)] self._change_messages_date_sent(msg_ids, date_sent) return msg_ids def _send_cross_realm_personal_message(self) -> int: # Send message from bot to users from different realm. bot_email = "notification-bot@zulip.com" internal_realm = get_realm(settings.SYSTEM_BOT_REALM) zulip_user = self.example_user("hamlet") msg_id = internal_send_private_message( sender=get_system_bot(bot_email, internal_realm.id), recipient_user=zulip_user, content="test message", ) assert msg_id is not None return msg_id def _make_expired_zulip_messages(self, message_quantity: int) -> List[int]: msg_ids = list( Message.objects.order_by("id") .filter(sender__realm=self.zulip_realm) .values_list("id", flat=True)[3 : 3 + message_quantity] ) self._change_messages_date_sent( msg_ids, timezone_now() - timedelta(ZULIP_REALM_DAYS + 1), ) return msg_ids def _send_messages_with_attachments(self) -> Dict[str, int]: user_profile = self.example_user("hamlet") sample_size = 10 host = user_profile.realm.host realm_id = get_realm("zulip").id dummy_files = [ ("zulip.txt", f"{realm_id}/31/4CBjtTLYZhk66pZrF8hnYGwc/zulip.txt", sample_size), ("temp_file.py", f"{realm_id}/31/4CBjtTLYZhk66pZrF8hnYGwc/temp_file.py", sample_size), ("abc.py", f"{realm_id}/31/4CBjtTLYZhk66pZrF8hnYGwc/abc.py", sample_size), ] for file_name, path_id, size in dummy_files: create_attachment(file_name, path_id, user_profile, size) self.subscribe(user_profile, "Denmark") body = ( "Some files here ... [zulip.txt](http://{host}/user_uploads/{id}/31/4CBjtTLYZhk66pZrF8hnYGwc/zulip.txt)" + " http://{host}/user_uploads/{id}/31/4CBjtTLYZhk66pZrF8hnYGwc/temp_file.py.... Some more...." + " http://{host}/user_uploads/{id}/31/4CBjtTLYZhk66pZrF8hnYGwc/abc.py" ).format(id=realm_id, host=host) expired_message_id = self.send_stream_message(user_profile, "Denmark", body) actual_message_id = self.send_stream_message(user_profile, "Denmark", body) othello = self.example_user("othello") other_message_id = self.send_stream_message(othello, "Denmark", body) self._change_messages_date_sent( [expired_message_id], timezone_now() - timedelta(days=MIT_REALM_DAYS + 1) ) return { "expired_message_id": expired_message_id, "actual_message_id": actual_message_id, "other_user_message_id": other_message_id, } class TestArchiveMessagesGeneral(ArchiveMessagesTestingBase): def test_no_expired_messages(self) -> None: archive_messages() self.assertEqual(ArchivedUserMessage.objects.count(), 0) self.assertEqual(ArchivedMessage.objects.count(), 0) def test_expired_messages_in_each_realm(self) -> None: """General test for archiving expired messages properly with multiple realms involved""" # Make some expired messages in MIT: expired_mit_msg_ids = self._make_mit_messages( 5, timezone_now() - timedelta(days=MIT_REALM_DAYS + 1), ) # Make some non-expired messages in MIT: self._make_mit_messages(4, timezone_now() - timedelta(days=MIT_REALM_DAYS - 1)) # Change some Zulip messages to be expired: expired_zulip_msg_ids = list( Message.objects.order_by("id") .filter(sender__realm=self.zulip_realm) .values_list("id", flat=True)[3:10] ) self._change_messages_date_sent( expired_zulip_msg_ids, timezone_now() - timedelta(ZULIP_REALM_DAYS + 1), ) expired_msg_ids = expired_mit_msg_ids + expired_zulip_msg_ids expired_usermsg_ids = self._get_usermessage_ids(expired_msg_ids) archive_messages() self._verify_archive_data(expired_msg_ids, expired_usermsg_ids) restore_all_data_from_archive() self._verify_restored_data(expired_msg_ids, expired_usermsg_ids) def test_expired_messages_in_one_realm(self) -> None: """Test with a retention policy set for only the MIT realm""" self._set_realm_message_retention_value(self.zulip_realm, -1) # Make some expired messages in MIT: expired_mit_msg_ids = self._make_mit_messages( 5, timezone_now() - timedelta(days=MIT_REALM_DAYS + 1), ) # Make some non-expired messages in MIT: self._make_mit_messages(4, timezone_now() - timedelta(days=MIT_REALM_DAYS - 1)) # Change some Zulip messages date_sent, but the realm has no retention policy, # so they shouldn't get archived zulip_msg_ids = list( Message.objects.order_by("id") .filter(sender__realm=self.zulip_realm) .values_list("id", flat=True)[3:10] ) self._change_messages_date_sent( zulip_msg_ids, timezone_now() - timedelta(ZULIP_REALM_DAYS + 1), ) # Only MIT has a retention policy: expired_msg_ids = expired_mit_msg_ids expired_usermsg_ids = self._get_usermessage_ids(expired_msg_ids) archive_messages() self._verify_archive_data(expired_msg_ids, expired_usermsg_ids) restore_all_data_from_archive() self._verify_restored_data(expired_msg_ids, expired_usermsg_ids) self._set_realm_message_retention_value(self.zulip_realm, ZULIP_REALM_DAYS) def test_different_stream_realm_policies(self) -> None: verona = get_stream("Verona", self.zulip_realm) hamlet = self.example_user("hamlet") msg_id = self.send_stream_message(hamlet, "Verona", "test") usermsg_ids = self._get_usermessage_ids([msg_id]) self._change_messages_date_sent([msg_id], timezone_now() - timedelta(days=2)) # Don't archive if stream's retention policy set to -1: self._set_realm_message_retention_value(self.zulip_realm, 1) self._set_stream_message_retention_value(verona, -1) archive_messages() self._verify_archive_data([], []) # Don't archive if stream and realm have no retention policy: self._set_realm_message_retention_value(self.zulip_realm, -1) self._set_stream_message_retention_value(verona, None) archive_messages() self._verify_archive_data([], []) # Archive if stream has a retention policy set: self._set_realm_message_retention_value(self.zulip_realm, -1) self._set_stream_message_retention_value(verona, 1) archive_messages() self._verify_archive_data([msg_id], usermsg_ids) def test_cross_realm_personal_message_archiving(self) -> None: """Check that cross-realm personal messages get correctly archived.""" msg_ids = [self._send_cross_realm_personal_message() for i in range(1, 7)] usermsg_ids = self._get_usermessage_ids(msg_ids) # Make the message expired on the recipient's realm: self._change_messages_date_sent(msg_ids, timezone_now() - timedelta(ZULIP_REALM_DAYS + 1)) archive_messages() self._verify_archive_data(msg_ids, usermsg_ids) def test_archiving_interrupted(self) -> None: """Check that queries get rolled back to a consistent state if archiving gets interrupted in the middle of processing a chunk.""" expired_msg_ids = self._make_expired_zulip_messages(7) expired_usermsg_ids = self._get_usermessage_ids(expired_msg_ids) # Insert an exception near the end of the archiving process of a chunk: with mock.patch("zerver.lib.retention.delete_messages", side_effect=Exception): with self.assertRaises(Exception): # Specify large chunk_size to ensure things happen in a single batch archive_messages(chunk_size=1000) # Archiving code has been executed, but because we got an exception, things should have been rolled back: self._verify_archive_data([], []) self.assertEqual( set(Message.objects.filter(id__in=expired_msg_ids).values_list("id", flat=True)), set(expired_msg_ids), ) self.assertEqual( set( UserMessage.objects.filter(id__in=expired_usermsg_ids).values_list( "id", flat=True ) ), set(expired_usermsg_ids), ) def test_archive_message_tool(self) -> None: """End-to-end test of the archiving tool, directly calling archive_messages.""" # Make some expired messages in MIT: expired_mit_msg_ids = self._make_mit_messages( 5, timezone_now() - timedelta(days=MIT_REALM_DAYS + 1), ) # Make some non-expired messages in MIT: self._make_mit_messages(4, timezone_now() - timedelta(days=MIT_REALM_DAYS - 1)) # Change some Zulip messages to be expired: expired_zulip_msg_ids = self._make_expired_zulip_messages(7) expired_crossrealm_msg_id = self._send_cross_realm_personal_message() # Make the message expired in the recipient's realm: self._change_messages_date_sent( [expired_crossrealm_msg_id], timezone_now() - timedelta(ZULIP_REALM_DAYS + 1), ) expired_msg_ids = [*expired_mit_msg_ids, *expired_zulip_msg_ids, expired_crossrealm_msg_id] expired_usermsg_ids = self._get_usermessage_ids(expired_msg_ids) archive_messages(chunk_size=2) # Specify low chunk_size to test batching. # Make sure we archived what needed: self._verify_archive_data(expired_msg_ids, expired_usermsg_ids) restore_all_data_from_archive() self._verify_restored_data(expired_msg_ids, expired_usermsg_ids) def test_archiving_attachments(self) -> None: """End-to-end test for the logic for archiving attachments. This test is hard to read without first reading _send_messages_with_attachments""" msgs_ids = self._send_messages_with_attachments() # First, confirm deleting the oldest message # (`expired_message_id`) creates ArchivedAttachment objects # and associates that message ID with them, but does not # delete the Attachment object. archive_messages() self.assertEqual(ArchivedAttachment.objects.count(), 3) self.assertEqual( list( ArchivedAttachment.objects.distinct("messages__id").values_list( "messages__id", flat=True ) ), [msgs_ids["expired_message_id"]], ) self.assertEqual(Attachment.objects.count(), 3) # Now make `actual_message_id` expired too. We still don't # delete the Attachment objects. self._change_messages_date_sent( [msgs_ids["actual_message_id"]], timezone_now() - timedelta(days=MIT_REALM_DAYS + 1) ) archive_messages() self.assertEqual(Attachment.objects.count(), 3) # Finally, make the last message mentioning those attachments # expired. We should now delete the Attachment objects and # each ArchivedAttachment object should list all 3 messages. self._change_messages_date_sent( [msgs_ids["other_user_message_id"]], timezone_now() - timedelta(days=MIT_REALM_DAYS + 1) ) archive_messages() self.assertEqual(Attachment.objects.count(), 0) self.assertEqual(ArchivedAttachment.objects.count(), 3) self.assertEqual( list( ArchivedAttachment.objects.distinct("messages__id") .order_by("messages__id") .values_list("messages__id", flat=True) ), sorted(msgs_ids.values()), ) restore_all_data_from_archive() # Attachments should have been restored: self.assertEqual(Attachment.objects.count(), 3) # Archived data doesn't get deleted by restoring. self.assertEqual(ArchivedAttachment.objects.count(), 3) self.assertEqual( list( Attachment.objects.distinct("messages__id") .order_by("messages__id") .values_list("messages__id", flat=True) ), sorted(msgs_ids.values()), ) def test_restoring_and_rearchiving(self) -> None: expired_msg_ids = self._make_mit_messages( 7, timezone_now() - timedelta(days=MIT_REALM_DAYS + 1), ) expired_usermsg_ids = self._get_usermessage_ids(expired_msg_ids) archive_messages(chunk_size=4) self._verify_archive_data(expired_msg_ids, expired_usermsg_ids) transactions = ArchiveTransaction.objects.all() self.assert_length(transactions, 2) # With chunk_size 4, there should be 2 transactions restore_all_data_from_archive() transactions[0].refresh_from_db() transactions[1].refresh_from_db() self.assertTrue(transactions[0].restored) self.assertTrue(transactions[1].restored) archive_messages(chunk_size=10) self._verify_archive_data(expired_msg_ids, expired_usermsg_ids) transactions = ArchiveTransaction.objects.order_by("id") self.assert_length(transactions, 3) archived_messages = ArchivedMessage.objects.filter(id__in=expired_msg_ids) # Check that the re-archived messages are correctly assigned to the new transaction: for message in archived_messages: self.assertEqual(message.archive_transaction_id, transactions[2].id) class TestArchivingSubMessages(ArchiveMessagesTestingBase): def test_archiving_submessages(self) -> None: expired_msg_ids = self._make_expired_zulip_messages(2) cordelia = self.example_user("cordelia") hamlet = self.example_user("hamlet") do_add_submessage( realm=self.zulip_realm, sender_id=cordelia.id, message_id=expired_msg_ids[0], msg_type="whatever", content='{"name": "alice", "salary": 20}', ) do_add_submessage( realm=self.zulip_realm, sender_id=hamlet.id, message_id=expired_msg_ids[0], msg_type="whatever", content='{"name": "john", "salary": 30}', ) do_add_submessage( realm=self.zulip_realm, sender_id=cordelia.id, message_id=expired_msg_ids[1], msg_type="whatever", content='{"name": "jack", "salary": 10}', ) submessage_ids = list( SubMessage.objects.filter(message_id__in=expired_msg_ids).values_list("id", flat=True), ) self.assert_length(submessage_ids, 3) self.assertEqual(SubMessage.objects.filter(id__in=submessage_ids).count(), 3) archive_messages() self.assertEqual(SubMessage.objects.filter(id__in=submessage_ids).count(), 0) self.assertEqual( set( ArchivedSubMessage.objects.filter(id__in=submessage_ids).values_list( "id", flat=True ) ), set(submessage_ids), ) restore_all_data_from_archive() self.assertEqual( set(SubMessage.objects.filter(id__in=submessage_ids).values_list("id", flat=True)), set(submessage_ids), ) class TestArchivingReactions(ArchiveMessagesTestingBase): def test_archiving_reactions(self) -> None: expired_msg_ids = self._make_expired_zulip_messages(2) hamlet = self.example_user("hamlet") cordelia = self.example_user("cordelia") for sender in [hamlet, cordelia]: self.api_post( sender, f"/api/v1/messages/{expired_msg_ids[0]}/reactions", zulip_reaction_info(), ) self.api_post( hamlet, f"/api/v1/messages/{expired_msg_ids[1]}/reactions", zulip_reaction_info(), ) reaction_ids = list( Reaction.objects.filter(message_id__in=expired_msg_ids).values_list("id", flat=True), ) self.assert_length(reaction_ids, 3) self.assertEqual(Reaction.objects.filter(id__in=reaction_ids).count(), 3) archive_messages() self.assertEqual(Reaction.objects.filter(id__in=reaction_ids).count(), 0) self.assertEqual( set(ArchivedReaction.objects.filter(id__in=reaction_ids).values_list("id", flat=True)), set(reaction_ids), ) restore_all_data_from_archive() self.assertEqual( set(Reaction.objects.filter(id__in=reaction_ids).values_list("id", flat=True)), set(reaction_ids), ) class MoveMessageToArchiveBase(RetentionTestingBase): def setUp(self) -> None: super().setUp() self.sender = self.example_user("hamlet") self.recipient = self.example_user("cordelia") def _create_attachments(self) -> None: sample_size = 10 realm_id = get_realm("zulip").id dummy_files = [ ("zulip.txt", f"{realm_id}/31/4CBjtTLYZhk66pZrF8hnYGwc/zulip.txt", sample_size), ("temp_file.py", f"{realm_id}/31/4CBjtTLYZhk66pZrF8hnYGwc/temp_file.py", sample_size), ("abc.py", f"{realm_id}/31/4CBjtTLYZhk66pZrF8hnYGwc/abc.py", sample_size), ("hello.txt", f"{realm_id}/31/4CBjtTLYZhk66pZrF8hnYGwc/hello.txt", sample_size), ("new.py", f"{realm_id}/31/4CBjtTLYZhk66pZrF8hnYGwc/new.py", sample_size), ] user_profile = self.example_user("hamlet") for file_name, path_id, size in dummy_files: create_attachment(file_name, path_id, user_profile, size) def _assert_archive_empty(self) -> None: self.assertFalse(ArchivedUserMessage.objects.exists()) self.assertFalse(ArchivedMessage.objects.exists()) self.assertFalse(ArchivedAttachment.objects.exists()) class MoveMessageToArchiveGeneral(MoveMessageToArchiveBase): def test_personal_messages_archiving(self) -> None: msg_ids = [self.send_personal_message(self.sender, self.recipient) for i in range(0, 3)] usermsg_ids = self._get_usermessage_ids(msg_ids) self._assert_archive_empty() move_messages_to_archive(message_ids=msg_ids) self._verify_archive_data(msg_ids, usermsg_ids) restore_all_data_from_archive() self._verify_restored_data(msg_ids, usermsg_ids) def test_move_messages_to_archive_with_realm_argument(self) -> None: realm = get_realm("zulip") msg_ids = [self.send_personal_message(self.sender, self.recipient) for i in range(0, 3)] usermsg_ids = self._get_usermessage_ids(msg_ids) self._assert_archive_empty() move_messages_to_archive(message_ids=msg_ids, realm=realm) self._verify_archive_data(msg_ids, usermsg_ids) archive_transaction = ArchiveTransaction.objects.last() assert archive_transaction is not None self.assertEqual(archive_transaction.realm, realm) def test_stream_messages_archiving(self) -> None: msg_ids = [self.send_stream_message(self.sender, "Verona") for i in range(0, 3)] usermsg_ids = self._get_usermessage_ids(msg_ids) self._assert_archive_empty() move_messages_to_archive(message_ids=msg_ids) self._verify_archive_data(msg_ids, usermsg_ids) restore_all_data_from_archive() self._verify_restored_data(msg_ids, usermsg_ids) def test_archiving_messages_second_time(self) -> None: msg_ids = [self.send_stream_message(self.sender, "Verona") for i in range(0, 3)] usermsg_ids = self._get_usermessage_ids(msg_ids) self._assert_archive_empty() move_messages_to_archive(message_ids=msg_ids) self._verify_archive_data(msg_ids, usermsg_ids) with self.assertRaises(Message.DoesNotExist): move_messages_to_archive(message_ids=msg_ids) def test_archiving_messages_multiple_realms(self) -> None: """ Verifies that move_messages_to_archive works correctly if called on messages in multiple realms. """ iago = self.example_user("iago") othello = self.example_user("othello") cordelia = self.lear_user("cordelia") king = self.lear_user("king") zulip_msg_ids = [self.send_personal_message(iago, othello) for i in range(0, 3)] leary_msg_ids = [self.send_personal_message(cordelia, king) for i in range(0, 3)] msg_ids = zulip_msg_ids + leary_msg_ids usermsg_ids = self._get_usermessage_ids(msg_ids) self._assert_archive_empty() move_messages_to_archive(message_ids=msg_ids) self._verify_archive_data(msg_ids, usermsg_ids) restore_all_data_from_archive() self._verify_restored_data(msg_ids, usermsg_ids) def test_archiving_messages_with_attachment(self) -> None: self._create_attachments() realm_id = get_realm("zulip").id host = get_realm("zulip").host body1 = """Some files here ...[zulip.txt]( http://{host}/user_uploads/{id}/31/4CBjtTLYZhk66pZrF8hnYGwc/zulip.txt) http://{host}/user_uploads/{id}/31/4CBjtTLYZhk66pZrF8hnYGwc/temp_file.py .... Some more.... http://{host}/user_uploads/{id}/31/4CBjtTLYZhk66pZrF8hnYGwc/abc.py """.format( id=realm_id, host=host ) body2 = """Some files here http://{host}/user_uploads/{id}/31/4CBjtTLYZhk66pZrF8hnYGwc/zulip.txt ... http://{host}/user_uploads/{id}/31/4CBjtTLYZhk66pZrF8hnYGwc/hello.txt .... http://{host}/user_uploads/{id}/31/4CBjtTLYZhk66pZrF8hnYGwc/new.py .... """.format( id=realm_id, host=host ) msg_ids = [ self.send_personal_message(self.sender, self.recipient, body1), self.send_personal_message(self.sender, self.recipient, body2), ] attachment_id_to_message_ids: Dict[int, List[int]] = {} attachment_ids = list( Attachment.objects.filter(messages__id__in=msg_ids).values_list("id", flat=True), ) for attachment_id in attachment_ids: attachment_id_to_message_ids[attachment_id] = list( Message.objects.filter(attachment__id=attachment_id).values_list("id", flat=True), ) usermsg_ids = self._get_usermessage_ids(msg_ids) self._assert_archive_empty() move_messages_to_archive(message_ids=msg_ids) self._verify_archive_data(msg_ids, usermsg_ids) self.assertFalse(Attachment.objects.exists()) archived_attachment_ids = list( ArchivedAttachment.objects.filter(messages__id__in=msg_ids).values_list( "id", flat=True ), ) self.assertEqual(set(attachment_ids), set(archived_attachment_ids)) for attachment_id in archived_attachment_ids: self.assertEqual( set(attachment_id_to_message_ids[attachment_id]), set( ArchivedMessage.objects.filter( archivedattachment__id=attachment_id ).values_list("id", flat=True) ), ) restore_all_data_from_archive() self._verify_restored_data(msg_ids, usermsg_ids) restored_attachment_ids = list( Attachment.objects.filter(messages__id__in=msg_ids).values_list("id", flat=True), ) self.assertEqual(set(attachment_ids), set(restored_attachment_ids)) for attachment_id in restored_attachment_ids: self.assertEqual( set(attachment_id_to_message_ids[attachment_id]), set( Message.objects.filter(attachment__id=attachment_id).values_list( "id", flat=True ) ), ) def test_archiving_message_with_shared_attachment(self) -> None: # Make sure that attachments still in use in other messages don't get deleted: self._create_attachments() realm_id = get_realm("zulip").id host = get_realm("zulip").host body = """Some files here ...[zulip.txt]( http://{host}/user_uploads/{id}/31/4CBjtTLYZhk66pZrF8hnYGwc/zulip.txt) http://{host}/user_uploads/{id}/31/4CBjtTLYZhk66pZrF8hnYGwc/temp_file.py .... Some more.... http://{host}/user_uploads/{id}/31/4CBjtTLYZhk66pZrF8hnYGwc/abc.py ... http://{host}/user_uploads/{id}/31/4CBjtTLYZhk66pZrF8hnYGwc/new.py .... http://{host}/user_uploads/{id}/31/4CBjtTLYZhk66pZrF8hnYGwc/hello.txt .... """.format( id=realm_id, host=host ) msg_id = self.send_personal_message(self.sender, self.recipient, body) # Simulate a reply with the same contents. reply_msg_id = self.send_personal_message( from_user=self.recipient, to_user=self.sender, content=body, ) usermsg_ids = self._get_usermessage_ids([msg_id]) attachment_ids = list( Attachment.objects.filter(messages__id=msg_id).values_list("id", flat=True), ) self._assert_archive_empty() # Archive one of the messages: move_messages_to_archive(message_ids=[msg_id]) self._verify_archive_data([msg_id], usermsg_ids) # Attachments shouldn't have been deleted, as the second message links to them: self.assertEqual(Attachment.objects.count(), 5) self.assertEqual( set( ArchivedAttachment.objects.filter(messages__id=msg_id).values_list("id", flat=True) ), set(attachment_ids), ) # Restore the first message: restore_all_data_from_archive() # Archive the second: move_messages_to_archive(message_ids=[reply_msg_id]) # The restored messages links to the Attachments, so they shouldn't be deleted: self.assertEqual(Attachment.objects.count(), 5) # Archive the first message again: move_messages_to_archive(message_ids=[msg_id]) # Now the attachment should have been deleted: self.assertEqual(Attachment.objects.count(), 0) # Restore everything: restore_all_data_from_archive() self.assertEqual( set(Attachment.objects.filter(messages__id=msg_id).values_list("id", flat=True)), set(attachment_ids), ) class MoveMessageToArchiveWithSubMessages(MoveMessageToArchiveBase): def test_archiving_message_with_submessages(self) -> None: msg_id = self.send_stream_message(self.sender, "Verona") cordelia = self.example_user("cordelia") hamlet = self.example_user("hamlet") do_add_submessage( realm=get_realm("zulip"), sender_id=cordelia.id, message_id=msg_id, msg_type="whatever", content='{"name": "alice", "salary": 20}', ) do_add_submessage( realm=get_realm("zulip"), sender_id=hamlet.id, message_id=msg_id, msg_type="whatever", content='{"name": "john", "salary": 30}', ) submessage_ids = list( SubMessage.objects.filter(message_id=msg_id).values_list("id", flat=True), ) self.assertEqual(SubMessage.objects.filter(id__in=submessage_ids).count(), 2) move_messages_to_archive(message_ids=[msg_id]) self.assertEqual( set(ArchivedSubMessage.objects.filter(message_id=msg_id).values_list("id", flat=True)), set(submessage_ids), ) self.assertEqual(SubMessage.objects.filter(id__in=submessage_ids).count(), 0) restore_all_data_from_archive() self.assertEqual( set(SubMessage.objects.filter(id__in=submessage_ids).values_list("id", flat=True)), set(submessage_ids), ) class MoveMessageToArchiveWithReactions(MoveMessageToArchiveBase): def test_archiving_message_with_reactions(self) -> None: msg_id = self.send_stream_message(self.sender, "Verona") for name in ["hamlet", "cordelia"]: self.api_post( self.example_user(name), f"/api/v1/messages/{msg_id}/reactions", zulip_reaction_info(), ) reaction_ids = list( Reaction.objects.filter(message_id=msg_id).values_list("id", flat=True), ) self.assertEqual(Reaction.objects.filter(id__in=reaction_ids).count(), 2) move_messages_to_archive(message_ids=[msg_id]) self.assertEqual( set(ArchivedReaction.objects.filter(message_id=msg_id).values_list("id", flat=True)), set(reaction_ids), ) self.assertEqual(Reaction.objects.filter(id__in=reaction_ids).count(), 0) restore_all_data_from_archive() self.assertEqual( set(Reaction.objects.filter(id__in=reaction_ids).values_list("id", flat=True)), set(reaction_ids), ) class TestCleaningArchive(ArchiveMessagesTestingBase): def test_clean_archived_data(self) -> None: self._make_expired_zulip_messages(7) archive_messages(chunk_size=2) # Small chunk size to have multiple transactions transactions = list(ArchiveTransaction.objects.all()) for transaction in transactions[0:-1]: transaction.timestamp = timezone_now() - timedelta( days=settings.ARCHIVED_DATA_VACUUMING_DELAY_DAYS + 1 ) transaction.save() message_ids_to_clean = list( ArchivedMessage.objects.filter(archive_transaction__in=transactions[0:-1]).values_list( "id", flat=True ) ) clean_archived_data() remaining_transactions = list(ArchiveTransaction.objects.all()) self.assert_length(remaining_transactions, 1) # All transactions except the last one were deleted: self.assertEqual(remaining_transactions[0].id, transactions[-1].id) # And corresponding ArchivedMessages should have been deleted: self.assertFalse(ArchivedMessage.objects.filter(id__in=message_ids_to_clean).exists()) self.assertFalse( ArchivedUserMessage.objects.filter(message_id__in=message_ids_to_clean).exists() ) for message in ArchivedMessage.objects.all(): self.assertEqual(message.archive_transaction_id, remaining_transactions[0].id) class TestGetRealmAndStreamsForArchiving(ZulipTestCase): def fix_ordering_of_result(self, result: List[Tuple[Realm, List[Stream]]]) -> None: """ This is a helper for giving the structure returned by get_realms_and_streams_for_archiving a consistent ordering. """ # Sort the list of tuples by realm id: result.sort(key=lambda x: x[0].id) # Now we sort the lists of streams in each tuple: for realm, streams_list in result: streams_list.sort(key=lambda stream: stream.id) def simple_get_realms_and_streams_for_archiving(self) -> List[Tuple[Realm, List[Stream]]]: """ This is an implementation of the function we're testing, but using the obvious, unoptimized algorithm. We can use this for additional verification of correctness, by comparing the output of the two implementations. """ result = [] for realm in Realm.objects.all(): if realm.message_retention_days != -1: streams = Stream.objects.filter(realm=realm).exclude(message_retention_days=-1) result.append((realm, list(streams))) else: streams = ( Stream.objects.filter(realm=realm) .exclude(message_retention_days__isnull=True) .exclude(message_retention_days=-1) ) if streams.exists(): result.append((realm, list(streams))) return result def test_get_realms_and_streams_for_archiving(self) -> None: zulip_realm = get_realm("zulip") zulip_realm.message_retention_days = 10 zulip_realm.save() verona = get_stream("Verona", zulip_realm) verona.message_retention_days = -1 # Block archiving for this stream verona.save() denmark = get_stream("Denmark", zulip_realm) denmark.message_retention_days = 1 denmark.save() zephyr_realm = get_realm("zephyr") zephyr_realm.message_retention_days = -1 zephyr_realm.save() self.make_stream("normal stream", realm=zephyr_realm) archiving_blocked_zephyr_stream = self.make_stream("no archiving", realm=zephyr_realm) archiving_blocked_zephyr_stream.message_retention_days = -1 archiving_blocked_zephyr_stream.save() archiving_enabled_zephyr_stream = self.make_stream("with archiving", realm=zephyr_realm) archiving_enabled_zephyr_stream.message_retention_days = 1 archiving_enabled_zephyr_stream.save() no_archiving_realm = do_create_realm(string_id="no_archiving", name="no_archiving") do_set_realm_property(no_archiving_realm, "invite_required", False, acting_user=None) do_set_realm_property(no_archiving_realm, "message_retention_days", -1, acting_user=None) # Realm for testing the edge case where it has a default retention policy, # but all streams disable it. realm_all_streams_archiving_disabled = do_create_realm( string_id="with_archiving", name="with_archiving" ) do_set_realm_property( realm_all_streams_archiving_disabled, "invite_required", False, acting_user=None ) do_set_realm_property( realm_all_streams_archiving_disabled, "message_retention_days", 1, acting_user=None ) Stream.objects.filter(realm=realm_all_streams_archiving_disabled).update( message_retention_days=-1 ) # We construct a list representing how the result of get_realms_and_streams_for_archiving should be. # One nuisance is that the ordering of the elements in the result structure is not deterministic, # so we use a helper to order both structures in a consistent manner. This wouldn't be necessary # if python had a true "unordered list" data structure. Set doesn't do the job, because it requires # elements to be hashable. expected_result = [ (zulip_realm, list(Stream.objects.filter(realm=zulip_realm).exclude(id=verona.id))), (zephyr_realm, [archiving_enabled_zephyr_stream]), (realm_all_streams_archiving_disabled, []), ] self.fix_ordering_of_result(expected_result) simple_algorithm_result = self.simple_get_realms_and_streams_for_archiving() self.fix_ordering_of_result(simple_algorithm_result) result = get_realms_and_streams_for_archiving() self.fix_ordering_of_result(result) self.assert_length(result, len(expected_result)) self.assertEqual(result, expected_result) self.assert_length(result, len(simple_algorithm_result)) self.assertEqual(result, simple_algorithm_result) class TestRestoreStreamMessages(ArchiveMessagesTestingBase): def test_restore_retention_policy_deletions_for_stream(self) -> None: cordelia = self.example_user("cordelia") hamlet = self.example_user("hamlet") realm = get_realm("zulip") stream_name = "Verona" stream = get_stream(stream_name, realm) message_ids_to_archive_manually = [ self.send_stream_message(cordelia, stream_name, str(i)) for i in range(0, 2) ] usermessage_ids_to_archive_manually = self._get_usermessage_ids( message_ids_to_archive_manually ) message_ids_to_archive_by_policy = [ self.send_stream_message(hamlet, stream_name, str(i)) for i in range(0, 2) ] usermessage_ids_to_archive_by_policy = self._get_usermessage_ids( message_ids_to_archive_by_policy ) expected_archived_message_ids = ( message_ids_to_archive_manually + message_ids_to_archive_by_policy ) expected_archived_usermessage_ids = ( usermessage_ids_to_archive_manually + usermessage_ids_to_archive_by_policy ) self._set_stream_message_retention_value(stream, 5) self._change_messages_date_sent( message_ids_to_archive_by_policy, timezone_now() - timedelta(days=6) ) move_messages_to_archive(message_ids_to_archive_manually) archive_messages() self._verify_archive_data(expected_archived_message_ids, expected_archived_usermessage_ids) restore_retention_policy_deletions_for_stream(stream) # Verify that we restore the stream messages that were archived due to retention policy, # but not the ones manually deleted. self.assert_length( Message.objects.filter(id__in=message_ids_to_archive_by_policy), len(message_ids_to_archive_by_policy), ) self.assertFalse(Message.objects.filter(id__in=message_ids_to_archive_manually)) class TestDoDeleteMessages(ZulipTestCase): def test_do_delete_messages_multiple(self) -> None: realm = get_realm("zulip") cordelia = self.example_user("cordelia") message_ids = [self.send_stream_message(cordelia, "Verona", str(i)) for i in range(0, 10)] messages = Message.objects.filter(id__in=message_ids) with queries_captured() as queries: do_delete_messages(realm, messages) self.assertFalse(Message.objects.filter(id__in=message_ids).exists()) self.assert_length(queries, 19) archived_messages = ArchivedMessage.objects.filter(id__in=message_ids) self.assertEqual(archived_messages.count(), len(message_ids)) self.assert_length({message.archive_transaction_id for message in archived_messages}, 1) def test_old_event_format_processed_correctly(self) -> None: """ do_delete_messages used to send events with users in dict format {"id": }. We have a block in process_notification to deal with that old format, that should be deleted in a later release. This test is meant to ensure correctness of that block. """ realm = get_realm("zulip") cordelia = self.example_user("cordelia") hamlet = self.example_user("hamlet") message_id = self.send_personal_message(cordelia, hamlet) message = Message.objects.get(id=message_id) event = { "type": "delete_message", "sender": message.sender.email, "sender_id": message.sender_id, "message_id": message.id, "message_type": "private", "recipient_id": message.recipient_id, } move_messages_to_archive([message_id]) # We only send the event to see no exception is thrown - as it would be if the block # in process_notification to handle this old format of "users to notify" wasn't correct. send_event(realm, event, [{"id": cordelia.id}, {"id": hamlet.id}])