zulip/zerver/tests/test_message_dict.py

691 lines
27 KiB
Python
Raw Normal View History

from typing import Any, Dict, List
from unittest import mock
from django.utils.timezone import now as timezone_now
from zerver.lib.cache import cache_delete, to_dict_cache_key_id
from zerver.lib.display_recipient import get_display_recipient
from zerver.lib.markdown import version as markdown_version
from zerver.lib.message import messages_for_ids
from zerver.lib.message_cache import MessageDict, sew_messages_and_reactions
per-request caches: Add per_request_cache library. We have historically cached two types of values on a per-request basis inside of memory: * linkifiers * display recipients Both of these caches were hand-written, and they both actually cache values that are also in memcached, so the per-request cache essentially only saves us from a few memcached hits. I think the linkifier per-request cache is a necessary evil. It's an important part of message rendering, and it's not super easy to structure the code to just get a single value up front and pass it down the stack. I'm not so sure we even need the display recipient per-request cache any more, as we are generally pretty smart now about hydrating recipient data in terms of how the code is organized. But I haven't done thorough research on that hypotheseis. Fortunately, it's not rocket science to just write a glorified memoize decorator and tie it into key places in the code: * middleware * tests (e.g. asserting db counts) * queue processors That's what I did in this commit. This commit definitely reduces the amount of code to maintain. I think it also gets us closer to possibly phasing out this whole technique, but that effort is beyond the scope of this PR. We could add some instrumentation to the decorator to see how often we get a non-trivial number of saved round trips to memcached. Note that when we flush linkifiers, we just use a big hammer and flush the entire per-request cache for linkifiers, since there is only ever one realm in the cache.
2023-07-14 19:46:50 +02:00
from zerver.lib.per_request_cache import flush_per_request_caches
from zerver.lib.test_classes import ZulipTestCase
from zerver.lib.test_helpers import make_client
from zerver.lib.topic import TOPIC_LINKS
from zerver.lib.types import DisplayRecipientT, UserDisplayRecipient
from zerver.models import Message, Reaction, Realm, RealmFilter, Recipient, Stream, UserProfile
from zerver.models.realms import get_realm
from zerver.models.streams import get_stream
class MessageDictTest(ZulipTestCase):
def test_both_codepaths(self) -> None:
"""
We have two different codepaths that
extract a particular shape of dictionary
for messages to send to clients:
events:
These are the events we send to MANY
clients when a message is originally
sent.
fetch:
These are the messages we send to ONE
client when they fetch messages via
some narrow/search in the UI.
Different clients have different needs
when it comes to things like generating avatar
hashes or including both rendered and unrendered
Markdown, so that explains the different shapes.
And then the two codepaths have different
performance needs. In the events codepath, we
have the Django view generate a single "wide"
dictionary that gets put on the event queue,
and then we send events to multiple clients,
finalizing the payload for each of them depending
on the "shape" they want. (We also avoid
doing extra work for any two clients who want
the same shape dictionary, but that's out of the
scope of this particular test).
In the fetch scenario, the single client only needs
a dictionary of one shape, but we need to re-hydrate
the sender information, since the sender details
may have changed since the message was originally
sent.
This test simply verifies that the two codepaths
ultimately provide the same result.
"""
def reload_message(msg_id: int) -> Message:
# Get a clean copy of the message, and
# clear the cache.
cache_delete(to_dict_cache_key_id(msg_id))
msg = Message.objects.get(id=msg_id)
return msg
def get_send_message_payload(
msg_id: int, apply_markdown: bool, client_gravatar: bool
) -> Dict[str, Any]:
msg = reload_message(msg_id)
wide_dict = MessageDict.wide_dict(msg)
narrow_dict = MessageDict.finalize_payload(
wide_dict,
apply_markdown=apply_markdown,
client_gravatar=client_gravatar,
)
return narrow_dict
def get_fetch_payload(
msg_id: int, apply_markdown: bool, client_gravatar: bool
) -> Dict[str, Any]:
msg = reload_message(msg_id)
unhydrated_dict = MessageDict.messages_to_encoded_cache_helper([msg])[0]
# The next step mutates the dict in place
# for performance reasons.
MessageDict.post_process_dicts(
[unhydrated_dict],
apply_markdown=apply_markdown,
client_gravatar=client_gravatar,
realm=get_realm("zulip"),
)
final_dict = unhydrated_dict
return final_dict
def test_message_id() -> int:
hamlet = self.example_user("hamlet")
self.login_user(hamlet)
msg_id = self.send_stream_message(
hamlet,
tests: Ensure stream senders get a UserMessage row. We now complain if a test author sends a stream message that does not result in the sender getting a UserMessage row for the message. This is basically 100% equivalent to complaining that the author failed to subscribe the sender to the stream as part of the test setup, as far as I can tell, so the AssertionError instructs the author to subscribe the sender to the stream. We exempt bots from this check, although it is plausible we should only exempt the system bots like the notification bot. I considered auto-subscribing the sender to the stream, but that can be a little more expensive than the current check, and we generally want test setup to be explicit. If there is some legitimate way than a subscribed human sender can't get a UserMessage, then we probably want an explicit test for that, or we may want to change the backend to just write a UserMessage row in that hypothetical situation. For most tests, including almost all the ones fixed here, the author just wants their test setup to realistically reflect normal operation, and often devs may not realize that Cordelia is not subscribed to Denmark or not realize that Hamlet is not subscribed to Scotland. Some of us don't remember our Shakespeare from high school, and our stream subscriptions don't even necessarily reflect which countries the Bard placed his characters in. There may also be some legitimate use case where an author wants to simulate sending a message to an unsubscribed stream, but for those edge cases, they can always set allow_unsubscribed_sender to True.
2021-12-10 13:55:48 +01:00
"Denmark",
topic_name="editing",
content="before edit",
)
return msg_id
flag_setups = [
[False, False],
[False, True],
[True, False],
[True, True],
]
msg_id = test_message_id()
for apply_markdown, client_gravatar in flag_setups:
send_message_payload = get_send_message_payload(
msg_id,
apply_markdown=apply_markdown,
client_gravatar=client_gravatar,
)
fetch_payload = get_fetch_payload(
msg_id,
apply_markdown=apply_markdown,
client_gravatar=client_gravatar,
)
self.assertEqual(send_message_payload, fetch_payload)
def test_bulk_message_fetching(self) -> None:
sender = self.example_user("othello")
receiver = self.example_user("hamlet")
realm = get_realm("zulip")
pm_recipient = Recipient.objects.get(type_id=receiver.id, type=Recipient.PERSONAL)
stream_name = "Çiğdem"
stream = self.make_stream(stream_name)
stream_recipient = Recipient.objects.get(type_id=stream.id, type=Recipient.STREAM)
sending_client = make_client(name="test suite")
ids = []
for i in range(300):
for recipient in [pm_recipient, stream_recipient]:
message = Message(
sender=sender,
recipient=recipient,
realm=realm,
content=f"whatever {i}",
rendered_content="DOES NOT MATTER",
rendered_content_version=markdown_version,
date_sent=timezone_now(),
sending_client=sending_client,
last_edit_time=timezone_now(),
edit_history="[]",
)
message.set_topic_name("whatever")
message.save()
ids.append(message.id)
Reaction.objects.create(
user_profile=sender, message=message, emoji_name="simple_smile"
)
num_ids = len(ids)
self.assertTrue(num_ids >= 600)
with self.assert_database_query_count(7):
objs = MessageDict.ids_to_dict(ids)
MessageDict.post_process_dicts(
objs, apply_markdown=False, client_gravatar=False, realm=realm
)
self.assert_length(objs, num_ids)
def test_applying_markdown(self) -> None:
sender = self.example_user("othello")
receiver = self.example_user("hamlet")
recipient = Recipient.objects.get(type_id=receiver.id, type=Recipient.PERSONAL)
sending_client = make_client(name="test suite")
message = Message(
sender=sender,
recipient=recipient,
realm=receiver.realm,
content="hello **world**",
date_sent=timezone_now(),
sending_client=sending_client,
last_edit_time=timezone_now(),
edit_history="[]",
)
message.set_topic_name("whatever")
message.save()
# An important part of this test is to get the message through this exact code path,
# because there is an ugly hack we need to cover. So don't just say "row = message".
dct = MessageDict.ids_to_dict([message.id])[0]
expected_content = "<p>hello <strong>world</strong></p>"
self.assertEqual(dct["rendered_content"], expected_content)
message = Message.objects.get(id=message.id)
self.assertEqual(message.rendered_content, expected_content)
self.assertEqual(message.rendered_content_version, markdown_version)
@mock.patch("zerver.lib.message_cache.render_message_markdown")
def test_applying_markdown_invalid_format(self, convert_mock: Any) -> None:
# pretend the converter returned an invalid message without raising an exception
convert_mock.return_value = None
sender = self.example_user("othello")
receiver = self.example_user("hamlet")
recipient = Recipient.objects.get(type_id=receiver.id, type=Recipient.PERSONAL)
sending_client = make_client(name="test suite")
message = Message(
sender=sender,
recipient=recipient,
realm=receiver.realm,
content="hello **world**",
date_sent=timezone_now(),
sending_client=sending_client,
last_edit_time=timezone_now(),
edit_history="[]",
)
message.set_topic_name("whatever")
message.save()
# An important part of this test is to get the message through this exact code path,
# because there is an ugly hack we need to cover. So don't just say "row = message".
dct = MessageDict.ids_to_dict([message.id])[0]
error_content = (
"<p>[Zulip note: Sorry, we could not understand the formatting of your message]</p>"
)
self.assertEqual(dct["rendered_content"], error_content)
def test_topic_links_use_stream_realm(self) -> None:
# Set up a realm filter on 'zulip' and assert that messages
# sent to a stream on 'zulip' have the topic linkified,
# and not linkified when sent to a stream in 'lear'.
zulip_realm = get_realm("zulip")
lear_realm = get_realm("lear")
linkifier: Support URL templates for linkifiers. This swaps out url_format_string from all of our APIs and replaces it with url_template. Note that the documentation changes in the following commits will be squashed with this commit. We change the "url_format" key to "url_template" for the realm_linkifiers events in event_schema, along with updating LinkifierDict. "url_template" is the name chosen to normalize mixed usages of "url_format_string" and "url_format" throughout the backend. The markdown processor is updated to stop handling the format string interpolation and delegate the task template expansion to the uri_template library instead. This change affects many test cases. We mostly just replace "%(name)s" with "{name}", "url_format_string" with "url_template" to make sure that they still pass. There are some test cases dedicated for testing "%" escaping, which aren't relevant anymore and are subject to removal. But for now we keep most of them as-is, and make sure that "%" is always escaped since we do not use it for variable substitution any more. Since url_format_string is not populated anymore, a migration is created to remove this field entirely, and make url_template non-nullable since we will always populate it. Note that it is possible to have url_template being null after migration 0422 and before 0424, but in practice, url_template will not be None after backfilling and the backend now is always setting url_template. With the removal of url_format_string, RealmFilter model will now be cleaned with URL template checks, and the old checks for escapes are removed. We also modified RealmFilter.clean to skip the validation when the url_template is invalid. This avoids raising mulitple ValidationError's when calling full_clean on a linkifier. But we might eventually want to have a more centric approach to data validation instead of having the same validation in both the clean method and the validator. Fixes #23124. Signed-off-by: Zixuan James Li <p359101898@gmail.com>
2022-10-05 20:55:31 +02:00
url_template = r"https://trac.example.com/ticket/{id}"
links = {"url": "https://trac.example.com/ticket/123", "text": "#123"}
topic_name = "test #123"
linkifier = RealmFilter(
linkifier: Support URL templates for linkifiers. This swaps out url_format_string from all of our APIs and replaces it with url_template. Note that the documentation changes in the following commits will be squashed with this commit. We change the "url_format" key to "url_template" for the realm_linkifiers events in event_schema, along with updating LinkifierDict. "url_template" is the name chosen to normalize mixed usages of "url_format_string" and "url_format" throughout the backend. The markdown processor is updated to stop handling the format string interpolation and delegate the task template expansion to the uri_template library instead. This change affects many test cases. We mostly just replace "%(name)s" with "{name}", "url_format_string" with "url_template" to make sure that they still pass. There are some test cases dedicated for testing "%" escaping, which aren't relevant anymore and are subject to removal. But for now we keep most of them as-is, and make sure that "%" is always escaped since we do not use it for variable substitution any more. Since url_format_string is not populated anymore, a migration is created to remove this field entirely, and make url_template non-nullable since we will always populate it. Note that it is possible to have url_template being null after migration 0422 and before 0424, but in practice, url_template will not be None after backfilling and the backend now is always setting url_template. With the removal of url_format_string, RealmFilter model will now be cleaned with URL template checks, and the old checks for escapes are removed. We also modified RealmFilter.clean to skip the validation when the url_template is invalid. This avoids raising mulitple ValidationError's when calling full_clean on a linkifier. But we might eventually want to have a more centric approach to data validation instead of having the same validation in both the clean method and the validator. Fixes #23124. Signed-off-by: Zixuan James Li <p359101898@gmail.com>
2022-10-05 20:55:31 +02:00
realm=zulip_realm, pattern=r"#(?P<id>[0-9]{2,8})", url_template=url_template
)
self.assertEqual(
repr(linkifier),
linkifier: Support URL templates for linkifiers. This swaps out url_format_string from all of our APIs and replaces it with url_template. Note that the documentation changes in the following commits will be squashed with this commit. We change the "url_format" key to "url_template" for the realm_linkifiers events in event_schema, along with updating LinkifierDict. "url_template" is the name chosen to normalize mixed usages of "url_format_string" and "url_format" throughout the backend. The markdown processor is updated to stop handling the format string interpolation and delegate the task template expansion to the uri_template library instead. This change affects many test cases. We mostly just replace "%(name)s" with "{name}", "url_format_string" with "url_template" to make sure that they still pass. There are some test cases dedicated for testing "%" escaping, which aren't relevant anymore and are subject to removal. But for now we keep most of them as-is, and make sure that "%" is always escaped since we do not use it for variable substitution any more. Since url_format_string is not populated anymore, a migration is created to remove this field entirely, and make url_template non-nullable since we will always populate it. Note that it is possible to have url_template being null after migration 0422 and before 0424, but in practice, url_template will not be None after backfilling and the backend now is always setting url_template. With the removal of url_format_string, RealmFilter model will now be cleaned with URL template checks, and the old checks for escapes are removed. We also modified RealmFilter.clean to skip the validation when the url_template is invalid. This avoids raising mulitple ValidationError's when calling full_clean on a linkifier. But we might eventually want to have a more centric approach to data validation instead of having the same validation in both the clean method and the validator. Fixes #23124. Signed-off-by: Zixuan James Li <p359101898@gmail.com>
2022-10-05 20:55:31 +02:00
"<RealmFilter: zulip: #(?P<id>[0-9]{2,8}) https://trac.example.com/ticket/{id}>",
)
def get_message(sender: UserProfile, realm: Realm) -> Message:
stream_name = "Denmark"
if not Stream.objects.filter(realm=realm, name=stream_name).exists():
self.make_stream(stream_name, realm)
tests: Ensure stream senders get a UserMessage row. We now complain if a test author sends a stream message that does not result in the sender getting a UserMessage row for the message. This is basically 100% equivalent to complaining that the author failed to subscribe the sender to the stream as part of the test setup, as far as I can tell, so the AssertionError instructs the author to subscribe the sender to the stream. We exempt bots from this check, although it is plausible we should only exempt the system bots like the notification bot. I considered auto-subscribing the sender to the stream, but that can be a little more expensive than the current check, and we generally want test setup to be explicit. If there is some legitimate way than a subscribed human sender can't get a UserMessage, then we probably want an explicit test for that, or we may want to change the backend to just write a UserMessage row in that hypothetical situation. For most tests, including almost all the ones fixed here, the author just wants their test setup to realistically reflect normal operation, and often devs may not realize that Cordelia is not subscribed to Denmark or not realize that Hamlet is not subscribed to Scotland. Some of us don't remember our Shakespeare from high school, and our stream subscriptions don't even necessarily reflect which countries the Bard placed his characters in. There may also be some legitimate use case where an author wants to simulate sending a message to an unsubscribed stream, but for those edge cases, they can always set allow_unsubscribed_sender to True.
2021-12-10 13:55:48 +01:00
self.subscribe(sender, stream_name)
msg_id = self.send_stream_message(sender, "Denmark", "hello world", topic_name, realm)
return Message.objects.get(id=msg_id)
def assert_topic_links(links: List[Dict[str, str]], msg: Message) -> None:
dct = MessageDict.messages_to_encoded_cache_helper([msg])[0]
self.assertEqual(dct[TOPIC_LINKS], links)
# Send messages before and after saving the realm filter from each user.
assert_topic_links([], get_message(self.example_user("othello"), zulip_realm))
assert_topic_links([], get_message(self.lear_user("cordelia"), lear_realm))
assert_topic_links([], get_message(self.notification_bot(zulip_realm), zulip_realm))
linkifier.save()
assert_topic_links([links], get_message(self.example_user("othello"), zulip_realm))
assert_topic_links([], get_message(self.lear_user("cordelia"), lear_realm))
assert_topic_links([links], get_message(self.notification_bot(zulip_realm), zulip_realm))
def test_reaction(self) -> None:
sender = self.example_user("othello")
receiver = self.example_user("hamlet")
recipient = Recipient.objects.get(type_id=receiver.id, type=Recipient.PERSONAL)
sending_client = make_client(name="test suite")
message = Message(
sender=sender,
recipient=recipient,
realm=receiver.realm,
content="hello **world**",
date_sent=timezone_now(),
sending_client=sending_client,
last_edit_time=timezone_now(),
edit_history="[]",
)
message.set_topic_name("whatever")
message.save()
reaction = Reaction.objects.create(
message=message, user_profile=sender, emoji_name="simple_smile"
)
msg_dict = MessageDict.ids_to_dict([message.id])[0]
self.assertEqual(msg_dict["reactions"][0]["emoji_name"], reaction.emoji_name)
self.assertEqual(msg_dict["reactions"][0]["user_id"], sender.id)
self.assertEqual(msg_dict["reactions"][0]["user"]["id"], sender.id)
self.assertEqual(msg_dict["reactions"][0]["user"]["email"], sender.email)
self.assertEqual(msg_dict["reactions"][0]["user"]["full_name"], sender.full_name)
def test_missing_anchor(self) -> None:
self.login("hamlet")
result = self.client_get(
"/json/messages",
{"use_first_unread_anchor": "false", "num_before": "1", "num_after": "1"},
)
self.assert_json_error(result, "Missing 'anchor' argument.")
def test_invalid_anchor(self) -> None:
self.login("hamlet")
result = self.client_get(
"/json/messages",
{
"use_first_unread_anchor": "false",
"num_before": "1",
"num_after": "1",
"anchor": "chocolate",
},
)
self.assert_json_error(result, "Invalid anchor")
class MessageHydrationTest(ZulipTestCase):
def test_hydrate_stream_recipient_info(self) -> None:
realm = get_realm("zulip")
cordelia = self.example_user("cordelia")
stream_id = get_stream("Verona", realm).id
obj = dict(
recipient_type=Recipient.STREAM,
recipient_type_id=stream_id,
sender_is_mirror_dummy=False,
sender_email=cordelia.email,
sender_full_name=cordelia.full_name,
sender_id=cordelia.id,
)
MessageDict.hydrate_recipient_info(obj, "Verona")
self.assertEqual(obj["display_recipient"], "Verona")
self.assertEqual(obj["type"], "stream")
def test_hydrate_pm_recipient_info(self) -> None:
cordelia = self.example_user("cordelia")
display_recipient: List[UserDisplayRecipient] = [
dict(
email="aaron@example.com",
full_name="Aaron Smith",
id=999,
is_mirror_dummy=False,
),
]
obj = dict(
recipient_type=Recipient.PERSONAL,
recipient_type_id=None,
sender_is_mirror_dummy=False,
sender_email=cordelia.email,
sender_full_name=cordelia.full_name,
sender_id=cordelia.id,
)
MessageDict.hydrate_recipient_info(obj, display_recipient)
self.assertEqual(
obj["display_recipient"],
[
dict(
email="aaron@example.com",
full_name="Aaron Smith",
id=999,
is_mirror_dummy=False,
),
dict(
email=cordelia.email,
full_name=cordelia.full_name,
id=cordelia.id,
is_mirror_dummy=False,
),
],
)
self.assertEqual(obj["type"], "private")
def test_messages_for_ids(self) -> None:
hamlet = self.example_user("hamlet")
cordelia = self.example_user("cordelia")
stream_name = "test stream"
self.subscribe(cordelia, stream_name)
old_message_id = self.send_stream_message(cordelia, stream_name, content="foo")
self.subscribe(hamlet, stream_name)
content = "hello @**King Hamlet**"
new_message_id = self.send_stream_message(cordelia, stream_name, content=content)
user_message_flags = {
old_message_id: ["read", "historical"],
new_message_id: ["mentioned"],
}
messages = messages_for_ids(
message_ids=[old_message_id, new_message_id],
user_message_flags=user_message_flags,
search_fields={},
apply_markdown=True,
client_gravatar=True,
allow_edit_history=False,
user_profile=cordelia,
realm=cordelia.realm,
)
self.assert_length(messages, 2)
for message in messages:
if message["id"] == old_message_id:
old_message = message
elif message["id"] == new_message_id:
new_message = message
self.assertEqual(old_message["content"], "<p>foo</p>")
self.assertEqual(old_message["flags"], ["read", "historical"])
self.assertIn('class="user-mention"', new_message["content"])
self.assertEqual(new_message["flags"], ["mentioned"])
def test_message_for_ids_for_restricted_user_access(self) -> None:
self.set_up_db_for_testing_user_access()
hamlet = self.example_user("hamlet")
self.send_stream_message(
hamlet,
"test_stream1",
topic_name="test",
content="test message again",
)
realm = get_realm("zulip")
stream = get_stream("test_stream1", realm)
assert stream.recipient_id is not None
message_ids = Message.objects.filter(
recipient_id=stream.recipient_id, realm=realm
).values_list("id", flat=True)
self.assert_length(message_ids, 2)
user_message_flags = {
message_ids[0]: ["read", "historical"],
message_ids[1]: ["read"],
}
messages = messages_for_ids(
message_ids=list(message_ids),
user_message_flags=user_message_flags,
search_fields={},
apply_markdown=True,
client_gravatar=True,
allow_edit_history=False,
user_profile=self.example_user("polonius"),
realm=realm,
)
(inaccessible_sender_msg,) = (msg for msg in messages if msg["sender_id"] != hamlet.id)
self.assertEqual(inaccessible_sender_msg["sender_id"], self.example_user("othello").id)
self.assertEqual(inaccessible_sender_msg["sender_full_name"], "Unknown user")
self.assertTrue(
inaccessible_sender_msg["avatar_url"].endswith("images/unknown-user-avatar.png")
)
def test_display_recipient_up_to_date(self) -> None:
"""
This is a test for a bug where due to caching of message_dicts,
after updating a user's information, fetching those cached messages
via messages_for_ids would return message_dicts with display_recipient
still having the old information. The returned message_dicts should have
up-to-date display_recipients and we check for that here.
"""
hamlet = self.example_user("hamlet")
cordelia = self.example_user("cordelia")
message_id = self.send_personal_message(hamlet, cordelia, "test")
cordelia_recipient = cordelia.recipient
# Cause the display_recipient to get cached:
assert cordelia_recipient is not None
get_display_recipient(cordelia_recipient)
# Change cordelia's email:
cordelia_new_email = "new-cordelia@zulip.com"
cordelia.email = cordelia_new_email
cordelia.save()
# Local display_recipient cache needs to be flushed.
# flush_per_request_caches() is called after every request,
# so it makes sense to run it here.
flush_per_request_caches()
messages = messages_for_ids(
message_ids=[message_id],
user_message_flags={message_id: ["read"]},
search_fields={},
apply_markdown=True,
client_gravatar=True,
allow_edit_history=False,
user_profile=cordelia,
realm=cordelia.realm,
)
message = messages[0]
# Find which display_recipient in the list is cordelia:
for display_recipient in message["display_recipient"]:
if display_recipient["id"] == cordelia.id:
cordelia_display_recipient = display_recipient
# Make sure the email is up-to-date.
self.assertEqual(cordelia_display_recipient["email"], cordelia_new_email)
class TestMessageForIdsDisplayRecipientFetching(ZulipTestCase):
def _verify_display_recipient(
self,
display_recipient: DisplayRecipientT,
expected_recipient_objects: List[UserProfile],
) -> None:
for user_profile in expected_recipient_objects:
recipient_dict: UserDisplayRecipient = {
"email": user_profile.email,
"full_name": user_profile.full_name,
"id": user_profile.id,
"is_mirror_dummy": user_profile.is_mirror_dummy,
}
self.assertTrue(recipient_dict in display_recipient)
def test_display_recipient_personal(self) -> None:
hamlet = self.example_user("hamlet")
cordelia = self.example_user("cordelia")
othello = self.example_user("othello")
message_ids = [
self.send_personal_message(hamlet, cordelia, "test"),
self.send_personal_message(cordelia, othello, "test"),
]
messages = messages_for_ids(
message_ids=message_ids,
user_message_flags={message_id: ["read"] for message_id in message_ids},
search_fields={},
apply_markdown=True,
client_gravatar=True,
allow_edit_history=False,
user_profile=cordelia,
realm=cordelia.realm,
)
self._verify_display_recipient(messages[0]["display_recipient"], [hamlet, cordelia])
self._verify_display_recipient(messages[1]["display_recipient"], [cordelia, othello])
def test_display_recipient_stream(self) -> None:
cordelia = self.example_user("cordelia")
tests: Ensure stream senders get a UserMessage row. We now complain if a test author sends a stream message that does not result in the sender getting a UserMessage row for the message. This is basically 100% equivalent to complaining that the author failed to subscribe the sender to the stream as part of the test setup, as far as I can tell, so the AssertionError instructs the author to subscribe the sender to the stream. We exempt bots from this check, although it is plausible we should only exempt the system bots like the notification bot. I considered auto-subscribing the sender to the stream, but that can be a little more expensive than the current check, and we generally want test setup to be explicit. If there is some legitimate way than a subscribed human sender can't get a UserMessage, then we probably want an explicit test for that, or we may want to change the backend to just write a UserMessage row in that hypothetical situation. For most tests, including almost all the ones fixed here, the author just wants their test setup to realistically reflect normal operation, and often devs may not realize that Cordelia is not subscribed to Denmark or not realize that Hamlet is not subscribed to Scotland. Some of us don't remember our Shakespeare from high school, and our stream subscriptions don't even necessarily reflect which countries the Bard placed his characters in. There may also be some legitimate use case where an author wants to simulate sending a message to an unsubscribed stream, but for those edge cases, they can always set allow_unsubscribed_sender to True.
2021-12-10 13:55:48 +01:00
self.subscribe(cordelia, "Denmark")
message_ids = [
self.send_stream_message(cordelia, "Verona", content="test"),
self.send_stream_message(cordelia, "Denmark", content="test"),
]
messages = messages_for_ids(
message_ids=message_ids,
user_message_flags={message_id: ["read"] for message_id in message_ids},
search_fields={},
apply_markdown=True,
client_gravatar=True,
allow_edit_history=False,
user_profile=cordelia,
realm=cordelia.realm,
)
self.assertEqual(messages[0]["display_recipient"], "Verona")
self.assertEqual(messages[1]["display_recipient"], "Denmark")
def test_display_recipient_direct_message_group(self) -> None:
hamlet = self.example_user("hamlet")
cordelia = self.example_user("cordelia")
othello = self.example_user("othello")
iago = self.example_user("iago")
message_ids = [
self.send_group_direct_message(hamlet, [cordelia, othello], "test"),
self.send_group_direct_message(cordelia, [hamlet, othello, iago], "test"),
]
messages = messages_for_ids(
message_ids=message_ids,
user_message_flags={message_id: ["read"] for message_id in message_ids},
search_fields={},
apply_markdown=True,
client_gravatar=True,
allow_edit_history=False,
user_profile=cordelia,
realm=cordelia.realm,
)
self._verify_display_recipient(
messages[0]["display_recipient"], [hamlet, cordelia, othello]
)
self._verify_display_recipient(
messages[1]["display_recipient"], [hamlet, cordelia, othello, iago]
)
def test_display_recipient_various_types(self) -> None:
hamlet = self.example_user("hamlet")
cordelia = self.example_user("cordelia")
othello = self.example_user("othello")
iago = self.example_user("iago")
tests: Ensure stream senders get a UserMessage row. We now complain if a test author sends a stream message that does not result in the sender getting a UserMessage row for the message. This is basically 100% equivalent to complaining that the author failed to subscribe the sender to the stream as part of the test setup, as far as I can tell, so the AssertionError instructs the author to subscribe the sender to the stream. We exempt bots from this check, although it is plausible we should only exempt the system bots like the notification bot. I considered auto-subscribing the sender to the stream, but that can be a little more expensive than the current check, and we generally want test setup to be explicit. If there is some legitimate way than a subscribed human sender can't get a UserMessage, then we probably want an explicit test for that, or we may want to change the backend to just write a UserMessage row in that hypothetical situation. For most tests, including almost all the ones fixed here, the author just wants their test setup to realistically reflect normal operation, and often devs may not realize that Cordelia is not subscribed to Denmark or not realize that Hamlet is not subscribed to Scotland. Some of us don't remember our Shakespeare from high school, and our stream subscriptions don't even necessarily reflect which countries the Bard placed his characters in. There may also be some legitimate use case where an author wants to simulate sending a message to an unsubscribed stream, but for those edge cases, they can always set allow_unsubscribed_sender to True.
2021-12-10 13:55:48 +01:00
self.subscribe(cordelia, "Denmark")
self.subscribe(hamlet, "Scotland")
message_ids = [
self.send_group_direct_message(hamlet, [cordelia, othello], "test"),
self.send_stream_message(cordelia, "Verona", content="test"),
self.send_personal_message(hamlet, cordelia, "test"),
self.send_stream_message(cordelia, "Denmark", content="test"),
self.send_group_direct_message(cordelia, [hamlet, othello, iago], "test"),
self.send_personal_message(cordelia, othello, "test"),
]
messages = messages_for_ids(
message_ids=message_ids,
user_message_flags={message_id: ["read"] for message_id in message_ids},
search_fields={},
apply_markdown=True,
client_gravatar=True,
allow_edit_history=False,
user_profile=cordelia,
realm=cordelia.realm,
)
self._verify_display_recipient(
messages[0]["display_recipient"], [hamlet, cordelia, othello]
)
self.assertEqual(messages[1]["display_recipient"], "Verona")
self._verify_display_recipient(messages[2]["display_recipient"], [hamlet, cordelia])
self.assertEqual(messages[3]["display_recipient"], "Denmark")
self._verify_display_recipient(
messages[4]["display_recipient"], [hamlet, cordelia, othello, iago]
)
self._verify_display_recipient(messages[5]["display_recipient"], [cordelia, othello])
class SewMessageAndReactionTest(ZulipTestCase):
def test_sew_messages_and_reaction(self) -> None:
sender = self.example_user("othello")
receiver = self.example_user("hamlet")
realm = get_realm("zulip")
pm_recipient = Recipient.objects.get(type_id=receiver.id, type=Recipient.PERSONAL)
stream_name = "Çiğdem"
stream = self.make_stream(stream_name)
stream_recipient = Recipient.objects.get(type_id=stream.id, type=Recipient.STREAM)
sending_client = make_client(name="test suite")
needed_ids = []
for i in range(5):
for recipient in [pm_recipient, stream_recipient]:
message = Message(
sender=sender,
recipient=recipient,
realm=realm,
content=f"whatever {i}",
date_sent=timezone_now(),
sending_client=sending_client,
last_edit_time=timezone_now(),
edit_history="[]",
)
message.set_topic_name("whatever")
message.save()
needed_ids.append(message.id)
reaction = Reaction(user_profile=sender, message=message, emoji_name="simple_smile")
reaction.save()
messages = Message.objects.filter(id__in=needed_ids).values(*["id", "content"])
reactions = Reaction.get_raw_db_rows(needed_ids)
tied_data = sew_messages_and_reactions(messages, reactions)
for data in tied_data:
self.assert_length(data["reactions"], 1)
self.assertEqual(data["reactions"][0]["emoji_name"], "simple_smile")
self.assertTrue(data["id"])
self.assertTrue(data["content"])