2020-07-07 01:00:03 +02:00
|
|
|
from typing import Any, Dict, List, Union
|
2020-07-06 01:55:16 +02:00
|
|
|
from unittest import mock
|
|
|
|
|
|
|
|
from django.utils.timezone import now as timezone_now
|
|
|
|
|
|
|
|
from zerver.lib.cache import cache_delete, to_dict_cache_key_id
|
|
|
|
from zerver.lib.markdown import version as markdown_version
|
2020-07-08 00:06:01 +02:00
|
|
|
from zerver.lib.message import MessageDict, messages_for_ids, sew_messages_and_reactions
|
2020-07-06 01:55:16 +02:00
|
|
|
from zerver.lib.test_classes import ZulipTestCase
|
2022-10-15 22:47:40 +02:00
|
|
|
from zerver.lib.test_helpers import make_client
|
2020-07-06 01:55:16 +02:00
|
|
|
from zerver.lib.topic import TOPIC_LINKS
|
2020-07-07 01:00:03 +02:00
|
|
|
from zerver.lib.types import DisplayRecipientT, UserDisplayRecipient
|
2020-07-06 01:55:16 +02:00
|
|
|
from zerver.models import (
|
|
|
|
Message,
|
|
|
|
Reaction,
|
2021-03-25 15:42:51 +01:00
|
|
|
Realm,
|
2020-07-06 01:55:16 +02:00
|
|
|
RealmFilter,
|
|
|
|
Recipient,
|
2020-07-07 01:00:03 +02:00
|
|
|
Stream,
|
2020-07-06 01:55:16 +02:00
|
|
|
UserProfile,
|
|
|
|
flush_per_request_caches,
|
2020-07-07 00:30:53 +02:00
|
|
|
get_display_recipient,
|
2020-07-06 01:55:16 +02:00
|
|
|
get_realm,
|
2020-07-07 00:30:53 +02:00
|
|
|
get_stream,
|
2020-07-06 01:55:16 +02:00
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
|
class MessageDictTest(ZulipTestCase):
|
|
|
|
def test_both_codepaths(self) -> None:
|
2021-02-12 08:19:30 +01:00
|
|
|
"""
|
2020-07-06 01:55:16 +02:00
|
|
|
We have two different codepaths that
|
|
|
|
extract a particular shape of dictionary
|
|
|
|
for messages to send to clients:
|
|
|
|
|
|
|
|
events:
|
|
|
|
|
|
|
|
These are the events we send to MANY
|
|
|
|
clients when a message is originally
|
|
|
|
sent.
|
|
|
|
|
|
|
|
fetch:
|
|
|
|
|
|
|
|
These are the messages we send to ONE
|
|
|
|
client when they fetch messages via
|
|
|
|
some narrow/search in the UI.
|
|
|
|
|
|
|
|
Different clients have different needs
|
|
|
|
when it comes to things like generating avatar
|
|
|
|
hashes or including both rendered and unrendered
|
2020-08-11 01:47:49 +02:00
|
|
|
Markdown, so that explains the different shapes.
|
2020-07-06 01:55:16 +02:00
|
|
|
|
|
|
|
And then the two codepaths have different
|
|
|
|
performance needs. In the events codepath, we
|
|
|
|
have the Django view generate a single "wide"
|
|
|
|
dictionary that gets put on the event queue,
|
|
|
|
and then we send events to multiple clients,
|
|
|
|
finalizing the payload for each of them depending
|
|
|
|
on the "shape" they want. (We also avoid
|
|
|
|
doing extra work for any two clients who want
|
|
|
|
the same shape dictionary, but that's out of the
|
|
|
|
scope of this particular test).
|
|
|
|
|
|
|
|
In the fetch scenario, the single client only needs
|
|
|
|
a dictionary of one shape, but we need to re-hydrate
|
|
|
|
the sender information, since the sender details
|
|
|
|
may have changed since the message was originally
|
|
|
|
sent.
|
|
|
|
|
|
|
|
This test simply verifies that the two codepaths
|
|
|
|
ultimately provide the same result.
|
2021-02-12 08:19:30 +01:00
|
|
|
"""
|
2020-07-06 01:55:16 +02:00
|
|
|
|
|
|
|
def reload_message(msg_id: int) -> Message:
|
|
|
|
# Get a clean copy of the message, and
|
|
|
|
# clear the cache.
|
|
|
|
cache_delete(to_dict_cache_key_id(msg_id))
|
|
|
|
msg = Message.objects.get(id=msg_id)
|
|
|
|
return msg
|
|
|
|
|
|
|
|
def get_send_message_payload(
|
2021-02-12 08:19:30 +01:00
|
|
|
msg_id: int, apply_markdown: bool, client_gravatar: bool
|
|
|
|
) -> Dict[str, Any]:
|
2020-07-06 01:55:16 +02:00
|
|
|
msg = reload_message(msg_id)
|
|
|
|
wide_dict = MessageDict.wide_dict(msg)
|
|
|
|
|
|
|
|
narrow_dict = MessageDict.finalize_payload(
|
|
|
|
wide_dict,
|
|
|
|
apply_markdown=apply_markdown,
|
|
|
|
client_gravatar=client_gravatar,
|
|
|
|
)
|
|
|
|
return narrow_dict
|
|
|
|
|
|
|
|
def get_fetch_payload(
|
2021-02-12 08:19:30 +01:00
|
|
|
msg_id: int, apply_markdown: bool, client_gravatar: bool
|
|
|
|
) -> Dict[str, Any]:
|
2020-07-06 01:55:16 +02:00
|
|
|
msg = reload_message(msg_id)
|
|
|
|
unhydrated_dict = MessageDict.to_dict_uncached_helper([msg])[0]
|
|
|
|
# The next step mutates the dict in place
|
|
|
|
# for performance reasons.
|
|
|
|
MessageDict.post_process_dicts(
|
|
|
|
[unhydrated_dict],
|
|
|
|
apply_markdown=apply_markdown,
|
|
|
|
client_gravatar=client_gravatar,
|
|
|
|
)
|
|
|
|
final_dict = unhydrated_dict
|
|
|
|
return final_dict
|
|
|
|
|
|
|
|
def test_message_id() -> int:
|
2021-02-12 08:20:45 +01:00
|
|
|
hamlet = self.example_user("hamlet")
|
2020-07-06 01:55:16 +02:00
|
|
|
self.login_user(hamlet)
|
|
|
|
msg_id = self.send_stream_message(
|
|
|
|
hamlet,
|
tests: Ensure stream senders get a UserMessage row.
We now complain if a test author sends a stream message
that does not result in the sender getting a
UserMessage row for the message.
This is basically 100% equivalent to complaining that
the author failed to subscribe the sender to the stream
as part of the test setup, as far as I can tell, so the
AssertionError instructs the author to subscribe the
sender to the stream.
We exempt bots from this check, although it is
plausible we should only exempt the system bots like
the notification bot.
I considered auto-subscribing the sender to the stream,
but that can be a little more expensive than the
current check, and we generally want test setup to be
explicit.
If there is some legitimate way than a subscribed human
sender can't get a UserMessage, then we probably want
an explicit test for that, or we may want to change the
backend to just write a UserMessage row in that
hypothetical situation.
For most tests, including almost all the ones fixed
here, the author just wants their test setup to
realistically reflect normal operation, and often devs
may not realize that Cordelia is not subscribed to
Denmark or not realize that Hamlet is not subscribed to
Scotland.
Some of us don't remember our Shakespeare from high
school, and our stream subscriptions don't even
necessarily reflect which countries the Bard placed his
characters in.
There may also be some legitimate use case where an
author wants to simulate sending a message to an
unsubscribed stream, but for those edge cases, they can
always set allow_unsubscribed_sender to True.
2021-12-10 13:55:48 +01:00
|
|
|
"Denmark",
|
2020-07-06 01:55:16 +02:00
|
|
|
topic_name="editing",
|
|
|
|
content="before edit",
|
|
|
|
)
|
|
|
|
return msg_id
|
|
|
|
|
|
|
|
flag_setups = [
|
|
|
|
[False, False],
|
|
|
|
[False, True],
|
|
|
|
[True, False],
|
|
|
|
[True, True],
|
|
|
|
]
|
|
|
|
|
|
|
|
msg_id = test_message_id()
|
|
|
|
|
2023-02-02 04:35:24 +01:00
|
|
|
for apply_markdown, client_gravatar in flag_setups:
|
2020-07-06 01:55:16 +02:00
|
|
|
send_message_payload = get_send_message_payload(
|
|
|
|
msg_id,
|
|
|
|
apply_markdown=apply_markdown,
|
|
|
|
client_gravatar=client_gravatar,
|
|
|
|
)
|
|
|
|
|
|
|
|
fetch_payload = get_fetch_payload(
|
|
|
|
msg_id,
|
|
|
|
apply_markdown=apply_markdown,
|
|
|
|
client_gravatar=client_gravatar,
|
|
|
|
)
|
|
|
|
|
|
|
|
self.assertEqual(send_message_payload, fetch_payload)
|
|
|
|
|
|
|
|
def test_bulk_message_fetching(self) -> None:
|
2021-02-12 08:20:45 +01:00
|
|
|
sender = self.example_user("othello")
|
|
|
|
receiver = self.example_user("hamlet")
|
2022-09-27 21:42:31 +02:00
|
|
|
realm = get_realm("zulip")
|
2020-07-06 01:55:16 +02:00
|
|
|
pm_recipient = Recipient.objects.get(type_id=receiver.id, type=Recipient.PERSONAL)
|
2021-02-12 08:20:45 +01:00
|
|
|
stream_name = "Çiğdem"
|
2020-07-06 01:55:16 +02:00
|
|
|
stream = self.make_stream(stream_name)
|
|
|
|
stream_recipient = Recipient.objects.get(type_id=stream.id, type=Recipient.STREAM)
|
|
|
|
sending_client = make_client(name="test suite")
|
|
|
|
|
|
|
|
ids = []
|
|
|
|
for i in range(300):
|
|
|
|
for recipient in [pm_recipient, stream_recipient]:
|
|
|
|
message = Message(
|
|
|
|
sender=sender,
|
|
|
|
recipient=recipient,
|
2022-09-27 21:42:31 +02:00
|
|
|
realm=realm,
|
2021-02-12 08:20:45 +01:00
|
|
|
content=f"whatever {i}",
|
|
|
|
rendered_content="DOES NOT MATTER",
|
2020-07-06 01:55:16 +02:00
|
|
|
rendered_content_version=markdown_version,
|
|
|
|
date_sent=timezone_now(),
|
|
|
|
sending_client=sending_client,
|
|
|
|
last_edit_time=timezone_now(),
|
2021-02-12 08:20:45 +01:00
|
|
|
edit_history="[]",
|
2020-07-06 01:55:16 +02:00
|
|
|
)
|
2021-02-12 08:20:45 +01:00
|
|
|
message.set_topic_name("whatever")
|
2020-07-06 01:55:16 +02:00
|
|
|
message.save()
|
|
|
|
ids.append(message.id)
|
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
Reaction.objects.create(
|
2021-02-12 08:20:45 +01:00
|
|
|
user_profile=sender, message=message, emoji_name="simple_smile"
|
2021-02-12 08:19:30 +01:00
|
|
|
)
|
2020-07-06 01:55:16 +02:00
|
|
|
|
|
|
|
num_ids = len(ids)
|
|
|
|
self.assertTrue(num_ids >= 600)
|
|
|
|
|
|
|
|
flush_per_request_caches()
|
2022-10-15 22:47:40 +02:00
|
|
|
with self.assert_database_query_count(7):
|
2020-07-06 01:55:16 +02:00
|
|
|
rows = list(MessageDict.get_raw_db_rows(ids))
|
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
objs = [MessageDict.build_dict_from_raw_db_row(row) for row in rows]
|
2020-07-06 01:55:16 +02:00
|
|
|
MessageDict.post_process_dicts(objs, apply_markdown=False, client_gravatar=False)
|
|
|
|
|
2021-05-17 05:41:32 +02:00
|
|
|
self.assert_length(rows, num_ids)
|
2020-07-06 01:55:16 +02:00
|
|
|
|
|
|
|
def test_applying_markdown(self) -> None:
|
2021-02-12 08:20:45 +01:00
|
|
|
sender = self.example_user("othello")
|
|
|
|
receiver = self.example_user("hamlet")
|
2020-07-06 01:55:16 +02:00
|
|
|
recipient = Recipient.objects.get(type_id=receiver.id, type=Recipient.PERSONAL)
|
|
|
|
sending_client = make_client(name="test suite")
|
|
|
|
message = Message(
|
|
|
|
sender=sender,
|
|
|
|
recipient=recipient,
|
2022-09-27 21:42:31 +02:00
|
|
|
realm=receiver.realm,
|
2021-02-12 08:20:45 +01:00
|
|
|
content="hello **world**",
|
2020-07-06 01:55:16 +02:00
|
|
|
date_sent=timezone_now(),
|
|
|
|
sending_client=sending_client,
|
|
|
|
last_edit_time=timezone_now(),
|
2021-02-12 08:20:45 +01:00
|
|
|
edit_history="[]",
|
2020-07-06 01:55:16 +02:00
|
|
|
)
|
2021-02-12 08:20:45 +01:00
|
|
|
message.set_topic_name("whatever")
|
2020-07-06 01:55:16 +02:00
|
|
|
message.save()
|
|
|
|
|
|
|
|
# An important part of this test is to get the message through this exact code path,
|
|
|
|
# because there is an ugly hack we need to cover. So don't just say "row = message".
|
|
|
|
row = MessageDict.get_raw_db_rows([message.id])[0]
|
|
|
|
dct = MessageDict.build_dict_from_raw_db_row(row)
|
2021-02-12 08:20:45 +01:00
|
|
|
expected_content = "<p>hello <strong>world</strong></p>"
|
|
|
|
self.assertEqual(dct["rendered_content"], expected_content)
|
2020-07-06 01:55:16 +02:00
|
|
|
message = Message.objects.get(id=message.id)
|
|
|
|
self.assertEqual(message.rendered_content, expected_content)
|
|
|
|
self.assertEqual(message.rendered_content_version, markdown_version)
|
|
|
|
|
|
|
|
@mock.patch("zerver.lib.message.markdown_convert")
|
|
|
|
def test_applying_markdown_invalid_format(self, convert_mock: Any) -> None:
|
|
|
|
# pretend the converter returned an invalid message without raising an exception
|
|
|
|
convert_mock.return_value = None
|
2021-02-12 08:20:45 +01:00
|
|
|
sender = self.example_user("othello")
|
|
|
|
receiver = self.example_user("hamlet")
|
2020-07-06 01:55:16 +02:00
|
|
|
recipient = Recipient.objects.get(type_id=receiver.id, type=Recipient.PERSONAL)
|
|
|
|
sending_client = make_client(name="test suite")
|
|
|
|
message = Message(
|
|
|
|
sender=sender,
|
|
|
|
recipient=recipient,
|
2022-09-27 21:42:31 +02:00
|
|
|
realm=receiver.realm,
|
2021-02-12 08:20:45 +01:00
|
|
|
content="hello **world**",
|
2020-07-06 01:55:16 +02:00
|
|
|
date_sent=timezone_now(),
|
|
|
|
sending_client=sending_client,
|
|
|
|
last_edit_time=timezone_now(),
|
2021-02-12 08:20:45 +01:00
|
|
|
edit_history="[]",
|
2020-07-06 01:55:16 +02:00
|
|
|
)
|
2021-02-12 08:20:45 +01:00
|
|
|
message.set_topic_name("whatever")
|
2020-07-06 01:55:16 +02:00
|
|
|
message.save()
|
|
|
|
|
|
|
|
# An important part of this test is to get the message through this exact code path,
|
|
|
|
# because there is an ugly hack we need to cover. So don't just say "row = message".
|
|
|
|
row = MessageDict.get_raw_db_rows([message.id])[0]
|
|
|
|
dct = MessageDict.build_dict_from_raw_db_row(row)
|
2021-02-12 08:19:30 +01:00
|
|
|
error_content = (
|
2021-02-12 08:20:45 +01:00
|
|
|
"<p>[Zulip note: Sorry, we could not understand the formatting of your message]</p>"
|
2021-02-12 08:19:30 +01:00
|
|
|
)
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertEqual(dct["rendered_content"], error_content)
|
2020-07-06 01:55:16 +02:00
|
|
|
|
|
|
|
def test_topic_links_use_stream_realm(self) -> None:
|
|
|
|
# Set up a realm filter on 'zulip' and assert that messages
|
2021-03-25 15:42:51 +01:00
|
|
|
# sent to a stream on 'zulip' have the topic linkified,
|
|
|
|
# and not linkified when sent to a stream in 'lear'.
|
2021-02-12 08:20:45 +01:00
|
|
|
zulip_realm = get_realm("zulip")
|
2021-03-25 15:42:51 +01:00
|
|
|
lear_realm = get_realm("lear")
|
linkifier: Support URL templates for linkifiers.
This swaps out url_format_string from all of our APIs and replaces it
with url_template. Note that the documentation changes in the following
commits will be squashed with this commit.
We change the "url_format" key to "url_template" for the
realm_linkifiers events in event_schema, along with updating
LinkifierDict. "url_template" is the name chosen to normalize
mixed usages of "url_format_string" and "url_format" throughout
the backend.
The markdown processor is updated to stop handling the format string
interpolation and delegate the task template expansion to the uri_template
library instead.
This change affects many test cases. We mostly just replace "%(name)s"
with "{name}", "url_format_string" with "url_template" to make sure that
they still pass. There are some test cases dedicated for testing "%"
escaping, which aren't relevant anymore and are subject to removal.
But for now we keep most of them as-is, and make sure that "%" is always
escaped since we do not use it for variable substitution any more.
Since url_format_string is not populated anymore, a migration is created
to remove this field entirely, and make url_template non-nullable since
we will always populate it. Note that it is possible to have
url_template being null after migration 0422 and before 0424, but
in practice, url_template will not be None after backfilling and the
backend now is always setting url_template.
With the removal of url_format_string, RealmFilter model will now be cleaned
with URL template checks, and the old checks for escapes are removed.
We also modified RealmFilter.clean to skip the validation when the
url_template is invalid. This avoids raising mulitple ValidationError's
when calling full_clean on a linkifier. But we might eventually want to
have a more centric approach to data validation instead of having
the same validation in both the clean method and the validator.
Fixes #23124.
Signed-off-by: Zixuan James Li <p359101898@gmail.com>
2022-10-05 20:55:31 +02:00
|
|
|
url_template = r"https://trac.example.com/ticket/{id}"
|
2021-01-26 07:32:29 +01:00
|
|
|
links = {"url": "https://trac.example.com/ticket/123", "text": "#123"}
|
2021-02-12 08:20:45 +01:00
|
|
|
topic_name = "test #123"
|
2020-07-06 01:55:16 +02:00
|
|
|
|
2021-03-30 12:08:03 +02:00
|
|
|
linkifier = RealmFilter(
|
linkifier: Support URL templates for linkifiers.
This swaps out url_format_string from all of our APIs and replaces it
with url_template. Note that the documentation changes in the following
commits will be squashed with this commit.
We change the "url_format" key to "url_template" for the
realm_linkifiers events in event_schema, along with updating
LinkifierDict. "url_template" is the name chosen to normalize
mixed usages of "url_format_string" and "url_format" throughout
the backend.
The markdown processor is updated to stop handling the format string
interpolation and delegate the task template expansion to the uri_template
library instead.
This change affects many test cases. We mostly just replace "%(name)s"
with "{name}", "url_format_string" with "url_template" to make sure that
they still pass. There are some test cases dedicated for testing "%"
escaping, which aren't relevant anymore and are subject to removal.
But for now we keep most of them as-is, and make sure that "%" is always
escaped since we do not use it for variable substitution any more.
Since url_format_string is not populated anymore, a migration is created
to remove this field entirely, and make url_template non-nullable since
we will always populate it. Note that it is possible to have
url_template being null after migration 0422 and before 0424, but
in practice, url_template will not be None after backfilling and the
backend now is always setting url_template.
With the removal of url_format_string, RealmFilter model will now be cleaned
with URL template checks, and the old checks for escapes are removed.
We also modified RealmFilter.clean to skip the validation when the
url_template is invalid. This avoids raising mulitple ValidationError's
when calling full_clean on a linkifier. But we might eventually want to
have a more centric approach to data validation instead of having
the same validation in both the clean method and the validator.
Fixes #23124.
Signed-off-by: Zixuan James Li <p359101898@gmail.com>
2022-10-05 20:55:31 +02:00
|
|
|
realm=zulip_realm, pattern=r"#(?P<id>[0-9]{2,8})", url_template=url_template
|
2021-02-12 08:19:30 +01:00
|
|
|
)
|
2020-07-06 01:55:16 +02:00
|
|
|
self.assertEqual(
|
2023-03-08 22:18:59 +01:00
|
|
|
repr(linkifier),
|
linkifier: Support URL templates for linkifiers.
This swaps out url_format_string from all of our APIs and replaces it
with url_template. Note that the documentation changes in the following
commits will be squashed with this commit.
We change the "url_format" key to "url_template" for the
realm_linkifiers events in event_schema, along with updating
LinkifierDict. "url_template" is the name chosen to normalize
mixed usages of "url_format_string" and "url_format" throughout
the backend.
The markdown processor is updated to stop handling the format string
interpolation and delegate the task template expansion to the uri_template
library instead.
This change affects many test cases. We mostly just replace "%(name)s"
with "{name}", "url_format_string" with "url_template" to make sure that
they still pass. There are some test cases dedicated for testing "%"
escaping, which aren't relevant anymore and are subject to removal.
But for now we keep most of them as-is, and make sure that "%" is always
escaped since we do not use it for variable substitution any more.
Since url_format_string is not populated anymore, a migration is created
to remove this field entirely, and make url_template non-nullable since
we will always populate it. Note that it is possible to have
url_template being null after migration 0422 and before 0424, but
in practice, url_template will not be None after backfilling and the
backend now is always setting url_template.
With the removal of url_format_string, RealmFilter model will now be cleaned
with URL template checks, and the old checks for escapes are removed.
We also modified RealmFilter.clean to skip the validation when the
url_template is invalid. This avoids raising mulitple ValidationError's
when calling full_clean on a linkifier. But we might eventually want to
have a more centric approach to data validation instead of having
the same validation in both the clean method and the validator.
Fixes #23124.
Signed-off-by: Zixuan James Li <p359101898@gmail.com>
2022-10-05 20:55:31 +02:00
|
|
|
"<RealmFilter: zulip: #(?P<id>[0-9]{2,8}) https://trac.example.com/ticket/{id}>",
|
2021-02-12 08:19:30 +01:00
|
|
|
)
|
2020-07-06 01:55:16 +02:00
|
|
|
|
2021-03-25 15:42:51 +01:00
|
|
|
def get_message(sender: UserProfile, realm: Realm) -> Message:
|
|
|
|
stream_name = "Denmark"
|
|
|
|
if not Stream.objects.filter(realm=realm, name=stream_name).exists():
|
|
|
|
self.make_stream(stream_name, realm)
|
tests: Ensure stream senders get a UserMessage row.
We now complain if a test author sends a stream message
that does not result in the sender getting a
UserMessage row for the message.
This is basically 100% equivalent to complaining that
the author failed to subscribe the sender to the stream
as part of the test setup, as far as I can tell, so the
AssertionError instructs the author to subscribe the
sender to the stream.
We exempt bots from this check, although it is
plausible we should only exempt the system bots like
the notification bot.
I considered auto-subscribing the sender to the stream,
but that can be a little more expensive than the
current check, and we generally want test setup to be
explicit.
If there is some legitimate way than a subscribed human
sender can't get a UserMessage, then we probably want
an explicit test for that, or we may want to change the
backend to just write a UserMessage row in that
hypothetical situation.
For most tests, including almost all the ones fixed
here, the author just wants their test setup to
realistically reflect normal operation, and often devs
may not realize that Cordelia is not subscribed to
Denmark or not realize that Hamlet is not subscribed to
Scotland.
Some of us don't remember our Shakespeare from high
school, and our stream subscriptions don't even
necessarily reflect which countries the Bard placed his
characters in.
There may also be some legitimate use case where an
author wants to simulate sending a message to an
unsubscribed stream, but for those edge cases, they can
always set allow_unsubscribed_sender to True.
2021-12-10 13:55:48 +01:00
|
|
|
self.subscribe(sender, stream_name)
|
2021-03-25 15:42:51 +01:00
|
|
|
msg_id = self.send_stream_message(sender, "Denmark", "hello world", topic_name, realm)
|
2020-07-06 01:55:16 +02:00
|
|
|
return Message.objects.get(id=msg_id)
|
|
|
|
|
2021-01-26 07:32:29 +01:00
|
|
|
def assert_topic_links(links: List[Dict[str, str]], msg: Message) -> None:
|
2020-07-06 01:55:16 +02:00
|
|
|
dct = MessageDict.to_dict_uncached_helper([msg])[0]
|
|
|
|
self.assertEqual(dct[TOPIC_LINKS], links)
|
|
|
|
|
|
|
|
# Send messages before and after saving the realm filter from each user.
|
2021-03-25 15:42:51 +01:00
|
|
|
assert_topic_links([], get_message(self.example_user("othello"), zulip_realm))
|
|
|
|
assert_topic_links([], get_message(self.lear_user("cordelia"), lear_realm))
|
2021-03-08 11:39:48 +01:00
|
|
|
assert_topic_links([], get_message(self.notification_bot(zulip_realm), zulip_realm))
|
2021-03-30 12:08:03 +02:00
|
|
|
linkifier.save()
|
2021-03-25 15:42:51 +01:00
|
|
|
assert_topic_links([links], get_message(self.example_user("othello"), zulip_realm))
|
|
|
|
assert_topic_links([], get_message(self.lear_user("cordelia"), lear_realm))
|
2021-03-08 11:39:48 +01:00
|
|
|
assert_topic_links([links], get_message(self.notification_bot(zulip_realm), zulip_realm))
|
2020-07-06 01:55:16 +02:00
|
|
|
|
|
|
|
def test_reaction(self) -> None:
|
2021-02-12 08:20:45 +01:00
|
|
|
sender = self.example_user("othello")
|
|
|
|
receiver = self.example_user("hamlet")
|
2020-07-06 01:55:16 +02:00
|
|
|
recipient = Recipient.objects.get(type_id=receiver.id, type=Recipient.PERSONAL)
|
|
|
|
sending_client = make_client(name="test suite")
|
|
|
|
message = Message(
|
|
|
|
sender=sender,
|
|
|
|
recipient=recipient,
|
2022-09-27 21:42:31 +02:00
|
|
|
realm=receiver.realm,
|
2021-02-12 08:20:45 +01:00
|
|
|
content="hello **world**",
|
2020-07-06 01:55:16 +02:00
|
|
|
date_sent=timezone_now(),
|
|
|
|
sending_client=sending_client,
|
|
|
|
last_edit_time=timezone_now(),
|
2021-02-12 08:20:45 +01:00
|
|
|
edit_history="[]",
|
2020-07-06 01:55:16 +02:00
|
|
|
)
|
2021-02-12 08:20:45 +01:00
|
|
|
message.set_topic_name("whatever")
|
2020-07-06 01:55:16 +02:00
|
|
|
message.save()
|
|
|
|
|
|
|
|
reaction = Reaction.objects.create(
|
2021-02-12 08:20:45 +01:00
|
|
|
message=message, user_profile=sender, emoji_name="simple_smile"
|
2021-02-12 08:19:30 +01:00
|
|
|
)
|
2020-07-06 01:55:16 +02:00
|
|
|
row = MessageDict.get_raw_db_rows([message.id])[0]
|
|
|
|
msg_dict = MessageDict.build_dict_from_raw_db_row(row)
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertEqual(msg_dict["reactions"][0]["emoji_name"], reaction.emoji_name)
|
|
|
|
self.assertEqual(msg_dict["reactions"][0]["user_id"], sender.id)
|
|
|
|
self.assertEqual(msg_dict["reactions"][0]["user"]["id"], sender.id)
|
|
|
|
self.assertEqual(msg_dict["reactions"][0]["user"]["email"], sender.email)
|
|
|
|
self.assertEqual(msg_dict["reactions"][0]["user"]["full_name"], sender.full_name)
|
2020-07-06 01:55:16 +02:00
|
|
|
|
|
|
|
def test_missing_anchor(self) -> None:
|
2021-02-12 08:20:45 +01:00
|
|
|
self.login("hamlet")
|
2020-07-06 01:55:16 +02:00
|
|
|
result = self.client_get(
|
2021-02-12 08:19:30 +01:00
|
|
|
"/json/messages",
|
|
|
|
{"use_first_unread_anchor": "false", "num_before": "1", "num_after": "1"},
|
|
|
|
)
|
2020-07-06 01:55:16 +02:00
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
self.assert_json_error(result, "Missing 'anchor' argument.")
|
2020-07-06 01:55:16 +02:00
|
|
|
|
|
|
|
def test_invalid_anchor(self) -> None:
|
2021-02-12 08:20:45 +01:00
|
|
|
self.login("hamlet")
|
2020-07-06 01:55:16 +02:00
|
|
|
result = self.client_get(
|
2021-02-12 08:19:30 +01:00
|
|
|
"/json/messages",
|
|
|
|
{
|
|
|
|
"use_first_unread_anchor": "false",
|
|
|
|
"num_before": "1",
|
|
|
|
"num_after": "1",
|
|
|
|
"anchor": "chocolate",
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
self.assert_json_error(result, "Invalid anchor")
|
2020-07-06 01:55:16 +02:00
|
|
|
|
2020-07-07 00:30:53 +02:00
|
|
|
|
|
|
|
class MessageHydrationTest(ZulipTestCase):
|
|
|
|
def test_hydrate_stream_recipient_info(self) -> None:
|
2021-02-12 08:20:45 +01:00
|
|
|
realm = get_realm("zulip")
|
|
|
|
cordelia = self.example_user("cordelia")
|
2020-07-07 00:30:53 +02:00
|
|
|
|
2021-02-12 08:20:45 +01:00
|
|
|
stream_id = get_stream("Verona", realm).id
|
2020-07-07 00:30:53 +02:00
|
|
|
|
|
|
|
obj = dict(
|
|
|
|
recipient_type=Recipient.STREAM,
|
|
|
|
recipient_type_id=stream_id,
|
|
|
|
sender_is_mirror_dummy=False,
|
|
|
|
sender_email=cordelia.email,
|
|
|
|
sender_full_name=cordelia.full_name,
|
|
|
|
sender_id=cordelia.id,
|
|
|
|
)
|
|
|
|
|
2021-02-12 08:20:45 +01:00
|
|
|
MessageDict.hydrate_recipient_info(obj, "Verona")
|
2020-07-07 00:30:53 +02:00
|
|
|
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertEqual(obj["display_recipient"], "Verona")
|
|
|
|
self.assertEqual(obj["type"], "stream")
|
2020-07-07 00:30:53 +02:00
|
|
|
|
|
|
|
def test_hydrate_pm_recipient_info(self) -> None:
|
2021-02-12 08:20:45 +01:00
|
|
|
cordelia = self.example_user("cordelia")
|
2020-07-07 00:30:53 +02:00
|
|
|
display_recipient: List[UserDisplayRecipient] = [
|
|
|
|
dict(
|
2021-02-12 08:20:45 +01:00
|
|
|
email="aaron@example.com",
|
|
|
|
full_name="Aaron Smith",
|
2020-07-07 00:30:53 +02:00
|
|
|
id=999,
|
|
|
|
is_mirror_dummy=False,
|
|
|
|
),
|
|
|
|
]
|
|
|
|
|
|
|
|
obj = dict(
|
|
|
|
recipient_type=Recipient.PERSONAL,
|
|
|
|
recipient_type_id=None,
|
|
|
|
sender_is_mirror_dummy=False,
|
|
|
|
sender_email=cordelia.email,
|
|
|
|
sender_full_name=cordelia.full_name,
|
|
|
|
sender_id=cordelia.id,
|
|
|
|
)
|
|
|
|
|
|
|
|
MessageDict.hydrate_recipient_info(obj, display_recipient)
|
|
|
|
|
|
|
|
self.assertEqual(
|
2021-02-12 08:20:45 +01:00
|
|
|
obj["display_recipient"],
|
2020-07-07 00:30:53 +02:00
|
|
|
[
|
|
|
|
dict(
|
2021-02-12 08:20:45 +01:00
|
|
|
email="aaron@example.com",
|
|
|
|
full_name="Aaron Smith",
|
2020-07-07 00:30:53 +02:00
|
|
|
id=999,
|
|
|
|
is_mirror_dummy=False,
|
|
|
|
),
|
|
|
|
dict(
|
|
|
|
email=cordelia.email,
|
|
|
|
full_name=cordelia.full_name,
|
|
|
|
id=cordelia.id,
|
|
|
|
is_mirror_dummy=False,
|
|
|
|
),
|
|
|
|
],
|
|
|
|
)
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertEqual(obj["type"], "private")
|
2020-07-07 00:30:53 +02:00
|
|
|
|
|
|
|
def test_messages_for_ids(self) -> None:
|
2021-02-12 08:20:45 +01:00
|
|
|
hamlet = self.example_user("hamlet")
|
|
|
|
cordelia = self.example_user("cordelia")
|
2020-07-07 00:30:53 +02:00
|
|
|
|
2021-02-12 08:20:45 +01:00
|
|
|
stream_name = "test stream"
|
2020-07-07 00:30:53 +02:00
|
|
|
self.subscribe(cordelia, stream_name)
|
|
|
|
|
2021-02-12 08:20:45 +01:00
|
|
|
old_message_id = self.send_stream_message(cordelia, stream_name, content="foo")
|
2020-07-07 00:30:53 +02:00
|
|
|
|
|
|
|
self.subscribe(hamlet, stream_name)
|
|
|
|
|
2021-02-12 08:20:45 +01:00
|
|
|
content = "hello @**King Hamlet**"
|
2020-07-07 00:30:53 +02:00
|
|
|
new_message_id = self.send_stream_message(cordelia, stream_name, content=content)
|
|
|
|
|
|
|
|
user_message_flags = {
|
2021-02-12 08:20:45 +01:00
|
|
|
old_message_id: ["read", "historical"],
|
|
|
|
new_message_id: ["mentioned"],
|
2020-07-07 00:30:53 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
messages = messages_for_ids(
|
|
|
|
message_ids=[old_message_id, new_message_id],
|
|
|
|
user_message_flags=user_message_flags,
|
|
|
|
search_fields={},
|
|
|
|
apply_markdown=True,
|
|
|
|
client_gravatar=True,
|
|
|
|
allow_edit_history=False,
|
|
|
|
)
|
|
|
|
|
2021-05-17 05:41:32 +02:00
|
|
|
self.assert_length(messages, 2)
|
2020-07-07 00:30:53 +02:00
|
|
|
|
|
|
|
for message in messages:
|
2021-02-12 08:20:45 +01:00
|
|
|
if message["id"] == old_message_id:
|
2020-07-07 00:30:53 +02:00
|
|
|
old_message = message
|
2021-02-12 08:20:45 +01:00
|
|
|
elif message["id"] == new_message_id:
|
2020-07-07 00:30:53 +02:00
|
|
|
new_message = message
|
|
|
|
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertEqual(old_message["content"], "<p>foo</p>")
|
|
|
|
self.assertEqual(old_message["flags"], ["read", "historical"])
|
2020-07-07 00:30:53 +02:00
|
|
|
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertIn('class="user-mention"', new_message["content"])
|
|
|
|
self.assertEqual(new_message["flags"], ["mentioned"])
|
2020-07-07 00:30:53 +02:00
|
|
|
|
|
|
|
def test_display_recipient_up_to_date(self) -> None:
|
|
|
|
"""
|
|
|
|
This is a test for a bug where due to caching of message_dicts,
|
|
|
|
after updating a user's information, fetching those cached messages
|
|
|
|
via messages_for_ids would return message_dicts with display_recipient
|
|
|
|
still having the old information. The returned message_dicts should have
|
|
|
|
up-to-date display_recipients and we check for that here.
|
|
|
|
"""
|
|
|
|
|
2021-02-12 08:20:45 +01:00
|
|
|
hamlet = self.example_user("hamlet")
|
|
|
|
cordelia = self.example_user("cordelia")
|
|
|
|
message_id = self.send_personal_message(hamlet, cordelia, "test")
|
2020-07-07 00:30:53 +02:00
|
|
|
|
|
|
|
cordelia_recipient = cordelia.recipient
|
|
|
|
# Cause the display_recipient to get cached:
|
2022-05-31 01:27:38 +02:00
|
|
|
assert cordelia_recipient is not None
|
2020-07-07 00:30:53 +02:00
|
|
|
get_display_recipient(cordelia_recipient)
|
|
|
|
|
|
|
|
# Change cordelia's email:
|
2021-02-12 08:20:45 +01:00
|
|
|
cordelia_new_email = "new-cordelia@zulip.com"
|
2020-07-07 00:30:53 +02:00
|
|
|
cordelia.email = cordelia_new_email
|
|
|
|
cordelia.save()
|
|
|
|
|
|
|
|
# Local display_recipient cache needs to be flushed.
|
|
|
|
# flush_per_request_caches() is called after every request,
|
|
|
|
# so it makes sense to run it here.
|
|
|
|
flush_per_request_caches()
|
|
|
|
|
|
|
|
messages = messages_for_ids(
|
|
|
|
message_ids=[message_id],
|
2021-02-12 08:20:45 +01:00
|
|
|
user_message_flags={message_id: ["read"]},
|
2020-07-07 00:30:53 +02:00
|
|
|
search_fields={},
|
|
|
|
apply_markdown=True,
|
|
|
|
client_gravatar=True,
|
|
|
|
allow_edit_history=False,
|
|
|
|
)
|
|
|
|
message = messages[0]
|
|
|
|
|
|
|
|
# Find which display_recipient in the list is cordelia:
|
2021-02-12 08:20:45 +01:00
|
|
|
for display_recipient in message["display_recipient"]:
|
|
|
|
if display_recipient["id"] == cordelia.id:
|
2020-07-07 00:30:53 +02:00
|
|
|
cordelia_display_recipient = display_recipient
|
|
|
|
|
|
|
|
# Make sure the email is up-to-date.
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertEqual(cordelia_display_recipient["email"], cordelia_new_email)
|
2020-07-07 01:00:03 +02:00
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
|
2020-07-07 01:00:03 +02:00
|
|
|
class TestMessageForIdsDisplayRecipientFetching(ZulipTestCase):
|
2021-02-12 08:19:30 +01:00
|
|
|
def _verify_display_recipient(
|
|
|
|
self,
|
|
|
|
display_recipient: DisplayRecipientT,
|
|
|
|
expected_recipient_objects: Union[Stream, List[UserProfile]],
|
|
|
|
) -> None:
|
2020-07-07 01:00:03 +02:00
|
|
|
if isinstance(expected_recipient_objects, Stream):
|
|
|
|
self.assertEqual(display_recipient, expected_recipient_objects.name)
|
|
|
|
|
|
|
|
else:
|
|
|
|
for user_profile in expected_recipient_objects:
|
|
|
|
recipient_dict: UserDisplayRecipient = {
|
2021-02-12 08:20:45 +01:00
|
|
|
"email": user_profile.email,
|
|
|
|
"full_name": user_profile.full_name,
|
|
|
|
"id": user_profile.id,
|
|
|
|
"is_mirror_dummy": user_profile.is_mirror_dummy,
|
2020-07-07 01:00:03 +02:00
|
|
|
}
|
|
|
|
self.assertTrue(recipient_dict in display_recipient)
|
|
|
|
|
|
|
|
def test_display_recipient_personal(self) -> None:
|
2021-02-12 08:20:45 +01:00
|
|
|
hamlet = self.example_user("hamlet")
|
|
|
|
cordelia = self.example_user("cordelia")
|
|
|
|
othello = self.example_user("othello")
|
2020-07-07 01:00:03 +02:00
|
|
|
message_ids = [
|
2021-02-12 08:20:45 +01:00
|
|
|
self.send_personal_message(hamlet, cordelia, "test"),
|
|
|
|
self.send_personal_message(cordelia, othello, "test"),
|
2020-07-07 01:00:03 +02:00
|
|
|
]
|
|
|
|
|
|
|
|
messages = messages_for_ids(
|
|
|
|
message_ids=message_ids,
|
2021-02-12 08:20:45 +01:00
|
|
|
user_message_flags={message_id: ["read"] for message_id in message_ids},
|
2020-07-07 01:00:03 +02:00
|
|
|
search_fields={},
|
|
|
|
apply_markdown=True,
|
|
|
|
client_gravatar=True,
|
|
|
|
allow_edit_history=False,
|
|
|
|
)
|
|
|
|
|
2021-02-12 08:20:45 +01:00
|
|
|
self._verify_display_recipient(messages[0]["display_recipient"], [hamlet, cordelia])
|
|
|
|
self._verify_display_recipient(messages[1]["display_recipient"], [cordelia, othello])
|
2020-07-07 01:00:03 +02:00
|
|
|
|
|
|
|
def test_display_recipient_stream(self) -> None:
|
2021-02-12 08:20:45 +01:00
|
|
|
cordelia = self.example_user("cordelia")
|
tests: Ensure stream senders get a UserMessage row.
We now complain if a test author sends a stream message
that does not result in the sender getting a
UserMessage row for the message.
This is basically 100% equivalent to complaining that
the author failed to subscribe the sender to the stream
as part of the test setup, as far as I can tell, so the
AssertionError instructs the author to subscribe the
sender to the stream.
We exempt bots from this check, although it is
plausible we should only exempt the system bots like
the notification bot.
I considered auto-subscribing the sender to the stream,
but that can be a little more expensive than the
current check, and we generally want test setup to be
explicit.
If there is some legitimate way than a subscribed human
sender can't get a UserMessage, then we probably want
an explicit test for that, or we may want to change the
backend to just write a UserMessage row in that
hypothetical situation.
For most tests, including almost all the ones fixed
here, the author just wants their test setup to
realistically reflect normal operation, and often devs
may not realize that Cordelia is not subscribed to
Denmark or not realize that Hamlet is not subscribed to
Scotland.
Some of us don't remember our Shakespeare from high
school, and our stream subscriptions don't even
necessarily reflect which countries the Bard placed his
characters in.
There may also be some legitimate use case where an
author wants to simulate sending a message to an
unsubscribed stream, but for those edge cases, they can
always set allow_unsubscribed_sender to True.
2021-12-10 13:55:48 +01:00
|
|
|
self.subscribe(cordelia, "Denmark")
|
|
|
|
|
2020-07-07 01:00:03 +02:00
|
|
|
message_ids = [
|
2021-02-12 08:20:45 +01:00
|
|
|
self.send_stream_message(cordelia, "Verona", content="test"),
|
|
|
|
self.send_stream_message(cordelia, "Denmark", content="test"),
|
2020-07-07 01:00:03 +02:00
|
|
|
]
|
|
|
|
|
|
|
|
messages = messages_for_ids(
|
|
|
|
message_ids=message_ids,
|
2021-02-12 08:20:45 +01:00
|
|
|
user_message_flags={message_id: ["read"] for message_id in message_ids},
|
2020-07-07 01:00:03 +02:00
|
|
|
search_fields={},
|
|
|
|
apply_markdown=True,
|
|
|
|
client_gravatar=True,
|
|
|
|
allow_edit_history=False,
|
|
|
|
)
|
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
self._verify_display_recipient(
|
2021-02-12 08:20:45 +01:00
|
|
|
messages[0]["display_recipient"], get_stream("Verona", cordelia.realm)
|
2021-02-12 08:19:30 +01:00
|
|
|
)
|
|
|
|
self._verify_display_recipient(
|
2021-02-12 08:20:45 +01:00
|
|
|
messages[1]["display_recipient"], get_stream("Denmark", cordelia.realm)
|
2021-02-12 08:19:30 +01:00
|
|
|
)
|
2020-07-07 01:00:03 +02:00
|
|
|
|
|
|
|
def test_display_recipient_huddle(self) -> None:
|
2021-02-12 08:20:45 +01:00
|
|
|
hamlet = self.example_user("hamlet")
|
|
|
|
cordelia = self.example_user("cordelia")
|
|
|
|
othello = self.example_user("othello")
|
|
|
|
iago = self.example_user("iago")
|
2020-07-07 01:00:03 +02:00
|
|
|
message_ids = [
|
2021-02-12 08:20:45 +01:00
|
|
|
self.send_huddle_message(hamlet, [cordelia, othello], "test"),
|
|
|
|
self.send_huddle_message(cordelia, [hamlet, othello, iago], "test"),
|
2020-07-07 01:00:03 +02:00
|
|
|
]
|
|
|
|
|
|
|
|
messages = messages_for_ids(
|
|
|
|
message_ids=message_ids,
|
2021-02-12 08:20:45 +01:00
|
|
|
user_message_flags={message_id: ["read"] for message_id in message_ids},
|
2020-07-07 01:00:03 +02:00
|
|
|
search_fields={},
|
|
|
|
apply_markdown=True,
|
|
|
|
client_gravatar=True,
|
|
|
|
allow_edit_history=False,
|
|
|
|
)
|
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
self._verify_display_recipient(
|
2021-02-12 08:20:45 +01:00
|
|
|
messages[0]["display_recipient"], [hamlet, cordelia, othello]
|
2021-02-12 08:19:30 +01:00
|
|
|
)
|
|
|
|
self._verify_display_recipient(
|
2021-02-12 08:20:45 +01:00
|
|
|
messages[1]["display_recipient"], [hamlet, cordelia, othello, iago]
|
2021-02-12 08:19:30 +01:00
|
|
|
)
|
2020-07-07 01:00:03 +02:00
|
|
|
|
|
|
|
def test_display_recipient_various_types(self) -> None:
|
2021-02-12 08:20:45 +01:00
|
|
|
hamlet = self.example_user("hamlet")
|
|
|
|
cordelia = self.example_user("cordelia")
|
|
|
|
othello = self.example_user("othello")
|
|
|
|
iago = self.example_user("iago")
|
tests: Ensure stream senders get a UserMessage row.
We now complain if a test author sends a stream message
that does not result in the sender getting a
UserMessage row for the message.
This is basically 100% equivalent to complaining that
the author failed to subscribe the sender to the stream
as part of the test setup, as far as I can tell, so the
AssertionError instructs the author to subscribe the
sender to the stream.
We exempt bots from this check, although it is
plausible we should only exempt the system bots like
the notification bot.
I considered auto-subscribing the sender to the stream,
but that can be a little more expensive than the
current check, and we generally want test setup to be
explicit.
If there is some legitimate way than a subscribed human
sender can't get a UserMessage, then we probably want
an explicit test for that, or we may want to change the
backend to just write a UserMessage row in that
hypothetical situation.
For most tests, including almost all the ones fixed
here, the author just wants their test setup to
realistically reflect normal operation, and often devs
may not realize that Cordelia is not subscribed to
Denmark or not realize that Hamlet is not subscribed to
Scotland.
Some of us don't remember our Shakespeare from high
school, and our stream subscriptions don't even
necessarily reflect which countries the Bard placed his
characters in.
There may also be some legitimate use case where an
author wants to simulate sending a message to an
unsubscribed stream, but for those edge cases, they can
always set allow_unsubscribed_sender to True.
2021-12-10 13:55:48 +01:00
|
|
|
|
|
|
|
self.subscribe(cordelia, "Denmark")
|
|
|
|
self.subscribe(hamlet, "Scotland")
|
|
|
|
|
2020-07-07 01:00:03 +02:00
|
|
|
message_ids = [
|
2021-02-12 08:20:45 +01:00
|
|
|
self.send_huddle_message(hamlet, [cordelia, othello], "test"),
|
|
|
|
self.send_stream_message(cordelia, "Verona", content="test"),
|
|
|
|
self.send_personal_message(hamlet, cordelia, "test"),
|
|
|
|
self.send_stream_message(cordelia, "Denmark", content="test"),
|
|
|
|
self.send_huddle_message(cordelia, [hamlet, othello, iago], "test"),
|
|
|
|
self.send_personal_message(cordelia, othello, "test"),
|
2020-07-07 01:00:03 +02:00
|
|
|
]
|
|
|
|
|
|
|
|
messages = messages_for_ids(
|
|
|
|
message_ids=message_ids,
|
2021-02-12 08:20:45 +01:00
|
|
|
user_message_flags={message_id: ["read"] for message_id in message_ids},
|
2020-07-07 01:00:03 +02:00
|
|
|
search_fields={},
|
|
|
|
apply_markdown=True,
|
|
|
|
client_gravatar=True,
|
|
|
|
allow_edit_history=False,
|
|
|
|
)
|
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
self._verify_display_recipient(
|
2021-02-12 08:20:45 +01:00
|
|
|
messages[0]["display_recipient"], [hamlet, cordelia, othello]
|
2021-02-12 08:19:30 +01:00
|
|
|
)
|
|
|
|
self._verify_display_recipient(
|
2021-02-12 08:20:45 +01:00
|
|
|
messages[1]["display_recipient"], get_stream("Verona", hamlet.realm)
|
2021-02-12 08:19:30 +01:00
|
|
|
)
|
2021-02-12 08:20:45 +01:00
|
|
|
self._verify_display_recipient(messages[2]["display_recipient"], [hamlet, cordelia])
|
2021-02-12 08:19:30 +01:00
|
|
|
self._verify_display_recipient(
|
2021-02-12 08:20:45 +01:00
|
|
|
messages[3]["display_recipient"], get_stream("Denmark", hamlet.realm)
|
2021-02-12 08:19:30 +01:00
|
|
|
)
|
|
|
|
self._verify_display_recipient(
|
2021-02-12 08:20:45 +01:00
|
|
|
messages[4]["display_recipient"], [hamlet, cordelia, othello, iago]
|
2021-02-12 08:19:30 +01:00
|
|
|
)
|
2021-02-12 08:20:45 +01:00
|
|
|
self._verify_display_recipient(messages[5]["display_recipient"], [cordelia, othello])
|
2020-07-08 00:06:01 +02:00
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
|
2020-07-08 00:06:01 +02:00
|
|
|
class SewMessageAndReactionTest(ZulipTestCase):
|
|
|
|
def test_sew_messages_and_reaction(self) -> None:
|
2021-02-12 08:20:45 +01:00
|
|
|
sender = self.example_user("othello")
|
|
|
|
receiver = self.example_user("hamlet")
|
2022-09-27 21:42:31 +02:00
|
|
|
realm = get_realm("zulip")
|
2020-07-08 00:06:01 +02:00
|
|
|
pm_recipient = Recipient.objects.get(type_id=receiver.id, type=Recipient.PERSONAL)
|
2021-02-12 08:20:45 +01:00
|
|
|
stream_name = "Çiğdem"
|
2020-07-08 00:06:01 +02:00
|
|
|
stream = self.make_stream(stream_name)
|
|
|
|
stream_recipient = Recipient.objects.get(type_id=stream.id, type=Recipient.STREAM)
|
|
|
|
sending_client = make_client(name="test suite")
|
|
|
|
|
|
|
|
needed_ids = []
|
|
|
|
for i in range(5):
|
|
|
|
for recipient in [pm_recipient, stream_recipient]:
|
|
|
|
message = Message(
|
|
|
|
sender=sender,
|
|
|
|
recipient=recipient,
|
2022-09-27 21:42:31 +02:00
|
|
|
realm=realm,
|
2021-02-12 08:20:45 +01:00
|
|
|
content=f"whatever {i}",
|
2020-07-08 00:06:01 +02:00
|
|
|
date_sent=timezone_now(),
|
|
|
|
sending_client=sending_client,
|
|
|
|
last_edit_time=timezone_now(),
|
2021-02-12 08:20:45 +01:00
|
|
|
edit_history="[]",
|
2020-07-08 00:06:01 +02:00
|
|
|
)
|
2021-02-12 08:20:45 +01:00
|
|
|
message.set_topic_name("whatever")
|
2020-07-08 00:06:01 +02:00
|
|
|
message.save()
|
|
|
|
needed_ids.append(message.id)
|
2021-02-12 08:20:45 +01:00
|
|
|
reaction = Reaction(user_profile=sender, message=message, emoji_name="simple_smile")
|
2020-07-08 00:06:01 +02:00
|
|
|
reaction.save()
|
|
|
|
|
2021-02-12 08:20:45 +01:00
|
|
|
messages = Message.objects.filter(id__in=needed_ids).values(*["id", "content"])
|
2020-07-08 00:06:01 +02:00
|
|
|
reactions = Reaction.get_raw_db_rows(needed_ids)
|
|
|
|
tied_data = sew_messages_and_reactions(messages, reactions)
|
|
|
|
for data in tied_data:
|
2021-05-17 05:41:32 +02:00
|
|
|
self.assert_length(data["reactions"], 1)
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertEqual(data["reactions"][0]["emoji_name"], "simple_smile")
|
|
|
|
self.assertTrue(data["id"])
|
|
|
|
self.assertTrue(data["content"])
|