2020-06-29 13:11:26 +02:00
|
|
|
import time
|
2024-02-14 20:27:17 +01:00
|
|
|
from typing import Any, Callable, Dict, List
|
2020-06-29 13:11:26 +02:00
|
|
|
from unittest import mock
|
2023-01-18 18:40:44 +01:00
|
|
|
from urllib.parse import urlsplit
|
2020-06-29 13:11:26 +02:00
|
|
|
|
2020-08-07 01:09:47 +02:00
|
|
|
import orjson
|
2020-06-29 13:11:26 +02:00
|
|
|
from django.conf import settings
|
|
|
|
from django.http import HttpRequest, HttpResponse
|
2022-08-24 21:42:44 +02:00
|
|
|
from django.test import override_settings
|
2021-05-05 13:15:13 +02:00
|
|
|
from django.utils.timezone import now as timezone_now
|
2023-10-12 19:43:45 +02:00
|
|
|
from typing_extensions import override
|
2020-06-29 13:11:26 +02:00
|
|
|
|
2022-10-27 19:05:10 +02:00
|
|
|
from zerver.actions.custom_profile_fields import try_update_realm_custom_profile_field
|
2022-04-14 23:50:10 +02:00
|
|
|
from zerver.actions.message_send import check_send_message
|
2022-04-14 23:44:33 +02:00
|
|
|
from zerver.actions.presence import do_update_user_presence
|
2021-10-26 09:15:16 +02:00
|
|
|
from zerver.actions.user_settings import do_change_user_setting
|
2022-04-14 23:48:28 +02:00
|
|
|
from zerver.actions.users import do_change_user_role
|
2024-02-10 04:19:08 +01:00
|
|
|
from zerver.lib.event_schema import check_web_reload_client_event
|
2021-07-16 22:11:10 +02:00
|
|
|
from zerver.lib.events import fetch_initial_state_data
|
2021-07-04 08:45:34 +02:00
|
|
|
from zerver.lib.exceptions import AccessDeniedError
|
2022-07-15 00:00:31 +02:00
|
|
|
from zerver.lib.request import RequestVariableMissingError
|
2020-06-29 13:11:26 +02:00
|
|
|
from zerver.lib.test_classes import ZulipTestCase
|
2021-10-26 09:15:16 +02:00
|
|
|
from zerver.lib.test_helpers import (
|
|
|
|
HostRequestMock,
|
|
|
|
dummy_handler,
|
2023-03-01 07:34:25 +01:00
|
|
|
reset_email_visibility_to_everyone_in_zulip_realm,
|
2021-10-26 09:15:16 +02:00
|
|
|
stub_event_queue_user_events,
|
|
|
|
)
|
2023-11-08 08:21:24 +01:00
|
|
|
from zerver.lib.users import get_api_key, get_users_for_api
|
2023-12-15 04:33:19 +01:00
|
|
|
from zerver.models import CustomProfileField, UserMessage, UserPresence, UserProfile
|
|
|
|
from zerver.models.clients import get_client
|
2023-12-15 02:14:24 +01:00
|
|
|
from zerver.models.realms import get_realm
|
2023-12-15 03:57:04 +01:00
|
|
|
from zerver.models.streams import get_stream
|
2023-12-15 01:16:00 +01:00
|
|
|
from zerver.models.users import get_system_bot
|
2020-06-29 13:11:26 +02:00
|
|
|
from zerver.tornado.event_queue import (
|
|
|
|
allocate_client_descriptor,
|
|
|
|
clear_client_event_queues_for_testing,
|
|
|
|
get_client_info_for_message_event,
|
2024-02-07 22:03:15 +01:00
|
|
|
mark_clients_to_reload,
|
2020-06-29 13:11:26 +02:00
|
|
|
process_message_event,
|
2024-02-10 04:19:08 +01:00
|
|
|
send_web_reload_client_events,
|
2020-06-29 13:11:26 +02:00
|
|
|
)
|
2022-09-17 21:55:15 +02:00
|
|
|
from zerver.tornado.exceptions import BadEventQueueIdError
|
2024-02-14 20:27:17 +01:00
|
|
|
from zerver.tornado.views import get_events
|
|
|
|
from zerver.views.events_register import _default_all_public_streams, _default_narrow
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
|
|
|
|
class EventsEndpointTest(ZulipTestCase):
|
2022-11-11 01:21:04 +01:00
|
|
|
def test_events_register_without_user_agent(self) -> None:
|
|
|
|
result = self.client_post("/json/register", skip_user_agent=True)
|
|
|
|
self.assert_json_success(result)
|
|
|
|
|
2023-06-30 01:03:42 +02:00
|
|
|
def test_narrows(self) -> None:
|
|
|
|
user = self.example_user("hamlet")
|
|
|
|
with mock.patch("zerver.lib.events.request_event_queue", return_value=None) as m:
|
|
|
|
munge = lambda obj: orjson.dumps(obj).decode()
|
|
|
|
narrow = [["stream", "devel"], ["is", "mentioned"]]
|
|
|
|
payload = dict(narrow=munge(narrow))
|
|
|
|
result = self.api_post(user, "/api/v1/register", payload)
|
|
|
|
|
|
|
|
# We want the test to abort before actually fetching data.
|
|
|
|
self.assert_json_error(result, "Could not allocate event queue")
|
|
|
|
|
|
|
|
self.assertEqual(m.call_args.kwargs["narrow"], [["stream", "devel"], ["is", "mentioned"]])
|
|
|
|
|
2020-06-29 13:11:26 +02:00
|
|
|
def test_events_register_endpoint(self) -> None:
|
|
|
|
# This test is intended to get minimal coverage on the
|
|
|
|
# events_register code paths
|
|
|
|
user = self.example_user("hamlet")
|
2021-02-12 08:20:45 +01:00
|
|
|
with mock.patch("zerver.views.events_register.do_events_register", return_value={}):
|
2022-11-04 20:26:18 +01:00
|
|
|
result = self.api_post(user, "/api/v1/register")
|
2020-06-29 13:11:26 +02:00
|
|
|
self.assert_json_success(result)
|
|
|
|
|
2021-02-12 08:20:45 +01:00
|
|
|
with mock.patch("zerver.lib.events.request_event_queue", return_value=None):
|
2022-11-04 20:26:18 +01:00
|
|
|
result = self.api_post(user, "/api/v1/register")
|
2020-06-29 13:11:26 +02:00
|
|
|
self.assert_json_error(result, "Could not allocate event queue")
|
|
|
|
|
2021-02-12 08:20:45 +01:00
|
|
|
return_event_queue = "15:11"
|
2020-06-29 13:11:26 +02:00
|
|
|
return_user_events: List[Dict[str, Any]] = []
|
|
|
|
|
|
|
|
# We choose realm_emoji somewhat randomly--we want
|
|
|
|
# a "boring" event type for the purpose of this test.
|
2021-02-12 08:20:45 +01:00
|
|
|
event_type = "realm_emoji"
|
2023-08-16 20:53:41 +02:00
|
|
|
empty_realm_emoji_dict: Dict[str, Any] = {}
|
|
|
|
test_event = dict(id=6, type=event_type, realm_emoji=empty_realm_emoji_dict)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
# Test that call is made to deal with a returning soft deactivated user.
|
2021-02-12 08:20:45 +01:00
|
|
|
with mock.patch("zerver.lib.events.reactivate_user_if_soft_deactivated") as fa:
|
2020-06-29 13:11:26 +02:00
|
|
|
with stub_event_queue_user_events(return_event_queue, return_user_events):
|
2021-02-12 08:19:30 +01:00
|
|
|
result = self.api_post(
|
2022-11-04 20:26:18 +01:00
|
|
|
user, "/api/v1/register", dict(event_types=orjson.dumps([event_type]).decode())
|
2021-02-12 08:19:30 +01:00
|
|
|
)
|
2020-06-29 13:11:26 +02:00
|
|
|
self.assertEqual(fa.call_count, 1)
|
|
|
|
|
|
|
|
with stub_event_queue_user_events(return_event_queue, return_user_events):
|
2021-02-12 08:19:30 +01:00
|
|
|
result = self.api_post(
|
2022-11-04 20:26:18 +01:00
|
|
|
user, "/api/v1/register", dict(event_types=orjson.dumps([event_type]).decode())
|
2021-02-12 08:19:30 +01:00
|
|
|
)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
2022-06-07 01:37:01 +02:00
|
|
|
result_dict = self.assert_json_success(result)
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertEqual(result_dict["last_event_id"], -1)
|
|
|
|
self.assertEqual(result_dict["queue_id"], "15:11")
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
# Now start simulating returning actual data
|
2021-02-12 08:20:45 +01:00
|
|
|
return_event_queue = "15:12"
|
2020-06-29 13:11:26 +02:00
|
|
|
return_user_events = [test_event]
|
|
|
|
|
|
|
|
with stub_event_queue_user_events(return_event_queue, return_user_events):
|
2021-02-12 08:19:30 +01:00
|
|
|
result = self.api_post(
|
2022-11-04 20:26:18 +01:00
|
|
|
user, "/api/v1/register", dict(event_types=orjson.dumps([event_type]).decode())
|
2021-02-12 08:19:30 +01:00
|
|
|
)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
2022-06-07 01:37:01 +02:00
|
|
|
result_dict = self.assert_json_success(result)
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertEqual(result_dict["last_event_id"], 6)
|
|
|
|
self.assertEqual(result_dict["queue_id"], "15:12")
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
# sanity check the data relevant to our event
|
2023-08-16 20:53:41 +02:00
|
|
|
self.assertEqual(result_dict["realm_emoji"], {})
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
# Now test with `fetch_event_types` not matching the event
|
2021-02-12 08:20:45 +01:00
|
|
|
return_event_queue = "15:13"
|
2020-06-29 13:11:26 +02:00
|
|
|
with stub_event_queue_user_events(return_event_queue, return_user_events):
|
2021-02-12 08:19:30 +01:00
|
|
|
result = self.api_post(
|
|
|
|
user,
|
2022-11-04 20:26:18 +01:00
|
|
|
"/api/v1/register",
|
2021-02-12 08:19:30 +01:00
|
|
|
dict(
|
|
|
|
event_types=orjson.dumps([event_type]).decode(),
|
2021-02-12 08:20:45 +01:00
|
|
|
fetch_event_types=orjson.dumps(["message"]).decode(),
|
2021-02-12 08:19:30 +01:00
|
|
|
),
|
|
|
|
)
|
2022-06-07 01:37:01 +02:00
|
|
|
result_dict = self.assert_json_success(result)
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertEqual(result_dict["last_event_id"], 6)
|
2020-06-29 13:11:26 +02:00
|
|
|
# Check that the message event types data is in there
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertIn("max_message_id", result_dict)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
# Check that our original event type is not there.
|
|
|
|
self.assertNotIn(event_type, result_dict)
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertEqual(result_dict["queue_id"], "15:13")
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
# Now test with `fetch_event_types` matching the event
|
|
|
|
with stub_event_queue_user_events(return_event_queue, return_user_events):
|
2021-02-12 08:19:30 +01:00
|
|
|
result = self.api_post(
|
|
|
|
user,
|
2022-11-04 20:26:18 +01:00
|
|
|
"/api/v1/register",
|
2021-02-12 08:19:30 +01:00
|
|
|
dict(
|
|
|
|
fetch_event_types=orjson.dumps([event_type]).decode(),
|
2021-02-12 08:20:45 +01:00
|
|
|
event_types=orjson.dumps(["message"]).decode(),
|
2021-02-12 08:19:30 +01:00
|
|
|
),
|
|
|
|
)
|
2022-06-07 01:37:01 +02:00
|
|
|
result_dict = self.assert_json_success(result)
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertEqual(result_dict["last_event_id"], 6)
|
2020-06-29 13:11:26 +02:00
|
|
|
# Check that we didn't fetch the messages data
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertNotIn("max_message_id", result_dict)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
# Check that the realm_emoji data is in there.
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertIn("realm_emoji", result_dict)
|
2023-08-16 20:53:41 +02:00
|
|
|
self.assertEqual(result_dict["realm_emoji"], {})
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertEqual(result_dict["queue_id"], "15:13")
|
2020-06-29 13:11:26 +02:00
|
|
|
|
2022-05-03 23:48:52 +02:00
|
|
|
def test_events_register_spectators(self) -> None:
|
|
|
|
# Verify that POST /register works for spectators, but not for
|
|
|
|
# normal users.
|
|
|
|
with self.settings(WEB_PUBLIC_STREAMS_ENABLED=False):
|
2022-09-26 17:37:04 +02:00
|
|
|
result = self.client_post("/json/register")
|
2022-05-03 23:48:52 +02:00
|
|
|
self.assert_json_error(
|
|
|
|
result,
|
|
|
|
"Not logged in: API authentication or user session required",
|
|
|
|
status_code=401,
|
|
|
|
)
|
|
|
|
|
2022-09-26 17:37:04 +02:00
|
|
|
result = self.client_post("/json/register")
|
2022-06-07 01:37:01 +02:00
|
|
|
result_dict = self.assert_json_success(result)
|
2022-05-03 23:48:52 +02:00
|
|
|
self.assertEqual(result_dict["queue_id"], None)
|
|
|
|
self.assertEqual(result_dict["realm_uri"], "http://zulip.testserver")
|
|
|
|
|
2022-09-26 17:37:04 +02:00
|
|
|
result = self.client_post("/json/register")
|
|
|
|
self.assertEqual(result.status_code, 200)
|
|
|
|
|
|
|
|
result = self.client_post("/json/register", dict(client_gravatar="false"))
|
|
|
|
self.assertEqual(result.status_code, 200)
|
|
|
|
|
|
|
|
result = self.client_post("/json/register", dict(client_gravatar="true"))
|
|
|
|
self.assert_json_error(
|
|
|
|
result,
|
|
|
|
"Invalid 'client_gravatar' parameter for anonymous request",
|
|
|
|
status_code=400,
|
|
|
|
)
|
|
|
|
|
|
|
|
result = self.client_post("/json/register", dict(include_subscribers="true"))
|
|
|
|
self.assert_json_error(
|
|
|
|
result,
|
|
|
|
"Invalid 'include_subscribers' parameter for anonymous request",
|
|
|
|
status_code=400,
|
|
|
|
)
|
|
|
|
|
2021-04-08 22:14:31 +02:00
|
|
|
def test_events_register_endpoint_all_public_streams_access(self) -> None:
|
|
|
|
guest_user = self.example_user("polonius")
|
|
|
|
normal_user = self.example_user("hamlet")
|
|
|
|
self.assertEqual(guest_user.role, UserProfile.ROLE_GUEST)
|
|
|
|
self.assertEqual(normal_user.role, UserProfile.ROLE_MEMBER)
|
|
|
|
|
|
|
|
with mock.patch("zerver.views.events_register.do_events_register", return_value={}):
|
2022-11-04 20:26:18 +01:00
|
|
|
result = self.api_post(normal_user, "/api/v1/register", dict(all_public_streams="true"))
|
2021-04-08 22:14:31 +02:00
|
|
|
self.assert_json_success(result)
|
|
|
|
|
|
|
|
with mock.patch("zerver.views.events_register.do_events_register", return_value={}):
|
2022-11-04 20:26:18 +01:00
|
|
|
result = self.api_post(guest_user, "/api/v1/register", dict(all_public_streams="true"))
|
2021-04-08 22:14:31 +02:00
|
|
|
self.assert_json_error(result, "User not authorized for this query")
|
|
|
|
|
|
|
|
def test_events_get_events_endpoint_guest_cant_use_all_public_streams_param(self) -> None:
|
|
|
|
"""
|
|
|
|
This test is meant to execute the very beginning of the codepath
|
|
|
|
to ensure guest users are immediately disallowed to use the
|
|
|
|
all_public_streams param. Deeper testing is hard (and not necessary for this case)
|
|
|
|
due to the codepath expecting AsyncDjangoHandler to be attached to the request,
|
|
|
|
which doesn't happen in our test setup.
|
|
|
|
"""
|
|
|
|
|
|
|
|
guest_user = self.example_user("polonius")
|
|
|
|
self.assertEqual(guest_user.role, UserProfile.ROLE_GUEST)
|
|
|
|
|
|
|
|
result = self.api_get(guest_user, "/api/v1/events", dict(all_public_streams="true"))
|
|
|
|
self.assert_json_error(result, "User not authorized for this query")
|
|
|
|
|
2020-06-29 13:11:26 +02:00
|
|
|
def test_tornado_endpoint(self) -> None:
|
2024-02-08 20:10:25 +01:00
|
|
|
# This test is mostly intended to get minimal coverage on the
|
|
|
|
# /api/internal/notify_tornado endpoint (only used in
|
|
|
|
# puppeteer tests), so we can have 100% URL coverage, but it
|
|
|
|
# does exercise a little bit of the codepath.
|
2020-06-29 13:11:26 +02:00
|
|
|
post_data = dict(
|
2020-08-07 01:09:47 +02:00
|
|
|
data=orjson.dumps(
|
2020-06-29 13:11:26 +02:00
|
|
|
dict(
|
|
|
|
event=dict(
|
2021-02-12 08:20:45 +01:00
|
|
|
type="other",
|
2020-06-29 13:11:26 +02:00
|
|
|
),
|
2021-02-12 08:20:45 +01:00
|
|
|
users=[self.example_user("hamlet").id],
|
2020-06-29 13:11:26 +02:00
|
|
|
),
|
2020-08-07 01:09:47 +02:00
|
|
|
).decode(),
|
2020-06-29 13:11:26 +02:00
|
|
|
)
|
2022-07-11 19:35:31 +02:00
|
|
|
req = HostRequestMock(post_data)
|
2021-02-12 08:20:45 +01:00
|
|
|
req.META["REMOTE_ADDR"] = "127.0.0.1"
|
2022-07-15 00:00:31 +02:00
|
|
|
with self.assertRaises(RequestVariableMissingError) as context:
|
2024-02-08 20:10:25 +01:00
|
|
|
result = self.client_post_request("/api/internal/notify_tornado", req)
|
2022-07-15 00:00:31 +02:00
|
|
|
self.assertEqual(str(context.exception), "Missing 'secret' argument")
|
|
|
|
self.assertEqual(context.exception.http_status_code, 400)
|
|
|
|
|
|
|
|
post_data["secret"] = "random"
|
|
|
|
req = HostRequestMock(post_data, user_profile=None)
|
|
|
|
req.META["REMOTE_ADDR"] = "127.0.0.1"
|
2022-07-28 17:17:55 +02:00
|
|
|
with self.assertRaises(AccessDeniedError) as access_denied_error:
|
2024-02-08 20:10:25 +01:00
|
|
|
result = self.client_post_request("/api/internal/notify_tornado", req)
|
2022-07-28 17:17:55 +02:00
|
|
|
self.assertEqual(str(access_denied_error.exception), "Access denied")
|
|
|
|
self.assertEqual(access_denied_error.exception.http_status_code, 403)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
2021-02-12 08:20:45 +01:00
|
|
|
post_data["secret"] = settings.SHARED_SECRET
|
2022-07-11 19:35:31 +02:00
|
|
|
req = HostRequestMock(post_data, tornado_handler=dummy_handler)
|
2021-02-12 08:20:45 +01:00
|
|
|
req.META["REMOTE_ADDR"] = "127.0.0.1"
|
2024-02-08 20:10:25 +01:00
|
|
|
result = self.client_post_request("/api/internal/notify_tornado", req)
|
2020-06-29 13:11:26 +02:00
|
|
|
self.assert_json_success(result)
|
|
|
|
|
2022-07-14 21:11:26 +02:00
|
|
|
post_data = dict(secret=settings.SHARED_SECRET)
|
|
|
|
req = HostRequestMock(post_data, tornado_handler=dummy_handler)
|
|
|
|
req.META["REMOTE_ADDR"] = "127.0.0.1"
|
|
|
|
with self.assertRaises(RequestVariableMissingError) as context:
|
2024-02-08 20:10:25 +01:00
|
|
|
result = self.client_post_request("/api/internal/notify_tornado", req)
|
2022-07-14 21:11:26 +02:00
|
|
|
self.assertEqual(str(context.exception), "Missing 'data' argument")
|
|
|
|
self.assertEqual(context.exception.http_status_code, 400)
|
|
|
|
|
2024-02-08 20:57:16 +01:00
|
|
|
def test_web_reload_clients(self) -> None:
|
|
|
|
# Minimal testing of the /api/internal/web_reload_clients endpoint
|
|
|
|
post_data = {
|
|
|
|
"client_count": "1",
|
|
|
|
"immediate": orjson.dumps(False).decode(),
|
|
|
|
"secret": settings.SHARED_SECRET,
|
|
|
|
}
|
|
|
|
req = HostRequestMock(post_data, tornado_handler=dummy_handler)
|
|
|
|
req.META["REMOTE_ADDR"] = "127.0.0.1"
|
|
|
|
result = self.client_post_request("/api/internal/web_reload_clients", req)
|
|
|
|
self.assert_json_success(result)
|
|
|
|
self.assertEqual(orjson.loads(result.content)["sent_events"], 0)
|
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
|
2020-06-29 13:11:26 +02:00
|
|
|
class GetEventsTest(ZulipTestCase):
|
2021-02-12 08:19:30 +01:00
|
|
|
def tornado_call(
|
|
|
|
self,
|
|
|
|
view_func: Callable[[HttpRequest, UserProfile], HttpResponse],
|
|
|
|
user_profile: UserProfile,
|
|
|
|
post_data: Dict[str, Any],
|
|
|
|
) -> HttpResponse:
|
2022-06-24 09:18:14 +02:00
|
|
|
request = HostRequestMock(post_data, user_profile, tornado_handler=dummy_handler)
|
2020-06-29 13:11:26 +02:00
|
|
|
return view_func(request, user_profile)
|
|
|
|
|
|
|
|
def test_get_events(self) -> None:
|
2021-02-12 08:20:45 +01:00
|
|
|
user_profile = self.example_user("hamlet")
|
2020-06-29 13:11:26 +02:00
|
|
|
email = user_profile.email
|
2021-02-12 08:20:45 +01:00
|
|
|
recipient_user_profile = self.example_user("othello")
|
2020-06-29 13:11:26 +02:00
|
|
|
recipient_email = recipient_user_profile.email
|
|
|
|
self.login_user(user_profile)
|
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
result = self.tornado_call(
|
|
|
|
get_events,
|
|
|
|
user_profile,
|
|
|
|
{
|
|
|
|
"apply_markdown": orjson.dumps(True).decode(),
|
|
|
|
"client_gravatar": orjson.dumps(True).decode(),
|
|
|
|
"event_types": orjson.dumps(["message"]).decode(),
|
|
|
|
"user_client": "website",
|
|
|
|
"dont_block": orjson.dumps(True).decode(),
|
|
|
|
},
|
|
|
|
)
|
2020-06-29 13:11:26 +02:00
|
|
|
self.assert_json_success(result)
|
2020-08-07 01:09:47 +02:00
|
|
|
queue_id = orjson.loads(result.content)["queue_id"]
|
2020-06-29 13:11:26 +02:00
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
recipient_result = self.tornado_call(
|
|
|
|
get_events,
|
|
|
|
recipient_user_profile,
|
|
|
|
{
|
|
|
|
"apply_markdown": orjson.dumps(True).decode(),
|
|
|
|
"client_gravatar": orjson.dumps(True).decode(),
|
|
|
|
"event_types": orjson.dumps(["message"]).decode(),
|
|
|
|
"user_client": "website",
|
|
|
|
"dont_block": orjson.dumps(True).decode(),
|
|
|
|
},
|
|
|
|
)
|
2020-06-29 13:11:26 +02:00
|
|
|
self.assert_json_success(recipient_result)
|
2020-08-07 01:09:47 +02:00
|
|
|
recipient_queue_id = orjson.loads(recipient_result.content)["queue_id"]
|
2020-06-29 13:11:26 +02:00
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
result = self.tornado_call(
|
|
|
|
get_events,
|
|
|
|
user_profile,
|
|
|
|
{
|
|
|
|
"queue_id": queue_id,
|
|
|
|
"user_client": "website",
|
|
|
|
"last_event_id": -1,
|
|
|
|
"dont_block": orjson.dumps(True).decode(),
|
|
|
|
},
|
|
|
|
)
|
2020-08-07 01:09:47 +02:00
|
|
|
events = orjson.loads(result.content)["events"]
|
2020-06-29 13:11:26 +02:00
|
|
|
self.assert_json_success(result)
|
|
|
|
self.assert_length(events, 0)
|
|
|
|
|
2021-02-12 08:20:45 +01:00
|
|
|
local_id = "10.01"
|
Revert "create_user: Use transaction.atomic decorator for do_create_user."
This reverts commit 851d68e0fc364d649175533c286c179cf38f89d6.
That commit widened how long the transaction is open, which made it
much more likely that after the user was created in the transaction,
and the memcached caches were flushed, some other request will fill
the `get_realm_user_dicts` cache with data which did not include the
new user (because it had not been committed yet).
If a user creation request lost this race, the user would, upon first
request to `/`, get a blank page and a Javascript error:
Unknown user_id in get_by_user_id: 12345
...where 12345 was their own user-id. This error would persist until
the cache expired (in 7 days) or something else expunged it.
Reverting this does not prevent the race, as the post_save hook's call
to flush_user_profile is still in a transaction (and has been since
168f241ff0a5), and thus leaves the potential race window open.
However, it much shortens the potential window of opportunity, and is
a reasonable short-term stopgap.
2023-02-18 02:44:51 +01:00
|
|
|
check_send_message(
|
|
|
|
sender=user_profile,
|
|
|
|
client=get_client("whatever"),
|
2023-04-18 17:23:58 +02:00
|
|
|
recipient_type_name="private",
|
Revert "create_user: Use transaction.atomic decorator for do_create_user."
This reverts commit 851d68e0fc364d649175533c286c179cf38f89d6.
That commit widened how long the transaction is open, which made it
much more likely that after the user was created in the transaction,
and the memcached caches were flushed, some other request will fill
the `get_realm_user_dicts` cache with data which did not include the
new user (because it had not been committed yet).
If a user creation request lost this race, the user would, upon first
request to `/`, get a blank page and a Javascript error:
Unknown user_id in get_by_user_id: 12345
...where 12345 was their own user-id. This error would persist until
the cache expired (in 7 days) or something else expunged it.
Reverting this does not prevent the race, as the post_save hook's call
to flush_user_profile is still in a transaction (and has been since
168f241ff0a5), and thus leaves the potential race window open.
However, it much shortens the potential window of opportunity, and is
a reasonable short-term stopgap.
2023-02-18 02:44:51 +01:00
|
|
|
message_to=[recipient_email],
|
|
|
|
topic_name=None,
|
|
|
|
message_content="hello",
|
|
|
|
local_id=local_id,
|
|
|
|
sender_queue_id=queue_id,
|
|
|
|
)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
result = self.tornado_call(
|
|
|
|
get_events,
|
|
|
|
user_profile,
|
|
|
|
{
|
|
|
|
"queue_id": queue_id,
|
|
|
|
"user_client": "website",
|
|
|
|
"last_event_id": -1,
|
|
|
|
"dont_block": orjson.dumps(True).decode(),
|
|
|
|
},
|
|
|
|
)
|
2020-08-07 01:09:47 +02:00
|
|
|
events = orjson.loads(result.content)["events"]
|
2020-06-29 13:11:26 +02:00
|
|
|
self.assert_json_success(result)
|
|
|
|
self.assert_length(events, 1)
|
|
|
|
self.assertEqual(events[0]["type"], "message")
|
|
|
|
self.assertEqual(events[0]["message"]["sender_email"], email)
|
|
|
|
self.assertEqual(events[0]["local_message_id"], local_id)
|
|
|
|
self.assertEqual(events[0]["message"]["display_recipient"][0]["is_mirror_dummy"], False)
|
|
|
|
self.assertEqual(events[0]["message"]["display_recipient"][1]["is_mirror_dummy"], False)
|
|
|
|
|
|
|
|
last_event_id = events[0]["id"]
|
2021-02-12 08:20:45 +01:00
|
|
|
local_id = "10.02"
|
2020-06-29 13:11:26 +02:00
|
|
|
|
Revert "create_user: Use transaction.atomic decorator for do_create_user."
This reverts commit 851d68e0fc364d649175533c286c179cf38f89d6.
That commit widened how long the transaction is open, which made it
much more likely that after the user was created in the transaction,
and the memcached caches were flushed, some other request will fill
the `get_realm_user_dicts` cache with data which did not include the
new user (because it had not been committed yet).
If a user creation request lost this race, the user would, upon first
request to `/`, get a blank page and a Javascript error:
Unknown user_id in get_by_user_id: 12345
...where 12345 was their own user-id. This error would persist until
the cache expired (in 7 days) or something else expunged it.
Reverting this does not prevent the race, as the post_save hook's call
to flush_user_profile is still in a transaction (and has been since
168f241ff0a5), and thus leaves the potential race window open.
However, it much shortens the potential window of opportunity, and is
a reasonable short-term stopgap.
2023-02-18 02:44:51 +01:00
|
|
|
check_send_message(
|
|
|
|
sender=user_profile,
|
|
|
|
client=get_client("whatever"),
|
2023-04-18 17:23:58 +02:00
|
|
|
recipient_type_name="private",
|
Revert "create_user: Use transaction.atomic decorator for do_create_user."
This reverts commit 851d68e0fc364d649175533c286c179cf38f89d6.
That commit widened how long the transaction is open, which made it
much more likely that after the user was created in the transaction,
and the memcached caches were flushed, some other request will fill
the `get_realm_user_dicts` cache with data which did not include the
new user (because it had not been committed yet).
If a user creation request lost this race, the user would, upon first
request to `/`, get a blank page and a Javascript error:
Unknown user_id in get_by_user_id: 12345
...where 12345 was their own user-id. This error would persist until
the cache expired (in 7 days) or something else expunged it.
Reverting this does not prevent the race, as the post_save hook's call
to flush_user_profile is still in a transaction (and has been since
168f241ff0a5), and thus leaves the potential race window open.
However, it much shortens the potential window of opportunity, and is
a reasonable short-term stopgap.
2023-02-18 02:44:51 +01:00
|
|
|
message_to=[recipient_email],
|
|
|
|
topic_name=None,
|
|
|
|
message_content="hello",
|
|
|
|
local_id=local_id,
|
|
|
|
sender_queue_id=queue_id,
|
|
|
|
)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
result = self.tornado_call(
|
|
|
|
get_events,
|
|
|
|
user_profile,
|
|
|
|
{
|
|
|
|
"queue_id": queue_id,
|
|
|
|
"user_client": "website",
|
|
|
|
"last_event_id": last_event_id,
|
|
|
|
"dont_block": orjson.dumps(True).decode(),
|
|
|
|
},
|
|
|
|
)
|
2020-08-07 01:09:47 +02:00
|
|
|
events = orjson.loads(result.content)["events"]
|
2020-06-29 13:11:26 +02:00
|
|
|
self.assert_json_success(result)
|
|
|
|
self.assert_length(events, 1)
|
|
|
|
self.assertEqual(events[0]["type"], "message")
|
|
|
|
self.assertEqual(events[0]["message"]["sender_email"], email)
|
|
|
|
self.assertEqual(events[0]["local_message_id"], local_id)
|
|
|
|
|
|
|
|
# Test that the received message in the receiver's event queue
|
|
|
|
# exists and does not contain a local id
|
2021-02-12 08:19:30 +01:00
|
|
|
recipient_result = self.tornado_call(
|
|
|
|
get_events,
|
|
|
|
recipient_user_profile,
|
|
|
|
{
|
|
|
|
"queue_id": recipient_queue_id,
|
|
|
|
"user_client": "website",
|
|
|
|
"last_event_id": -1,
|
|
|
|
"dont_block": orjson.dumps(True).decode(),
|
|
|
|
},
|
|
|
|
)
|
2020-08-07 01:09:47 +02:00
|
|
|
recipient_events = orjson.loads(recipient_result.content)["events"]
|
2020-06-29 13:11:26 +02:00
|
|
|
self.assert_json_success(recipient_result)
|
2021-05-17 05:41:32 +02:00
|
|
|
self.assert_length(recipient_events, 2)
|
2020-06-29 13:11:26 +02:00
|
|
|
self.assertEqual(recipient_events[0]["type"], "message")
|
|
|
|
self.assertEqual(recipient_events[0]["message"]["sender_email"], email)
|
|
|
|
self.assertTrue("local_message_id" not in recipient_events[0])
|
|
|
|
self.assertEqual(recipient_events[1]["type"], "message")
|
|
|
|
self.assertEqual(recipient_events[1]["message"]["sender_email"], email)
|
|
|
|
self.assertTrue("local_message_id" not in recipient_events[1])
|
|
|
|
|
|
|
|
def test_get_events_narrow(self) -> None:
|
2021-02-12 08:20:45 +01:00
|
|
|
user_profile = self.example_user("hamlet")
|
2020-06-29 13:11:26 +02:00
|
|
|
self.login_user(user_profile)
|
|
|
|
|
|
|
|
def get_message(apply_markdown: bool, client_gravatar: bool) -> Dict[str, Any]:
|
|
|
|
result = self.tornado_call(
|
|
|
|
get_events,
|
|
|
|
user_profile,
|
|
|
|
dict(
|
2020-08-07 01:09:47 +02:00
|
|
|
apply_markdown=orjson.dumps(apply_markdown).decode(),
|
|
|
|
client_gravatar=orjson.dumps(client_gravatar).decode(),
|
|
|
|
event_types=orjson.dumps(["message"]).decode(),
|
|
|
|
narrow=orjson.dumps([["stream", "denmark"]]).decode(),
|
2020-06-29 13:11:26 +02:00
|
|
|
user_client="website",
|
2020-08-07 01:09:47 +02:00
|
|
|
dont_block=orjson.dumps(True).decode(),
|
2020-06-29 13:11:26 +02:00
|
|
|
),
|
|
|
|
)
|
|
|
|
|
|
|
|
self.assert_json_success(result)
|
2020-08-07 01:09:47 +02:00
|
|
|
queue_id = orjson.loads(result.content)["queue_id"]
|
2020-06-29 13:11:26 +02:00
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
result = self.tornado_call(
|
|
|
|
get_events,
|
|
|
|
user_profile,
|
|
|
|
{
|
|
|
|
"queue_id": queue_id,
|
|
|
|
"user_client": "website",
|
|
|
|
"last_event_id": -1,
|
|
|
|
"dont_block": orjson.dumps(True).decode(),
|
|
|
|
},
|
|
|
|
)
|
2020-08-07 01:09:47 +02:00
|
|
|
events = orjson.loads(result.content)["events"]
|
2020-06-29 13:11:26 +02:00
|
|
|
self.assert_json_success(result)
|
|
|
|
self.assert_length(events, 0)
|
|
|
|
|
|
|
|
self.send_personal_message(user_profile, self.example_user("othello"), "hello")
|
|
|
|
self.send_stream_message(user_profile, "Denmark", "**hello**")
|
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
result = self.tornado_call(
|
|
|
|
get_events,
|
|
|
|
user_profile,
|
|
|
|
{
|
|
|
|
"queue_id": queue_id,
|
|
|
|
"user_client": "website",
|
|
|
|
"last_event_id": -1,
|
|
|
|
"dont_block": orjson.dumps(True).decode(),
|
|
|
|
},
|
|
|
|
)
|
2020-08-07 01:09:47 +02:00
|
|
|
events = orjson.loads(result.content)["events"]
|
2020-06-29 13:11:26 +02:00
|
|
|
self.assert_json_success(result)
|
|
|
|
self.assert_length(events, 1)
|
|
|
|
self.assertEqual(events[0]["type"], "message")
|
2021-02-12 08:20:45 +01:00
|
|
|
return events[0]["message"]
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
message = get_message(apply_markdown=False, client_gravatar=False)
|
|
|
|
self.assertEqual(message["display_recipient"], "Denmark")
|
|
|
|
self.assertEqual(message["content"], "**hello**")
|
2023-01-18 18:40:44 +01:00
|
|
|
self.assertEqual(urlsplit(message["avatar_url"]).hostname, "secure.gravatar.com")
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
message = get_message(apply_markdown=True, client_gravatar=False)
|
|
|
|
self.assertEqual(message["display_recipient"], "Denmark")
|
|
|
|
self.assertEqual(message["content"], "<p><strong>hello</strong></p>")
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertIn("gravatar.com", message["avatar_url"])
|
2020-06-29 13:11:26 +02:00
|
|
|
|
2021-10-26 09:15:16 +02:00
|
|
|
do_change_user_setting(
|
|
|
|
user_profile,
|
|
|
|
"email_address_visibility",
|
|
|
|
UserProfile.EMAIL_ADDRESS_VISIBILITY_EVERYONE,
|
|
|
|
acting_user=None,
|
|
|
|
)
|
|
|
|
|
2020-06-29 13:11:26 +02:00
|
|
|
message = get_message(apply_markdown=False, client_gravatar=True)
|
|
|
|
self.assertEqual(message["display_recipient"], "Denmark")
|
|
|
|
self.assertEqual(message["content"], "**hello**")
|
|
|
|
self.assertEqual(message["avatar_url"], None)
|
|
|
|
|
|
|
|
message = get_message(apply_markdown=True, client_gravatar=True)
|
|
|
|
self.assertEqual(message["display_recipient"], "Denmark")
|
|
|
|
self.assertEqual(message["content"], "<p><strong>hello</strong></p>")
|
|
|
|
self.assertEqual(message["avatar_url"], None)
|
|
|
|
|
2022-09-17 21:55:15 +02:00
|
|
|
def test_bogus_queue_id(self) -> None:
|
|
|
|
user = self.example_user("hamlet")
|
|
|
|
|
|
|
|
with self.assertRaises(BadEventQueueIdError):
|
|
|
|
self.tornado_call(
|
|
|
|
get_events,
|
|
|
|
user,
|
|
|
|
{
|
|
|
|
"queue_id": "hamster",
|
|
|
|
"user_client": "website",
|
|
|
|
"last_event_id": -1,
|
|
|
|
"dont_block": orjson.dumps(True).decode(),
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
def test_wrong_user_queue_id(self) -> None:
|
|
|
|
user = self.example_user("hamlet")
|
|
|
|
wrong_user = self.example_user("othello")
|
|
|
|
|
|
|
|
result = self.tornado_call(
|
|
|
|
get_events,
|
|
|
|
user,
|
|
|
|
{
|
|
|
|
"apply_markdown": orjson.dumps(True).decode(),
|
|
|
|
"client_gravatar": orjson.dumps(True).decode(),
|
|
|
|
"event_types": orjson.dumps(["message"]).decode(),
|
|
|
|
"user_client": "website",
|
|
|
|
"dont_block": orjson.dumps(True).decode(),
|
|
|
|
},
|
|
|
|
)
|
|
|
|
self.assert_json_success(result)
|
|
|
|
queue_id = orjson.loads(result.content)["queue_id"]
|
|
|
|
|
|
|
|
with self.assertLogs(level="WARNING") as cm, self.assertRaises(BadEventQueueIdError):
|
|
|
|
self.tornado_call(
|
|
|
|
get_events,
|
|
|
|
wrong_user,
|
|
|
|
{
|
|
|
|
"queue_id": queue_id,
|
|
|
|
"user_client": "website",
|
|
|
|
"last_event_id": -1,
|
|
|
|
"dont_block": orjson.dumps(True).decode(),
|
|
|
|
},
|
|
|
|
)
|
|
|
|
self.assertIn("not authorized for queue", cm.output[0])
|
|
|
|
|
2022-10-27 19:05:10 +02:00
|
|
|
def test_get_events_custom_profile_fields(self) -> None:
|
|
|
|
user_profile = self.example_user("iago")
|
|
|
|
self.login_user(user_profile)
|
|
|
|
profile_field = CustomProfileField.objects.get(realm=user_profile.realm, name="Pronouns")
|
|
|
|
|
|
|
|
def check_pronouns_type_field_supported(
|
|
|
|
pronouns_field_type_supported: bool, new_name: str
|
|
|
|
) -> None:
|
|
|
|
clear_client_event_queues_for_testing()
|
|
|
|
|
|
|
|
queue_data = dict(
|
|
|
|
apply_markdown=True,
|
|
|
|
all_public_streams=True,
|
|
|
|
client_type_name="ZulipMobile",
|
|
|
|
event_types=["custom_profile_fields"],
|
|
|
|
last_connection_time=time.time(),
|
|
|
|
queue_timeout=0,
|
|
|
|
realm_id=user_profile.realm.id,
|
|
|
|
user_profile_id=user_profile.id,
|
|
|
|
pronouns_field_type_supported=pronouns_field_type_supported,
|
|
|
|
)
|
|
|
|
|
|
|
|
client = allocate_client_descriptor(queue_data)
|
|
|
|
|
|
|
|
try_update_realm_custom_profile_field(
|
|
|
|
realm=user_profile.realm, field=profile_field, name=new_name
|
|
|
|
)
|
|
|
|
result = self.tornado_call(
|
|
|
|
get_events,
|
|
|
|
user_profile,
|
|
|
|
{
|
|
|
|
"queue_id": client.event_queue.id,
|
|
|
|
"user_client": "ZulipAndroid",
|
|
|
|
"last_event_id": -1,
|
|
|
|
"dont_block": orjson.dumps(True).decode(),
|
|
|
|
},
|
|
|
|
)
|
|
|
|
events = orjson.loads(result.content)["events"]
|
|
|
|
self.assert_json_success(result)
|
|
|
|
self.assert_length(events, 1)
|
|
|
|
|
2023-07-22 00:34:11 +02:00
|
|
|
[pronouns_field] = (
|
2022-10-27 19:05:10 +02:00
|
|
|
field for field in events[0]["fields"] if field["id"] == profile_field.id
|
2023-07-22 00:34:11 +02:00
|
|
|
)
|
2022-10-27 19:05:10 +02:00
|
|
|
if pronouns_field_type_supported:
|
|
|
|
expected_type = CustomProfileField.PRONOUNS
|
|
|
|
else:
|
|
|
|
expected_type = CustomProfileField.SHORT_TEXT
|
|
|
|
self.assertEqual(pronouns_field["type"], expected_type)
|
|
|
|
|
|
|
|
check_pronouns_type_field_supported(False, "Pronouns field")
|
|
|
|
check_pronouns_type_field_supported(True, "Pronouns")
|
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
|
2020-06-29 13:11:26 +02:00
|
|
|
class FetchInitialStateDataTest(ZulipTestCase):
|
|
|
|
# Non-admin users don't have access to all bots
|
|
|
|
def test_realm_bots_non_admin(self) -> None:
|
2021-02-12 08:20:45 +01:00
|
|
|
user_profile = self.example_user("cordelia")
|
2020-06-29 13:11:26 +02:00
|
|
|
self.assertFalse(user_profile.is_realm_admin)
|
2021-01-17 17:58:50 +01:00
|
|
|
result = fetch_initial_state_data(user_profile)
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assert_length(result["realm_bots"], 0)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
# additionally the API key for a random bot is not present in the data
|
2021-03-08 11:39:48 +01:00
|
|
|
api_key = get_api_key(self.notification_bot(user_profile.realm))
|
2020-06-29 13:11:26 +02:00
|
|
|
self.assertNotIn(api_key, str(result))
|
|
|
|
|
|
|
|
# Admin users have access to all bots in the realm_bots field
|
|
|
|
def test_realm_bots_admin(self) -> None:
|
2021-02-12 08:20:45 +01:00
|
|
|
user_profile = self.example_user("hamlet")
|
2021-03-27 05:13:46 +01:00
|
|
|
do_change_user_role(user_profile, UserProfile.ROLE_REALM_ADMINISTRATOR, acting_user=None)
|
2020-06-29 13:11:26 +02:00
|
|
|
self.assertTrue(user_profile.is_realm_admin)
|
2021-01-17 17:58:50 +01:00
|
|
|
result = fetch_initial_state_data(user_profile)
|
2021-07-13 19:42:37 +02:00
|
|
|
self.assertGreater(len(result["realm_bots"]), 2)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
def test_max_message_id_with_no_history(self) -> None:
|
2021-02-12 08:20:45 +01:00
|
|
|
user_profile = self.example_user("aaron")
|
2020-06-29 13:11:26 +02:00
|
|
|
# Delete all historical messages for this user
|
|
|
|
UserMessage.objects.filter(user_profile=user_profile).delete()
|
2021-01-17 17:58:50 +01:00
|
|
|
result = fetch_initial_state_data(user_profile)
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertEqual(result["max_message_id"], -1)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
def test_delivery_email_presence_for_non_admins(self) -> None:
|
2021-02-12 08:20:45 +01:00
|
|
|
user_profile = self.example_user("aaron")
|
2021-10-26 09:15:16 +02:00
|
|
|
hamlet = self.example_user("hamlet")
|
2020-06-29 13:11:26 +02:00
|
|
|
self.assertFalse(user_profile.is_realm_admin)
|
2022-05-25 13:13:31 +02:00
|
|
|
hamlet = self.example_user("hamlet")
|
2020-06-29 13:11:26 +02:00
|
|
|
|
2021-10-26 09:15:16 +02:00
|
|
|
do_change_user_setting(
|
|
|
|
hamlet,
|
2021-03-01 11:33:24 +01:00
|
|
|
"email_address_visibility",
|
2021-10-26 09:15:16 +02:00
|
|
|
UserProfile.EMAIL_ADDRESS_VISIBILITY_EVERYONE,
|
2021-03-01 11:33:24 +01:00
|
|
|
acting_user=None,
|
2021-02-12 08:19:30 +01:00
|
|
|
)
|
2021-01-17 17:58:50 +01:00
|
|
|
result = fetch_initial_state_data(user_profile)
|
|
|
|
|
2022-05-25 13:13:31 +02:00
|
|
|
(hamlet_obj,) = (value for key, value in result["raw_users"].items() if key == hamlet.id)
|
|
|
|
self.assertEqual(hamlet_obj["delivery_email"], hamlet.delivery_email)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
2021-10-26 09:15:16 +02:00
|
|
|
do_change_user_setting(
|
|
|
|
hamlet,
|
2021-03-01 11:33:24 +01:00
|
|
|
"email_address_visibility",
|
2021-10-26 09:15:16 +02:00
|
|
|
UserProfile.EMAIL_ADDRESS_VISIBILITY_ADMINS,
|
2021-03-01 11:33:24 +01:00
|
|
|
acting_user=None,
|
2021-02-12 08:19:30 +01:00
|
|
|
)
|
2021-01-17 17:58:50 +01:00
|
|
|
result = fetch_initial_state_data(user_profile)
|
|
|
|
|
2022-05-25 13:13:31 +02:00
|
|
|
(hamlet_obj,) = (value for key, value in result["raw_users"].items() if key == hamlet.id)
|
|
|
|
self.assertIsNone(hamlet_obj["delivery_email"])
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
def test_delivery_email_presence_for_admins(self) -> None:
|
2021-02-12 08:20:45 +01:00
|
|
|
user_profile = self.example_user("iago")
|
2021-10-26 09:15:16 +02:00
|
|
|
hamlet = self.example_user("hamlet")
|
2020-06-29 13:11:26 +02:00
|
|
|
self.assertTrue(user_profile.is_realm_admin)
|
2022-05-25 13:13:31 +02:00
|
|
|
hamlet = self.example_user("hamlet")
|
2020-06-29 13:11:26 +02:00
|
|
|
|
2021-10-26 09:15:16 +02:00
|
|
|
do_change_user_setting(
|
|
|
|
hamlet,
|
2021-03-01 11:33:24 +01:00
|
|
|
"email_address_visibility",
|
2021-10-26 09:15:16 +02:00
|
|
|
UserProfile.EMAIL_ADDRESS_VISIBILITY_EVERYONE,
|
2021-03-01 11:33:24 +01:00
|
|
|
acting_user=None,
|
2021-02-12 08:19:30 +01:00
|
|
|
)
|
2021-01-17 17:58:50 +01:00
|
|
|
result = fetch_initial_state_data(user_profile)
|
2021-12-11 08:17:57 +01:00
|
|
|
|
2022-05-25 13:13:31 +02:00
|
|
|
(hamlet_obj,) = (value for key, value in result["raw_users"].items() if key == hamlet.id)
|
|
|
|
self.assertEqual(hamlet_obj["delivery_email"], hamlet.delivery_email)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
2021-10-26 09:15:16 +02:00
|
|
|
do_change_user_setting(
|
|
|
|
hamlet,
|
2021-03-01 11:33:24 +01:00
|
|
|
"email_address_visibility",
|
2021-10-26 09:15:16 +02:00
|
|
|
UserProfile.EMAIL_ADDRESS_VISIBILITY_ADMINS,
|
2021-03-01 11:33:24 +01:00
|
|
|
acting_user=None,
|
2021-02-12 08:19:30 +01:00
|
|
|
)
|
2021-01-17 17:58:50 +01:00
|
|
|
result = fetch_initial_state_data(user_profile)
|
2021-10-26 09:15:16 +02:00
|
|
|
(hamlet_obj,) = (value for key, value in result["raw_users"].items() if key == hamlet.id)
|
|
|
|
self.assertIn("delivery_email", hamlet_obj)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
def test_user_avatar_url_field_optional(self) -> None:
|
2021-02-12 08:20:45 +01:00
|
|
|
hamlet = self.example_user("hamlet")
|
2020-06-29 13:11:26 +02:00
|
|
|
users = [
|
2021-02-12 08:20:45 +01:00
|
|
|
self.example_user("iago"),
|
|
|
|
self.example_user("cordelia"),
|
|
|
|
self.example_user("ZOE"),
|
|
|
|
self.example_user("othello"),
|
2020-06-29 13:11:26 +02:00
|
|
|
]
|
|
|
|
|
|
|
|
for user in users:
|
|
|
|
user.long_term_idle = True
|
|
|
|
user.save()
|
|
|
|
|
|
|
|
long_term_idle_users_ids = [user.id for user in users]
|
|
|
|
|
2021-01-17 17:58:50 +01:00
|
|
|
result = fetch_initial_state_data(
|
|
|
|
user_profile=hamlet,
|
|
|
|
user_avatar_url_field_optional=True,
|
|
|
|
)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
2021-02-12 08:20:45 +01:00
|
|
|
raw_users = result["raw_users"]
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
for user_dict in raw_users.values():
|
2021-02-12 08:20:45 +01:00
|
|
|
if user_dict["user_id"] in long_term_idle_users_ids:
|
|
|
|
self.assertFalse("avatar_url" in user_dict)
|
2020-06-29 13:11:26 +02:00
|
|
|
else:
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertIsNotNone(user_dict["avatar_url"])
|
2020-06-29 13:11:26 +02:00
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
gravatar_users_id = [
|
2021-02-12 08:20:45 +01:00
|
|
|
user_dict["user_id"]
|
2021-02-12 08:19:30 +01:00
|
|
|
for user_dict in raw_users.values()
|
2023-01-18 18:40:44 +01:00
|
|
|
if "avatar_url" in user_dict
|
|
|
|
and urlsplit(user_dict["avatar_url"]).hostname == "secure.gravatar.com"
|
2021-02-12 08:19:30 +01:00
|
|
|
]
|
2020-06-29 13:11:26 +02:00
|
|
|
|
2023-03-01 07:34:25 +01:00
|
|
|
reset_email_visibility_to_everyone_in_zulip_realm()
|
2021-10-26 09:15:16 +02:00
|
|
|
|
2020-06-29 13:11:26 +02:00
|
|
|
# Test again with client_gravatar = True
|
2021-01-17 17:58:50 +01:00
|
|
|
result = fetch_initial_state_data(
|
|
|
|
user_profile=hamlet,
|
|
|
|
client_gravatar=True,
|
|
|
|
user_avatar_url_field_optional=True,
|
|
|
|
)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
2021-02-12 08:20:45 +01:00
|
|
|
raw_users = result["raw_users"]
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
for user_dict in raw_users.values():
|
2021-02-12 08:20:45 +01:00
|
|
|
if user_dict["user_id"] in gravatar_users_id:
|
|
|
|
self.assertIsNone(user_dict["avatar_url"])
|
2020-06-29 13:11:26 +02:00
|
|
|
else:
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertFalse("avatar_url" in user_dict)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
2021-07-30 17:44:42 +02:00
|
|
|
def test_user_settings_based_on_client_capabilities(self) -> None:
|
|
|
|
hamlet = self.example_user("hamlet")
|
|
|
|
result = fetch_initial_state_data(
|
|
|
|
user_profile=hamlet,
|
|
|
|
user_settings_object=True,
|
|
|
|
)
|
|
|
|
self.assertIn("user_settings", result)
|
|
|
|
for prop in UserProfile.property_types:
|
|
|
|
self.assertNotIn(prop, result)
|
|
|
|
self.assertIn(prop, result["user_settings"])
|
|
|
|
|
|
|
|
result = fetch_initial_state_data(
|
|
|
|
user_profile=hamlet,
|
|
|
|
user_settings_object=False,
|
|
|
|
)
|
|
|
|
self.assertIn("user_settings", result)
|
|
|
|
for prop in UserProfile.property_types:
|
2021-10-03 08:40:03 +02:00
|
|
|
if prop in {
|
|
|
|
**UserProfile.display_settings_legacy,
|
|
|
|
**UserProfile.notification_settings_legacy,
|
|
|
|
}:
|
|
|
|
# Only legacy settings are included in the top level.
|
|
|
|
self.assertIn(prop, result)
|
2021-07-30 17:44:42 +02:00
|
|
|
self.assertIn(prop, result["user_settings"])
|
|
|
|
|
linkifier: Support URL templates for linkifiers.
This swaps out url_format_string from all of our APIs and replaces it
with url_template. Note that the documentation changes in the following
commits will be squashed with this commit.
We change the "url_format" key to "url_template" for the
realm_linkifiers events in event_schema, along with updating
LinkifierDict. "url_template" is the name chosen to normalize
mixed usages of "url_format_string" and "url_format" throughout
the backend.
The markdown processor is updated to stop handling the format string
interpolation and delegate the task template expansion to the uri_template
library instead.
This change affects many test cases. We mostly just replace "%(name)s"
with "{name}", "url_format_string" with "url_template" to make sure that
they still pass. There are some test cases dedicated for testing "%"
escaping, which aren't relevant anymore and are subject to removal.
But for now we keep most of them as-is, and make sure that "%" is always
escaped since we do not use it for variable substitution any more.
Since url_format_string is not populated anymore, a migration is created
to remove this field entirely, and make url_template non-nullable since
we will always populate it. Note that it is possible to have
url_template being null after migration 0422 and before 0424, but
in practice, url_template will not be None after backfilling and the
backend now is always setting url_template.
With the removal of url_format_string, RealmFilter model will now be cleaned
with URL template checks, and the old checks for escapes are removed.
We also modified RealmFilter.clean to skip the validation when the
url_template is invalid. This avoids raising mulitple ValidationError's
when calling full_clean on a linkifier. But we might eventually want to
have a more centric approach to data validation instead of having
the same validation in both the clean method and the validator.
Fixes #23124.
Signed-off-by: Zixuan James Li <p359101898@gmail.com>
2022-10-05 20:55:31 +02:00
|
|
|
def test_realm_linkifiers_based_on_client_capabilities(self) -> None:
|
|
|
|
user = self.example_user("iago")
|
|
|
|
self.login_user(user)
|
|
|
|
|
|
|
|
data = {
|
|
|
|
"pattern": "#(?P<id>[123])",
|
|
|
|
"url_template": "https://realm.com/my_realm_filter/{id}",
|
|
|
|
}
|
|
|
|
post_result = self.client_post("/json/realm/filters", info=data)
|
|
|
|
self.assert_json_success(post_result)
|
|
|
|
|
|
|
|
result = fetch_initial_state_data(
|
|
|
|
user_profile=user,
|
|
|
|
linkifier_url_template=True,
|
|
|
|
)
|
|
|
|
self.assertEqual(result["realm_filters"], [])
|
2023-08-11 01:10:21 +02:00
|
|
|
self.assertEqual(result["realm_linkifiers"][-1]["pattern"], "#(?P<id>[123])")
|
linkifier: Support URL templates for linkifiers.
This swaps out url_format_string from all of our APIs and replaces it
with url_template. Note that the documentation changes in the following
commits will be squashed with this commit.
We change the "url_format" key to "url_template" for the
realm_linkifiers events in event_schema, along with updating
LinkifierDict. "url_template" is the name chosen to normalize
mixed usages of "url_format_string" and "url_format" throughout
the backend.
The markdown processor is updated to stop handling the format string
interpolation and delegate the task template expansion to the uri_template
library instead.
This change affects many test cases. We mostly just replace "%(name)s"
with "{name}", "url_format_string" with "url_template" to make sure that
they still pass. There are some test cases dedicated for testing "%"
escaping, which aren't relevant anymore and are subject to removal.
But for now we keep most of them as-is, and make sure that "%" is always
escaped since we do not use it for variable substitution any more.
Since url_format_string is not populated anymore, a migration is created
to remove this field entirely, and make url_template non-nullable since
we will always populate it. Note that it is possible to have
url_template being null after migration 0422 and before 0424, but
in practice, url_template will not be None after backfilling and the
backend now is always setting url_template.
With the removal of url_format_string, RealmFilter model will now be cleaned
with URL template checks, and the old checks for escapes are removed.
We also modified RealmFilter.clean to skip the validation when the
url_template is invalid. This avoids raising mulitple ValidationError's
when calling full_clean on a linkifier. But we might eventually want to
have a more centric approach to data validation instead of having
the same validation in both the clean method and the validator.
Fixes #23124.
Signed-off-by: Zixuan James Li <p359101898@gmail.com>
2022-10-05 20:55:31 +02:00
|
|
|
self.assertEqual(
|
2023-08-11 01:10:21 +02:00
|
|
|
result["realm_linkifiers"][-1]["url_template"],
|
linkifier: Support URL templates for linkifiers.
This swaps out url_format_string from all of our APIs and replaces it
with url_template. Note that the documentation changes in the following
commits will be squashed with this commit.
We change the "url_format" key to "url_template" for the
realm_linkifiers events in event_schema, along with updating
LinkifierDict. "url_template" is the name chosen to normalize
mixed usages of "url_format_string" and "url_format" throughout
the backend.
The markdown processor is updated to stop handling the format string
interpolation and delegate the task template expansion to the uri_template
library instead.
This change affects many test cases. We mostly just replace "%(name)s"
with "{name}", "url_format_string" with "url_template" to make sure that
they still pass. There are some test cases dedicated for testing "%"
escaping, which aren't relevant anymore and are subject to removal.
But for now we keep most of them as-is, and make sure that "%" is always
escaped since we do not use it for variable substitution any more.
Since url_format_string is not populated anymore, a migration is created
to remove this field entirely, and make url_template non-nullable since
we will always populate it. Note that it is possible to have
url_template being null after migration 0422 and before 0424, but
in practice, url_template will not be None after backfilling and the
backend now is always setting url_template.
With the removal of url_format_string, RealmFilter model will now be cleaned
with URL template checks, and the old checks for escapes are removed.
We also modified RealmFilter.clean to skip the validation when the
url_template is invalid. This avoids raising mulitple ValidationError's
when calling full_clean on a linkifier. But we might eventually want to
have a more centric approach to data validation instead of having
the same validation in both the clean method and the validator.
Fixes #23124.
Signed-off-by: Zixuan James Li <p359101898@gmail.com>
2022-10-05 20:55:31 +02:00
|
|
|
"https://realm.com/my_realm_filter/{id}",
|
|
|
|
)
|
|
|
|
|
|
|
|
# The default behavior should be `linkifier_url_template=False`
|
|
|
|
result = fetch_initial_state_data(
|
|
|
|
user_profile=user,
|
|
|
|
)
|
|
|
|
self.assertEqual(result["realm_filters"], [])
|
|
|
|
self.assertEqual(result["realm_linkifiers"], [])
|
|
|
|
|
2022-10-27 19:05:10 +02:00
|
|
|
def test_pronouns_field_type_support(self) -> None:
|
|
|
|
hamlet = self.example_user("hamlet")
|
|
|
|
result = fetch_initial_state_data(
|
|
|
|
user_profile=hamlet,
|
|
|
|
pronouns_field_type_supported=False,
|
|
|
|
)
|
|
|
|
self.assertIn("custom_profile_fields", result)
|
|
|
|
custom_profile_fields = result["custom_profile_fields"]
|
2023-07-22 00:34:11 +02:00
|
|
|
[pronouns_field] = (field for field in custom_profile_fields if field["name"] == "Pronouns")
|
2022-10-27 19:05:10 +02:00
|
|
|
self.assertEqual(pronouns_field["type"], CustomProfileField.SHORT_TEXT)
|
|
|
|
|
|
|
|
result = fetch_initial_state_data(
|
|
|
|
user_profile=hamlet,
|
|
|
|
pronouns_field_type_supported=True,
|
|
|
|
)
|
|
|
|
self.assertIn("custom_profile_fields", result)
|
|
|
|
custom_profile_fields = result["custom_profile_fields"]
|
2023-07-22 00:34:11 +02:00
|
|
|
[pronouns_field] = (field for field in custom_profile_fields if field["name"] == "Pronouns")
|
2022-10-27 19:05:10 +02:00
|
|
|
self.assertEqual(pronouns_field["type"], CustomProfileField.PRONOUNS)
|
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
|
2020-06-29 13:11:26 +02:00
|
|
|
class ClientDescriptorsTest(ZulipTestCase):
|
|
|
|
def test_get_client_info_for_all_public_streams(self) -> None:
|
2021-02-12 08:20:45 +01:00
|
|
|
hamlet = self.example_user("hamlet")
|
2020-06-29 13:11:26 +02:00
|
|
|
realm = hamlet.realm
|
|
|
|
|
|
|
|
queue_data = dict(
|
|
|
|
all_public_streams=True,
|
|
|
|
apply_markdown=True,
|
|
|
|
client_gravatar=True,
|
2021-02-12 08:20:45 +01:00
|
|
|
client_type_name="website",
|
|
|
|
event_types=["message"],
|
2020-06-29 13:11:26 +02:00
|
|
|
last_connection_time=time.time(),
|
|
|
|
queue_timeout=0,
|
|
|
|
realm_id=realm.id,
|
|
|
|
user_profile_id=hamlet.id,
|
|
|
|
)
|
|
|
|
|
|
|
|
client = allocate_client_descriptor(queue_data)
|
|
|
|
|
|
|
|
message_event = dict(
|
|
|
|
realm_id=realm.id,
|
2021-02-12 08:20:45 +01:00
|
|
|
stream_name="whatever",
|
2020-06-29 13:11:26 +02:00
|
|
|
)
|
|
|
|
|
|
|
|
client_info = get_client_info_for_message_event(
|
|
|
|
message_event,
|
|
|
|
users=[],
|
|
|
|
)
|
|
|
|
|
2021-05-17 05:41:32 +02:00
|
|
|
self.assert_length(client_info, 1)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
dct = client_info[client.event_queue.id]
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertEqual(dct["client"].apply_markdown, True)
|
|
|
|
self.assertEqual(dct["client"].client_gravatar, True)
|
|
|
|
self.assertEqual(dct["client"].user_profile_id, hamlet.id)
|
|
|
|
self.assertEqual(dct["flags"], [])
|
|
|
|
self.assertEqual(dct["is_sender"], False)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
message_event = dict(
|
|
|
|
realm_id=realm.id,
|
2021-02-12 08:20:45 +01:00
|
|
|
stream_name="whatever",
|
2020-06-29 13:11:26 +02:00
|
|
|
sender_queue_id=client.event_queue.id,
|
|
|
|
)
|
|
|
|
|
|
|
|
client_info = get_client_info_for_message_event(
|
|
|
|
message_event,
|
|
|
|
users=[],
|
|
|
|
)
|
|
|
|
dct = client_info[client.event_queue.id]
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertEqual(dct["is_sender"], True)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
def test_get_client_info_for_normal_users(self) -> None:
|
2021-02-12 08:20:45 +01:00
|
|
|
hamlet = self.example_user("hamlet")
|
|
|
|
cordelia = self.example_user("cordelia")
|
2020-06-29 13:11:26 +02:00
|
|
|
realm = hamlet.realm
|
|
|
|
|
|
|
|
def test_get_info(apply_markdown: bool, client_gravatar: bool) -> None:
|
|
|
|
clear_client_event_queues_for_testing()
|
|
|
|
|
|
|
|
queue_data = dict(
|
|
|
|
all_public_streams=False,
|
|
|
|
apply_markdown=apply_markdown,
|
|
|
|
client_gravatar=client_gravatar,
|
2021-02-12 08:20:45 +01:00
|
|
|
client_type_name="website",
|
|
|
|
event_types=["message"],
|
2020-06-29 13:11:26 +02:00
|
|
|
last_connection_time=time.time(),
|
|
|
|
queue_timeout=0,
|
|
|
|
realm_id=realm.id,
|
|
|
|
user_profile_id=hamlet.id,
|
|
|
|
)
|
|
|
|
|
|
|
|
client = allocate_client_descriptor(queue_data)
|
|
|
|
message_event = dict(
|
|
|
|
realm_id=realm.id,
|
2021-02-12 08:20:45 +01:00
|
|
|
stream_name="whatever",
|
2020-06-29 13:11:26 +02:00
|
|
|
)
|
|
|
|
|
|
|
|
client_info = get_client_info_for_message_event(
|
|
|
|
message_event,
|
|
|
|
users=[
|
|
|
|
dict(id=cordelia.id),
|
|
|
|
],
|
|
|
|
)
|
|
|
|
|
2021-05-17 05:41:32 +02:00
|
|
|
self.assert_length(client_info, 0)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
client_info = get_client_info_for_message_event(
|
|
|
|
message_event,
|
|
|
|
users=[
|
|
|
|
dict(id=cordelia.id),
|
2021-02-12 08:20:45 +01:00
|
|
|
dict(id=hamlet.id, flags=["mentioned"]),
|
2020-06-29 13:11:26 +02:00
|
|
|
],
|
|
|
|
)
|
2021-05-17 05:41:32 +02:00
|
|
|
self.assert_length(client_info, 1)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
dct = client_info[client.event_queue.id]
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertEqual(dct["client"].apply_markdown, apply_markdown)
|
|
|
|
self.assertEqual(dct["client"].client_gravatar, client_gravatar)
|
|
|
|
self.assertEqual(dct["client"].user_profile_id, hamlet.id)
|
|
|
|
self.assertEqual(dct["flags"], ["mentioned"])
|
|
|
|
self.assertEqual(dct["is_sender"], False)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
test_get_info(apply_markdown=False, client_gravatar=False)
|
|
|
|
test_get_info(apply_markdown=True, client_gravatar=False)
|
|
|
|
|
|
|
|
test_get_info(apply_markdown=False, client_gravatar=True)
|
|
|
|
test_get_info(apply_markdown=True, client_gravatar=True)
|
|
|
|
|
|
|
|
def test_process_message_event_with_mocked_client_info(self) -> None:
|
|
|
|
hamlet = self.example_user("hamlet")
|
|
|
|
|
|
|
|
class MockClient:
|
2021-02-12 08:19:30 +01:00
|
|
|
def __init__(
|
|
|
|
self, user_profile_id: int, apply_markdown: bool, client_gravatar: bool
|
|
|
|
) -> None:
|
2020-06-29 13:11:26 +02:00
|
|
|
self.user_profile_id = user_profile_id
|
|
|
|
self.apply_markdown = apply_markdown
|
|
|
|
self.client_gravatar = client_gravatar
|
2021-02-12 08:20:45 +01:00
|
|
|
self.client_type_name = "whatever"
|
2020-06-29 13:11:26 +02:00
|
|
|
self.events: List[Dict[str, Any]] = []
|
|
|
|
|
|
|
|
def accepts_messages(self) -> bool:
|
|
|
|
return True
|
|
|
|
|
|
|
|
def accepts_event(self, event: Dict[str, Any]) -> bool:
|
2021-02-12 08:20:45 +01:00
|
|
|
assert event["type"] == "message"
|
2020-06-29 13:11:26 +02:00
|
|
|
return True
|
|
|
|
|
|
|
|
def add_event(self, event: Dict[str, Any]) -> None:
|
|
|
|
self.events.append(event)
|
|
|
|
|
|
|
|
client1 = MockClient(
|
|
|
|
user_profile_id=hamlet.id,
|
|
|
|
apply_markdown=True,
|
|
|
|
client_gravatar=False,
|
|
|
|
)
|
|
|
|
|
|
|
|
client2 = MockClient(
|
|
|
|
user_profile_id=hamlet.id,
|
|
|
|
apply_markdown=False,
|
|
|
|
client_gravatar=False,
|
|
|
|
)
|
|
|
|
|
|
|
|
client3 = MockClient(
|
|
|
|
user_profile_id=hamlet.id,
|
|
|
|
apply_markdown=True,
|
|
|
|
client_gravatar=True,
|
|
|
|
)
|
|
|
|
|
|
|
|
client4 = MockClient(
|
|
|
|
user_profile_id=hamlet.id,
|
|
|
|
apply_markdown=False,
|
|
|
|
client_gravatar=True,
|
|
|
|
)
|
|
|
|
|
|
|
|
client_info = {
|
2021-02-12 08:20:45 +01:00
|
|
|
"client:1": dict(
|
2020-06-29 13:11:26 +02:00
|
|
|
client=client1,
|
2021-02-12 08:20:45 +01:00
|
|
|
flags=["starred"],
|
2020-06-29 13:11:26 +02:00
|
|
|
),
|
2021-02-12 08:20:45 +01:00
|
|
|
"client:2": dict(
|
2020-06-29 13:11:26 +02:00
|
|
|
client=client2,
|
2021-02-12 08:20:45 +01:00
|
|
|
flags=["has_alert_word"],
|
2020-06-29 13:11:26 +02:00
|
|
|
),
|
2021-02-12 08:20:45 +01:00
|
|
|
"client:3": dict(
|
2020-06-29 13:11:26 +02:00
|
|
|
client=client3,
|
|
|
|
flags=[],
|
|
|
|
),
|
2021-02-12 08:20:45 +01:00
|
|
|
"client:4": dict(
|
2020-06-29 13:11:26 +02:00
|
|
|
client=client4,
|
|
|
|
flags=[],
|
|
|
|
),
|
|
|
|
}
|
|
|
|
|
|
|
|
sender = hamlet
|
|
|
|
|
|
|
|
message_event = dict(
|
|
|
|
message_dict=dict(
|
|
|
|
id=999,
|
2021-02-12 08:20:45 +01:00
|
|
|
content="**hello**",
|
|
|
|
rendered_content="<b>hello</b>",
|
2020-06-29 13:11:26 +02:00
|
|
|
sender_id=sender.id,
|
2021-02-12 08:20:45 +01:00
|
|
|
type="stream",
|
|
|
|
client="website",
|
2020-06-29 13:11:26 +02:00
|
|
|
# NOTE: Some of these fields are clutter, but some
|
|
|
|
# will be useful when we let clients specify
|
|
|
|
# that they can compute their own gravatar URLs.
|
|
|
|
sender_email=sender.email,
|
|
|
|
sender_delivery_email=sender.delivery_email,
|
|
|
|
sender_realm_id=sender.realm_id,
|
|
|
|
sender_avatar_source=UserProfile.AVATAR_FROM_GRAVATAR,
|
|
|
|
sender_avatar_version=1,
|
|
|
|
sender_is_mirror_dummy=None,
|
2021-10-26 09:15:16 +02:00
|
|
|
sender_email_address_visibility=UserProfile.EMAIL_ADDRESS_VISIBILITY_EVERYONE,
|
2020-06-29 13:11:26 +02:00
|
|
|
recipient_type=None,
|
|
|
|
recipient_type_id=None,
|
|
|
|
),
|
|
|
|
)
|
|
|
|
|
|
|
|
# Setting users to `[]` bypasses code we don't care about
|
|
|
|
# for this test--we assume client_info is correct in our mocks,
|
|
|
|
# and we are interested in how messages are put on event queue.
|
|
|
|
users: List[Dict[str, Any]] = []
|
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
with mock.patch(
|
2021-02-12 08:20:45 +01:00
|
|
|
"zerver.tornado.event_queue.get_client_info_for_message_event", return_value=client_info
|
2021-02-12 08:19:30 +01:00
|
|
|
):
|
2020-06-29 13:11:26 +02:00
|
|
|
process_message_event(message_event, users)
|
|
|
|
|
|
|
|
# We are not closely examining avatar_url at this point, so
|
|
|
|
# just sanity check them and then delete the keys so that
|
|
|
|
# upcoming comparisons work.
|
|
|
|
for client in [client1, client2]:
|
2021-02-12 08:20:45 +01:00
|
|
|
message = client.events[0]["message"]
|
|
|
|
self.assertIn("gravatar.com", message["avatar_url"])
|
|
|
|
message.pop("avatar_url")
|
2020-06-29 13:11:26 +02:00
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
self.assertEqual(
|
|
|
|
client1.events,
|
|
|
|
[
|
|
|
|
dict(
|
2021-02-12 08:20:45 +01:00
|
|
|
type="message",
|
2021-02-12 08:19:30 +01:00
|
|
|
message=dict(
|
2021-02-12 08:20:45 +01:00
|
|
|
type="stream",
|
2021-02-12 08:19:30 +01:00
|
|
|
sender_id=sender.id,
|
|
|
|
sender_email=sender.email,
|
|
|
|
id=999,
|
2021-02-12 08:20:45 +01:00
|
|
|
content="<b>hello</b>",
|
|
|
|
content_type="text/html",
|
|
|
|
client="website",
|
2021-02-12 08:19:30 +01:00
|
|
|
),
|
2021-02-12 08:20:45 +01:00
|
|
|
flags=["starred"],
|
2020-06-29 13:11:26 +02:00
|
|
|
),
|
2021-02-12 08:19:30 +01:00
|
|
|
],
|
|
|
|
)
|
|
|
|
|
|
|
|
self.assertEqual(
|
|
|
|
client2.events,
|
|
|
|
[
|
|
|
|
dict(
|
2021-02-12 08:20:45 +01:00
|
|
|
type="message",
|
2021-02-12 08:19:30 +01:00
|
|
|
message=dict(
|
2021-02-12 08:20:45 +01:00
|
|
|
type="stream",
|
2021-02-12 08:19:30 +01:00
|
|
|
sender_id=sender.id,
|
|
|
|
sender_email=sender.email,
|
|
|
|
id=999,
|
2021-02-12 08:20:45 +01:00
|
|
|
content="**hello**",
|
|
|
|
content_type="text/x-markdown",
|
|
|
|
client="website",
|
2021-02-12 08:19:30 +01:00
|
|
|
),
|
2021-02-12 08:20:45 +01:00
|
|
|
flags=["has_alert_word"],
|
2020-06-29 13:11:26 +02:00
|
|
|
),
|
2021-02-12 08:19:30 +01:00
|
|
|
],
|
|
|
|
)
|
|
|
|
|
|
|
|
self.assertEqual(
|
|
|
|
client3.events,
|
|
|
|
[
|
|
|
|
dict(
|
2021-02-12 08:20:45 +01:00
|
|
|
type="message",
|
2021-02-12 08:19:30 +01:00
|
|
|
message=dict(
|
2021-02-12 08:20:45 +01:00
|
|
|
type="stream",
|
2021-02-12 08:19:30 +01:00
|
|
|
sender_id=sender.id,
|
|
|
|
sender_email=sender.email,
|
|
|
|
avatar_url=None,
|
|
|
|
id=999,
|
2021-02-12 08:20:45 +01:00
|
|
|
content="<b>hello</b>",
|
|
|
|
content_type="text/html",
|
|
|
|
client="website",
|
2021-02-12 08:19:30 +01:00
|
|
|
),
|
|
|
|
flags=[],
|
2020-06-29 13:11:26 +02:00
|
|
|
),
|
2021-02-12 08:19:30 +01:00
|
|
|
],
|
|
|
|
)
|
|
|
|
|
|
|
|
self.assertEqual(
|
|
|
|
client4.events,
|
|
|
|
[
|
|
|
|
dict(
|
2021-02-12 08:20:45 +01:00
|
|
|
type="message",
|
2021-02-12 08:19:30 +01:00
|
|
|
message=dict(
|
2021-02-12 08:20:45 +01:00
|
|
|
type="stream",
|
2021-02-12 08:19:30 +01:00
|
|
|
sender_id=sender.id,
|
|
|
|
sender_email=sender.email,
|
|
|
|
avatar_url=None,
|
|
|
|
id=999,
|
2021-02-12 08:20:45 +01:00
|
|
|
content="**hello**",
|
|
|
|
content_type="text/x-markdown",
|
|
|
|
client="website",
|
2021-02-12 08:19:30 +01:00
|
|
|
),
|
|
|
|
flags=[],
|
2020-06-29 13:11:26 +02:00
|
|
|
),
|
2021-02-12 08:19:30 +01:00
|
|
|
],
|
|
|
|
)
|
|
|
|
|
2020-06-29 13:11:26 +02:00
|
|
|
|
2024-02-14 20:27:17 +01:00
|
|
|
class ReloadWebClientsTest(ZulipTestCase):
|
2024-02-10 04:19:08 +01:00
|
|
|
def test_web_reload_clients(self) -> None:
|
2021-03-21 15:12:24 +01:00
|
|
|
hamlet = self.example_user("hamlet")
|
|
|
|
realm = hamlet.realm
|
|
|
|
|
|
|
|
clear_client_event_queues_for_testing()
|
|
|
|
|
|
|
|
queue_data = dict(
|
|
|
|
all_public_streams=False,
|
|
|
|
apply_markdown=True,
|
|
|
|
client_gravatar=True,
|
|
|
|
client_type_name="website",
|
|
|
|
event_types=None,
|
|
|
|
last_connection_time=time.time(),
|
|
|
|
queue_timeout=0,
|
|
|
|
realm_id=realm.id,
|
|
|
|
user_profile_id=hamlet.id,
|
|
|
|
)
|
|
|
|
client = allocate_client_descriptor(queue_data)
|
|
|
|
|
2024-02-10 04:19:08 +01:00
|
|
|
send_web_reload_client_events()
|
2024-02-07 22:03:15 +01:00
|
|
|
self.assert_length(client.event_queue.queue, 0)
|
2021-03-21 15:12:24 +01:00
|
|
|
|
2024-02-07 22:03:15 +01:00
|
|
|
mark_clients_to_reload([client.event_queue.id])
|
|
|
|
send_web_reload_client_events()
|
2024-02-09 21:29:01 +01:00
|
|
|
self.assert_length(client.event_queue.queue, 1)
|
2024-02-10 04:19:08 +01:00
|
|
|
reload_event = client.event_queue.queue[0]
|
2021-03-21 15:12:24 +01:00
|
|
|
|
2024-02-10 04:19:08 +01:00
|
|
|
check_web_reload_client_event("web_reload_client_event", reload_event)
|
2021-03-21 15:12:24 +01:00
|
|
|
self.assertEqual(
|
2024-02-10 04:19:08 +01:00
|
|
|
reload_event,
|
2021-03-21 15:12:24 +01:00
|
|
|
dict(
|
2024-02-10 04:19:08 +01:00
|
|
|
type="web_reload_client",
|
|
|
|
immediate=False,
|
2021-03-21 15:12:24 +01:00
|
|
|
id=0,
|
|
|
|
),
|
|
|
|
)
|
|
|
|
|
|
|
|
|
2020-06-29 13:11:26 +02:00
|
|
|
class FetchQueriesTest(ZulipTestCase):
|
|
|
|
def test_queries(self) -> None:
|
|
|
|
user = self.example_user("hamlet")
|
|
|
|
|
|
|
|
self.login_user(user)
|
|
|
|
|
2024-01-26 14:45:37 +01:00
|
|
|
with self.assert_database_query_count(41):
|
2021-02-12 08:20:45 +01:00
|
|
|
with mock.patch("zerver.lib.events.always_want") as want_mock:
|
2021-01-17 17:58:50 +01:00
|
|
|
fetch_initial_state_data(user)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
expected_counts = dict(
|
|
|
|
alert_words=1,
|
|
|
|
custom_profile_fields=1,
|
|
|
|
default_streams=1,
|
|
|
|
default_stream_groups=1,
|
2021-07-24 06:56:56 +02:00
|
|
|
drafts=1,
|
2020-06-29 13:11:26 +02:00
|
|
|
message=1,
|
|
|
|
muted_topics=1,
|
2021-03-27 12:23:32 +01:00
|
|
|
muted_users=1,
|
2023-12-02 11:30:35 +01:00
|
|
|
onboarding_steps=1,
|
2020-06-29 13:11:26 +02:00
|
|
|
presence=1,
|
2023-04-17 22:14:14 +02:00
|
|
|
realm=1,
|
2020-06-29 13:11:26 +02:00
|
|
|
realm_bot=1,
|
|
|
|
realm_domains=1,
|
|
|
|
realm_embedded_bots=0,
|
|
|
|
realm_incoming_webhook_bots=0,
|
|
|
|
realm_emoji=1,
|
linkifier: Support URL templates for linkifiers.
This swaps out url_format_string from all of our APIs and replaces it
with url_template. Note that the documentation changes in the following
commits will be squashed with this commit.
We change the "url_format" key to "url_template" for the
realm_linkifiers events in event_schema, along with updating
LinkifierDict. "url_template" is the name chosen to normalize
mixed usages of "url_format_string" and "url_format" throughout
the backend.
The markdown processor is updated to stop handling the format string
interpolation and delegate the task template expansion to the uri_template
library instead.
This change affects many test cases. We mostly just replace "%(name)s"
with "{name}", "url_format_string" with "url_template" to make sure that
they still pass. There are some test cases dedicated for testing "%"
escaping, which aren't relevant anymore and are subject to removal.
But for now we keep most of them as-is, and make sure that "%" is always
escaped since we do not use it for variable substitution any more.
Since url_format_string is not populated anymore, a migration is created
to remove this field entirely, and make url_template non-nullable since
we will always populate it. Note that it is possible to have
url_template being null after migration 0422 and before 0424, but
in practice, url_template will not be None after backfilling and the
backend now is always setting url_template.
With the removal of url_format_string, RealmFilter model will now be cleaned
with URL template checks, and the old checks for escapes are removed.
We also modified RealmFilter.clean to skip the validation when the
url_template is invalid. This avoids raising mulitple ValidationError's
when calling full_clean on a linkifier. But we might eventually want to
have a more centric approach to data validation instead of having
the same validation in both the clean method and the validator.
Fixes #23124.
Signed-off-by: Zixuan James Li <p359101898@gmail.com>
2022-10-05 20:55:31 +02:00
|
|
|
realm_filters=0,
|
|
|
|
realm_linkifiers=0,
|
2020-10-28 04:00:46 +01:00
|
|
|
realm_playgrounds=1,
|
2020-06-29 13:11:26 +02:00
|
|
|
realm_user=3,
|
2022-02-28 11:50:33 +01:00
|
|
|
realm_user_groups=3,
|
2021-07-21 13:40:46 +02:00
|
|
|
realm_user_settings_defaults=1,
|
2020-06-29 13:11:26 +02:00
|
|
|
recent_private_conversations=1,
|
2023-04-20 04:40:41 +02:00
|
|
|
scheduled_messages=1,
|
2020-06-29 13:11:26 +02:00
|
|
|
starred_messages=1,
|
2023-07-27 16:42:21 +02:00
|
|
|
stream=3,
|
2020-06-29 13:11:26 +02:00
|
|
|
stop_words=0,
|
2020-10-18 14:51:25 +02:00
|
|
|
subscription=4,
|
2020-06-29 13:11:26 +02:00
|
|
|
update_display_settings=0,
|
|
|
|
update_global_notifications=0,
|
|
|
|
update_message_flags=5,
|
2021-07-26 19:55:14 +02:00
|
|
|
user_settings=0,
|
2020-06-29 13:11:26 +02:00
|
|
|
user_status=1,
|
2022-02-25 21:48:56 +01:00
|
|
|
user_topic=1,
|
2020-06-29 13:11:26 +02:00
|
|
|
video_calls=0,
|
2021-03-19 13:21:18 +01:00
|
|
|
giphy=0,
|
2020-06-29 13:11:26 +02:00
|
|
|
)
|
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
wanted_event_types = {item[0][0] for item in want_mock.call_args_list}
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
self.assertEqual(wanted_event_types, set(expected_counts))
|
|
|
|
|
|
|
|
for event_type in sorted(wanted_event_types):
|
|
|
|
count = expected_counts[event_type]
|
2022-10-15 22:47:40 +02:00
|
|
|
with self.assert_database_query_count(count):
|
2021-02-12 08:20:45 +01:00
|
|
|
if event_type == "update_message_flags":
|
|
|
|
event_types = ["update_message_flags", "message"]
|
2020-06-29 13:11:26 +02:00
|
|
|
else:
|
|
|
|
event_types = [event_type]
|
|
|
|
|
2021-01-17 17:58:50 +01:00
|
|
|
fetch_initial_state_data(user, event_types=event_types)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
|
|
|
|
class TestEventsRegisterAllPublicStreamsDefaults(ZulipTestCase):
|
2023-10-12 19:43:45 +02:00
|
|
|
@override
|
2020-06-29 13:11:26 +02:00
|
|
|
def setUp(self) -> None:
|
|
|
|
super().setUp()
|
2021-02-12 08:20:45 +01:00
|
|
|
self.user_profile = self.example_user("hamlet")
|
2020-06-29 13:11:26 +02:00
|
|
|
self.email = self.user_profile.email
|
|
|
|
|
|
|
|
def test_use_passed_all_public_true_default_false(self) -> None:
|
|
|
|
self.user_profile.default_all_public_streams = False
|
|
|
|
self.user_profile.save()
|
|
|
|
result = _default_all_public_streams(self.user_profile, True)
|
|
|
|
self.assertTrue(result)
|
|
|
|
|
|
|
|
def test_use_passed_all_public_true_default(self) -> None:
|
|
|
|
self.user_profile.default_all_public_streams = True
|
|
|
|
self.user_profile.save()
|
|
|
|
result = _default_all_public_streams(self.user_profile, True)
|
|
|
|
self.assertTrue(result)
|
|
|
|
|
|
|
|
def test_use_passed_all_public_false_default_false(self) -> None:
|
|
|
|
self.user_profile.default_all_public_streams = False
|
|
|
|
self.user_profile.save()
|
|
|
|
result = _default_all_public_streams(self.user_profile, False)
|
|
|
|
self.assertFalse(result)
|
|
|
|
|
|
|
|
def test_use_passed_all_public_false_default_true(self) -> None:
|
|
|
|
self.user_profile.default_all_public_streams = True
|
|
|
|
self.user_profile.save()
|
|
|
|
result = _default_all_public_streams(self.user_profile, False)
|
|
|
|
self.assertFalse(result)
|
|
|
|
|
|
|
|
def test_use_true_default_for_none(self) -> None:
|
|
|
|
self.user_profile.default_all_public_streams = True
|
|
|
|
self.user_profile.save()
|
|
|
|
result = _default_all_public_streams(self.user_profile, None)
|
|
|
|
self.assertTrue(result)
|
|
|
|
|
|
|
|
def test_use_false_default_for_none(self) -> None:
|
|
|
|
self.user_profile.default_all_public_streams = False
|
|
|
|
self.user_profile.save()
|
|
|
|
result = _default_all_public_streams(self.user_profile, None)
|
|
|
|
self.assertFalse(result)
|
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
|
2020-06-29 13:11:26 +02:00
|
|
|
class TestEventsRegisterNarrowDefaults(ZulipTestCase):
|
2023-10-12 19:43:45 +02:00
|
|
|
@override
|
2020-06-29 13:11:26 +02:00
|
|
|
def setUp(self) -> None:
|
|
|
|
super().setUp()
|
2021-02-12 08:20:45 +01:00
|
|
|
self.user_profile = self.example_user("hamlet")
|
2020-06-29 13:11:26 +02:00
|
|
|
self.email = self.user_profile.email
|
2021-02-12 08:20:45 +01:00
|
|
|
self.stream = get_stream("Verona", self.user_profile.realm)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
def test_use_passed_narrow_no_default(self) -> None:
|
|
|
|
self.user_profile.default_events_register_stream_id = None
|
|
|
|
self.user_profile.save()
|
2021-02-12 08:20:45 +01:00
|
|
|
result = _default_narrow(self.user_profile, [["stream", "my_stream"]])
|
|
|
|
self.assertEqual(result, [["stream", "my_stream"]])
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
def test_use_passed_narrow_with_default(self) -> None:
|
|
|
|
self.user_profile.default_events_register_stream_id = self.stream.id
|
|
|
|
self.user_profile.save()
|
2021-02-12 08:20:45 +01:00
|
|
|
result = _default_narrow(self.user_profile, [["stream", "my_stream"]])
|
|
|
|
self.assertEqual(result, [["stream", "my_stream"]])
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
def test_use_default_if_narrow_is_empty(self) -> None:
|
|
|
|
self.user_profile.default_events_register_stream_id = self.stream.id
|
|
|
|
self.user_profile.save()
|
|
|
|
result = _default_narrow(self.user_profile, [])
|
2021-02-12 08:20:45 +01:00
|
|
|
self.assertEqual(result, [["stream", "Verona"]])
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
def test_use_narrow_if_default_is_none(self) -> None:
|
|
|
|
self.user_profile.default_events_register_stream_id = None
|
|
|
|
self.user_profile.save()
|
|
|
|
result = _default_narrow(self.user_profile, [])
|
|
|
|
self.assertEqual(result, [])
|
|
|
|
|
2021-02-12 08:19:30 +01:00
|
|
|
|
2023-11-08 08:21:24 +01:00
|
|
|
class TestGetUserAPIDataSystemBotRealm(ZulipTestCase):
|
|
|
|
def test_get_users_api_data_on_system_bot_realm(self) -> None:
|
2021-03-08 11:54:39 +01:00
|
|
|
realm = get_realm(settings.SYSTEM_BOT_REALM)
|
2023-11-08 08:21:24 +01:00
|
|
|
result = get_users_for_api(
|
2021-03-08 11:54:39 +01:00
|
|
|
realm,
|
2021-02-12 08:20:45 +01:00
|
|
|
self.example_user("hamlet"),
|
2021-02-12 08:19:30 +01:00
|
|
|
client_gravatar=True,
|
|
|
|
user_avatar_url_field_optional=True,
|
|
|
|
)
|
2020-06-29 13:11:26 +02:00
|
|
|
|
|
|
|
for bot_email in settings.CROSS_REALM_BOT_EMAILS:
|
2021-03-08 11:54:39 +01:00
|
|
|
bot_profile = get_system_bot(bot_email, realm.id)
|
2020-06-29 13:11:26 +02:00
|
|
|
self.assertTrue(bot_profile.id in result)
|
2021-03-07 15:51:55 +01:00
|
|
|
self.assertTrue(result[bot_profile.id]["is_system_bot"])
|
2021-05-05 13:15:13 +02:00
|
|
|
|
|
|
|
|
|
|
|
class TestUserPresenceUpdatesDisabled(ZulipTestCase):
|
2022-08-24 21:42:44 +02:00
|
|
|
# For this test, we verify do_update_user_presence doesn't send
|
|
|
|
# events for organizations with more than
|
|
|
|
# USER_LIMIT_FOR_SENDING_PRESENCE_UPDATE_EVENTS users, unless
|
|
|
|
# force_send_update is passed.
|
|
|
|
@override_settings(USER_LIMIT_FOR_SENDING_PRESENCE_UPDATE_EVENTS=3)
|
2022-02-08 00:13:33 +01:00
|
|
|
def test_presence_events_disabled_on_larger_realm(self) -> None:
|
2023-04-05 13:36:01 +02:00
|
|
|
with self.capture_send_event_calls(expected_num_events=1):
|
2021-05-05 13:15:13 +02:00
|
|
|
do_update_user_presence(
|
|
|
|
self.example_user("cordelia"),
|
|
|
|
get_client("website"),
|
|
|
|
timezone_now(),
|
2020-06-11 16:03:47 +02:00
|
|
|
UserPresence.LEGACY_STATUS_ACTIVE_INT,
|
2022-08-24 21:42:44 +02:00
|
|
|
force_send_update=True,
|
2021-05-05 13:15:13 +02:00
|
|
|
)
|
|
|
|
|
2023-04-05 13:36:01 +02:00
|
|
|
with self.capture_send_event_calls(expected_num_events=0):
|
2022-08-24 21:42:44 +02:00
|
|
|
do_update_user_presence(
|
|
|
|
self.example_user("hamlet"),
|
|
|
|
get_client("website"),
|
|
|
|
timezone_now(),
|
2020-06-11 16:03:47 +02:00
|
|
|
UserPresence.LEGACY_STATUS_ACTIVE_INT,
|
2022-08-24 21:42:44 +02:00
|
|
|
force_send_update=False,
|
|
|
|
)
|