2022-08-03 20:06:59 +02:00
|
|
|
import logging
|
2022-07-27 23:33:49 +02:00
|
|
|
from email.headerregistry import Address
|
2022-10-12 09:14:27 +02:00
|
|
|
from typing import Any, Dict, Literal, Optional, Tuple, Union
|
2022-04-14 23:57:15 +02:00
|
|
|
|
|
|
|
import orjson
|
|
|
|
from django.conf import settings
|
|
|
|
from django.db import transaction
|
2023-03-04 01:52:14 +01:00
|
|
|
from django.db.models import QuerySet
|
2022-04-14 23:57:15 +02:00
|
|
|
from django.utils.timezone import now as timezone_now
|
|
|
|
|
|
|
|
from confirmation.models import Confirmation, create_confirmation_link, generate_key
|
|
|
|
from zerver.actions.custom_profile_fields import do_remove_realm_custom_profile_fields
|
2022-07-17 13:00:21 +02:00
|
|
|
from zerver.actions.message_delete import do_delete_messages_by_sender
|
2022-04-14 23:57:15 +02:00
|
|
|
from zerver.actions.user_groups import update_users_in_full_members_system_group
|
2021-10-27 20:39:45 +02:00
|
|
|
from zerver.actions.user_settings import do_delete_avatar_image
|
2022-12-23 14:56:27 +01:00
|
|
|
from zerver.lib.message import parse_message_time_limit_setting, update_first_visible_message_id
|
2020-05-13 18:09:17 +02:00
|
|
|
from zerver.lib.retention import move_messages_to_archive
|
2022-04-14 23:57:15 +02:00
|
|
|
from zerver.lib.send_email import FromAddress, send_email_to_admins
|
|
|
|
from zerver.lib.sessions import delete_user_sessions
|
2023-02-28 04:44:29 +01:00
|
|
|
from zerver.lib.upload import delete_message_attachments
|
2022-04-14 23:57:15 +02:00
|
|
|
from zerver.lib.user_counts import realm_user_count_by_role
|
|
|
|
from zerver.models import (
|
2023-02-28 03:35:01 +01:00
|
|
|
ArchivedAttachment,
|
2022-04-14 23:57:15 +02:00
|
|
|
Attachment,
|
2020-05-13 18:09:17 +02:00
|
|
|
Message,
|
2022-04-14 23:57:15 +02:00
|
|
|
Realm,
|
|
|
|
RealmAuditLog,
|
2023-04-16 21:53:22 +02:00
|
|
|
RealmAuthenticationMethod,
|
2022-07-26 15:48:26 +02:00
|
|
|
RealmReactivationStatus,
|
2022-04-14 23:57:15 +02:00
|
|
|
RealmUserDefault,
|
2020-05-13 18:09:17 +02:00
|
|
|
Recipient,
|
2022-04-14 23:57:15 +02:00
|
|
|
ScheduledEmail,
|
|
|
|
Stream,
|
2020-05-13 18:09:17 +02:00
|
|
|
Subscription,
|
2022-04-14 23:57:15 +02:00
|
|
|
UserProfile,
|
|
|
|
active_user_ids,
|
2020-05-13 18:09:17 +02:00
|
|
|
get_realm,
|
2022-04-14 23:57:15 +02:00
|
|
|
)
|
django_api: Extract send_event_on_commit helper.
django-stubs 4.2.1 gives transaction.on_commit a more accurate type
annotation, but this exposed that mypy can’t handle the lambda default
parameters that we use to recapture loop variables such as
for stream_id in public_stream_ids:
peer_user_ids = …
event = …
transaction.on_commit(
lambda event=event, peer_user_ids=peer_user_ids: send_event(
realm, event, peer_user_ids
)
)
https://github.com/python/mypy/issues/15459
A workaround that mypy accepts is
transaction.on_commit(
(
lambda event, peer_user_ids: lambda: send_event(
realm, event, peer_user_ids
)
)(event, peer_user_ids)
)
But that’s kind of ugly and potentially error-prone, so let’s make a
helper function for this very common pattern.
send_event_on_commit(realm, event, peer_user_ids)
Signed-off-by: Anders Kaseorg <anders@zulip.com>
2023-06-17 20:53:07 +02:00
|
|
|
from zerver.tornado.django_api import send_event, send_event_on_commit
|
2022-04-14 23:57:15 +02:00
|
|
|
|
|
|
|
if settings.BILLING_ENABLED:
|
|
|
|
from corporate.lib.stripe import downgrade_now_without_creating_additional_invoices
|
|
|
|
|
|
|
|
|
2022-06-23 20:24:25 +02:00
|
|
|
def active_humans_in_realm(realm: Realm) -> QuerySet[UserProfile]:
|
2022-04-14 23:57:15 +02:00
|
|
|
return UserProfile.objects.filter(realm=realm, is_active=True, is_bot=False)
|
|
|
|
|
|
|
|
|
|
|
|
@transaction.atomic(savepoint=False)
|
|
|
|
def do_set_realm_property(
|
|
|
|
realm: Realm, name: str, value: Any, *, acting_user: Optional[UserProfile]
|
|
|
|
) -> None:
|
|
|
|
"""Takes in a realm object, the name of an attribute to update, the
|
|
|
|
value to update and and the user who initiated the update.
|
|
|
|
"""
|
|
|
|
property_type = Realm.property_types[name]
|
|
|
|
assert isinstance(
|
|
|
|
value, property_type
|
|
|
|
), f"Cannot update {name}: {value} is not an instance of {property_type}"
|
|
|
|
|
|
|
|
old_value = getattr(realm, name)
|
|
|
|
setattr(realm, name, value)
|
|
|
|
realm.save(update_fields=[name])
|
|
|
|
|
|
|
|
event = dict(
|
|
|
|
type="realm",
|
|
|
|
op="update",
|
|
|
|
property=name,
|
|
|
|
value=value,
|
|
|
|
)
|
2022-09-22 10:53:37 +02:00
|
|
|
|
|
|
|
# These settings have a different event format due to their history.
|
|
|
|
message_edit_settings = [
|
|
|
|
"allow_message_editing",
|
|
|
|
"edit_topic_policy",
|
|
|
|
"message_content_edit_limit_seconds",
|
|
|
|
]
|
|
|
|
if name in message_edit_settings:
|
|
|
|
event = dict(
|
|
|
|
type="realm",
|
|
|
|
op="update_dict",
|
|
|
|
property="default",
|
|
|
|
data={name: value},
|
|
|
|
)
|
|
|
|
|
django_api: Extract send_event_on_commit helper.
django-stubs 4.2.1 gives transaction.on_commit a more accurate type
annotation, but this exposed that mypy can’t handle the lambda default
parameters that we use to recapture loop variables such as
for stream_id in public_stream_ids:
peer_user_ids = …
event = …
transaction.on_commit(
lambda event=event, peer_user_ids=peer_user_ids: send_event(
realm, event, peer_user_ids
)
)
https://github.com/python/mypy/issues/15459
A workaround that mypy accepts is
transaction.on_commit(
(
lambda event, peer_user_ids: lambda: send_event(
realm, event, peer_user_ids
)
)(event, peer_user_ids)
)
But that’s kind of ugly and potentially error-prone, so let’s make a
helper function for this very common pattern.
send_event_on_commit(realm, event, peer_user_ids)
Signed-off-by: Anders Kaseorg <anders@zulip.com>
2023-06-17 20:53:07 +02:00
|
|
|
send_event_on_commit(realm, event, active_user_ids(realm.id))
|
2022-04-14 23:57:15 +02:00
|
|
|
|
|
|
|
event_time = timezone_now()
|
|
|
|
RealmAuditLog.objects.create(
|
|
|
|
realm=realm,
|
|
|
|
event_type=RealmAuditLog.REALM_PROPERTY_CHANGED,
|
|
|
|
event_time=event_time,
|
|
|
|
acting_user=acting_user,
|
|
|
|
extra_data=orjson.dumps(
|
|
|
|
{
|
|
|
|
RealmAuditLog.OLD_VALUE: old_value,
|
|
|
|
RealmAuditLog.NEW_VALUE: value,
|
|
|
|
"property": name,
|
|
|
|
}
|
|
|
|
).decode(),
|
|
|
|
)
|
|
|
|
|
|
|
|
if name == "waiting_period_threshold":
|
2022-11-21 03:30:07 +01:00
|
|
|
update_users_in_full_members_system_group(realm, acting_user=acting_user)
|
2022-04-14 23:57:15 +02:00
|
|
|
|
|
|
|
|
2022-10-12 09:14:27 +02:00
|
|
|
def parse_and_set_setting_value_if_required(
|
|
|
|
realm: Realm, setting_name: str, value: Union[int, str], *, acting_user: Optional[UserProfile]
|
|
|
|
) -> Tuple[Optional[int], bool]:
|
2022-12-23 14:56:27 +01:00
|
|
|
parsed_value = parse_message_time_limit_setting(
|
2022-10-12 09:14:27 +02:00
|
|
|
value,
|
2022-12-23 14:56:27 +01:00
|
|
|
Realm.MESSAGE_TIME_LIMIT_SETTING_SPECIAL_VALUES_MAP,
|
2022-10-12 09:14:27 +02:00
|
|
|
setting_name=setting_name,
|
|
|
|
)
|
|
|
|
|
|
|
|
setting_value_changed = False
|
|
|
|
if parsed_value is None and getattr(realm, setting_name) is not None:
|
|
|
|
# We handle "None" here separately, since in the update_realm view
|
|
|
|
# function, do_set_realm_property is called only if setting value is
|
|
|
|
# not "None". For values other than "None", the view function itself
|
|
|
|
# sets the value by calling "do_set_realm_property".
|
|
|
|
do_set_realm_property(
|
|
|
|
realm,
|
|
|
|
setting_name,
|
|
|
|
parsed_value,
|
|
|
|
acting_user=acting_user,
|
|
|
|
)
|
|
|
|
setting_value_changed = True
|
|
|
|
|
|
|
|
return parsed_value, setting_value_changed
|
|
|
|
|
|
|
|
|
2022-04-14 23:57:15 +02:00
|
|
|
def do_set_realm_authentication_methods(
|
|
|
|
realm: Realm, authentication_methods: Dict[str, bool], *, acting_user: Optional[UserProfile]
|
|
|
|
) -> None:
|
|
|
|
old_value = realm.authentication_methods_dict()
|
|
|
|
with transaction.atomic():
|
|
|
|
for key, value in list(authentication_methods.items()):
|
2023-04-16 21:53:22 +02:00
|
|
|
# This does queries in a loop, but this isn't a performance sensitive
|
|
|
|
# path and is only run rarely.
|
|
|
|
if value:
|
|
|
|
RealmAuthenticationMethod.objects.get_or_create(realm=realm, name=key)
|
|
|
|
else:
|
|
|
|
RealmAuthenticationMethod.objects.filter(realm=realm, name=key).delete()
|
|
|
|
|
2022-04-14 23:57:15 +02:00
|
|
|
updated_value = realm.authentication_methods_dict()
|
|
|
|
RealmAuditLog.objects.create(
|
|
|
|
realm=realm,
|
|
|
|
event_type=RealmAuditLog.REALM_PROPERTY_CHANGED,
|
|
|
|
event_time=timezone_now(),
|
|
|
|
acting_user=acting_user,
|
|
|
|
extra_data=orjson.dumps(
|
|
|
|
{
|
|
|
|
RealmAuditLog.OLD_VALUE: old_value,
|
|
|
|
RealmAuditLog.NEW_VALUE: updated_value,
|
|
|
|
"property": "authentication_methods",
|
|
|
|
}
|
|
|
|
).decode(),
|
|
|
|
)
|
|
|
|
|
|
|
|
event = dict(
|
|
|
|
type="realm",
|
|
|
|
op="update_dict",
|
|
|
|
property="default",
|
|
|
|
data=dict(authentication_methods=updated_value),
|
|
|
|
)
|
|
|
|
send_event(realm, event, active_user_ids(realm.id))
|
|
|
|
|
|
|
|
|
2022-06-20 01:39:18 +02:00
|
|
|
def do_set_realm_stream(
|
|
|
|
realm: Realm,
|
|
|
|
field: Literal["notifications_stream", "signup_notifications_stream"],
|
|
|
|
stream: Optional[Stream],
|
|
|
|
stream_id: int,
|
|
|
|
*,
|
|
|
|
acting_user: Optional[UserProfile],
|
2022-04-14 23:57:15 +02:00
|
|
|
) -> None:
|
2022-06-20 01:39:18 +02:00
|
|
|
# We could calculate more of these variables from `field`, but
|
|
|
|
# it's probably more readable to not do so.
|
|
|
|
if field == "notifications_stream":
|
|
|
|
old_value = realm.notifications_stream_id
|
|
|
|
realm.notifications_stream = stream
|
|
|
|
property = "notifications_stream_id"
|
|
|
|
elif field == "signup_notifications_stream":
|
|
|
|
old_value = realm.signup_notifications_stream_id
|
|
|
|
realm.signup_notifications_stream = stream
|
|
|
|
property = "signup_notifications_stream_id"
|
|
|
|
else:
|
|
|
|
raise AssertionError("Invalid realm stream field.")
|
|
|
|
|
2022-04-14 23:57:15 +02:00
|
|
|
with transaction.atomic():
|
2022-06-20 01:39:18 +02:00
|
|
|
realm.save(update_fields=[field])
|
2022-04-14 23:57:15 +02:00
|
|
|
|
|
|
|
event_time = timezone_now()
|
|
|
|
RealmAuditLog.objects.create(
|
|
|
|
realm=realm,
|
|
|
|
event_type=RealmAuditLog.REALM_PROPERTY_CHANGED,
|
|
|
|
event_time=event_time,
|
|
|
|
acting_user=acting_user,
|
|
|
|
extra_data=orjson.dumps(
|
|
|
|
{
|
|
|
|
RealmAuditLog.OLD_VALUE: old_value,
|
|
|
|
RealmAuditLog.NEW_VALUE: stream_id,
|
2022-06-20 01:39:18 +02:00
|
|
|
"property": field,
|
2022-04-14 23:57:15 +02:00
|
|
|
}
|
|
|
|
).decode(),
|
|
|
|
)
|
|
|
|
|
|
|
|
event = dict(
|
|
|
|
type="realm",
|
|
|
|
op="update",
|
2022-06-20 01:39:18 +02:00
|
|
|
property=property,
|
2022-04-14 23:57:15 +02:00
|
|
|
value=stream_id,
|
|
|
|
)
|
|
|
|
send_event(realm, event, active_user_ids(realm.id))
|
|
|
|
|
|
|
|
|
2022-06-20 01:39:18 +02:00
|
|
|
def do_set_realm_notifications_stream(
|
2022-04-14 23:57:15 +02:00
|
|
|
realm: Realm, stream: Optional[Stream], stream_id: int, *, acting_user: Optional[UserProfile]
|
|
|
|
) -> None:
|
2022-06-20 01:39:18 +02:00
|
|
|
do_set_realm_stream(realm, "notifications_stream", stream, stream_id, acting_user=acting_user)
|
2022-04-14 23:57:15 +02:00
|
|
|
|
2022-06-20 01:39:18 +02:00
|
|
|
|
|
|
|
def do_set_realm_signup_notifications_stream(
|
|
|
|
realm: Realm, stream: Optional[Stream], stream_id: int, *, acting_user: Optional[UserProfile]
|
|
|
|
) -> None:
|
|
|
|
do_set_realm_stream(
|
|
|
|
realm, "signup_notifications_stream", stream, stream_id, acting_user=acting_user
|
2022-04-14 23:57:15 +02:00
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
|
def do_set_realm_user_default_setting(
|
|
|
|
realm_user_default: RealmUserDefault,
|
|
|
|
name: str,
|
|
|
|
value: Any,
|
|
|
|
*,
|
|
|
|
acting_user: Optional[UserProfile],
|
|
|
|
) -> None:
|
|
|
|
old_value = getattr(realm_user_default, name)
|
|
|
|
realm = realm_user_default.realm
|
|
|
|
event_time = timezone_now()
|
|
|
|
|
|
|
|
with transaction.atomic(savepoint=False):
|
|
|
|
setattr(realm_user_default, name, value)
|
|
|
|
realm_user_default.save(update_fields=[name])
|
|
|
|
|
|
|
|
RealmAuditLog.objects.create(
|
|
|
|
realm=realm,
|
|
|
|
event_type=RealmAuditLog.REALM_DEFAULT_USER_SETTINGS_CHANGED,
|
|
|
|
event_time=event_time,
|
|
|
|
acting_user=acting_user,
|
|
|
|
extra_data=orjson.dumps(
|
|
|
|
{
|
|
|
|
RealmAuditLog.OLD_VALUE: old_value,
|
|
|
|
RealmAuditLog.NEW_VALUE: value,
|
|
|
|
"property": name,
|
|
|
|
}
|
|
|
|
).decode(),
|
|
|
|
)
|
|
|
|
|
|
|
|
event = dict(
|
|
|
|
type="realm_user_settings_defaults",
|
|
|
|
op="update",
|
|
|
|
property=name,
|
|
|
|
value=value,
|
|
|
|
)
|
|
|
|
send_event(realm, event, active_user_ids(realm.id))
|
|
|
|
|
|
|
|
|
|
|
|
def do_deactivate_realm(realm: Realm, *, acting_user: Optional[UserProfile]) -> None:
|
|
|
|
"""
|
|
|
|
Deactivate this realm. Do NOT deactivate the users -- we need to be able to
|
|
|
|
tell the difference between users that were intentionally deactivated,
|
|
|
|
e.g. by a realm admin, and users who can't currently use Zulip because their
|
|
|
|
realm has been deactivated.
|
|
|
|
"""
|
|
|
|
if realm.deactivated:
|
|
|
|
return
|
|
|
|
|
|
|
|
realm.deactivated = True
|
|
|
|
realm.save(update_fields=["deactivated"])
|
|
|
|
|
|
|
|
if settings.BILLING_ENABLED:
|
|
|
|
downgrade_now_without_creating_additional_invoices(realm)
|
|
|
|
|
|
|
|
event_time = timezone_now()
|
|
|
|
RealmAuditLog.objects.create(
|
|
|
|
realm=realm,
|
|
|
|
event_type=RealmAuditLog.REALM_DEACTIVATED,
|
|
|
|
event_time=event_time,
|
|
|
|
acting_user=acting_user,
|
|
|
|
extra_data=orjson.dumps(
|
|
|
|
{
|
|
|
|
RealmAuditLog.ROLE_COUNT: realm_user_count_by_role(realm),
|
|
|
|
}
|
|
|
|
).decode(),
|
|
|
|
)
|
|
|
|
|
|
|
|
ScheduledEmail.objects.filter(realm=realm).delete()
|
|
|
|
for user in active_humans_in_realm(realm):
|
|
|
|
# Don't deactivate the users, but do delete their sessions so they get
|
|
|
|
# bumped to the login screen, where they'll get a realm deactivation
|
|
|
|
# notice when they try to log in.
|
|
|
|
delete_user_sessions(user)
|
|
|
|
|
|
|
|
# This event will only ever be received by clients with an active
|
|
|
|
# longpoll connection, because by this point clients will be
|
|
|
|
# unable to authenticate again to their event queue (triggering an
|
|
|
|
# immediate reload into the page explaining the realm was
|
|
|
|
# deactivated). So the purpose of sending this is to flush all
|
|
|
|
# active longpoll connections for the realm.
|
|
|
|
event = dict(type="realm", op="deactivated", realm_id=realm.id)
|
|
|
|
send_event(realm, event, active_user_ids(realm.id))
|
|
|
|
|
|
|
|
|
|
|
|
def do_reactivate_realm(realm: Realm) -> None:
|
2022-08-03 20:06:59 +02:00
|
|
|
if not realm.deactivated:
|
|
|
|
logging.warning("Realm %s cannot be reactivated because it is already active.", realm.id)
|
|
|
|
return
|
|
|
|
|
2022-04-14 23:57:15 +02:00
|
|
|
realm.deactivated = False
|
|
|
|
with transaction.atomic():
|
|
|
|
realm.save(update_fields=["deactivated"])
|
|
|
|
|
|
|
|
event_time = timezone_now()
|
|
|
|
RealmAuditLog.objects.create(
|
2022-04-16 00:24:37 +02:00
|
|
|
# We hardcode acting_user=None, since realm reactivation
|
|
|
|
# uses an email authentication mechanism that will never
|
|
|
|
# know which user initiated the change.
|
|
|
|
acting_user=None,
|
2022-04-14 23:57:15 +02:00
|
|
|
realm=realm,
|
|
|
|
event_type=RealmAuditLog.REALM_REACTIVATED,
|
|
|
|
event_time=event_time,
|
|
|
|
extra_data=orjson.dumps(
|
|
|
|
{
|
|
|
|
RealmAuditLog.ROLE_COUNT: realm_user_count_by_role(realm),
|
|
|
|
}
|
|
|
|
).decode(),
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
|
def do_add_deactivated_redirect(realm: Realm, redirect_url: str) -> None:
|
|
|
|
realm.deactivated_redirect = redirect_url
|
|
|
|
realm.save(update_fields=["deactivated_redirect"])
|
|
|
|
|
|
|
|
|
2023-02-28 04:44:29 +01:00
|
|
|
def do_delete_all_realm_attachments(realm: Realm, *, batch_size: int = 1000) -> None:
|
2023-02-28 03:35:01 +01:00
|
|
|
# Delete attachment files from the storage backend, so that we
|
|
|
|
# don't leave them dangling.
|
|
|
|
for obj_class in Attachment, ArchivedAttachment:
|
2023-02-28 04:44:29 +01:00
|
|
|
last_id = 0
|
|
|
|
while True:
|
|
|
|
to_delete = (
|
|
|
|
obj_class.objects.filter(realm_id=realm.id, id__gt=last_id) # type: ignore[misc] # Does not recognize shared 'id' PK column
|
|
|
|
.order_by("id")
|
|
|
|
.values_list("id", "path_id")[:batch_size]
|
|
|
|
)
|
|
|
|
if len(to_delete) > 0:
|
|
|
|
delete_message_attachments([row[1] for row in to_delete])
|
|
|
|
last_id = to_delete[len(to_delete) - 1][0]
|
|
|
|
if len(to_delete) < batch_size:
|
|
|
|
break
|
2023-02-28 03:35:01 +01:00
|
|
|
obj_class.objects.filter(realm=realm).delete()
|
|
|
|
|
|
|
|
|
2022-04-14 23:57:15 +02:00
|
|
|
def do_scrub_realm(realm: Realm, *, acting_user: Optional[UserProfile]) -> None:
|
|
|
|
if settings.BILLING_ENABLED:
|
|
|
|
downgrade_now_without_creating_additional_invoices(realm)
|
|
|
|
|
|
|
|
users = UserProfile.objects.filter(realm=realm)
|
|
|
|
for user in users:
|
|
|
|
do_delete_messages_by_sender(user)
|
|
|
|
do_delete_avatar_image(user, acting_user=acting_user)
|
|
|
|
user.full_name = f"Scrubbed {generate_key()[:15]}"
|
2022-07-27 23:33:49 +02:00
|
|
|
scrubbed_email = Address(
|
|
|
|
username=f"scrubbed-{generate_key()[:15]}", domain=realm.host
|
|
|
|
).addr_spec
|
2022-04-14 23:57:15 +02:00
|
|
|
user.email = scrubbed_email
|
|
|
|
user.delivery_email = scrubbed_email
|
|
|
|
user.save(update_fields=["full_name", "email", "delivery_email"])
|
|
|
|
|
2020-05-13 18:09:17 +02:00
|
|
|
internal_realm = get_realm(settings.SYSTEM_BOT_REALM)
|
|
|
|
# We could more simply obtain the Message list by just doing
|
|
|
|
# Message.objects.filter(sender__realm=internal_realm, realm=realm), but it's
|
|
|
|
# more secure against bugs that may cause Message.realm to be incorrect for some
|
|
|
|
# cross-realm messages to also determine the actual Recipients - to prevent
|
|
|
|
# deletion of excessive messages.
|
|
|
|
all_recipient_ids_in_realm = (
|
|
|
|
list(Stream.objects.filter(realm=realm).values_list("recipient_id", flat=True))
|
|
|
|
+ list(UserProfile.objects.filter(realm=realm).values_list("recipient_id", flat=True))
|
|
|
|
+ list(
|
|
|
|
Subscription.objects.filter(
|
|
|
|
recipient__type=Recipient.HUDDLE, user_profile__realm=realm
|
|
|
|
).values_list("recipient_id", flat=True)
|
|
|
|
)
|
|
|
|
)
|
|
|
|
cross_realm_bot_message_ids = list(
|
|
|
|
Message.objects.filter(
|
|
|
|
# Filtering by both message.recipient and message.realm is more robust for ensuring
|
|
|
|
# no messages belonging to another realm will be deleted due to some bugs.
|
|
|
|
sender__realm=internal_realm,
|
|
|
|
recipient_id__in=all_recipient_ids_in_realm,
|
|
|
|
realm=realm,
|
|
|
|
).values_list("id", flat=True)
|
|
|
|
)
|
|
|
|
move_messages_to_archive(cross_realm_bot_message_ids)
|
|
|
|
|
2022-04-14 23:57:15 +02:00
|
|
|
do_remove_realm_custom_profile_fields(realm)
|
2023-02-28 03:35:01 +01:00
|
|
|
do_delete_all_realm_attachments(realm)
|
2022-04-14 23:57:15 +02:00
|
|
|
|
|
|
|
RealmAuditLog.objects.create(
|
|
|
|
realm=realm,
|
|
|
|
event_time=timezone_now(),
|
|
|
|
acting_user=acting_user,
|
|
|
|
event_type=RealmAuditLog.REALM_SCRUBBED,
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
|
@transaction.atomic(durable=True)
|
|
|
|
def do_change_realm_org_type(
|
|
|
|
realm: Realm,
|
|
|
|
org_type: int,
|
|
|
|
acting_user: Optional[UserProfile],
|
|
|
|
) -> None:
|
|
|
|
old_value = realm.org_type
|
|
|
|
realm.org_type = org_type
|
|
|
|
realm.save(update_fields=["org_type"])
|
|
|
|
|
|
|
|
RealmAuditLog.objects.create(
|
|
|
|
event_type=RealmAuditLog.REALM_ORG_TYPE_CHANGED,
|
|
|
|
realm=realm,
|
|
|
|
event_time=timezone_now(),
|
|
|
|
acting_user=acting_user,
|
2022-07-25 14:54:24 +02:00
|
|
|
extra_data=str({"old_value": old_value, "new_value": org_type}),
|
2022-04-14 23:57:15 +02:00
|
|
|
)
|
|
|
|
|
2022-04-11 19:26:16 +02:00
|
|
|
event = dict(type="realm", op="update", property="org_type", value=org_type)
|
django_api: Extract send_event_on_commit helper.
django-stubs 4.2.1 gives transaction.on_commit a more accurate type
annotation, but this exposed that mypy can’t handle the lambda default
parameters that we use to recapture loop variables such as
for stream_id in public_stream_ids:
peer_user_ids = …
event = …
transaction.on_commit(
lambda event=event, peer_user_ids=peer_user_ids: send_event(
realm, event, peer_user_ids
)
)
https://github.com/python/mypy/issues/15459
A workaround that mypy accepts is
transaction.on_commit(
(
lambda event, peer_user_ids: lambda: send_event(
realm, event, peer_user_ids
)
)(event, peer_user_ids)
)
But that’s kind of ugly and potentially error-prone, so let’s make a
helper function for this very common pattern.
send_event_on_commit(realm, event, peer_user_ids)
Signed-off-by: Anders Kaseorg <anders@zulip.com>
2023-06-17 20:53:07 +02:00
|
|
|
send_event_on_commit(realm, event, active_user_ids(realm.id))
|
2022-04-11 19:26:16 +02:00
|
|
|
|
2022-04-14 23:57:15 +02:00
|
|
|
|
|
|
|
@transaction.atomic(savepoint=False)
|
|
|
|
def do_change_realm_plan_type(
|
|
|
|
realm: Realm, plan_type: int, *, acting_user: Optional[UserProfile]
|
|
|
|
) -> None:
|
|
|
|
old_value = realm.plan_type
|
2022-06-02 15:56:30 +02:00
|
|
|
|
|
|
|
if plan_type == Realm.PLAN_TYPE_LIMITED:
|
|
|
|
# We do not allow public access on limited plans.
|
|
|
|
do_set_realm_property(realm, "enable_spectator_access", False, acting_user=acting_user)
|
|
|
|
|
2022-04-14 23:57:15 +02:00
|
|
|
realm.plan_type = plan_type
|
|
|
|
realm.save(update_fields=["plan_type"])
|
|
|
|
RealmAuditLog.objects.create(
|
|
|
|
event_type=RealmAuditLog.REALM_PLAN_TYPE_CHANGED,
|
|
|
|
realm=realm,
|
|
|
|
event_time=timezone_now(),
|
|
|
|
acting_user=acting_user,
|
2022-07-25 14:54:24 +02:00
|
|
|
extra_data=str({"old_value": old_value, "new_value": plan_type}),
|
2022-04-14 23:57:15 +02:00
|
|
|
)
|
|
|
|
|
|
|
|
if plan_type == Realm.PLAN_TYPE_PLUS:
|
|
|
|
realm.max_invites = Realm.INVITES_STANDARD_REALM_DAILY_MAX
|
|
|
|
realm.message_visibility_limit = None
|
|
|
|
realm.upload_quota_gb = Realm.UPLOAD_QUOTA_STANDARD
|
|
|
|
elif plan_type == Realm.PLAN_TYPE_STANDARD:
|
|
|
|
realm.max_invites = Realm.INVITES_STANDARD_REALM_DAILY_MAX
|
|
|
|
realm.message_visibility_limit = None
|
|
|
|
realm.upload_quota_gb = Realm.UPLOAD_QUOTA_STANDARD
|
|
|
|
elif plan_type == Realm.PLAN_TYPE_SELF_HOSTED:
|
2022-07-05 21:21:44 +02:00
|
|
|
realm.max_invites = None # type: ignore[assignment] # https://github.com/python/mypy/issues/3004
|
2022-04-14 23:57:15 +02:00
|
|
|
realm.message_visibility_limit = None
|
|
|
|
realm.upload_quota_gb = None
|
|
|
|
elif plan_type == Realm.PLAN_TYPE_STANDARD_FREE:
|
|
|
|
realm.max_invites = Realm.INVITES_STANDARD_REALM_DAILY_MAX
|
|
|
|
realm.message_visibility_limit = None
|
|
|
|
realm.upload_quota_gb = Realm.UPLOAD_QUOTA_STANDARD
|
|
|
|
elif plan_type == Realm.PLAN_TYPE_LIMITED:
|
|
|
|
realm.max_invites = settings.INVITES_DEFAULT_REALM_DAILY_MAX
|
|
|
|
realm.message_visibility_limit = Realm.MESSAGE_VISIBILITY_LIMITED
|
|
|
|
realm.upload_quota_gb = Realm.UPLOAD_QUOTA_LIMITED
|
|
|
|
else:
|
|
|
|
raise AssertionError("Invalid plan type")
|
|
|
|
|
|
|
|
update_first_visible_message_id(realm)
|
|
|
|
|
2022-06-02 15:56:30 +02:00
|
|
|
realm.save(
|
|
|
|
update_fields=[
|
|
|
|
"_max_invites",
|
|
|
|
"enable_spectator_access",
|
|
|
|
"message_visibility_limit",
|
|
|
|
"upload_quota_gb",
|
|
|
|
]
|
|
|
|
)
|
2022-04-14 23:57:15 +02:00
|
|
|
|
|
|
|
event = {
|
|
|
|
"type": "realm",
|
|
|
|
"op": "update",
|
|
|
|
"property": "plan_type",
|
|
|
|
"value": plan_type,
|
|
|
|
"extra_data": {"upload_quota": realm.upload_quota_bytes()},
|
|
|
|
}
|
django_api: Extract send_event_on_commit helper.
django-stubs 4.2.1 gives transaction.on_commit a more accurate type
annotation, but this exposed that mypy can’t handle the lambda default
parameters that we use to recapture loop variables such as
for stream_id in public_stream_ids:
peer_user_ids = …
event = …
transaction.on_commit(
lambda event=event, peer_user_ids=peer_user_ids: send_event(
realm, event, peer_user_ids
)
)
https://github.com/python/mypy/issues/15459
A workaround that mypy accepts is
transaction.on_commit(
(
lambda event, peer_user_ids: lambda: send_event(
realm, event, peer_user_ids
)
)(event, peer_user_ids)
)
But that’s kind of ugly and potentially error-prone, so let’s make a
helper function for this very common pattern.
send_event_on_commit(realm, event, peer_user_ids)
Signed-off-by: Anders Kaseorg <anders@zulip.com>
2023-06-17 20:53:07 +02:00
|
|
|
send_event_on_commit(realm, event, active_user_ids(realm.id))
|
2022-04-14 23:57:15 +02:00
|
|
|
|
|
|
|
|
|
|
|
def do_send_realm_reactivation_email(realm: Realm, *, acting_user: Optional[UserProfile]) -> None:
|
2022-07-26 15:48:26 +02:00
|
|
|
obj = RealmReactivationStatus.objects.create(realm=realm)
|
|
|
|
|
|
|
|
url = create_confirmation_link(obj, Confirmation.REALM_REACTIVATION)
|
2022-04-14 23:57:15 +02:00
|
|
|
RealmAuditLog.objects.create(
|
|
|
|
realm=realm,
|
|
|
|
acting_user=acting_user,
|
|
|
|
event_type=RealmAuditLog.REALM_REACTIVATION_EMAIL_SENT,
|
|
|
|
event_time=timezone_now(),
|
|
|
|
)
|
2023-05-01 12:47:38 +02:00
|
|
|
context = {
|
|
|
|
"confirmation_url": url,
|
|
|
|
"realm_uri": realm.uri,
|
|
|
|
"realm_name": realm.name,
|
|
|
|
"corporate_enabled": settings.CORPORATE_ENABLED,
|
|
|
|
}
|
2022-04-14 23:57:15 +02:00
|
|
|
language = realm.default_language
|
|
|
|
send_email_to_admins(
|
|
|
|
"zerver/emails/realm_reactivation",
|
|
|
|
realm,
|
|
|
|
from_address=FromAddress.tokenized_no_reply_address(),
|
|
|
|
from_name=FromAddress.security_email_from_name(language=language),
|
|
|
|
language=language,
|
|
|
|
context=context,
|
|
|
|
)
|