mirror of https://github.com/zulip/zulip.git
1066 lines
40 KiB
Python
1066 lines
40 KiB
Python
import logging
|
|
from collections import Counter
|
|
from datetime import datetime, timezone
|
|
from typing import Any, Dict, List, Optional, Type, TypedDict, TypeVar, Union
|
|
from uuid import UUID
|
|
|
|
import orjson
|
|
from django.conf import settings
|
|
from django.core.exceptions import ValidationError
|
|
from django.core.validators import URLValidator, validate_email
|
|
from django.db import IntegrityError, transaction
|
|
from django.db.models import Model
|
|
from django.http import HttpRequest, HttpResponse
|
|
from django.utils.crypto import constant_time_compare
|
|
from django.utils.timezone import now as timezone_now
|
|
from django.utils.translation import gettext as _
|
|
from django.utils.translation import gettext as err_
|
|
from django.views.decorators.csrf import csrf_exempt
|
|
from pydantic import BaseModel, ConfigDict, Json
|
|
|
|
from analytics.lib.counts import (
|
|
BOUNCER_ONLY_REMOTE_COUNT_STAT_PROPERTIES,
|
|
COUNT_STATS,
|
|
REMOTE_INSTALLATION_COUNT_STATS,
|
|
do_increment_logging_stat,
|
|
)
|
|
from corporate.lib.stripe import (
|
|
RemoteRealmBillingSession,
|
|
RemoteServerBillingSession,
|
|
do_deactivate_remote_server,
|
|
)
|
|
from corporate.models import CustomerPlan, get_current_plan_by_customer
|
|
from zerver.decorator import require_post
|
|
from zerver.lib.exceptions import JsonableError, RemoteRealmServerMismatchError
|
|
from zerver.lib.push_notifications import (
|
|
InvalidRemotePushDeviceTokenError,
|
|
UserPushIdentityCompat,
|
|
send_android_push_notification,
|
|
send_apple_push_notification,
|
|
send_test_push_notification_directly_to_devices,
|
|
)
|
|
from zerver.lib.remote_server import (
|
|
InstallationCountDataForAnalytics,
|
|
RealmAuditLogDataForAnalytics,
|
|
RealmCountDataForAnalytics,
|
|
RealmDataForAnalytics,
|
|
)
|
|
from zerver.lib.request import REQ, has_request_variables
|
|
from zerver.lib.response import json_success
|
|
from zerver.lib.timestamp import timestamp_to_datetime
|
|
from zerver.lib.typed_endpoint import JsonBodyPayload, typed_endpoint
|
|
from zerver.lib.types import RemoteRealmDictValue
|
|
from zerver.lib.validator import check_capped_string, check_int, check_string_fixed_length
|
|
from zerver.views.push_notifications import check_app_id, validate_token
|
|
from zilencer.auth import InvalidZulipServerKeyError
|
|
from zilencer.models import (
|
|
RemoteInstallationCount,
|
|
RemotePushDeviceToken,
|
|
RemoteRealm,
|
|
RemoteRealmAuditLog,
|
|
RemoteRealmCount,
|
|
RemoteZulipServer,
|
|
RemoteZulipServerAuditLog,
|
|
)
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
def validate_uuid(uuid: str) -> None:
|
|
try:
|
|
uuid_object = UUID(uuid, version=4)
|
|
# The UUID initialization under some circumstances will modify the uuid
|
|
# string to create a valid UUIDv4, instead of raising a ValueError.
|
|
# The submitted uuid needing to be modified means it's invalid, so
|
|
# we need to check for that condition.
|
|
if str(uuid_object) != uuid:
|
|
raise ValidationError(err_("Invalid UUID"))
|
|
except ValueError:
|
|
raise ValidationError(err_("Invalid UUID"))
|
|
|
|
|
|
def validate_bouncer_token_request(token: str, kind: int) -> None:
|
|
if kind not in [RemotePushDeviceToken.APNS, RemotePushDeviceToken.GCM]:
|
|
raise JsonableError(err_("Invalid token type"))
|
|
validate_token(token, kind)
|
|
|
|
|
|
@csrf_exempt
|
|
@require_post
|
|
@has_request_variables
|
|
def deactivate_remote_server(
|
|
request: HttpRequest,
|
|
remote_server: RemoteZulipServer,
|
|
) -> HttpResponse:
|
|
billing_session = RemoteServerBillingSession(remote_server)
|
|
do_deactivate_remote_server(remote_server, billing_session)
|
|
return json_success(request)
|
|
|
|
|
|
@csrf_exempt
|
|
@require_post
|
|
@has_request_variables
|
|
def register_remote_server(
|
|
request: HttpRequest,
|
|
zulip_org_id: str = REQ(str_validator=check_string_fixed_length(RemoteZulipServer.UUID_LENGTH)),
|
|
zulip_org_key: str = REQ(
|
|
str_validator=check_string_fixed_length(RemoteZulipServer.API_KEY_LENGTH)
|
|
),
|
|
hostname: str = REQ(str_validator=check_capped_string(RemoteZulipServer.HOSTNAME_MAX_LENGTH)),
|
|
contact_email: str = REQ(),
|
|
new_org_key: Optional[str] = REQ(
|
|
str_validator=check_string_fixed_length(RemoteZulipServer.API_KEY_LENGTH), default=None
|
|
),
|
|
) -> HttpResponse:
|
|
# REQ validated the the field lengths, but we still need to
|
|
# validate the format of these fields.
|
|
try:
|
|
# TODO: Ideally we'd not abuse the URL validator this way
|
|
url_validator = URLValidator()
|
|
url_validator("http://" + hostname)
|
|
except ValidationError:
|
|
raise JsonableError(_("{hostname} is not a valid hostname").format(hostname=hostname))
|
|
|
|
try:
|
|
validate_email(contact_email)
|
|
except ValidationError as e:
|
|
raise JsonableError(e.message)
|
|
|
|
try:
|
|
validate_uuid(zulip_org_id)
|
|
except ValidationError as e:
|
|
raise JsonableError(e.message)
|
|
|
|
with transaction.atomic():
|
|
remote_server, created = RemoteZulipServer.objects.get_or_create(
|
|
uuid=zulip_org_id,
|
|
defaults={
|
|
"hostname": hostname,
|
|
"contact_email": contact_email,
|
|
"api_key": zulip_org_key,
|
|
},
|
|
)
|
|
if created:
|
|
RemoteZulipServerAuditLog.objects.create(
|
|
event_type=RemoteZulipServerAuditLog.REMOTE_SERVER_CREATED,
|
|
server=remote_server,
|
|
event_time=remote_server.last_updated,
|
|
)
|
|
else:
|
|
if not constant_time_compare(remote_server.api_key, zulip_org_key):
|
|
raise InvalidZulipServerKeyError(zulip_org_id)
|
|
else:
|
|
remote_server.hostname = hostname
|
|
remote_server.contact_email = contact_email
|
|
if new_org_key is not None:
|
|
remote_server.api_key = new_org_key
|
|
remote_server.save()
|
|
|
|
return json_success(request, data={"created": created})
|
|
|
|
|
|
@has_request_variables
|
|
def register_remote_push_device(
|
|
request: HttpRequest,
|
|
server: RemoteZulipServer,
|
|
user_id: Optional[int] = REQ(json_validator=check_int, default=None),
|
|
user_uuid: Optional[str] = REQ(default=None),
|
|
realm_uuid: Optional[str] = REQ(default=None),
|
|
token: str = REQ(),
|
|
token_kind: int = REQ(json_validator=check_int),
|
|
ios_app_id: Optional[str] = REQ(str_validator=check_app_id, default=None),
|
|
) -> HttpResponse:
|
|
validate_bouncer_token_request(token, token_kind)
|
|
if token_kind == RemotePushDeviceToken.APNS and ios_app_id is None:
|
|
raise JsonableError(_("Missing ios_app_id"))
|
|
|
|
if user_id is None and user_uuid is None:
|
|
raise JsonableError(_("Missing user_id or user_uuid"))
|
|
if user_id is not None and user_uuid is not None:
|
|
kwargs: Dict[str, object] = {"user_uuid": user_uuid, "user_id": None}
|
|
# Delete pre-existing user_id registration for this user+device to avoid
|
|
# duplication. Further down, uuid registration will be created.
|
|
RemotePushDeviceToken.objects.filter(
|
|
server=server, token=token, kind=token_kind, user_id=user_id
|
|
).delete()
|
|
else:
|
|
# One of these is None, so these kwargs will lead to a proper registration
|
|
# of either user_id or user_uuid type
|
|
kwargs = {"user_id": user_id, "user_uuid": user_uuid}
|
|
|
|
if realm_uuid is not None:
|
|
# Servers 8.0+ also send the realm.uuid of the user.
|
|
assert isinstance(
|
|
user_uuid, str
|
|
), "Servers new enough to send realm_uuid, should also have user_uuid"
|
|
remote_realm = get_remote_realm_helper(request, server, realm_uuid, user_uuid)
|
|
if remote_realm is not None:
|
|
# We want to associate the RemotePushDeviceToken with the RemoteRealm.
|
|
kwargs["remote_realm_id"] = remote_realm.id
|
|
|
|
try:
|
|
with transaction.atomic():
|
|
RemotePushDeviceToken.objects.create(
|
|
server=server,
|
|
kind=token_kind,
|
|
token=token,
|
|
ios_app_id=ios_app_id,
|
|
# last_updated is to be renamed to date_created.
|
|
last_updated=timezone_now(),
|
|
**kwargs,
|
|
)
|
|
except IntegrityError:
|
|
pass
|
|
|
|
return json_success(request)
|
|
|
|
|
|
@has_request_variables
|
|
def unregister_remote_push_device(
|
|
request: HttpRequest,
|
|
server: RemoteZulipServer,
|
|
token: str = REQ(),
|
|
token_kind: int = REQ(json_validator=check_int),
|
|
user_id: Optional[int] = REQ(json_validator=check_int, default=None),
|
|
user_uuid: Optional[str] = REQ(default=None),
|
|
) -> HttpResponse:
|
|
validate_bouncer_token_request(token, token_kind)
|
|
user_identity = UserPushIdentityCompat(user_id=user_id, user_uuid=user_uuid)
|
|
|
|
(num_deleted, ignored) = RemotePushDeviceToken.objects.filter(
|
|
user_identity.filter_q(), token=token, kind=token_kind, server=server
|
|
).delete()
|
|
if num_deleted == 0:
|
|
raise JsonableError(err_("Token does not exist"))
|
|
|
|
return json_success(request)
|
|
|
|
|
|
@has_request_variables
|
|
def unregister_all_remote_push_devices(
|
|
request: HttpRequest,
|
|
server: RemoteZulipServer,
|
|
user_id: Optional[int] = REQ(json_validator=check_int, default=None),
|
|
user_uuid: Optional[str] = REQ(default=None),
|
|
) -> HttpResponse:
|
|
user_identity = UserPushIdentityCompat(user_id=user_id, user_uuid=user_uuid)
|
|
|
|
RemotePushDeviceToken.objects.filter(user_identity.filter_q(), server=server).delete()
|
|
return json_success(request)
|
|
|
|
|
|
def delete_duplicate_registrations(
|
|
registrations: List[RemotePushDeviceToken], server_id: int, user_id: int, user_uuid: str
|
|
) -> List[RemotePushDeviceToken]:
|
|
"""
|
|
When migrating to support registration by UUID, we introduced a bug where duplicate
|
|
registrations for the same device+user could be created - one by user_id and one by
|
|
user_uuid. Given no good way of detecting these duplicates at database level, we need to
|
|
take advantage of the fact that when a remote server sends a push notification request
|
|
to us, it sends both user_id and user_uuid of the user.
|
|
See https://github.com/zulip/zulip/issues/24969 for reference.
|
|
|
|
This function, knowing the user_id and user_uuid of the user, can detect duplicates
|
|
and delete the legacy user_id registration if appropriate.
|
|
|
|
Return the list of registrations with the user_id-based duplicates removed.
|
|
"""
|
|
|
|
# All registrations passed here should be of the same kind (apple vs android).
|
|
assert len({registration.kind for registration in registrations}) == 1
|
|
kind = registrations[0].kind
|
|
|
|
tokens_counter = Counter(device.token for device in registrations)
|
|
|
|
tokens_to_deduplicate = []
|
|
for key in tokens_counter:
|
|
if tokens_counter[key] <= 1:
|
|
continue
|
|
if tokens_counter[key] > 2:
|
|
raise AssertionError(
|
|
f"More than two registrations for token {key} for user id:{user_id} uuid:{user_uuid}, shouldn't be possible"
|
|
)
|
|
assert tokens_counter[key] == 2
|
|
tokens_to_deduplicate.append(key)
|
|
|
|
if not tokens_to_deduplicate:
|
|
return registrations
|
|
|
|
logger.info(
|
|
"Deduplicating push registrations for server id:%s user id:%s uuid:%s and tokens:%s",
|
|
server_id,
|
|
user_id,
|
|
user_uuid,
|
|
sorted(tokens_to_deduplicate),
|
|
)
|
|
RemotePushDeviceToken.objects.filter(
|
|
token__in=tokens_to_deduplicate, kind=kind, server_id=server_id, user_id=user_id
|
|
).delete()
|
|
|
|
deduplicated_registrations_to_return = []
|
|
for registration in registrations:
|
|
if registration.token in tokens_to_deduplicate and registration.user_id is not None:
|
|
# user_id registrations are the ones we deleted
|
|
continue
|
|
deduplicated_registrations_to_return.append(registration)
|
|
|
|
return deduplicated_registrations_to_return
|
|
|
|
|
|
class TestNotificationPayload(BaseModel):
|
|
token: str
|
|
token_kind: int
|
|
user_id: int
|
|
user_uuid: str
|
|
base_payload: Dict[str, Any]
|
|
|
|
model_config = ConfigDict(extra="forbid")
|
|
|
|
|
|
@typed_endpoint
|
|
def remote_server_send_test_notification(
|
|
request: HttpRequest,
|
|
server: RemoteZulipServer,
|
|
*,
|
|
payload: JsonBodyPayload[TestNotificationPayload],
|
|
) -> HttpResponse:
|
|
token = payload.token
|
|
token_kind = payload.token_kind
|
|
|
|
user_id = payload.user_id
|
|
user_uuid = payload.user_uuid
|
|
|
|
# The remote server only sends the base payload with basic user and server info,
|
|
# and the actual format of the test notification is defined on the bouncer, as that
|
|
# gives us the flexibility to modify it freely, without relying on other servers
|
|
# upgrading.
|
|
base_payload = payload.base_payload
|
|
|
|
# This is a new endpoint, so it can assume it will only be used by newer
|
|
# servers that will send user both UUID and ID.
|
|
user_identity = UserPushIdentityCompat(user_id=user_id, user_uuid=user_uuid)
|
|
|
|
try:
|
|
device = RemotePushDeviceToken.objects.get(
|
|
user_identity.filter_q(), token=token, kind=token_kind, server=server
|
|
)
|
|
except RemotePushDeviceToken.DoesNotExist:
|
|
raise InvalidRemotePushDeviceTokenError
|
|
|
|
send_test_push_notification_directly_to_devices(
|
|
user_identity, [device], base_payload, remote=server
|
|
)
|
|
return json_success(request)
|
|
|
|
|
|
def get_remote_realm_helper(
|
|
request: HttpRequest, server: RemoteZulipServer, realm_uuid: str, user_uuid: str
|
|
) -> Optional[RemoteRealm]:
|
|
"""
|
|
Tries to fetch RemoteRealm for the given realm_uuid and server. Otherwise,
|
|
returns None and logs what happened using request and user_uuid args to make
|
|
the output more informative.
|
|
"""
|
|
|
|
try:
|
|
remote_realm = RemoteRealm.objects.get(uuid=realm_uuid)
|
|
except RemoteRealm.DoesNotExist:
|
|
logger.info(
|
|
"%s: Received request for unknown realm %s, server %s, user %s",
|
|
request.path,
|
|
realm_uuid,
|
|
server.id,
|
|
user_uuid,
|
|
)
|
|
return None
|
|
|
|
if remote_realm.server_id != server.id:
|
|
logger.warning(
|
|
"%s: Realm %s exists, but not registered to server %s",
|
|
request.path,
|
|
realm_uuid,
|
|
server.id,
|
|
)
|
|
raise RemoteRealmServerMismatchError
|
|
|
|
return remote_realm
|
|
|
|
|
|
@has_request_variables
|
|
def remote_server_notify_push(
|
|
request: HttpRequest,
|
|
server: RemoteZulipServer,
|
|
payload: Dict[str, Any] = REQ(argument_type="body"),
|
|
) -> HttpResponse:
|
|
user_id = payload.get("user_id")
|
|
user_uuid = payload.get("user_uuid")
|
|
user_identity = UserPushIdentityCompat(user_id, user_uuid)
|
|
|
|
gcm_payload = payload["gcm_payload"]
|
|
apns_payload = payload["apns_payload"]
|
|
gcm_options = payload.get("gcm_options", {})
|
|
|
|
realm_uuid = payload.get("realm_uuid")
|
|
remote_realm = None
|
|
if realm_uuid is not None:
|
|
assert isinstance(
|
|
user_uuid, str
|
|
), "Servers new enough to send realm_uuid, should also have user_uuid"
|
|
remote_realm = get_remote_realm_helper(request, server, realm_uuid, user_uuid)
|
|
|
|
android_devices = list(
|
|
RemotePushDeviceToken.objects.filter(
|
|
user_identity.filter_q(),
|
|
kind=RemotePushDeviceToken.GCM,
|
|
server=server,
|
|
)
|
|
)
|
|
if android_devices and user_id is not None and user_uuid is not None:
|
|
android_devices = delete_duplicate_registrations(
|
|
android_devices, server.id, user_id, user_uuid
|
|
)
|
|
|
|
apple_devices = list(
|
|
RemotePushDeviceToken.objects.filter(
|
|
user_identity.filter_q(),
|
|
kind=RemotePushDeviceToken.APNS,
|
|
server=server,
|
|
)
|
|
)
|
|
if apple_devices and user_id is not None and user_uuid is not None:
|
|
apple_devices = delete_duplicate_registrations(apple_devices, server.id, user_id, user_uuid)
|
|
|
|
remote_queue_latency: Optional[str] = None
|
|
sent_time: Optional[Union[float, int]] = gcm_payload.get(
|
|
# TODO/compatibility: This could be a lot simpler if not for pre-5.0 Zulip servers
|
|
# that had an older format. Future implementation:
|
|
# "time", apns_payload["custom"]["zulip"].get("time")
|
|
"time",
|
|
apns_payload.get("custom", {}).get("zulip", {}).get("time"),
|
|
)
|
|
if sent_time is not None:
|
|
if isinstance(sent_time, int):
|
|
# The 'time' field only used to have whole-integer
|
|
# granularity, so if so we only report with
|
|
# whole-second granularity
|
|
remote_queue_latency = str(int(timezone_now().timestamp()) - sent_time)
|
|
else:
|
|
remote_queue_latency = f"{timezone_now().timestamp() - sent_time:.3f}"
|
|
logger.info(
|
|
"Remote queuing latency for %s:%s is %s seconds",
|
|
server.uuid,
|
|
user_identity,
|
|
remote_queue_latency,
|
|
)
|
|
|
|
logger.info(
|
|
"Sending mobile push notifications for remote user %s:%s: %s via FCM devices, %s via APNs devices",
|
|
server.uuid,
|
|
user_identity,
|
|
len(android_devices),
|
|
len(apple_devices),
|
|
)
|
|
do_increment_logging_stat(
|
|
server,
|
|
REMOTE_INSTALLATION_COUNT_STATS["mobile_pushes_received::day"],
|
|
None,
|
|
timezone_now(),
|
|
increment=len(android_devices) + len(apple_devices),
|
|
)
|
|
if remote_realm is not None:
|
|
do_increment_logging_stat(
|
|
remote_realm,
|
|
COUNT_STATS["mobile_pushes_received::day"],
|
|
None,
|
|
timezone_now(),
|
|
increment=len(android_devices) + len(apple_devices),
|
|
)
|
|
|
|
# Truncate incoming pushes to 200, due to APNs maximum message
|
|
# sizes; see handle_remove_push_notification for the version of
|
|
# this for notifications generated natively on the server. We
|
|
# apply this to remote-server pushes in case they predate that
|
|
# commit.
|
|
def truncate_payload(payload: Dict[str, Any]) -> Dict[str, Any]:
|
|
MAX_MESSAGE_IDS = 200
|
|
if payload and payload.get("event") == "remove" and payload.get("zulip_message_ids"):
|
|
ids = [int(id) for id in payload["zulip_message_ids"].split(",")]
|
|
truncated_ids = sorted(ids)[-MAX_MESSAGE_IDS:]
|
|
payload["zulip_message_ids"] = ",".join(str(id) for id in truncated_ids)
|
|
return payload
|
|
|
|
# The full request must complete within 30s, the timeout set by
|
|
# Zulip remote hosts for push notification requests (see
|
|
# PushBouncerSession). The timeouts in the FCM and APNS codepaths
|
|
# must be set accordingly; see send_android_push_notification and
|
|
# send_apple_push_notification.
|
|
|
|
gcm_payload = truncate_payload(gcm_payload)
|
|
android_successfully_delivered = send_android_push_notification(
|
|
user_identity, android_devices, gcm_payload, gcm_options, remote=server
|
|
)
|
|
|
|
if isinstance(apns_payload.get("custom"), dict) and isinstance(
|
|
apns_payload["custom"].get("zulip"), dict
|
|
):
|
|
apns_payload["custom"]["zulip"] = truncate_payload(apns_payload["custom"]["zulip"])
|
|
apple_successfully_delivered = send_apple_push_notification(
|
|
user_identity, apple_devices, apns_payload, remote=server
|
|
)
|
|
|
|
do_increment_logging_stat(
|
|
server,
|
|
REMOTE_INSTALLATION_COUNT_STATS["mobile_pushes_forwarded::day"],
|
|
None,
|
|
timezone_now(),
|
|
increment=android_successfully_delivered + apple_successfully_delivered,
|
|
)
|
|
|
|
remote_realm_dict: Optional[RemoteRealmDictValue] = None
|
|
if remote_realm is not None:
|
|
do_increment_logging_stat(
|
|
remote_realm,
|
|
COUNT_STATS["mobile_pushes_forwarded::day"],
|
|
None,
|
|
timezone_now(),
|
|
increment=android_successfully_delivered + apple_successfully_delivered,
|
|
)
|
|
billing_session = RemoteRealmBillingSession(remote_realm)
|
|
remote_realm_dict = billing_session.get_push_service_validity_dict()
|
|
|
|
deleted_devices = get_deleted_devices(
|
|
user_identity,
|
|
server,
|
|
android_devices=payload.get("android_devices", []),
|
|
apple_devices=payload.get("apple_devices", []),
|
|
)
|
|
|
|
return json_success(
|
|
request,
|
|
data={
|
|
"total_android_devices": len(android_devices),
|
|
"total_apple_devices": len(apple_devices),
|
|
"deleted_devices": deleted_devices,
|
|
"realm": remote_realm_dict,
|
|
},
|
|
)
|
|
|
|
|
|
class DevicesToCleanUpDict(TypedDict):
|
|
android_devices: List[str]
|
|
apple_devices: List[str]
|
|
|
|
|
|
def get_deleted_devices(
|
|
user_identity: UserPushIdentityCompat,
|
|
server: RemoteZulipServer,
|
|
android_devices: List[str],
|
|
apple_devices: List[str],
|
|
) -> DevicesToCleanUpDict:
|
|
"""The remote server sends us a list of (tokens of) devices that it
|
|
believes it has registered. However some of them may have been
|
|
deleted by us due to errors received in the low level code
|
|
responsible for directly sending push notifications.
|
|
|
|
Query the database for the RemotePushDeviceTokens from these lists
|
|
that we do indeed have and return a list of the ones that we don't
|
|
have and thus presumably have already deleted - the remote server
|
|
will want to delete them too.
|
|
"""
|
|
|
|
android_devices_we_have = RemotePushDeviceToken.objects.filter(
|
|
user_identity.filter_q(),
|
|
token__in=android_devices,
|
|
kind=RemotePushDeviceToken.GCM,
|
|
server=server,
|
|
).values_list("token", flat=True)
|
|
apple_devices_we_have = RemotePushDeviceToken.objects.filter(
|
|
user_identity.filter_q(),
|
|
token__in=apple_devices,
|
|
kind=RemotePushDeviceToken.APNS,
|
|
server=server,
|
|
).values_list("token", flat=True)
|
|
|
|
return DevicesToCleanUpDict(
|
|
android_devices=list(set(android_devices) - set(android_devices_we_have)),
|
|
apple_devices=list(set(apple_devices) - set(apple_devices_we_have)),
|
|
)
|
|
|
|
|
|
def validate_incoming_table_data(
|
|
server: RemoteZulipServer,
|
|
model: Any,
|
|
rows: List[Dict[str, Any]],
|
|
*,
|
|
is_count_stat: bool,
|
|
) -> None:
|
|
last_id = get_last_id_from_server(server, model)
|
|
for row in rows:
|
|
if is_count_stat and (
|
|
row["property"] not in COUNT_STATS
|
|
or row["property"] in BOUNCER_ONLY_REMOTE_COUNT_STAT_PROPERTIES
|
|
):
|
|
raise JsonableError(_("Invalid property {property}").format(property=row["property"]))
|
|
|
|
if not is_count_stat and row["event_type"] not in RemoteRealmAuditLog.SYNCED_BILLING_EVENTS:
|
|
raise JsonableError(_("Invalid event type."))
|
|
|
|
if row.get("id") is None:
|
|
# This shouldn't be possible, as submitting data like this should be
|
|
# prevented by our param validators.
|
|
raise AssertionError(f"Missing id field in row {row}")
|
|
if row["id"] <= last_id:
|
|
raise JsonableError(_("Data is out of order."))
|
|
last_id = row["id"]
|
|
|
|
|
|
ModelT = TypeVar("ModelT", bound=Model)
|
|
|
|
|
|
def batch_create_table_data(
|
|
server: RemoteZulipServer,
|
|
model: Type[ModelT],
|
|
row_objects: List[ModelT],
|
|
) -> None:
|
|
# We ignore previously-existing data, in case it was truncated and
|
|
# re-created on the remote server. `ignore_concflicts=True`
|
|
# cannot return the ids, or count thereof, of the new inserts,
|
|
# (see https://code.djangoproject.com/ticket/0138) so we rely on
|
|
# having a lock to accurately count them before and after. This
|
|
# query is also well-indexed.
|
|
before_count = model._default_manager.filter(server=server).count()
|
|
model._default_manager.bulk_create(row_objects, batch_size=1000, ignore_conflicts=True)
|
|
after_count = model._default_manager.filter(server=server).count()
|
|
inserted_count = after_count - before_count
|
|
if inserted_count < len(row_objects):
|
|
logging.warning(
|
|
"Dropped %d duplicated rows while saving %d rows of %s for server %s/%s",
|
|
len(row_objects) - inserted_count,
|
|
len(row_objects),
|
|
model._meta.db_table,
|
|
server.hostname,
|
|
server.uuid,
|
|
)
|
|
|
|
|
|
def update_remote_realm_data_for_server(
|
|
server: RemoteZulipServer, server_realms_info: List[RealmDataForAnalytics]
|
|
) -> None:
|
|
uuids = [realm.uuid for realm in server_realms_info]
|
|
already_registered_remote_realms = RemoteRealm.objects.filter(uuid__in=uuids, server=server)
|
|
already_registered_uuids = {
|
|
remote_realm.uuid for remote_realm in already_registered_remote_realms
|
|
}
|
|
|
|
new_remote_realms = [
|
|
RemoteRealm(
|
|
server=server,
|
|
uuid=realm.uuid,
|
|
uuid_owner_secret=realm.uuid_owner_secret,
|
|
host=realm.host,
|
|
realm_deactivated=realm.deactivated,
|
|
realm_date_created=timestamp_to_datetime(realm.date_created),
|
|
org_type=realm.org_type,
|
|
name=realm.name,
|
|
authentication_methods=realm.authentication_methods,
|
|
is_system_bot_realm=realm.is_system_bot_realm,
|
|
)
|
|
for realm in server_realms_info
|
|
if realm.uuid not in already_registered_uuids
|
|
]
|
|
|
|
try:
|
|
RemoteRealm.objects.bulk_create(new_remote_realms)
|
|
except IntegrityError:
|
|
raise JsonableError(_("Duplicate registration detected."))
|
|
|
|
uuid_to_realm_dict = {str(realm.uuid): realm for realm in server_realms_info}
|
|
remote_realms_to_update = []
|
|
remote_realm_audit_logs = []
|
|
now = timezone_now()
|
|
|
|
# Update RemoteRealm entries, for which the corresponding realm's info has changed
|
|
# (for the attributes that make sense to sync like this).
|
|
for remote_realm in already_registered_remote_realms:
|
|
modified = False
|
|
realm = uuid_to_realm_dict[str(remote_realm.uuid)]
|
|
for remote_realm_attr, realm_dict_key in [
|
|
("host", "host"),
|
|
("org_type", "org_type"),
|
|
("name", "name"),
|
|
("authentication_methods", "authentication_methods"),
|
|
("realm_deactivated", "deactivated"),
|
|
("is_system_bot_realm", "is_system_bot_realm"),
|
|
]:
|
|
old_value = getattr(remote_realm, remote_realm_attr)
|
|
new_value = getattr(realm, realm_dict_key)
|
|
|
|
if old_value == new_value:
|
|
continue
|
|
|
|
setattr(remote_realm, remote_realm_attr, new_value)
|
|
remote_realm_audit_logs.append(
|
|
RemoteRealmAuditLog(
|
|
server=server,
|
|
remote_id=None,
|
|
remote_realm=remote_realm,
|
|
realm_id=realm.id,
|
|
event_type=RemoteRealmAuditLog.REMOTE_REALM_VALUE_UPDATED,
|
|
event_time=now,
|
|
extra_data={
|
|
"attr_name": remote_realm_attr,
|
|
"old_value": old_value,
|
|
"new_value": new_value,
|
|
},
|
|
)
|
|
)
|
|
modified = True
|
|
|
|
if modified:
|
|
remote_realms_to_update.append(remote_realm)
|
|
|
|
RemoteRealm.objects.bulk_update(
|
|
remote_realms_to_update,
|
|
[
|
|
"host",
|
|
"realm_deactivated",
|
|
"name",
|
|
"authentication_methods",
|
|
"org_type",
|
|
"is_system_bot_realm",
|
|
],
|
|
)
|
|
RemoteRealmAuditLog.objects.bulk_create(remote_realm_audit_logs)
|
|
|
|
|
|
def get_human_user_realm_uuids(realms: List[RealmDataForAnalytics]) -> List[UUID]: # nocoverage
|
|
billable_realm_uuids = []
|
|
for realm in realms:
|
|
# TODO: Remove the `zulipinternal` string_id check once no server is on 8.0-beta.
|
|
if (
|
|
realm.is_system_bot_realm
|
|
or realm.host.startswith("zulipinternal.")
|
|
or (settings.DEVELOPMENT and realm.host.startswith("analytics."))
|
|
):
|
|
continue
|
|
billable_realm_uuids.append(realm.uuid)
|
|
|
|
return billable_realm_uuids
|
|
|
|
|
|
@transaction.atomic
|
|
def handle_customer_migration_from_server_to_realms(
|
|
server: RemoteZulipServer, realms: List[RealmDataForAnalytics]
|
|
) -> None: # nocoverage
|
|
server_billing_session = RemoteServerBillingSession(server)
|
|
server_customer = server_billing_session.get_customer()
|
|
if server_customer is None:
|
|
return
|
|
|
|
server_plan = get_current_plan_by_customer(server_customer)
|
|
if server_plan is None:
|
|
# If the server has no current plan, either because it never
|
|
# had one or because a previous legacy plan was migrated to
|
|
# the RemoteRealm object, there's nothing to potentially
|
|
# migrate.
|
|
return
|
|
|
|
realm_uuids = get_human_user_realm_uuids(realms)
|
|
if not realm_uuids:
|
|
return
|
|
|
|
event_time = timezone_now()
|
|
remote_realm_audit_logs = []
|
|
|
|
if (
|
|
server_plan.tier == CustomerPlan.TIER_SELF_HOSTED_LEGACY
|
|
and server_plan.status == CustomerPlan.ACTIVE
|
|
):
|
|
assert server.plan_type == RemoteZulipServer.PLAN_TYPE_SELF_MANAGED_LEGACY
|
|
assert server_plan.end_date is not None
|
|
remote_realms = RemoteRealm.objects.filter(
|
|
uuid__in=realm_uuids, server=server, plan_type=RemoteRealm.PLAN_TYPE_SELF_MANAGED
|
|
)
|
|
|
|
# Verify that all the realms are on self hosted plan.
|
|
assert remote_realms.count() == len(realm_uuids)
|
|
|
|
# End existing plan for server.
|
|
server_plan.status = CustomerPlan.ENDED
|
|
server_plan.save(update_fields=["status"])
|
|
|
|
server.plan_type = RemoteZulipServer.PLAN_TYPE_SELF_MANAGED
|
|
server.save(update_fields=["plan_type"])
|
|
|
|
# Create new legacy plan for each remote realm.
|
|
for remote_realm in remote_realms:
|
|
RemoteRealmBillingSession(remote_realm).migrate_customer_to_legacy_plan(
|
|
server_plan.billing_cycle_anchor, server_plan.end_date
|
|
)
|
|
remote_realm_audit_logs.append(
|
|
RemoteRealmAuditLog(
|
|
server=server,
|
|
remote_realm=remote_realm,
|
|
event_type=RemoteRealmAuditLog.REMOTE_PLAN_TRANSFERRED_SERVER_TO_REALM,
|
|
event_time=event_time,
|
|
# No extra_data since there was no real change in any RemoteRealm attribute.
|
|
)
|
|
)
|
|
|
|
elif len(realm_uuids) == 1:
|
|
# Here, we have exactly one non-system-bot realm, and some
|
|
# sort of plan on the server; move it to the realm.
|
|
remote_realm = RemoteRealm.objects.get(
|
|
uuid=realm_uuids[0], plan_type=RemoteRealm.PLAN_TYPE_SELF_MANAGED
|
|
)
|
|
# Migrate customer from server to remote realm if there is only one realm.
|
|
server_customer.remote_realm = remote_realm
|
|
server_customer.remote_server = None
|
|
server_customer.save(update_fields=["remote_realm", "remote_server"])
|
|
# TODO: Might be better to call do_change_plan_type here.
|
|
remote_realm.plan_type = server.plan_type
|
|
remote_realm.save(update_fields=["plan_type"])
|
|
server.plan_type = RemoteZulipServer.PLAN_TYPE_SELF_MANAGED
|
|
server.save(update_fields=["plan_type"])
|
|
remote_realm_audit_logs.append(
|
|
RemoteRealmAuditLog(
|
|
server=server,
|
|
remote_realm=remote_realm,
|
|
event_type=RemoteRealmAuditLog.REMOTE_PLAN_TRANSFERRED_SERVER_TO_REALM,
|
|
event_time=event_time,
|
|
extra_data={
|
|
"attr_name": "plan_type",
|
|
"old_value": RemoteRealm.PLAN_TYPE_SELF_MANAGED,
|
|
"new_value": remote_realm.plan_type,
|
|
},
|
|
)
|
|
)
|
|
|
|
RemoteRealmAuditLog.objects.bulk_create(remote_realm_audit_logs)
|
|
|
|
|
|
@typed_endpoint
|
|
@transaction.atomic
|
|
def remote_server_post_analytics(
|
|
request: HttpRequest,
|
|
server: RemoteZulipServer,
|
|
*,
|
|
realm_counts: Json[List[RealmCountDataForAnalytics]],
|
|
installation_counts: Json[List[InstallationCountDataForAnalytics]],
|
|
realmauditlog_rows: Optional[Json[List[RealmAuditLogDataForAnalytics]]] = None,
|
|
realms: Optional[Json[List[RealmDataForAnalytics]]] = None,
|
|
version: Optional[Json[str]] = None,
|
|
api_feature_level: Optional[Json[int]] = None,
|
|
) -> HttpResponse:
|
|
# Lock the server, preventing this from racing with other
|
|
# duplicate submissions of the data
|
|
server = RemoteZulipServer.objects.select_for_update().get(id=server.id)
|
|
|
|
remote_server_version_updated = False
|
|
if version is not None:
|
|
version = version[0 : RemoteZulipServer.VERSION_MAX_LENGTH]
|
|
if version != server.last_version or api_feature_level != server.last_api_feature_level:
|
|
server.last_version = version
|
|
server.last_api_feature_level = api_feature_level
|
|
server.save(update_fields=["last_version", "last_api_feature_level"])
|
|
remote_server_version_updated = True
|
|
|
|
validate_incoming_table_data(
|
|
server,
|
|
RemoteRealmCount,
|
|
[dict(count) for count in realm_counts],
|
|
is_count_stat=True,
|
|
)
|
|
validate_incoming_table_data(
|
|
server,
|
|
RemoteInstallationCount,
|
|
[dict(count) for count in installation_counts],
|
|
is_count_stat=True,
|
|
)
|
|
|
|
if realmauditlog_rows is not None:
|
|
validate_incoming_table_data(
|
|
server,
|
|
RemoteRealmAuditLog,
|
|
[dict(row) for row in realmauditlog_rows],
|
|
is_count_stat=False,
|
|
)
|
|
|
|
if realms is not None:
|
|
update_remote_realm_data_for_server(server, realms)
|
|
if remote_server_version_updated:
|
|
fix_remote_realm_foreign_keys(server, realms)
|
|
|
|
try:
|
|
handle_customer_migration_from_server_to_realms(server, realms)
|
|
except Exception: # nocoverage
|
|
logger.exception(
|
|
"%s: Failed to migrate customer from server (id: %s) to realms",
|
|
request.path,
|
|
server.id,
|
|
stack_info=True,
|
|
)
|
|
raise JsonableError(
|
|
_(
|
|
"Failed to migrate customer from server to realms. Please contact support for assistance."
|
|
)
|
|
)
|
|
|
|
realm_id_to_remote_realm = build_realm_id_to_remote_realm_dict(server, realms)
|
|
|
|
remote_realm_counts = [
|
|
RemoteRealmCount(
|
|
remote_realm=realm_id_to_remote_realm.get(row.realm),
|
|
property=row.property,
|
|
realm_id=row.realm,
|
|
remote_id=row.id,
|
|
server=server,
|
|
end_time=datetime.fromtimestamp(row.end_time, tz=timezone.utc),
|
|
subgroup=row.subgroup,
|
|
value=row.value,
|
|
)
|
|
for row in realm_counts
|
|
]
|
|
batch_create_table_data(server, RemoteRealmCount, remote_realm_counts)
|
|
|
|
remote_installation_counts = [
|
|
RemoteInstallationCount(
|
|
property=row.property,
|
|
remote_id=row.id,
|
|
server=server,
|
|
end_time=datetime.fromtimestamp(row.end_time, tz=timezone.utc),
|
|
subgroup=row.subgroup,
|
|
value=row.value,
|
|
)
|
|
for row in installation_counts
|
|
]
|
|
batch_create_table_data(server, RemoteInstallationCount, remote_installation_counts)
|
|
|
|
if realmauditlog_rows is not None:
|
|
# Creating audit logs, syncing license ledger, and updating
|
|
# 'last_audit_log_update' needs to be an atomic operation.
|
|
# This helps to rely on 'last_audit_log_update' to assume
|
|
# RemoteRealmAuditLog and LicenseLedger are up-to-date.
|
|
with transaction.atomic():
|
|
# Important: Do not return early if we receive 0 rows; we must
|
|
# updated last_audit_log_update even if there are no new rows,
|
|
# to help identify server whose ability to connect to this
|
|
# endpoint is broken by a networking problem.
|
|
remote_realms_set = set()
|
|
remote_realm_audit_logs = []
|
|
for row in realmauditlog_rows:
|
|
extra_data = {}
|
|
if isinstance(row.extra_data, str):
|
|
try:
|
|
extra_data = orjson.loads(row.extra_data)
|
|
except orjson.JSONDecodeError:
|
|
raise JsonableError(_("Malformed audit log data"))
|
|
elif row.extra_data is not None:
|
|
assert isinstance(row.extra_data, dict)
|
|
extra_data = row.extra_data
|
|
remote_realms_set.add(realm_id_to_remote_realm.get(row.realm))
|
|
remote_realm_audit_logs.append(
|
|
RemoteRealmAuditLog(
|
|
remote_realm=realm_id_to_remote_realm.get(row.realm),
|
|
realm_id=row.realm,
|
|
remote_id=row.id,
|
|
server=server,
|
|
event_time=datetime.fromtimestamp(row.event_time, tz=timezone.utc),
|
|
backfilled=row.backfilled,
|
|
extra_data=extra_data,
|
|
event_type=row.event_type,
|
|
)
|
|
)
|
|
batch_create_table_data(server, RemoteRealmAuditLog, remote_realm_audit_logs)
|
|
|
|
# We need to update 'last_audit_log_update' before calling the
|
|
# 'sync_license_ledger_if_needed' method to avoid 'MissingDataError'
|
|
# due to 'has_stale_audit_log' being True.
|
|
RemoteZulipServer.objects.filter(uuid=server.uuid).update(
|
|
last_audit_log_update=timezone_now()
|
|
)
|
|
|
|
# Update LicenseLedger for remote_realm customers using logs in RemoteRealmAuditlog.
|
|
for remote_realm in remote_realms_set:
|
|
if remote_realm:
|
|
billing_session = RemoteRealmBillingSession(remote_realm=remote_realm)
|
|
billing_session.sync_license_ledger_if_needed()
|
|
|
|
# Update LicenseLedger for remote_server customer using logs in RemoteRealmAuditlog.
|
|
remote_server_billing_session = RemoteServerBillingSession(remote_server=server)
|
|
remote_server_billing_session.sync_license_ledger_if_needed()
|
|
|
|
remote_realm_dict: Dict[str, RemoteRealmDictValue] = {}
|
|
remote_realms = RemoteRealm.objects.filter(server=server)
|
|
for remote_realm in remote_realms:
|
|
uuid = str(remote_realm.uuid)
|
|
billing_session = RemoteRealmBillingSession(remote_realm)
|
|
remote_realm_dict[uuid] = billing_session.get_push_service_validity_dict()
|
|
|
|
return json_success(request, data={"realms": remote_realm_dict})
|
|
|
|
|
|
def build_realm_id_to_remote_realm_dict(
|
|
server: RemoteZulipServer, realms: Optional[List[RealmDataForAnalytics]]
|
|
) -> Dict[int, Optional[RemoteRealm]]:
|
|
if realms is None:
|
|
return {}
|
|
|
|
realm_uuids = [realm.uuid for realm in realms]
|
|
remote_realms = RemoteRealm.objects.filter(uuid__in=realm_uuids, server=server)
|
|
|
|
uuid_to_remote_realm_dict = {
|
|
str(remote_realm.uuid): remote_realm for remote_realm in remote_realms
|
|
}
|
|
return {realm.id: uuid_to_remote_realm_dict[str(realm.uuid)] for realm in realms}
|
|
|
|
|
|
def fix_remote_realm_foreign_keys(
|
|
server: RemoteZulipServer, realms: List[RealmDataForAnalytics]
|
|
) -> None:
|
|
"""
|
|
Finds the RemoteRealmCount and RemoteRealmAuditLog entries without .remote_realm
|
|
set and sets it based on the "realms" data received from the remote server,
|
|
if possible.
|
|
"""
|
|
|
|
if (
|
|
not RemoteRealmCount.objects.filter(server=server, remote_realm=None).exists()
|
|
and not RemoteRealmAuditLog.objects.filter(server=server, remote_realm=None).exists()
|
|
):
|
|
return
|
|
|
|
realm_id_to_remote_realm = build_realm_id_to_remote_realm_dict(server, realms)
|
|
for realm_id in realm_id_to_remote_realm:
|
|
RemoteRealmCount.objects.filter(server=server, remote_realm=None, realm_id=realm_id).update(
|
|
remote_realm=realm_id_to_remote_realm[realm_id]
|
|
)
|
|
RemoteRealmAuditLog.objects.filter(
|
|
server=server, remote_realm=None, realm_id=realm_id
|
|
).update(remote_realm=realm_id_to_remote_realm[realm_id])
|
|
|
|
|
|
def get_last_id_from_server(server: RemoteZulipServer, model: Any) -> int:
|
|
last_count = (
|
|
model.objects.filter(server=server)
|
|
# Rows with remote_id=None are managed by the bouncer service itself,
|
|
# and thus aren't meant for syncing and should be ignored here.
|
|
.exclude(remote_id=None)
|
|
.order_by("remote_id")
|
|
.only("remote_id")
|
|
.last()
|
|
)
|
|
if last_count is not None:
|
|
return last_count.remote_id
|
|
return 0
|
|
|
|
|
|
@has_request_variables
|
|
def remote_server_check_analytics(request: HttpRequest, server: RemoteZulipServer) -> HttpResponse:
|
|
result = {
|
|
"last_realm_count_id": get_last_id_from_server(server, RemoteRealmCount),
|
|
"last_installation_count_id": get_last_id_from_server(server, RemoteInstallationCount),
|
|
"last_realmauditlog_id": get_last_id_from_server(server, RemoteRealmAuditLog),
|
|
}
|
|
return json_success(request, data=result)
|