import logging import os from concurrent.futures import ProcessPoolExecutor, as_completed from mimetypes import guess_type import bmemcached from django.conf import settings from django.core.cache import cache from django.db import connection from zerver.lib.avatar_hash import user_avatar_path from zerver.lib.upload import S3UploadBackend, upload_image_to_s3 from zerver.models import Attachment, RealmEmoji, UserProfile s3backend = S3UploadBackend() def transfer_uploads_to_s3(processes: int) -> None: # TODO: Eventually, we'll want to add realm icon and logo transfer_avatars_to_s3(processes) transfer_message_files_to_s3(processes) transfer_emoji_to_s3(processes) def _transfer_avatar_to_s3(user: UserProfile) -> None: avatar_path = user_avatar_path(user) assert settings.LOCAL_UPLOADS_DIR is not None file_path = os.path.join(settings.LOCAL_UPLOADS_DIR, "avatars", avatar_path) + ".original" try: with open(file_path, "rb") as f: s3backend.upload_avatar_image(f, user, user) logging.info("Uploaded avatar for %s in realm %s", user.id, user.realm.name) except FileNotFoundError: pass def transfer_avatars_to_s3(processes: int) -> None: users = list(UserProfile.objects.all()) if processes == 1: for user in users: _transfer_avatar_to_s3(user) else: # nocoverage connection.close() _cache = cache._cache # type: ignore[attr-defined] # not in stubs assert isinstance(_cache, bmemcached.Client) _cache.disconnect_all() with ProcessPoolExecutor(max_workers=processes) as executor: for future in as_completed( executor.submit(_transfer_avatar_to_s3, user) for user in users ): future.result() def _transfer_message_files_to_s3(attachment: Attachment) -> None: assert settings.LOCAL_UPLOADS_DIR is not None file_path = os.path.join(settings.LOCAL_UPLOADS_DIR, "files", attachment.path_id) try: with open(file_path, "rb") as f: guessed_type = guess_type(attachment.file_name)[0] upload_image_to_s3( s3backend.uploads_bucket, attachment.path_id, guessed_type, attachment.owner, f.read(), ) logging.info("Uploaded message file in path %s", file_path) except FileNotFoundError: # nocoverage pass def transfer_message_files_to_s3(processes: int) -> None: attachments = list(Attachment.objects.all()) if processes == 1: for attachment in attachments: _transfer_message_files_to_s3(attachment) else: # nocoverage connection.close() _cache = cache._cache # type: ignore[attr-defined] # not in stubs assert isinstance(_cache, bmemcached.Client) _cache.disconnect_all() with ProcessPoolExecutor(max_workers=processes) as executor: for future in as_completed( executor.submit(_transfer_message_files_to_s3, attachment) for attachment in attachments ): future.result() def _transfer_emoji_to_s3(realm_emoji: RealmEmoji) -> None: if not realm_emoji.file_name or not realm_emoji.author: return # nocoverage emoji_path = RealmEmoji.PATH_ID_TEMPLATE.format( realm_id=realm_emoji.realm.id, emoji_file_name=realm_emoji.file_name, ) assert settings.LOCAL_UPLOADS_DIR is not None emoji_path = os.path.join(settings.LOCAL_UPLOADS_DIR, "avatars", emoji_path) + ".original" try: with open(emoji_path, "rb") as f: s3backend.upload_emoji_image(f, realm_emoji.file_name, realm_emoji.author) logging.info("Uploaded emoji file in path %s", emoji_path) except FileNotFoundError: # nocoverage pass def transfer_emoji_to_s3(processes: int) -> None: realm_emojis = list(RealmEmoji.objects.filter()) if processes == 1: for realm_emoji in realm_emojis: _transfer_emoji_to_s3(realm_emoji) else: # nocoverage connection.close() _cache = cache._cache # type: ignore[attr-defined] # not in stubs assert isinstance(_cache, bmemcached.Client) _cache.disconnect_all() with ProcessPoolExecutor(max_workers=processes) as executor: for future in as_completed( executor.submit(_transfer_emoji_to_s3, realm_emoji) for realm_emoji in realm_emojis ): future.result()