mirror of https://github.com/zulip/zulip.git
329 lines
13 KiB
Python
329 lines
13 KiB
Python
import base64
|
|
import binascii
|
|
import os
|
|
from datetime import timedelta
|
|
from typing import List, Optional, Union
|
|
from urllib.parse import quote, urlsplit
|
|
|
|
from django.conf import settings
|
|
from django.contrib.auth.models import AnonymousUser
|
|
from django.core.files.uploadedfile import UploadedFile
|
|
from django.core.signing import BadSignature, TimestampSigner
|
|
from django.http import (
|
|
FileResponse,
|
|
HttpRequest,
|
|
HttpResponse,
|
|
HttpResponseBase,
|
|
HttpResponseForbidden,
|
|
HttpResponseNotFound,
|
|
)
|
|
from django.shortcuts import redirect
|
|
from django.urls import reverse
|
|
from django.utils.cache import patch_cache_control, patch_vary_headers
|
|
from django.utils.http import content_disposition_header
|
|
from django.utils.translation import gettext as _
|
|
|
|
from zerver.context_processors import get_valid_realm_from_request
|
|
from zerver.decorator import zulip_redirect_to_login
|
|
from zerver.lib.attachments import validate_attachment_request
|
|
from zerver.lib.exceptions import JsonableError
|
|
from zerver.lib.mime_types import guess_type
|
|
from zerver.lib.response import json_success
|
|
from zerver.lib.storage import static_path
|
|
from zerver.lib.upload import (
|
|
check_upload_within_quota,
|
|
get_public_upload_root_url,
|
|
upload_message_attachment_from_request,
|
|
)
|
|
from zerver.lib.upload.base import INLINE_MIME_TYPES
|
|
from zerver.lib.upload.local import assert_is_local_storage_path
|
|
from zerver.lib.upload.s3 import get_signed_upload_url
|
|
from zerver.models import UserProfile
|
|
|
|
|
|
def patch_disposition_header(response: HttpResponse, url: str, is_attachment: bool) -> None:
|
|
filename = os.path.basename(urlsplit(url).path)
|
|
content_disposition = content_disposition_header(is_attachment, filename)
|
|
|
|
if content_disposition is not None:
|
|
response.headers["Content-Disposition"] = content_disposition
|
|
|
|
|
|
def internal_nginx_redirect(internal_path: str, content_type: Optional[str] = None) -> HttpResponse:
|
|
# The following headers from this initial response are
|
|
# _preserved_, if present, and sent unmodified to the client;
|
|
# all other headers are overridden by the redirected URL:
|
|
# - Content-Type
|
|
# - Content-Disposition
|
|
# - Accept-Ranges
|
|
# - Set-Cookie
|
|
# - Cache-Control
|
|
# - Expires
|
|
# As such, we default to unsetting the Content-type header to
|
|
# allow nginx to set it from the static file; the caller can set
|
|
# Content-Disposition and Cache-Control on this response as they
|
|
# desire, and the client will see those values. In some cases
|
|
# (local files) we do wish to control the Content-Type, so also
|
|
# support setting it explicitly.
|
|
response = HttpResponse(content_type=content_type)
|
|
response["X-Accel-Redirect"] = internal_path
|
|
if content_type is None:
|
|
del response["Content-Type"]
|
|
return response
|
|
|
|
|
|
def serve_s3(request: HttpRequest, path_id: str, force_download: bool = False) -> HttpResponse:
|
|
url = get_signed_upload_url(path_id, force_download=force_download)
|
|
assert url.startswith("https://")
|
|
|
|
if settings.DEVELOPMENT:
|
|
# In development, we do not have the nginx server to offload
|
|
# the response to; serve a redirect to the short-lived S3 URL.
|
|
# This means the content cannot be cached by the browser, but
|
|
# this is acceptable in development.
|
|
return redirect(url)
|
|
|
|
# We over-escape the path, to work around it being impossible to
|
|
# get the _unescaped_ new internal request URI in nginx.
|
|
parsed_url = urlsplit(url)
|
|
assert parsed_url.hostname is not None
|
|
assert parsed_url.path is not None
|
|
assert parsed_url.query is not None
|
|
escaped_path_parts = parsed_url.hostname + quote(parsed_url.path) + "?" + parsed_url.query
|
|
response = internal_nginx_redirect("/internal/s3/" + escaped_path_parts)
|
|
|
|
# It is important that S3 generate both the Content-Type and
|
|
# Content-Disposition headers; when the file was uploaded, we
|
|
# stored the browser-provided value for the former, and set
|
|
# Content-Disposition according to if that was safe. As such,
|
|
# only S3 knows if a given attachment is safe to inline; we only
|
|
# override Content-Disposition to "attachment", and do so by
|
|
# telling S3 that is what we want in the signed URL.
|
|
patch_cache_control(response, private=True, immutable=True)
|
|
return response
|
|
|
|
|
|
def serve_local(
|
|
request: HttpRequest, path_id: str, force_download: bool = False
|
|
) -> HttpResponseBase:
|
|
assert settings.LOCAL_FILES_DIR is not None
|
|
local_path = os.path.join(settings.LOCAL_FILES_DIR, path_id)
|
|
assert_is_local_storage_path("files", local_path)
|
|
if not os.path.isfile(local_path):
|
|
return HttpResponseNotFound("<p>File not found</p>")
|
|
|
|
mimetype, encoding = guess_type(path_id)
|
|
download = force_download or mimetype not in INLINE_MIME_TYPES
|
|
|
|
if settings.DEVELOPMENT:
|
|
# In development, we do not have the nginx server to offload
|
|
# the response to; serve it directly ourselves. FileResponse
|
|
# handles setting Content-Type, Content-Disposition, etc.
|
|
response: HttpResponseBase = FileResponse(
|
|
open(local_path, "rb"), # noqa: SIM115
|
|
as_attachment=download,
|
|
)
|
|
patch_cache_control(response, private=True, immutable=True)
|
|
return response
|
|
|
|
# For local responses, we are in charge of generating both
|
|
# Content-Type and Content-Disposition headers; unlike with S3
|
|
# storage, the Content-Type is not stored with the file in any
|
|
# way, so Django makes the determination of it, and thus as well
|
|
# if that type is safe to have a Content-Disposition of "inline".
|
|
# nginx respects the values we send.
|
|
response = internal_nginx_redirect(
|
|
quote(f"/internal/local/uploads/{path_id}"), content_type=mimetype
|
|
)
|
|
patch_disposition_header(response, local_path, download)
|
|
patch_cache_control(response, private=True, immutable=True)
|
|
return response
|
|
|
|
|
|
def serve_file_download_backend(
|
|
request: HttpRequest,
|
|
maybe_user_profile: Union[UserProfile, AnonymousUser],
|
|
realm_id_str: str,
|
|
filename: str,
|
|
) -> HttpResponseBase:
|
|
return serve_file(
|
|
request, maybe_user_profile, realm_id_str, filename, url_only=False, force_download=True
|
|
)
|
|
|
|
|
|
def serve_file_backend(
|
|
request: HttpRequest,
|
|
maybe_user_profile: Union[UserProfile, AnonymousUser],
|
|
realm_id_str: str,
|
|
filename: str,
|
|
) -> HttpResponseBase:
|
|
return serve_file(request, maybe_user_profile, realm_id_str, filename, url_only=False)
|
|
|
|
|
|
def serve_file_url_backend(
|
|
request: HttpRequest, user_profile: UserProfile, realm_id_str: str, filename: str
|
|
) -> HttpResponseBase:
|
|
"""
|
|
We should return a signed, short-lived URL
|
|
that the client can use for native mobile download, rather than serving a redirect.
|
|
"""
|
|
|
|
return serve_file(request, user_profile, realm_id_str, filename, url_only=True)
|
|
|
|
|
|
def preferred_accept(request: HttpRequest, served_types: List[str]) -> Optional[str]:
|
|
# Returns the first of the served_types which the browser will
|
|
# accept, based on the browser's stated quality preferences.
|
|
# Returns None if none of the served_types are accepted by the
|
|
# browser.
|
|
accepted_types = sorted(
|
|
request.accepted_types,
|
|
key=lambda e: float(e.params.get("q", "1.0")),
|
|
reverse=True,
|
|
)
|
|
for potential_type in accepted_types:
|
|
for served_type in served_types:
|
|
if potential_type.match(served_type):
|
|
return served_type
|
|
return None
|
|
|
|
|
|
def serve_file(
|
|
request: HttpRequest,
|
|
maybe_user_profile: Union[UserProfile, AnonymousUser],
|
|
realm_id_str: str,
|
|
filename: str,
|
|
url_only: bool = False,
|
|
force_download: bool = False,
|
|
) -> HttpResponseBase:
|
|
path_id = f"{realm_id_str}/{filename}"
|
|
realm = get_valid_realm_from_request(request)
|
|
is_authorized = validate_attachment_request(maybe_user_profile, path_id, realm)
|
|
|
|
def serve_image_error(status: int, image_path: str) -> HttpResponseBase:
|
|
# We cannot use X-Accel-Redirect to offload the serving of
|
|
# this image to nginx, because it does not preserve the status
|
|
# code of this response, nor the Vary: header.
|
|
return FileResponse(open(static_path(image_path), "rb"), status=status) # noqa: SIM115
|
|
|
|
if is_authorized is None:
|
|
if preferred_accept(request, ["text/html", "image/png"]) == "image/png":
|
|
response = serve_image_error(404, "images/errors/image-not-exist.png")
|
|
else:
|
|
response = HttpResponseNotFound(
|
|
_("<p>This file does not exist or has been deleted.</p>")
|
|
)
|
|
patch_vary_headers(response, ("Accept",))
|
|
return response
|
|
if not is_authorized:
|
|
if preferred_accept(request, ["text/html", "image/png"]) == "image/png":
|
|
response = serve_image_error(403, "images/errors/image-no-auth.png")
|
|
elif isinstance(maybe_user_profile, AnonymousUser):
|
|
response = zulip_redirect_to_login(request)
|
|
else:
|
|
response = HttpResponseForbidden(_("<p>You are not authorized to view this file.</p>"))
|
|
patch_vary_headers(response, ("Accept",))
|
|
return response
|
|
if url_only:
|
|
url = generate_unauthed_file_access_url(path_id)
|
|
return json_success(request, data=dict(url=url))
|
|
|
|
if settings.LOCAL_UPLOADS_DIR is not None:
|
|
return serve_local(request, path_id, force_download=force_download)
|
|
else:
|
|
return serve_s3(request, path_id, force_download=force_download)
|
|
|
|
|
|
USER_UPLOADS_ACCESS_TOKEN_SALT = "user_uploads_"
|
|
|
|
|
|
def generate_unauthed_file_access_url(path_id: str) -> str:
|
|
signed_data = TimestampSigner(salt=USER_UPLOADS_ACCESS_TOKEN_SALT).sign(path_id)
|
|
token = base64.b16encode(signed_data.encode()).decode()
|
|
|
|
filename = path_id.split("/")[-1]
|
|
return reverse("file_unauthed_from_token", args=[token, filename])
|
|
|
|
|
|
def get_file_path_id_from_token(token: str) -> Optional[str]:
|
|
signer = TimestampSigner(salt=USER_UPLOADS_ACCESS_TOKEN_SALT)
|
|
try:
|
|
signed_data = base64.b16decode(token).decode()
|
|
path_id = signer.unsign(
|
|
signed_data, max_age=timedelta(seconds=settings.SIGNED_ACCESS_TOKEN_VALIDITY_IN_SECONDS)
|
|
)
|
|
except (BadSignature, binascii.Error):
|
|
return None
|
|
|
|
return path_id
|
|
|
|
|
|
def serve_file_unauthed_from_token(
|
|
request: HttpRequest, token: str, filename: str
|
|
) -> HttpResponseBase:
|
|
path_id = get_file_path_id_from_token(token)
|
|
if path_id is None:
|
|
raise JsonableError(_("Invalid token"))
|
|
if path_id.split("/")[-1] != filename:
|
|
raise JsonableError(_("Invalid filename"))
|
|
|
|
if settings.LOCAL_UPLOADS_DIR is not None:
|
|
return serve_local(request, path_id)
|
|
else:
|
|
return serve_s3(request, path_id)
|
|
|
|
|
|
def serve_local_avatar_unauthed(request: HttpRequest, path: str) -> HttpResponseBase:
|
|
"""Serves avatar images off disk, via nginx (or directly in dev), with no auth.
|
|
|
|
This is done unauthed because these need to be accessed from HTML
|
|
emails, where the client does not have any auth. We rely on the
|
|
URL being generated using the AVATAR_SALT secret.
|
|
|
|
"""
|
|
if settings.LOCAL_AVATARS_DIR is None:
|
|
# We do not expect clients to hit this URL when using the S3
|
|
# backend; however, there is no reason to not serve the
|
|
# redirect to S3 where the content lives.
|
|
return redirect(
|
|
get_public_upload_root_url() + path + "?" + request.GET.urlencode(), permanent=True
|
|
)
|
|
|
|
local_path = os.path.join(settings.LOCAL_AVATARS_DIR, path)
|
|
assert_is_local_storage_path("avatars", local_path)
|
|
if not os.path.isfile(local_path):
|
|
return HttpResponseNotFound("<p>File not found</p>")
|
|
|
|
if settings.DEVELOPMENT:
|
|
response: HttpResponseBase = FileResponse(open(local_path, "rb")) # noqa: SIM115
|
|
else:
|
|
response = internal_nginx_redirect(quote(f"/internal/local/user_avatars/{path}"))
|
|
|
|
# We do _not_ mark the contents as immutable for caching purposes,
|
|
# since the path for avatar images is hashed only by their user-id
|
|
# and a salt, and as such are reused when a user's avatar is
|
|
# updated.
|
|
return response
|
|
|
|
|
|
def upload_file_backend(request: HttpRequest, user_profile: UserProfile) -> HttpResponse:
|
|
if len(request.FILES) == 0:
|
|
raise JsonableError(_("You must specify a file to upload"))
|
|
if len(request.FILES) != 1:
|
|
raise JsonableError(_("You may only upload one file at a time"))
|
|
|
|
[user_file] = request.FILES.values()
|
|
assert isinstance(user_file, UploadedFile)
|
|
file_size = user_file.size
|
|
assert file_size is not None
|
|
if file_size > settings.MAX_FILE_UPLOAD_SIZE * 1024 * 1024:
|
|
raise JsonableError(
|
|
_("Uploaded file is larger than the allowed limit of {max_size} MiB").format(
|
|
max_size=settings.MAX_FILE_UPLOAD_SIZE,
|
|
)
|
|
)
|
|
check_upload_within_quota(user_profile.realm, file_size)
|
|
|
|
uri = upload_message_attachment_from_request(user_file, user_profile, file_size)
|
|
return json_success(request, data={"uri": uri})
|