2020-06-11 00:54:34 +02:00
|
|
|
from typing import Any, Dict, List, Optional
|
2020-05-26 07:16:25 +02:00
|
|
|
from unittest.mock import Mock, patch
|
2020-06-11 00:54:34 +02:00
|
|
|
|
|
|
|
from django.conf import settings
|
2018-05-14 18:48:18 +02:00
|
|
|
|
|
|
|
from zerver.apps import flush_cache
|
2020-06-11 00:54:34 +02:00
|
|
|
from zerver.lib.cache import (
|
|
|
|
MEMCACHED_MAX_KEY_LENGTH,
|
|
|
|
InvalidCacheKeyException,
|
|
|
|
NotFoundInCache,
|
2020-07-01 03:29:31 +02:00
|
|
|
bulk_cached_fetch,
|
2020-06-11 00:54:34 +02:00
|
|
|
cache_delete,
|
|
|
|
cache_delete_many,
|
|
|
|
cache_get,
|
|
|
|
cache_get_many,
|
|
|
|
cache_set,
|
|
|
|
cache_set_many,
|
|
|
|
cache_with_key,
|
|
|
|
get_cache_with_key,
|
|
|
|
safe_cache_get_many,
|
|
|
|
safe_cache_set_many,
|
|
|
|
user_profile_by_email_cache_key,
|
|
|
|
validate_cache_key,
|
|
|
|
)
|
2018-05-14 18:48:18 +02:00
|
|
|
from zerver.lib.test_classes import ZulipTestCase
|
2019-12-16 05:53:54 +01:00
|
|
|
from zerver.lib.test_helpers import queries_captured
|
2020-06-11 00:54:34 +02:00
|
|
|
from zerver.models import UserProfile, get_system_bot, get_user_profile_by_email
|
|
|
|
|
2018-05-14 18:48:18 +02:00
|
|
|
|
|
|
|
class AppsTest(ZulipTestCase):
|
|
|
|
def test_cache_gets_flushed(self) -> None:
|
|
|
|
with patch('zerver.apps.logging.info') as mock_logging:
|
|
|
|
with patch('zerver.apps.cache.clear') as mock:
|
|
|
|
# The argument to flush_cache doesn't matter
|
|
|
|
flush_cache(Mock())
|
|
|
|
mock.assert_called_once()
|
|
|
|
mock_logging.assert_called_once()
|
2019-07-13 00:11:58 +02:00
|
|
|
|
2019-12-16 05:53:54 +01:00
|
|
|
class CacheKeyValidationTest(ZulipTestCase):
|
|
|
|
def test_validate_cache_key(self) -> None:
|
|
|
|
validate_cache_key('nice_Ascii:string!~')
|
|
|
|
with self.assertRaises(InvalidCacheKeyException):
|
|
|
|
validate_cache_key('utf8_character:ą')
|
|
|
|
with self.assertRaises(InvalidCacheKeyException):
|
|
|
|
validate_cache_key('new_line_character:\n')
|
|
|
|
with self.assertRaises(InvalidCacheKeyException):
|
|
|
|
validate_cache_key('control_character:\r')
|
|
|
|
with self.assertRaises(InvalidCacheKeyException):
|
|
|
|
validate_cache_key('whitespace_character: ')
|
|
|
|
with self.assertRaises(InvalidCacheKeyException):
|
|
|
|
validate_cache_key('too_long:' + 'X'*MEMCACHED_MAX_KEY_LENGTH)
|
|
|
|
|
|
|
|
with self.assertRaises(InvalidCacheKeyException):
|
|
|
|
# validate_cache_key does validation on a key with the
|
|
|
|
# KEY_PREFIX appended to the start, so even though we're
|
|
|
|
# passing something "short enough" here, it becomes too
|
|
|
|
# long after appending KEY_PREFIX.
|
|
|
|
validate_cache_key('X' * (MEMCACHED_MAX_KEY_LENGTH - 2))
|
|
|
|
|
|
|
|
def test_cache_functions_raise_exception(self) -> None:
|
|
|
|
invalid_key = 'invalid_character:\n'
|
|
|
|
good_key = "good_key"
|
|
|
|
with self.assertRaises(InvalidCacheKeyException):
|
|
|
|
cache_get(invalid_key)
|
|
|
|
with self.assertRaises(InvalidCacheKeyException):
|
|
|
|
cache_set(invalid_key, 0)
|
|
|
|
with self.assertRaises(InvalidCacheKeyException):
|
|
|
|
cache_delete(invalid_key)
|
|
|
|
|
|
|
|
with self.assertRaises(InvalidCacheKeyException):
|
|
|
|
cache_get_many([good_key, invalid_key])
|
|
|
|
with self.assertRaises(InvalidCacheKeyException):
|
|
|
|
cache_set_many({good_key: 0, invalid_key: 1})
|
|
|
|
with self.assertRaises(InvalidCacheKeyException):
|
|
|
|
cache_delete_many([good_key, invalid_key])
|
|
|
|
|
|
|
|
class CacheWithKeyDecoratorTest(ZulipTestCase):
|
|
|
|
def test_cache_with_key_invalid_character(self) -> None:
|
|
|
|
def invalid_characters_cache_key_function(user_id: int) -> str:
|
2020-06-09 00:25:09 +02:00
|
|
|
return f'CacheWithKeyDecoratorTest:invalid_character:ą:{user_id}'
|
2019-12-16 05:53:54 +01:00
|
|
|
|
|
|
|
@cache_with_key(invalid_characters_cache_key_function, timeout=1000)
|
|
|
|
def get_user_function_with_bad_cache_keys(user_id: int) -> UserProfile:
|
|
|
|
return UserProfile.objects.get(id=user_id)
|
|
|
|
|
|
|
|
hamlet = self.example_user('hamlet')
|
|
|
|
with patch('zerver.lib.cache.cache_set') as mock_set, \
|
|
|
|
patch('zerver.lib.cache.logger.warning') as mock_warn:
|
|
|
|
with queries_captured() as queries:
|
|
|
|
result = get_user_function_with_bad_cache_keys(hamlet.id)
|
|
|
|
|
|
|
|
self.assertEqual(result, hamlet)
|
|
|
|
self.assert_length(queries, 1)
|
|
|
|
mock_set.assert_not_called()
|
|
|
|
mock_warn.assert_called_once()
|
|
|
|
|
|
|
|
def test_cache_with_key_key_too_long(self) -> None:
|
|
|
|
def too_long_cache_key_function(user_id: int) -> str:
|
|
|
|
return 'CacheWithKeyDecoratorTest:very_long_key:{}:{}'.format('a'*250, user_id)
|
|
|
|
|
|
|
|
@cache_with_key(too_long_cache_key_function, timeout=1000)
|
|
|
|
def get_user_function_with_bad_cache_keys(user_id: int) -> UserProfile:
|
|
|
|
return UserProfile.objects.get(id=user_id)
|
|
|
|
|
|
|
|
hamlet = self.example_user('hamlet')
|
|
|
|
|
|
|
|
with patch('zerver.lib.cache.cache_set') as mock_set, \
|
|
|
|
patch('zerver.lib.cache.logger.warning') as mock_warn:
|
|
|
|
with queries_captured() as queries:
|
|
|
|
result = get_user_function_with_bad_cache_keys(hamlet.id)
|
|
|
|
|
|
|
|
self.assertEqual(result, hamlet)
|
|
|
|
self.assert_length(queries, 1)
|
|
|
|
mock_set.assert_not_called()
|
|
|
|
mock_warn.assert_called_once()
|
|
|
|
|
|
|
|
def test_cache_with_key_good_key(self) -> None:
|
|
|
|
def good_cache_key_function(user_id: int) -> str:
|
2020-06-09 00:25:09 +02:00
|
|
|
return f'CacheWithKeyDecoratorTest:good_cache_key:{user_id}'
|
2019-12-16 05:53:54 +01:00
|
|
|
|
|
|
|
@cache_with_key(good_cache_key_function, timeout=1000)
|
|
|
|
def get_user_function_with_good_cache_keys(user_id: int) -> UserProfile:
|
|
|
|
return UserProfile.objects.get(id=user_id)
|
|
|
|
|
|
|
|
hamlet = self.example_user('hamlet')
|
|
|
|
|
|
|
|
with queries_captured() as queries:
|
|
|
|
result = get_user_function_with_good_cache_keys(hamlet.id)
|
|
|
|
|
|
|
|
self.assertEqual(result, hamlet)
|
|
|
|
self.assert_length(queries, 1)
|
|
|
|
|
|
|
|
# The previous function call should have cached the result correctly, so now
|
|
|
|
# no database queries should happen:
|
|
|
|
with queries_captured() as queries_two:
|
|
|
|
result_two = get_user_function_with_good_cache_keys(hamlet.id)
|
|
|
|
|
|
|
|
self.assertEqual(result_two, hamlet)
|
|
|
|
self.assert_length(queries_two, 0)
|
|
|
|
|
2020-02-21 15:40:59 +01:00
|
|
|
def test_cache_with_key_none_values(self) -> None:
|
|
|
|
def cache_key_function(user_id: int) -> str:
|
2020-06-09 00:25:09 +02:00
|
|
|
return f'CacheWithKeyDecoratorTest:test_cache_with_key_none_values:{user_id}'
|
2020-02-21 15:40:59 +01:00
|
|
|
|
|
|
|
@cache_with_key(cache_key_function, timeout=1000)
|
|
|
|
def get_user_function_can_return_none(user_id: int) -> Optional[UserProfile]:
|
|
|
|
try:
|
|
|
|
return UserProfile.objects.get(id=user_id)
|
|
|
|
except UserProfile.DoesNotExist:
|
|
|
|
return None
|
|
|
|
|
|
|
|
last_user_id = UserProfile.objects.last().id
|
|
|
|
with queries_captured() as queries:
|
|
|
|
result = get_user_function_can_return_none(last_user_id + 1)
|
|
|
|
|
|
|
|
self.assertEqual(result, None)
|
|
|
|
self.assert_length(queries, 1)
|
|
|
|
|
|
|
|
with queries_captured() as queries:
|
|
|
|
result_two = get_user_function_can_return_none(last_user_id + 1)
|
|
|
|
|
|
|
|
self.assertEqual(result_two, None)
|
|
|
|
self.assert_length(queries, 0)
|
|
|
|
|
2019-12-16 05:53:54 +01:00
|
|
|
class GetCacheWithKeyDecoratorTest(ZulipTestCase):
|
|
|
|
def test_get_cache_with_good_key(self) -> None:
|
|
|
|
# Test with a good cache key function, but a get_user function
|
|
|
|
# that always returns None just to make it convenient to tell
|
|
|
|
# whether the cache was used (whatever we put in the cache) or
|
|
|
|
# we got the result from calling the function (None)
|
|
|
|
|
|
|
|
def good_cache_key_function(user_id: int) -> str:
|
2020-06-09 00:25:09 +02:00
|
|
|
return f'CacheWithKeyDecoratorTest:good_cache_key:{user_id}'
|
2019-12-16 05:53:54 +01:00
|
|
|
|
|
|
|
@get_cache_with_key(good_cache_key_function)
|
|
|
|
def get_user_function_with_good_cache_keys(user_id: int) -> Any: # nocoverage
|
|
|
|
return
|
|
|
|
|
|
|
|
hamlet = self.example_user('hamlet')
|
|
|
|
with patch('zerver.lib.cache.logger.warning') as mock_warn:
|
|
|
|
with self.assertRaises(NotFoundInCache):
|
|
|
|
get_user_function_with_good_cache_keys(hamlet.id)
|
|
|
|
mock_warn.assert_not_called()
|
|
|
|
|
|
|
|
cache_set(good_cache_key_function(hamlet.id), hamlet)
|
|
|
|
result = get_user_function_with_good_cache_keys(hamlet.id)
|
|
|
|
self.assertEqual(result, hamlet)
|
|
|
|
|
|
|
|
def test_get_cache_with_bad_key(self) -> None:
|
|
|
|
def bad_cache_key_function(user_id: int) -> str:
|
2020-06-09 00:25:09 +02:00
|
|
|
return f'CacheWithKeyDecoratorTest:invalid_character:ą:{user_id}'
|
2019-12-16 05:53:54 +01:00
|
|
|
|
|
|
|
@get_cache_with_key(bad_cache_key_function)
|
|
|
|
def get_user_function_with_bad_cache_keys(user_id: int) -> Any: # nocoverage
|
|
|
|
return
|
|
|
|
|
|
|
|
hamlet = self.example_user('hamlet')
|
|
|
|
with patch('zerver.lib.cache.logger.warning') as mock_warn:
|
|
|
|
with self.assertRaises(NotFoundInCache):
|
|
|
|
get_user_function_with_bad_cache_keys(hamlet.id)
|
|
|
|
mock_warn.assert_called_once()
|
|
|
|
|
|
|
|
class SafeCacheFunctionsTest(ZulipTestCase):
|
|
|
|
def test_safe_cache_functions_with_all_good_keys(self) -> None:
|
|
|
|
items = {"SafeFunctionsTest:key1": 1, "SafeFunctionsTest:key2": 2, "SafeFunctionsTest:key3": 3}
|
|
|
|
safe_cache_set_many(items)
|
|
|
|
|
|
|
|
result = safe_cache_get_many(list(items.keys()))
|
|
|
|
for key, value in result.items():
|
|
|
|
self.assertEqual(value, items[key])
|
|
|
|
|
|
|
|
def test_safe_cache_functions_with_all_bad_keys(self) -> None:
|
|
|
|
items = {"SafeFunctionsTest:\nbadkey1": 1, "SafeFunctionsTest:\nbadkey2": 2}
|
|
|
|
with patch('zerver.lib.cache.logger.warning') as mock_warn:
|
|
|
|
safe_cache_set_many(items)
|
|
|
|
mock_warn.assert_called_once()
|
2020-05-02 08:44:14 +02:00
|
|
|
self.assertEqual(
|
|
|
|
mock_warn.call_args[0][1],
|
|
|
|
['SafeFunctionsTest:\nbadkey1', 'SafeFunctionsTest:\nbadkey2'],
|
|
|
|
)
|
2019-12-16 05:53:54 +01:00
|
|
|
|
|
|
|
with patch('zerver.lib.cache.logger.warning') as mock_warn:
|
|
|
|
result = safe_cache_get_many(list(items.keys()))
|
|
|
|
mock_warn.assert_called_once()
|
2020-05-02 08:44:14 +02:00
|
|
|
self.assertEqual(
|
|
|
|
mock_warn.call_args[0][1],
|
|
|
|
['SafeFunctionsTest:\nbadkey1', 'SafeFunctionsTest:\nbadkey2'],
|
|
|
|
)
|
2019-12-16 05:53:54 +01:00
|
|
|
|
|
|
|
self.assertEqual(result, {})
|
|
|
|
|
|
|
|
def test_safe_cache_functions_with_good_and_bad_keys(self) -> None:
|
|
|
|
bad_items = {"SafeFunctionsTest:\nbadkey1": 1, "SafeFunctionsTest:\nbadkey2": 2}
|
|
|
|
good_items = {"SafeFunctionsTest:goodkey1": 3, "SafeFunctionsTest:goodkey2": 4}
|
|
|
|
items = {**good_items, **bad_items}
|
|
|
|
|
|
|
|
with patch('zerver.lib.cache.logger.warning') as mock_warn:
|
|
|
|
safe_cache_set_many(items)
|
|
|
|
mock_warn.assert_called_once()
|
2020-05-02 08:44:14 +02:00
|
|
|
self.assertEqual(
|
|
|
|
mock_warn.call_args[0][1],
|
|
|
|
['SafeFunctionsTest:\nbadkey1', 'SafeFunctionsTest:\nbadkey2'],
|
|
|
|
)
|
2019-12-16 05:53:54 +01:00
|
|
|
|
|
|
|
with patch('zerver.lib.cache.logger.warning') as mock_warn:
|
|
|
|
result = safe_cache_get_many(list(items.keys()))
|
|
|
|
mock_warn.assert_called_once()
|
2020-05-02 08:44:14 +02:00
|
|
|
self.assertEqual(
|
|
|
|
mock_warn.call_args[0][1],
|
|
|
|
['SafeFunctionsTest:\nbadkey1', 'SafeFunctionsTest:\nbadkey2'],
|
|
|
|
)
|
2019-12-16 05:53:54 +01:00
|
|
|
|
|
|
|
self.assertEqual(result, good_items)
|
|
|
|
|
2019-07-13 00:11:58 +02:00
|
|
|
class BotCacheKeyTest(ZulipTestCase):
|
|
|
|
def test_bot_profile_key_deleted_on_save(self) -> None:
|
|
|
|
# Get the profile cached on both cache keys:
|
|
|
|
user_profile = get_user_profile_by_email(settings.EMAIL_GATEWAY_BOT)
|
|
|
|
bot_profile = get_system_bot(settings.EMAIL_GATEWAY_BOT)
|
|
|
|
self.assertEqual(user_profile, bot_profile)
|
|
|
|
|
|
|
|
# Flip the setting and save:
|
|
|
|
flipped_setting = not bot_profile.is_api_super_user
|
|
|
|
bot_profile.is_api_super_user = flipped_setting
|
|
|
|
bot_profile.save()
|
|
|
|
|
|
|
|
# The .save() should have deleted cache keys, so if we fetch again,
|
|
|
|
# the returned objects should have is_api_super_user set correctly.
|
|
|
|
bot_profile2 = get_system_bot(settings.EMAIL_GATEWAY_BOT)
|
|
|
|
self.assertEqual(bot_profile2.is_api_super_user, flipped_setting)
|
|
|
|
|
|
|
|
user_profile2 = get_user_profile_by_email(settings.EMAIL_GATEWAY_BOT)
|
|
|
|
self.assertEqual(user_profile2.is_api_super_user, flipped_setting)
|
2019-08-10 23:31:14 +02:00
|
|
|
|
2020-07-01 03:29:31 +02:00
|
|
|
def get_user_email(user: UserProfile) -> str:
|
|
|
|
return user.email # nocoverage
|
|
|
|
|
2019-08-10 23:31:14 +02:00
|
|
|
class GenericBulkCachedFetchTest(ZulipTestCase):
|
|
|
|
def test_query_function_called_only_if_needed(self) -> None:
|
|
|
|
# Get the user cached:
|
|
|
|
hamlet = get_user_profile_by_email(self.example_email("hamlet"))
|
|
|
|
|
|
|
|
class CustomException(Exception):
|
|
|
|
pass
|
|
|
|
|
|
|
|
def query_function(emails: List[str]) -> List[UserProfile]:
|
|
|
|
raise CustomException("The query function was called")
|
|
|
|
|
|
|
|
# query_function shouldn't be called, because the only requested object
|
|
|
|
# is already cached:
|
2020-07-01 03:29:31 +02:00
|
|
|
result: Dict[str, UserProfile] = bulk_cached_fetch(
|
2019-08-10 23:31:14 +02:00
|
|
|
cache_key_function=user_profile_by_email_cache_key,
|
|
|
|
query_function=query_function,
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
object_ids=[self.example_email("hamlet")],
|
2020-07-01 03:29:31 +02:00
|
|
|
id_fetcher=get_user_email,
|
python: Convert assignment type annotations to Python 3.6 style.
This commit was split by tabbott; this piece covers the vast majority
of files in Zulip, but excludes scripts/, tools/, and puppet/ to help
ensure we at least show the right error messages for Xenial systems.
We can likely further refine the remaining pieces with some testing.
Generated by com2ann, with whitespace fixes and various manual fixes
for runtime issues:
- invoiced_through: Optional[LicenseLedger] = models.ForeignKey(
+ invoiced_through: Optional["LicenseLedger"] = models.ForeignKey(
-_apns_client: Optional[APNsClient] = None
+_apns_client: Optional["APNsClient"] = None
- notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- signup_notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ signup_notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- author: Optional[UserProfile] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
+ author: Optional["UserProfile"] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
- bot_owner: Optional[UserProfile] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
+ bot_owner: Optional["UserProfile"] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
- default_sending_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
- default_events_register_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_sending_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_events_register_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
-descriptors_by_handler_id: Dict[int, ClientDescriptor] = {}
+descriptors_by_handler_id: Dict[int, "ClientDescriptor"] = {}
-worker_classes: Dict[str, Type[QueueProcessingWorker]] = {}
-queues: Dict[str, Dict[str, Type[QueueProcessingWorker]]] = {}
+worker_classes: Dict[str, Type["QueueProcessingWorker"]] = {}
+queues: Dict[str, Dict[str, Type["QueueProcessingWorker"]]] = {}
-AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional[LDAPSearch] = None
+AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional["LDAPSearch"] = None
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-22 01:09:50 +02:00
|
|
|
)
|
2020-03-12 14:17:25 +01:00
|
|
|
self.assertEqual(result, {hamlet.delivery_email: hamlet})
|
2019-08-10 23:31:14 +02:00
|
|
|
|
|
|
|
flush_cache(Mock())
|
|
|
|
# With the cache flushed, the query_function should get called:
|
|
|
|
with self.assertRaises(CustomException):
|
2020-07-01 03:29:31 +02:00
|
|
|
result = bulk_cached_fetch(
|
2019-08-10 23:31:14 +02:00
|
|
|
cache_key_function=user_profile_by_email_cache_key,
|
|
|
|
query_function=query_function,
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
object_ids=[self.example_email("hamlet")],
|
2020-07-01 03:29:31 +02:00
|
|
|
id_fetcher=get_user_email,
|
2019-08-10 23:31:14 +02:00
|
|
|
)
|
|
|
|
|
|
|
|
def test_empty_object_ids_list(self) -> None:
|
|
|
|
class CustomException(Exception):
|
|
|
|
pass
|
|
|
|
|
|
|
|
def cache_key_function(email: str) -> str: # nocoverage -- this is just here to make sure it's not called
|
|
|
|
raise CustomException("The cache key function was called")
|
|
|
|
|
|
|
|
def query_function(emails: List[str]) -> List[UserProfile]: # nocoverage -- this is just here to make sure it's not called
|
|
|
|
raise CustomException("The query function was called")
|
|
|
|
|
|
|
|
# query_function and cache_key_function shouldn't be called, because
|
|
|
|
# objects_ids is empty, so there's nothing to do.
|
2020-07-01 03:29:31 +02:00
|
|
|
result: Dict[str, UserProfile] = bulk_cached_fetch(
|
2019-08-10 23:31:14 +02:00
|
|
|
cache_key_function=cache_key_function,
|
|
|
|
query_function=query_function,
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
object_ids=[],
|
2020-07-01 03:29:31 +02:00
|
|
|
id_fetcher=get_user_email,
|
python: Convert assignment type annotations to Python 3.6 style.
This commit was split by tabbott; this piece covers the vast majority
of files in Zulip, but excludes scripts/, tools/, and puppet/ to help
ensure we at least show the right error messages for Xenial systems.
We can likely further refine the remaining pieces with some testing.
Generated by com2ann, with whitespace fixes and various manual fixes
for runtime issues:
- invoiced_through: Optional[LicenseLedger] = models.ForeignKey(
+ invoiced_through: Optional["LicenseLedger"] = models.ForeignKey(
-_apns_client: Optional[APNsClient] = None
+_apns_client: Optional["APNsClient"] = None
- notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- signup_notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ signup_notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- author: Optional[UserProfile] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
+ author: Optional["UserProfile"] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
- bot_owner: Optional[UserProfile] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
+ bot_owner: Optional["UserProfile"] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
- default_sending_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
- default_events_register_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_sending_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_events_register_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
-descriptors_by_handler_id: Dict[int, ClientDescriptor] = {}
+descriptors_by_handler_id: Dict[int, "ClientDescriptor"] = {}
-worker_classes: Dict[str, Type[QueueProcessingWorker]] = {}
-queues: Dict[str, Dict[str, Type[QueueProcessingWorker]]] = {}
+worker_classes: Dict[str, Type["QueueProcessingWorker"]] = {}
+queues: Dict[str, Dict[str, Type["QueueProcessingWorker"]]] = {}
-AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional[LDAPSearch] = None
+AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional["LDAPSearch"] = None
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-22 01:09:50 +02:00
|
|
|
)
|
2019-08-10 23:31:14 +02:00
|
|
|
self.assertEqual(result, {})
|