2020-06-11 00:54:34 +02:00
|
|
|
import inspect
|
2019-07-04 18:12:53 +02:00
|
|
|
import re
|
2019-07-10 13:23:25 +02:00
|
|
|
import sys
|
2020-06-11 00:54:34 +02:00
|
|
|
from typing import (
|
|
|
|
Any,
|
|
|
|
Callable,
|
|
|
|
Dict,
|
|
|
|
Iterable,
|
|
|
|
List,
|
|
|
|
Mapping,
|
|
|
|
Optional,
|
|
|
|
Sequence,
|
|
|
|
Set,
|
|
|
|
Tuple,
|
|
|
|
Union,
|
|
|
|
)
|
2020-05-26 07:16:25 +02:00
|
|
|
from unittest import mock
|
2020-06-11 00:54:34 +02:00
|
|
|
from unittest.mock import MagicMock, patch
|
2018-05-31 19:41:17 +02:00
|
|
|
|
2019-07-10 13:23:25 +02:00
|
|
|
from django.http import HttpResponse
|
2019-06-06 22:22:21 +02:00
|
|
|
|
2020-06-11 00:54:34 +02:00
|
|
|
from zerver.lib.request import _REQ, arguments_map
|
2018-05-31 19:41:17 +02:00
|
|
|
from zerver.lib.test_classes import ZulipTestCase
|
2020-06-11 00:54:34 +02:00
|
|
|
from zerver.openapi import openapi as openapi
|
|
|
|
from zerver.openapi.markdown_extension import (
|
|
|
|
generate_curl_example,
|
|
|
|
parse_language_and_options,
|
|
|
|
render_curl_example,
|
|
|
|
)
|
2020-02-23 18:10:42 +01:00
|
|
|
from zerver.openapi.openapi import (
|
2020-06-11 00:54:34 +02:00
|
|
|
OPENAPI_SPEC_PATH,
|
|
|
|
OpenAPISpec,
|
|
|
|
SchemaError,
|
|
|
|
get_openapi_fixture,
|
|
|
|
get_openapi_parameters,
|
|
|
|
get_openapi_paths,
|
|
|
|
openapi_spec,
|
|
|
|
to_python_type,
|
|
|
|
validate_against_openapi_schema,
|
2018-05-31 19:41:17 +02:00
|
|
|
)
|
|
|
|
|
|
|
|
TEST_ENDPOINT = '/messages/{message_id}'
|
|
|
|
TEST_METHOD = 'patch'
|
2018-06-18 16:32:30 +02:00
|
|
|
TEST_RESPONSE_BAD_REQ = '400'
|
|
|
|
TEST_RESPONSE_SUCCESS = '200'
|
2018-05-31 19:41:17 +02:00
|
|
|
|
2019-07-10 13:23:25 +02:00
|
|
|
VARMAP = {
|
|
|
|
'integer': int,
|
|
|
|
'string': str,
|
|
|
|
'boolean': bool,
|
|
|
|
'array': list,
|
|
|
|
'Typing.List': list,
|
2019-08-04 15:55:32 +02:00
|
|
|
'object': dict,
|
2019-08-10 00:30:34 +02:00
|
|
|
'NoneType': type(None),
|
2019-07-10 13:23:25 +02:00
|
|
|
}
|
2018-05-31 19:41:17 +02:00
|
|
|
|
|
|
|
class OpenAPIToolsTest(ZulipTestCase):
|
|
|
|
"""Make sure that the tools we use to handle our OpenAPI specification
|
2020-02-23 18:10:42 +01:00
|
|
|
(located in zerver/openapi/openapi.py) work as expected.
|
2018-05-31 19:41:17 +02:00
|
|
|
|
|
|
|
These tools are mostly dedicated to fetching parts of the -already parsed-
|
|
|
|
specification, and comparing them to objects returned by our REST API.
|
|
|
|
"""
|
2020-04-22 01:45:30 +02:00
|
|
|
|
2018-05-31 19:41:17 +02:00
|
|
|
def test_get_openapi_fixture(self) -> None:
|
2018-06-18 16:32:30 +02:00
|
|
|
actual = get_openapi_fixture(TEST_ENDPOINT, TEST_METHOD,
|
|
|
|
TEST_RESPONSE_BAD_REQ)
|
2018-05-31 19:41:17 +02:00
|
|
|
expected = {
|
|
|
|
'code': 'BAD_REQUEST',
|
|
|
|
'msg': 'You don\'t have permission to edit this message',
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
'result': 'error',
|
2018-05-31 19:41:17 +02:00
|
|
|
}
|
|
|
|
self.assertEqual(actual, expected)
|
|
|
|
|
|
|
|
def test_get_openapi_parameters(self) -> None:
|
|
|
|
actual = get_openapi_parameters(TEST_ENDPOINT, TEST_METHOD)
|
|
|
|
expected_item = {
|
|
|
|
'name': 'message_id',
|
|
|
|
'in': 'path',
|
|
|
|
'description':
|
2020-04-20 16:46:14 +02:00
|
|
|
'The target message\'s ID.\n',
|
2018-05-31 19:41:17 +02:00
|
|
|
'example': 42,
|
|
|
|
'required': True,
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
'schema': {'type': 'integer'},
|
2018-05-31 19:41:17 +02:00
|
|
|
}
|
|
|
|
assert(expected_item in actual)
|
|
|
|
|
|
|
|
def test_validate_against_openapi_schema(self) -> None:
|
|
|
|
with self.assertRaises(SchemaError,
|
|
|
|
msg=('Extraneous key "foo" in '
|
|
|
|
'the response\'scontent')):
|
python: Convert assignment type annotations to Python 3.6 style.
This commit was split by tabbott; this piece covers the vast majority
of files in Zulip, but excludes scripts/, tools/, and puppet/ to help
ensure we at least show the right error messages for Xenial systems.
We can likely further refine the remaining pieces with some testing.
Generated by com2ann, with whitespace fixes and various manual fixes
for runtime issues:
- invoiced_through: Optional[LicenseLedger] = models.ForeignKey(
+ invoiced_through: Optional["LicenseLedger"] = models.ForeignKey(
-_apns_client: Optional[APNsClient] = None
+_apns_client: Optional["APNsClient"] = None
- notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- signup_notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ signup_notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- author: Optional[UserProfile] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
+ author: Optional["UserProfile"] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
- bot_owner: Optional[UserProfile] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
+ bot_owner: Optional["UserProfile"] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
- default_sending_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
- default_events_register_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_sending_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_events_register_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
-descriptors_by_handler_id: Dict[int, ClientDescriptor] = {}
+descriptors_by_handler_id: Dict[int, "ClientDescriptor"] = {}
-worker_classes: Dict[str, Type[QueueProcessingWorker]] = {}
-queues: Dict[str, Dict[str, Type[QueueProcessingWorker]]] = {}
+worker_classes: Dict[str, Type["QueueProcessingWorker"]] = {}
+queues: Dict[str, Dict[str, Type["QueueProcessingWorker"]]] = {}
-AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional[LDAPSearch] = None
+AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional["LDAPSearch"] = None
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-22 01:09:50 +02:00
|
|
|
bad_content: Dict[str, object] = {
|
2018-05-31 19:41:17 +02:00
|
|
|
'msg': '',
|
|
|
|
'result': 'success',
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
'foo': 'bar',
|
python: Convert assignment type annotations to Python 3.6 style.
This commit was split by tabbott; this piece covers the vast majority
of files in Zulip, but excludes scripts/, tools/, and puppet/ to help
ensure we at least show the right error messages for Xenial systems.
We can likely further refine the remaining pieces with some testing.
Generated by com2ann, with whitespace fixes and various manual fixes
for runtime issues:
- invoiced_through: Optional[LicenseLedger] = models.ForeignKey(
+ invoiced_through: Optional["LicenseLedger"] = models.ForeignKey(
-_apns_client: Optional[APNsClient] = None
+_apns_client: Optional["APNsClient"] = None
- notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- signup_notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ signup_notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- author: Optional[UserProfile] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
+ author: Optional["UserProfile"] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
- bot_owner: Optional[UserProfile] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
+ bot_owner: Optional["UserProfile"] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
- default_sending_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
- default_events_register_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_sending_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_events_register_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
-descriptors_by_handler_id: Dict[int, ClientDescriptor] = {}
+descriptors_by_handler_id: Dict[int, "ClientDescriptor"] = {}
-worker_classes: Dict[str, Type[QueueProcessingWorker]] = {}
-queues: Dict[str, Dict[str, Type[QueueProcessingWorker]]] = {}
+worker_classes: Dict[str, Type["QueueProcessingWorker"]] = {}
+queues: Dict[str, Dict[str, Type["QueueProcessingWorker"]]] = {}
-AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional[LDAPSearch] = None
+AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional["LDAPSearch"] = None
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-22 01:09:50 +02:00
|
|
|
}
|
2018-05-31 19:41:17 +02:00
|
|
|
validate_against_openapi_schema(bad_content,
|
|
|
|
TEST_ENDPOINT,
|
2018-06-18 16:32:30 +02:00
|
|
|
TEST_METHOD,
|
|
|
|
TEST_RESPONSE_SUCCESS)
|
2018-05-31 19:41:17 +02:00
|
|
|
|
|
|
|
with self.assertRaises(SchemaError,
|
|
|
|
msg=("Expected type <class 'str'> for key "
|
|
|
|
"\"msg\", but actually got "
|
|
|
|
"<class 'int'>")):
|
|
|
|
bad_content = {
|
|
|
|
'msg': 42,
|
|
|
|
'result': 'success',
|
|
|
|
}
|
|
|
|
validate_against_openapi_schema(bad_content,
|
|
|
|
TEST_ENDPOINT,
|
2018-06-18 16:32:30 +02:00
|
|
|
TEST_METHOD,
|
|
|
|
TEST_RESPONSE_SUCCESS)
|
2018-05-31 19:41:17 +02:00
|
|
|
|
|
|
|
with self.assertRaises(SchemaError,
|
|
|
|
msg='Expected to find the "msg" required key'):
|
|
|
|
bad_content = {
|
|
|
|
'result': 'success',
|
|
|
|
}
|
|
|
|
validate_against_openapi_schema(bad_content,
|
|
|
|
TEST_ENDPOINT,
|
2018-06-18 16:32:30 +02:00
|
|
|
TEST_METHOD,
|
|
|
|
TEST_RESPONSE_SUCCESS)
|
2018-05-31 19:41:17 +02:00
|
|
|
|
|
|
|
# No exceptions should be raised here.
|
|
|
|
good_content = {
|
|
|
|
'msg': '',
|
|
|
|
'result': 'success',
|
|
|
|
}
|
|
|
|
validate_against_openapi_schema(good_content,
|
|
|
|
TEST_ENDPOINT,
|
2018-06-18 16:32:30 +02:00
|
|
|
TEST_METHOD,
|
|
|
|
TEST_RESPONSE_SUCCESS)
|
2018-05-31 19:41:17 +02:00
|
|
|
|
2018-06-20 19:31:24 +02:00
|
|
|
# Overwrite the exception list with a mocked one
|
|
|
|
openapi.EXCLUDE_PROPERTIES = {
|
|
|
|
TEST_ENDPOINT: {
|
|
|
|
TEST_METHOD: {
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
TEST_RESPONSE_SUCCESS: ['foo'],
|
|
|
|
},
|
|
|
|
},
|
2018-06-20 19:31:24 +02:00
|
|
|
}
|
|
|
|
good_content = {
|
|
|
|
'msg': '',
|
|
|
|
'result': 'success',
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
'foo': 'bar',
|
2018-06-20 19:31:24 +02:00
|
|
|
}
|
|
|
|
validate_against_openapi_schema(good_content,
|
|
|
|
TEST_ENDPOINT,
|
|
|
|
TEST_METHOD,
|
|
|
|
TEST_RESPONSE_SUCCESS)
|
|
|
|
|
2018-05-31 19:41:17 +02:00
|
|
|
def test_to_python_type(self) -> None:
|
|
|
|
TYPES = {
|
|
|
|
'string': str,
|
|
|
|
'number': float,
|
|
|
|
'integer': int,
|
|
|
|
'boolean': bool,
|
|
|
|
'array': list,
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
'object': dict,
|
2018-05-31 19:41:17 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
for oa_type, py_type in TYPES.items():
|
|
|
|
self.assertEqual(to_python_type(oa_type), py_type)
|
2018-08-07 23:40:07 +02:00
|
|
|
|
|
|
|
def test_live_reload(self) -> None:
|
|
|
|
# Force the reload by making the last update date < the file's last
|
|
|
|
# modified date
|
|
|
|
openapi_spec.last_update = 0
|
|
|
|
get_openapi_fixture(TEST_ENDPOINT, TEST_METHOD)
|
|
|
|
|
|
|
|
# Check that the file has been reloaded by verifying that the last
|
|
|
|
# update date isn't zero anymore
|
|
|
|
self.assertNotEqual(openapi_spec.last_update, 0)
|
2018-08-08 01:35:41 +02:00
|
|
|
|
|
|
|
# Now verify calling it again doesn't call reload
|
2020-02-23 18:10:42 +01:00
|
|
|
with mock.patch('zerver.openapi.openapi.openapi_spec.reload') as mock_reload:
|
2018-08-08 01:35:41 +02:00
|
|
|
get_openapi_fixture(TEST_ENDPOINT, TEST_METHOD)
|
|
|
|
self.assertFalse(mock_reload.called)
|
2019-06-06 22:22:21 +02:00
|
|
|
|
|
|
|
class OpenAPIArgumentsTest(ZulipTestCase):
|
2019-07-07 08:54:19 +02:00
|
|
|
# This will be filled during test_openapi_arguments:
|
python: Convert assignment type annotations to Python 3.6 style.
This commit was split by tabbott; this piece covers the vast majority
of files in Zulip, but excludes scripts/, tools/, and puppet/ to help
ensure we at least show the right error messages for Xenial systems.
We can likely further refine the remaining pieces with some testing.
Generated by com2ann, with whitespace fixes and various manual fixes
for runtime issues:
- invoiced_through: Optional[LicenseLedger] = models.ForeignKey(
+ invoiced_through: Optional["LicenseLedger"] = models.ForeignKey(
-_apns_client: Optional[APNsClient] = None
+_apns_client: Optional["APNsClient"] = None
- notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- signup_notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ signup_notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- author: Optional[UserProfile] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
+ author: Optional["UserProfile"] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
- bot_owner: Optional[UserProfile] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
+ bot_owner: Optional["UserProfile"] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
- default_sending_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
- default_events_register_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_sending_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_events_register_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
-descriptors_by_handler_id: Dict[int, ClientDescriptor] = {}
+descriptors_by_handler_id: Dict[int, "ClientDescriptor"] = {}
-worker_classes: Dict[str, Type[QueueProcessingWorker]] = {}
-queues: Dict[str, Dict[str, Type[QueueProcessingWorker]]] = {}
+worker_classes: Dict[str, Type["QueueProcessingWorker"]] = {}
+queues: Dict[str, Dict[str, Type["QueueProcessingWorker"]]] = {}
-AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional[LDAPSearch] = None
+AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional["LDAPSearch"] = None
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-22 01:09:50 +02:00
|
|
|
checked_endpoints: Set[str] = set()
|
2020-04-09 21:51:58 +02:00
|
|
|
pending_endpoints = {
|
2020-01-31 22:51:57 +01:00
|
|
|
#### TODO: These endpoints are a priority to document:
|
|
|
|
'/messages/matches_narrow',
|
|
|
|
'/realm/presence',
|
|
|
|
'/streams/{stream_id}/members',
|
|
|
|
'/streams/{stream_id}/delete_topic',
|
2019-07-07 08:54:19 +02:00
|
|
|
'/users/me/presence',
|
|
|
|
'/users/me/alert_words',
|
|
|
|
'/users/me/status',
|
2020-01-31 22:51:57 +01:00
|
|
|
|
|
|
|
#### These realm administration settings are valuable to document:
|
|
|
|
# List all files uploaded by current user. May want to add support
|
|
|
|
# for a larger list available to administrators?
|
2019-07-07 08:54:19 +02:00
|
|
|
'/attachments',
|
2020-01-31 22:51:57 +01:00
|
|
|
# Delete a file uploaded by current user.
|
|
|
|
'/attachments/{attachment_id}',
|
|
|
|
# List data exports for organization (GET) or request one (POST)
|
2019-07-07 08:54:19 +02:00
|
|
|
'/export/realm',
|
2020-01-31 22:51:57 +01:00
|
|
|
# Delete a data export.
|
2019-08-01 19:59:36 +02:00
|
|
|
'/export/realm/{export_id}',
|
2020-01-31 22:51:57 +01:00
|
|
|
# Manage default streams and default stream groups
|
|
|
|
'/default_streams',
|
|
|
|
'/default_stream_groups/create',
|
|
|
|
'/default_stream_groups/{group_id}',
|
|
|
|
'/default_stream_groups/{group_id}/streams',
|
|
|
|
# Administer invitations
|
2019-07-07 08:54:19 +02:00
|
|
|
'/invites',
|
|
|
|
'/invites/multiuse',
|
2020-01-31 22:51:57 +01:00
|
|
|
'/invites/{prereg_id}',
|
|
|
|
'/invites/{prereg_id}/resend',
|
|
|
|
'/invites/multiuse/{invite_id}',
|
|
|
|
# Single-stream settings alternative to the bulk endpoint
|
|
|
|
# users/me/subscriptions/properties; probably should just be a
|
|
|
|
# section of the same page.
|
|
|
|
'/users/me/subscriptions/{stream_id}',
|
|
|
|
|
2020-03-18 02:40:44 +01:00
|
|
|
# Real-time-events endpoint
|
|
|
|
'/real-time',
|
|
|
|
|
2020-04-02 02:15:28 +02:00
|
|
|
# Rest error handling endpoint
|
|
|
|
'/rest-error-handling',
|
|
|
|
|
2020-04-02 03:42:59 +02:00
|
|
|
# Zulip outgoing webhook payload
|
|
|
|
'/zulip-outgoing-webhook',
|
|
|
|
|
2020-01-31 22:51:57 +01:00
|
|
|
#### Mobile-app only endpoints; important for mobile developers.
|
|
|
|
# Mobile interface for fetching API keys
|
|
|
|
'/fetch_api_key',
|
|
|
|
# Already documented; need to fix tracking bug
|
|
|
|
'/dev_fetch_api_key',
|
|
|
|
# Mobile interface for development environment login
|
|
|
|
'/dev_list_users',
|
|
|
|
# Registration for iOS/Android mobile push notifications.
|
2019-07-07 08:54:19 +02:00
|
|
|
'/users/me/android_gcm_reg_id',
|
|
|
|
'/users/me/apns_device_token',
|
2020-01-31 22:51:57 +01:00
|
|
|
|
|
|
|
#### These personal settings endpoints have modest value to document:
|
|
|
|
'/settings',
|
|
|
|
'/users/me/avatar',
|
|
|
|
'/users/me/api_key/regenerate',
|
|
|
|
# Not very useful outside the UI
|
|
|
|
'/settings/display',
|
|
|
|
# Much more valuable would be an org admin bulk-upload feature.
|
|
|
|
'/users/me/profile_data',
|
|
|
|
# To be deprecated and deleted.
|
|
|
|
'/users/me/pointer',
|
|
|
|
|
|
|
|
#### Should be documented as part of interactive bots documentation
|
|
|
|
'/bot_storage',
|
|
|
|
'/submessage',
|
|
|
|
'/zcommand',
|
|
|
|
|
|
|
|
#### These "organization settings" endpoint have modest value to document:
|
|
|
|
'/realm',
|
|
|
|
'/realm/domains',
|
2019-07-11 12:45:26 +02:00
|
|
|
'/realm/domains/{domain}',
|
2020-01-31 22:51:57 +01:00
|
|
|
'/bots',
|
|
|
|
'/bots/{bot_id}',
|
|
|
|
'/bots/{bot_id}/api_key/regenerate',
|
|
|
|
#### These "organization settings" endpoints have low value to document:
|
|
|
|
'/realm/profile_fields',
|
2019-07-11 12:45:26 +02:00
|
|
|
'/realm/profile_fields/{field_id}',
|
2020-01-31 22:51:57 +01:00
|
|
|
'/realm/icon',
|
|
|
|
'/realm/logo',
|
|
|
|
'/realm/deactivate',
|
2019-08-20 00:33:09 +02:00
|
|
|
'/realm/subdomain/{subdomain}',
|
2020-01-31 22:51:57 +01:00
|
|
|
|
|
|
|
#### Other low value endpoints
|
|
|
|
# Used for dead desktop app to test connectivity. To delete.
|
|
|
|
'/generate_204',
|
|
|
|
# Used for failed approach with dead Android app.
|
|
|
|
'/fetch_google_client_id',
|
|
|
|
# API for video calls we're planning to remove/replace.
|
2019-11-16 09:26:28 +01:00
|
|
|
'/calls/zoom/create',
|
2020-01-31 22:51:57 +01:00
|
|
|
|
|
|
|
#### Documented endpoints not properly detected by tooling.
|
|
|
|
# E.g. '/user_groups/<user_group_id>' in urls.py but fails the
|
|
|
|
# reverse mapping test because of the variable name
|
|
|
|
# mismatch.
|
|
|
|
'/user_groups/{group_id}', # Equivalent of what's in urls.py
|
|
|
|
'/user_groups/{user_group_id}', # What's in the OpenAPI docs
|
2019-07-11 12:45:26 +02:00
|
|
|
'/user_groups/{user_group_id}/members',
|
2019-07-07 08:54:19 +02:00
|
|
|
# Regex with an unnamed capturing group.
|
|
|
|
'/users/(?!me/)(?P<email>[^/]*)/presence',
|
2020-04-09 21:51:58 +02:00
|
|
|
}
|
2019-10-22 01:43:54 +02:00
|
|
|
|
|
|
|
# Endpoints where the documentation is currently failing our
|
|
|
|
# consistency tests. We aim to keep this list empty.
|
python: Convert assignment type annotations to Python 3.6 style.
This commit was split by tabbott; this piece covers the vast majority
of files in Zulip, but excludes scripts/, tools/, and puppet/ to help
ensure we at least show the right error messages for Xenial systems.
We can likely further refine the remaining pieces with some testing.
Generated by com2ann, with whitespace fixes and various manual fixes
for runtime issues:
- invoiced_through: Optional[LicenseLedger] = models.ForeignKey(
+ invoiced_through: Optional["LicenseLedger"] = models.ForeignKey(
-_apns_client: Optional[APNsClient] = None
+_apns_client: Optional["APNsClient"] = None
- notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- signup_notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ signup_notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- author: Optional[UserProfile] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
+ author: Optional["UserProfile"] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
- bot_owner: Optional[UserProfile] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
+ bot_owner: Optional["UserProfile"] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
- default_sending_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
- default_events_register_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_sending_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_events_register_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
-descriptors_by_handler_id: Dict[int, ClientDescriptor] = {}
+descriptors_by_handler_id: Dict[int, "ClientDescriptor"] = {}
-worker_classes: Dict[str, Type[QueueProcessingWorker]] = {}
-queues: Dict[str, Dict[str, Type[QueueProcessingWorker]]] = {}
+worker_classes: Dict[str, Type["QueueProcessingWorker"]] = {}
+queues: Dict[str, Dict[str, Type["QueueProcessingWorker"]]] = {}
-AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional[LDAPSearch] = None
+AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional["LDAPSearch"] = None
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-22 01:09:50 +02:00
|
|
|
buggy_documentation_endpoints: Set[str] = set([
|
|
|
|
])
|
2019-07-07 08:54:19 +02:00
|
|
|
|
2019-07-09 08:28:29 +02:00
|
|
|
def convert_regex_to_url_pattern(self, regex_pattern: str) -> str:
|
|
|
|
""" Convert regular expressions style URL patterns to their
|
|
|
|
corresponding OpenAPI style formats. All patterns are
|
|
|
|
expected to start with ^ and end with $.
|
|
|
|
Examples:
|
|
|
|
1. /messages/{message_id} <-> r'^messages/(?P<message_id>[0-9]+)$'
|
|
|
|
2. /events <-> r'^events$'
|
2020-05-07 13:56:49 +02:00
|
|
|
3. '/realm/domains' <-> r'/realm\\/domains$'
|
2019-07-09 08:28:29 +02:00
|
|
|
"""
|
2019-08-20 00:15:11 +02:00
|
|
|
|
|
|
|
# TODO: Probably we should be able to address the below
|
|
|
|
# through alternative solutions (e.g. reordering urls.py
|
|
|
|
# entries or similar url organization, but for now these let
|
|
|
|
# us test more endpoints and so are worth doing).
|
|
|
|
me_pattern = '/(?!me/)'
|
|
|
|
if me_pattern in regex_pattern:
|
|
|
|
# Remove the exclude-me pattern if present.
|
|
|
|
regex_pattern = regex_pattern.replace(me_pattern, "/")
|
2020-05-07 13:56:49 +02:00
|
|
|
|
|
|
|
# Handle the presence-email code which has a non-slashes syntax.
|
|
|
|
regex_pattern = regex_pattern.replace('[^/]*', '.*').replace('[^/]+', '.*')
|
2019-08-20 00:15:11 +02:00
|
|
|
|
2019-07-09 08:28:29 +02:00
|
|
|
self.assertTrue(regex_pattern.startswith("^"))
|
|
|
|
self.assertTrue(regex_pattern.endswith("$"))
|
|
|
|
url_pattern = '/' + regex_pattern[1:][:-1]
|
|
|
|
url_pattern = re.sub(r"\(\?P<(\w+)>[^/]+\)", r"{\1}", url_pattern)
|
2020-05-07 13:56:49 +02:00
|
|
|
url_pattern = url_pattern.replace('\\', '')
|
2019-07-09 08:28:29 +02:00
|
|
|
return url_pattern
|
|
|
|
|
|
|
|
def ensure_no_documentation_if_intentionally_undocumented(self, url_pattern: str,
|
2019-07-20 20:16:47 +02:00
|
|
|
method: str,
|
|
|
|
msg: Optional[str]=None) -> None:
|
2019-07-09 08:28:29 +02:00
|
|
|
try:
|
|
|
|
get_openapi_parameters(url_pattern, method)
|
2019-07-19 07:02:10 +02:00
|
|
|
if not msg: # nocoverage
|
2019-07-20 20:16:47 +02:00
|
|
|
msg = """
|
|
|
|
We found some OpenAPI documentation for {method} {url_pattern},
|
|
|
|
so maybe we shouldn't mark it as intentionally undocumented in the urls.
|
|
|
|
""".format(method=method, url_pattern=url_pattern)
|
|
|
|
raise AssertionError(msg) # nocoverage
|
2019-07-09 08:28:29 +02:00
|
|
|
except KeyError:
|
|
|
|
return
|
|
|
|
|
|
|
|
def check_for_non_existant_openapi_endpoints(self) -> None:
|
|
|
|
""" Here, we check to see if every endpoint documented in the openapi
|
|
|
|
documentation actually exists in urls.py and thus in actual code.
|
|
|
|
Note: We define this as a helper called at the end of
|
|
|
|
test_openapi_arguments instead of as a separate test to ensure that
|
|
|
|
this test is only executed after test_openapi_arguments so that it's
|
|
|
|
results can be used here in the set operations. """
|
|
|
|
openapi_paths = set(get_openapi_paths())
|
|
|
|
undocumented_paths = openapi_paths - self.checked_endpoints
|
|
|
|
undocumented_paths -= self.buggy_documentation_endpoints
|
|
|
|
undocumented_paths -= self.pending_endpoints
|
|
|
|
try:
|
|
|
|
self.assertEqual(len(undocumented_paths), 0)
|
|
|
|
except AssertionError: # nocoverage
|
|
|
|
msg = "The following endpoints have been documented but can't be found in urls.py:"
|
|
|
|
for undocumented_path in undocumented_paths:
|
2020-06-09 00:25:09 +02:00
|
|
|
msg += f"\n + {undocumented_path}"
|
2019-07-09 08:28:29 +02:00
|
|
|
raise AssertionError(msg)
|
|
|
|
|
2019-08-10 00:30:34 +02:00
|
|
|
def get_type_by_priority(self, types: Sequence[Union[type, Tuple[type, object]]]) -> Union[type, Tuple[type, object]]:
|
2019-08-04 15:55:32 +02:00
|
|
|
priority = {list: 1, dict: 2, str: 3, int: 4, bool: 5}
|
|
|
|
tyiroirp = {1: list, 2: dict, 3: str, 4: int, 5: bool}
|
|
|
|
val = 6
|
2019-07-10 13:23:25 +02:00
|
|
|
for t in types:
|
2019-08-10 00:30:34 +02:00
|
|
|
if isinstance(t, tuple):
|
2019-08-04 15:55:32 +02:00
|
|
|
return t # e.g. (list, dict) or (list ,str)
|
2019-08-10 00:30:34 +02:00
|
|
|
v = priority.get(t, 6)
|
2019-07-10 13:23:25 +02:00
|
|
|
if v < val:
|
|
|
|
val = v
|
|
|
|
return tyiroirp.get(val, types[0])
|
|
|
|
|
2019-08-10 00:30:34 +02:00
|
|
|
def get_standardized_argument_type(self, t: Any) -> Union[type, Tuple[type, object]]:
|
2019-07-10 13:23:25 +02:00
|
|
|
""" Given a type from the typing module such as List[str] or Union[str, int],
|
|
|
|
convert it into a corresponding Python type. Unions are mapped to a canonical
|
|
|
|
choice among the options.
|
|
|
|
E.g. typing.Union[typing.List[typing.Dict[str, typing.Any]], NoneType]
|
|
|
|
needs to be mapped to list."""
|
|
|
|
|
2020-02-11 21:39:37 +01:00
|
|
|
if sys.version_info < (3, 7): # nocoverage # python 3.5-3.6
|
2020-04-22 01:59:09 +02:00
|
|
|
if sys.version_info < (3, 6) and isinstance(t, type(Union)): # python 3.5 has special consideration for Union
|
2020-02-11 21:39:37 +01:00
|
|
|
origin = Union
|
|
|
|
else:
|
|
|
|
origin = getattr(t, "__origin__", None)
|
|
|
|
else: # nocoverage # python3.7+
|
2020-02-11 07:20:25 +01:00
|
|
|
origin = getattr(t, "__origin__", None)
|
|
|
|
t_name = getattr(t, "_name", None)
|
|
|
|
if origin == list:
|
|
|
|
origin = List
|
|
|
|
elif origin == dict:
|
|
|
|
origin = Dict
|
|
|
|
elif t_name == "Iterable":
|
|
|
|
origin = Iterable
|
|
|
|
elif t_name == "Mapping":
|
|
|
|
origin = Mapping
|
|
|
|
|
2019-07-10 13:23:25 +02:00
|
|
|
if not origin:
|
|
|
|
# Then it's most likely one of the fundamental data types
|
|
|
|
# I.E. Not one of the data types from the "typing" module.
|
|
|
|
return t
|
|
|
|
elif origin == Union:
|
|
|
|
subtypes = []
|
2019-08-10 00:30:34 +02:00
|
|
|
if sys.version_info < (3, 6): # nocoverage # in python3.6+
|
2019-07-10 13:23:25 +02:00
|
|
|
args = t.__union_params__
|
|
|
|
else: # nocoverage # in python3.5
|
|
|
|
args = t.__args__
|
|
|
|
for st in args:
|
2019-08-04 15:55:32 +02:00
|
|
|
subtypes.append(self.get_standardized_argument_type(st))
|
2019-07-10 13:23:25 +02:00
|
|
|
return self.get_type_by_priority(subtypes)
|
2019-08-04 15:55:32 +02:00
|
|
|
elif origin in [List, Iterable]:
|
|
|
|
subtypes = [self.get_standardized_argument_type(st) for st in t.__args__]
|
2019-08-10 00:30:34 +02:00
|
|
|
return (list, self.get_type_by_priority(subtypes))
|
2019-08-04 15:55:32 +02:00
|
|
|
elif origin in [Dict, Mapping]:
|
|
|
|
return dict
|
|
|
|
return self.get_standardized_argument_type(t.__args__[0])
|
2019-07-10 13:23:25 +02:00
|
|
|
|
|
|
|
def render_openapi_type_exception(self, function: Callable[..., HttpResponse],
|
2019-08-10 00:30:34 +02:00
|
|
|
openapi_params: Set[Tuple[str, Union[type, Tuple[type, object]]]],
|
|
|
|
function_params: Set[Tuple[str, Union[type, Tuple[type, object]]]],
|
|
|
|
diff: Set[Tuple[str, Union[type, Tuple[type, object]]]]) -> None: # nocoverage
|
2019-07-10 13:23:25 +02:00
|
|
|
""" Print a *VERY* clear and verbose error message for when the types
|
|
|
|
(between the OpenAPI documentation and the function declaration) don't match. """
|
|
|
|
|
|
|
|
msg = """
|
|
|
|
The types for the request parameters in zerver/openapi/zulip.yaml
|
|
|
|
do not match the types declared in the implementation of {}.\n""".format(function.__name__)
|
|
|
|
msg += '='*65 + '\n'
|
|
|
|
msg += "{:<10s}{:^30s}{:>10s}\n".format("Parameter", "OpenAPI Type",
|
|
|
|
"Function Declaration Type")
|
|
|
|
msg += '='*65 + '\n'
|
|
|
|
opvtype = None
|
|
|
|
fdvtype = None
|
|
|
|
for element in diff:
|
|
|
|
vname = element[0]
|
|
|
|
for element in openapi_params:
|
|
|
|
if element[0] == vname:
|
|
|
|
opvtype = element[1]
|
|
|
|
break
|
|
|
|
for element in function_params:
|
|
|
|
if element[0] == vname:
|
|
|
|
fdvtype = element[1]
|
|
|
|
break
|
2020-06-09 00:25:09 +02:00
|
|
|
msg += f"{vname:<10s}{str(opvtype):^30s}{str(fdvtype):>10s}\n"
|
2019-07-10 13:23:25 +02:00
|
|
|
raise AssertionError(msg)
|
|
|
|
|
|
|
|
def check_argument_types(self, function: Callable[..., HttpResponse],
|
|
|
|
openapi_parameters: List[Dict[str, Any]]) -> None:
|
|
|
|
""" We construct for both the OpenAPI data and the function's definition a set of
|
|
|
|
tuples of the form (var_name, type) and then compare those sets to see if the
|
|
|
|
OpenAPI data defines a different type than that actually accepted by the function.
|
|
|
|
Otherwise, we print out the exact differences for convenient debugging and raise an
|
|
|
|
AssertionError. """
|
python: Convert assignment type annotations to Python 3.6 style.
This commit was split by tabbott; this piece covers the vast majority
of files in Zulip, but excludes scripts/, tools/, and puppet/ to help
ensure we at least show the right error messages for Xenial systems.
We can likely further refine the remaining pieces with some testing.
Generated by com2ann, with whitespace fixes and various manual fixes
for runtime issues:
- invoiced_through: Optional[LicenseLedger] = models.ForeignKey(
+ invoiced_through: Optional["LicenseLedger"] = models.ForeignKey(
-_apns_client: Optional[APNsClient] = None
+_apns_client: Optional["APNsClient"] = None
- notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- signup_notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ signup_notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- author: Optional[UserProfile] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
+ author: Optional["UserProfile"] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
- bot_owner: Optional[UserProfile] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
+ bot_owner: Optional["UserProfile"] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
- default_sending_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
- default_events_register_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_sending_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_events_register_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
-descriptors_by_handler_id: Dict[int, ClientDescriptor] = {}
+descriptors_by_handler_id: Dict[int, "ClientDescriptor"] = {}
-worker_classes: Dict[str, Type[QueueProcessingWorker]] = {}
-queues: Dict[str, Dict[str, Type[QueueProcessingWorker]]] = {}
+worker_classes: Dict[str, Type["QueueProcessingWorker"]] = {}
+queues: Dict[str, Dict[str, Type["QueueProcessingWorker"]]] = {}
-AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional[LDAPSearch] = None
+AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional["LDAPSearch"] = None
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-22 01:09:50 +02:00
|
|
|
openapi_params: Set[Tuple[str, Union[type, Tuple[type, object]]]] = set()
|
2019-08-04 15:55:32 +02:00
|
|
|
for element in openapi_parameters:
|
python: Convert assignment type annotations to Python 3.6 style.
This commit was split by tabbott; this piece covers the vast majority
of files in Zulip, but excludes scripts/, tools/, and puppet/ to help
ensure we at least show the right error messages for Xenial systems.
We can likely further refine the remaining pieces with some testing.
Generated by com2ann, with whitespace fixes and various manual fixes
for runtime issues:
- invoiced_through: Optional[LicenseLedger] = models.ForeignKey(
+ invoiced_through: Optional["LicenseLedger"] = models.ForeignKey(
-_apns_client: Optional[APNsClient] = None
+_apns_client: Optional["APNsClient"] = None
- notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- signup_notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ signup_notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- author: Optional[UserProfile] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
+ author: Optional["UserProfile"] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
- bot_owner: Optional[UserProfile] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
+ bot_owner: Optional["UserProfile"] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
- default_sending_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
- default_events_register_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_sending_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_events_register_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
-descriptors_by_handler_id: Dict[int, ClientDescriptor] = {}
+descriptors_by_handler_id: Dict[int, "ClientDescriptor"] = {}
-worker_classes: Dict[str, Type[QueueProcessingWorker]] = {}
-queues: Dict[str, Dict[str, Type[QueueProcessingWorker]]] = {}
+worker_classes: Dict[str, Type["QueueProcessingWorker"]] = {}
+queues: Dict[str, Dict[str, Type["QueueProcessingWorker"]]] = {}
-AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional[LDAPSearch] = None
+AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional["LDAPSearch"] = None
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-22 01:09:50 +02:00
|
|
|
name: str = element["name"]
|
2020-05-11 16:26:33 +02:00
|
|
|
schema = {}
|
|
|
|
if "content" in element:
|
|
|
|
schema = element["content"]["application/json"]["schema"]
|
|
|
|
# If content_type is application/json then the
|
|
|
|
# data type is essentially string.
|
|
|
|
openapi_params.add((name, VARMAP["string"]))
|
|
|
|
continue
|
|
|
|
|
|
|
|
else:
|
|
|
|
schema = element["schema"]
|
2020-01-28 07:28:22 +01:00
|
|
|
if 'oneOf' in schema:
|
|
|
|
# Hack: Just use the type of the first value
|
|
|
|
# Ideally, we'd turn this into a Union type.
|
|
|
|
_type = VARMAP[schema['oneOf'][0]['type']]
|
|
|
|
else:
|
|
|
|
_type = VARMAP[schema["type"]]
|
2019-08-04 15:55:32 +02:00
|
|
|
if _type == list:
|
2020-01-28 07:28:22 +01:00
|
|
|
items = schema["items"]
|
2019-08-04 15:55:32 +02:00
|
|
|
if "anyOf" in items.keys():
|
|
|
|
subtypes = []
|
|
|
|
for st in items["anyOf"]:
|
|
|
|
st = st["type"]
|
|
|
|
subtypes.append(VARMAP[st])
|
|
|
|
self.assertTrue(len(subtypes) > 1)
|
2019-08-10 00:30:34 +02:00
|
|
|
sub_type = self.get_type_by_priority(subtypes)
|
2020-04-09 19:07:57 +02:00
|
|
|
elif "oneOf" in items.keys():
|
|
|
|
sub_type = VARMAP[element["schema"]["items"]["oneOf"][0]["type"]]
|
|
|
|
self.assertIsNotNone(sub_type)
|
2019-08-04 15:55:32 +02:00
|
|
|
else:
|
2020-05-11 16:26:33 +02:00
|
|
|
sub_type = VARMAP[schema["items"]["type"]]
|
2019-08-04 15:55:32 +02:00
|
|
|
self.assertIsNotNone(sub_type)
|
2019-08-10 00:30:34 +02:00
|
|
|
openapi_params.add((name, (_type, sub_type)))
|
2019-08-04 15:55:32 +02:00
|
|
|
else:
|
2019-08-10 00:30:34 +02:00
|
|
|
openapi_params.add((name, _type))
|
2019-08-04 15:55:32 +02:00
|
|
|
|
python: Convert assignment type annotations to Python 3.6 style.
This commit was split by tabbott; this piece covers the vast majority
of files in Zulip, but excludes scripts/, tools/, and puppet/ to help
ensure we at least show the right error messages for Xenial systems.
We can likely further refine the remaining pieces with some testing.
Generated by com2ann, with whitespace fixes and various manual fixes
for runtime issues:
- invoiced_through: Optional[LicenseLedger] = models.ForeignKey(
+ invoiced_through: Optional["LicenseLedger"] = models.ForeignKey(
-_apns_client: Optional[APNsClient] = None
+_apns_client: Optional["APNsClient"] = None
- notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- signup_notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ signup_notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- author: Optional[UserProfile] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
+ author: Optional["UserProfile"] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
- bot_owner: Optional[UserProfile] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
+ bot_owner: Optional["UserProfile"] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
- default_sending_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
- default_events_register_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_sending_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_events_register_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
-descriptors_by_handler_id: Dict[int, ClientDescriptor] = {}
+descriptors_by_handler_id: Dict[int, "ClientDescriptor"] = {}
-worker_classes: Dict[str, Type[QueueProcessingWorker]] = {}
-queues: Dict[str, Dict[str, Type[QueueProcessingWorker]]] = {}
+worker_classes: Dict[str, Type["QueueProcessingWorker"]] = {}
+queues: Dict[str, Dict[str, Type["QueueProcessingWorker"]]] = {}
-AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional[LDAPSearch] = None
+AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional["LDAPSearch"] = None
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-22 01:09:50 +02:00
|
|
|
function_params: Set[Tuple[str, Union[type, Tuple[type, object]]]] = set()
|
2019-07-10 13:23:25 +02:00
|
|
|
|
|
|
|
# Iterate through the decorators to find the original
|
|
|
|
# function, wrapped by has_request_variables, so we can parse
|
|
|
|
# its arguments.
|
|
|
|
while getattr(function, "__wrapped__", None):
|
|
|
|
function = getattr(function, "__wrapped__", None)
|
|
|
|
# Tell mypy this is never None.
|
|
|
|
assert function is not None
|
|
|
|
|
|
|
|
# Now, we do inference mapping each REQ parameter's
|
|
|
|
# declaration details to the Python/mypy types for the
|
|
|
|
# arguments passed to it.
|
|
|
|
#
|
|
|
|
# Because the mypy types are the types used inside the inner
|
|
|
|
# function (after the original data is processed by any
|
|
|
|
# validators, converters, etc.), they will not always match
|
|
|
|
# the API-level argument types. The main case where this
|
|
|
|
# happens is when a `converter` is used that changes the types
|
|
|
|
# of its parameters.
|
|
|
|
for vname, defval in inspect.signature(function).parameters.items():
|
|
|
|
defval = defval.default
|
2019-08-07 11:15:46 +02:00
|
|
|
if defval.__class__ is _REQ:
|
2019-07-10 13:23:25 +02:00
|
|
|
# TODO: The below inference logic in cases where
|
|
|
|
# there's a converter function declared is incorrect.
|
|
|
|
# Theoretically, we could restructure the converter
|
|
|
|
# function model so that we can check what type it
|
|
|
|
# excepts to be passed to make validation here
|
|
|
|
# possible.
|
|
|
|
|
|
|
|
vtype = self.get_standardized_argument_type(function.__annotations__[vname])
|
2020-04-22 04:13:37 +02:00
|
|
|
vname = defval.post_var_name # type: ignore[attr-defined] # See zerver/lib/request.py
|
2019-07-10 13:23:25 +02:00
|
|
|
function_params.add((vname, vtype))
|
|
|
|
|
|
|
|
diff = openapi_params - function_params
|
|
|
|
if diff: # nocoverage
|
|
|
|
self.render_openapi_type_exception(function, openapi_params, function_params, diff)
|
|
|
|
|
2019-06-06 22:22:21 +02:00
|
|
|
def test_openapi_arguments(self) -> None:
|
2019-07-07 08:54:19 +02:00
|
|
|
"""This end-to-end API documentation test compares the arguments
|
|
|
|
defined in the actual code using @has_request_variables and
|
|
|
|
REQ(), with the arguments declared in our API documentation
|
|
|
|
for every API endpoint in Zulip.
|
|
|
|
|
|
|
|
First, we import the fancy-Django version of zproject/urls.py
|
|
|
|
by doing this, each has_request_variables wrapper around each
|
|
|
|
imported view function gets called to generate the wrapped
|
|
|
|
view function and thus filling the global arguments_map variable.
|
|
|
|
Basically, we're exploiting code execution during import.
|
|
|
|
|
|
|
|
Then we need to import some view modules not already imported in
|
|
|
|
urls.py. We use this different syntax because of the linters complaining
|
|
|
|
of an unused import (which is correct, but we do this for triggering the
|
|
|
|
has_request_variables decorator).
|
2019-07-08 14:08:02 +02:00
|
|
|
|
|
|
|
At the end, we perform a reverse mapping test that verifies that
|
|
|
|
every url pattern defined in the openapi documentation actually exists
|
|
|
|
in code.
|
2019-07-07 08:54:19 +02:00
|
|
|
"""
|
|
|
|
|
2020-06-11 00:54:34 +02:00
|
|
|
from zproject import urls as urlconf
|
2019-06-06 22:22:21 +02:00
|
|
|
|
|
|
|
# We loop through all the API patterns, looking in particular
|
2019-07-07 08:54:19 +02:00
|
|
|
# for those using the rest_dispatch decorator; we then parse
|
|
|
|
# its mapping of (HTTP_METHOD -> FUNCTION).
|
2019-08-20 00:33:09 +02:00
|
|
|
for p in urlconf.v1_api_and_json_patterns + urlconf.v1_api_mobile_patterns:
|
2019-06-06 22:22:21 +02:00
|
|
|
if p.lookup_str != 'zerver.lib.rest.rest_dispatch':
|
2019-08-20 00:33:09 +02:00
|
|
|
# Endpoints not using rest_dispatch don't have extra data.
|
|
|
|
methods_endpoints = dict(
|
|
|
|
GET=p.lookup_str,
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
methods_endpoints = p.default_args
|
2019-07-10 13:23:25 +02:00
|
|
|
|
|
|
|
# since the module was already imported and is now residing in
|
|
|
|
# memory, we won't actually face any performance penalties here.
|
2019-08-20 00:33:09 +02:00
|
|
|
for method, value in methods_endpoints.items():
|
2019-06-06 22:22:21 +02:00
|
|
|
if isinstance(value, str):
|
2019-07-15 17:53:51 +02:00
|
|
|
function_name = value
|
python: Convert assignment type annotations to Python 3.6 style.
This commit was split by tabbott; this piece covers the vast majority
of files in Zulip, but excludes scripts/, tools/, and puppet/ to help
ensure we at least show the right error messages for Xenial systems.
We can likely further refine the remaining pieces with some testing.
Generated by com2ann, with whitespace fixes and various manual fixes
for runtime issues:
- invoiced_through: Optional[LicenseLedger] = models.ForeignKey(
+ invoiced_through: Optional["LicenseLedger"] = models.ForeignKey(
-_apns_client: Optional[APNsClient] = None
+_apns_client: Optional["APNsClient"] = None
- notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- signup_notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ signup_notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- author: Optional[UserProfile] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
+ author: Optional["UserProfile"] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
- bot_owner: Optional[UserProfile] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
+ bot_owner: Optional["UserProfile"] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
- default_sending_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
- default_events_register_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_sending_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_events_register_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
-descriptors_by_handler_id: Dict[int, ClientDescriptor] = {}
+descriptors_by_handler_id: Dict[int, "ClientDescriptor"] = {}
-worker_classes: Dict[str, Type[QueueProcessingWorker]] = {}
-queues: Dict[str, Dict[str, Type[QueueProcessingWorker]]] = {}
+worker_classes: Dict[str, Type["QueueProcessingWorker"]] = {}
+queues: Dict[str, Dict[str, Type["QueueProcessingWorker"]]] = {}
-AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional[LDAPSearch] = None
+AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional["LDAPSearch"] = None
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-22 01:09:50 +02:00
|
|
|
tags: Set[str] = set()
|
2019-06-06 22:22:21 +02:00
|
|
|
else:
|
2019-07-15 17:53:51 +02:00
|
|
|
function_name, tags = value
|
2019-07-07 08:54:19 +02:00
|
|
|
|
2019-10-22 01:43:54 +02:00
|
|
|
if function_name == 'zerver.tornado.views.get_events':
|
|
|
|
# Work around the fact that the registered
|
|
|
|
# get_events view function isn't where we do
|
|
|
|
# @has_request_variables.
|
|
|
|
#
|
|
|
|
# TODO: Make this configurable via an optional argument
|
|
|
|
# to has_request_variables, e.g.
|
|
|
|
# @has_request_variables(view_func_name="zerver.tornado.views.get_events")
|
|
|
|
function_name = 'zerver.tornado.views.get_events_backend'
|
|
|
|
|
2019-07-10 13:23:25 +02:00
|
|
|
lookup_parts = function_name.split('.')
|
|
|
|
module = __import__('.'.join(lookup_parts[:-1]), {}, {}, [''])
|
|
|
|
function = getattr(module, lookup_parts[-1])
|
|
|
|
|
2019-06-06 22:22:21 +02:00
|
|
|
# Our accounting logic in the `has_request_variables()`
|
|
|
|
# code means we have the list of all arguments
|
|
|
|
# accepted by every view function in arguments_map.
|
2019-07-15 17:53:51 +02:00
|
|
|
accepted_arguments = set(arguments_map[function_name])
|
2019-06-06 22:22:21 +02:00
|
|
|
|
2018-02-02 05:43:18 +01:00
|
|
|
regex_pattern = p.pattern.regex.pattern
|
2019-07-09 08:28:29 +02:00
|
|
|
url_pattern = self.convert_regex_to_url_pattern(regex_pattern)
|
2019-07-04 18:12:53 +02:00
|
|
|
|
2019-07-20 20:16:47 +02:00
|
|
|
if "intentionally_undocumented" in tags:
|
2019-07-09 08:28:29 +02:00
|
|
|
self.ensure_no_documentation_if_intentionally_undocumented(url_pattern, method)
|
|
|
|
continue
|
2019-06-06 22:22:21 +02:00
|
|
|
|
2019-07-20 20:16:47 +02:00
|
|
|
if url_pattern in self.pending_endpoints:
|
|
|
|
# HACK: After all pending_endpoints have been resolved, we should remove
|
|
|
|
# this segment and the "msg" part of the `ensure_no_...` method.
|
|
|
|
msg = """
|
|
|
|
We found some OpenAPI documentation for {method} {url_pattern},
|
|
|
|
so maybe we shouldn't include it in pending_endpoints.
|
|
|
|
""".format(method=method, url_pattern=url_pattern)
|
|
|
|
self.ensure_no_documentation_if_intentionally_undocumented(url_pattern,
|
|
|
|
method, msg)
|
|
|
|
continue
|
|
|
|
|
2019-06-06 22:22:21 +02:00
|
|
|
try:
|
2019-08-17 01:21:08 +02:00
|
|
|
# Don't include OpenAPI parameters that live in
|
|
|
|
# the path; these are not extracted by REQ.
|
|
|
|
openapi_parameters = get_openapi_parameters(url_pattern, method,
|
|
|
|
include_url_parameters=False)
|
2019-06-06 22:22:21 +02:00
|
|
|
except Exception: # nocoverage
|
2019-07-11 19:05:48 +02:00
|
|
|
raise AssertionError("Could not find OpenAPI docs for %s %s" %
|
|
|
|
(method, url_pattern))
|
2019-06-06 22:22:21 +02:00
|
|
|
|
|
|
|
# We now have everything we need to understand the
|
2019-07-07 08:54:19 +02:00
|
|
|
# function as defined in our urls.py:
|
2019-06-06 22:22:21 +02:00
|
|
|
#
|
|
|
|
# * method is the HTTP method, e.g. GET, POST, or PATCH
|
|
|
|
#
|
2018-02-02 05:43:18 +01:00
|
|
|
# * p.pattern.regex.pattern is the URL pattern; might require
|
2019-06-06 22:22:21 +02:00
|
|
|
# some processing to match with OpenAPI rules
|
|
|
|
#
|
2019-07-07 08:54:19 +02:00
|
|
|
# * accepted_arguments is the full set of arguments
|
|
|
|
# this method accepts (from the REQ declarations in
|
|
|
|
# code).
|
2019-06-06 22:22:21 +02:00
|
|
|
#
|
|
|
|
# * The documented parameters for the endpoint as recorded in our
|
|
|
|
# OpenAPI data in zerver/openapi/zulip.yaml.
|
|
|
|
#
|
|
|
|
# We now compare these to confirm that the documented
|
|
|
|
# argument list matches what actually appears in the
|
|
|
|
# codebase.
|
|
|
|
|
2020-04-09 21:51:58 +02:00
|
|
|
openapi_parameter_names = {
|
|
|
|
parameter['name'] for parameter in openapi_parameters
|
|
|
|
}
|
2019-06-06 22:22:21 +02:00
|
|
|
|
2019-10-22 01:43:54 +02:00
|
|
|
if len(accepted_arguments - openapi_parameter_names) > 0: # nocoverage
|
2019-07-15 17:53:51 +02:00
|
|
|
print("Undocumented parameters for",
|
|
|
|
url_pattern, method, function_name)
|
2019-06-06 22:22:21 +02:00
|
|
|
print(" +", openapi_parameter_names)
|
|
|
|
print(" -", accepted_arguments)
|
2019-07-11 19:05:48 +02:00
|
|
|
assert(url_pattern in self.buggy_documentation_endpoints)
|
2019-10-22 01:26:08 +02:00
|
|
|
elif len(openapi_parameter_names - accepted_arguments) > 0: # nocoverage
|
2019-07-15 17:53:51 +02:00
|
|
|
print("Documented invalid parameters for",
|
|
|
|
url_pattern, method, function_name)
|
2019-06-06 22:22:21 +02:00
|
|
|
print(" -", openapi_parameter_names)
|
|
|
|
print(" +", accepted_arguments)
|
2019-07-11 19:05:48 +02:00
|
|
|
assert(url_pattern in self.buggy_documentation_endpoints)
|
2019-06-06 22:22:21 +02:00
|
|
|
else:
|
|
|
|
self.assertEqual(openapi_parameter_names, accepted_arguments)
|
2019-07-10 13:23:25 +02:00
|
|
|
self.check_argument_types(function, openapi_parameters)
|
2019-07-11 19:05:48 +02:00
|
|
|
self.checked_endpoints.add(url_pattern)
|
2019-07-08 14:08:02 +02:00
|
|
|
|
2019-07-09 08:28:29 +02:00
|
|
|
self.check_for_non_existant_openapi_endpoints()
|
2019-07-29 15:46:48 +02:00
|
|
|
|
2019-08-04 08:14:08 +02:00
|
|
|
|
|
|
|
class ModifyExampleGenerationTestCase(ZulipTestCase):
|
|
|
|
|
|
|
|
def test_no_mod_argument(self) -> None:
|
|
|
|
res = parse_language_and_options("python")
|
|
|
|
self.assertEqual(res, ("python", {}))
|
|
|
|
|
|
|
|
def test_single_simple_mod_argument(self) -> None:
|
|
|
|
res = parse_language_and_options("curl, mod=1")
|
|
|
|
self.assertEqual(res, ("curl", {"mod": 1}))
|
|
|
|
|
|
|
|
res = parse_language_and_options("curl, mod='somevalue'")
|
|
|
|
self.assertEqual(res, ("curl", {"mod": "somevalue"}))
|
|
|
|
|
|
|
|
res = parse_language_and_options("curl, mod=\"somevalue\"")
|
|
|
|
self.assertEqual(res, ("curl", {"mod": "somevalue"}))
|
|
|
|
|
|
|
|
def test_multiple_simple_mod_argument(self) -> None:
|
|
|
|
res = parse_language_and_options("curl, mod1=1, mod2='a'")
|
|
|
|
self.assertEqual(res, ("curl", {"mod1": 1, "mod2": "a"}))
|
|
|
|
|
|
|
|
res = parse_language_and_options("curl, mod1=\"asdf\", mod2='thing', mod3=3")
|
|
|
|
self.assertEqual(res, ("curl", {"mod1": "asdf", "mod2": "thing", "mod3": 3}))
|
|
|
|
|
|
|
|
def test_single_list_mod_argument(self) -> None:
|
|
|
|
res = parse_language_and_options("curl, exclude=['param1', 'param2']")
|
|
|
|
self.assertEqual(res, ("curl", {"exclude": ["param1", "param2"]}))
|
|
|
|
|
|
|
|
res = parse_language_and_options("curl, exclude=[\"param1\", \"param2\"]")
|
|
|
|
self.assertEqual(res, ("curl", {"exclude": ["param1", "param2"]}))
|
|
|
|
|
|
|
|
res = parse_language_and_options("curl, exclude=['param1', \"param2\"]")
|
|
|
|
self.assertEqual(res, ("curl", {"exclude": ["param1", "param2"]}))
|
|
|
|
|
|
|
|
def test_multiple_list_mod_argument(self) -> None:
|
|
|
|
res = parse_language_and_options("curl, exclude=['param1', \"param2\"], special=['param3']")
|
|
|
|
self.assertEqual(res, ("curl", {"exclude": ["param1", "param2"], "special": ["param3"]}))
|
|
|
|
|
|
|
|
def test_multiple_mixed_mod_arguments(self) -> None:
|
|
|
|
res = parse_language_and_options("curl, exclude=[\"asdf\", 'sdfg'], other_key='asdf', more_things=\"asdf\", another_list=[1, \"2\"]")
|
|
|
|
self.assertEqual(res, ("curl", {"exclude": ["asdf", "sdfg"], "other_key": "asdf", "more_things": "asdf", "another_list": [1, "2"]}))
|
|
|
|
|
|
|
|
|
2019-07-29 15:46:48 +02:00
|
|
|
class TestCurlExampleGeneration(ZulipTestCase):
|
|
|
|
|
|
|
|
spec_mock_without_examples = {
|
2019-12-04 12:27:15 +01:00
|
|
|
"security": [{"basicAuth": []}],
|
2019-07-29 15:46:48 +02:00
|
|
|
"paths": {
|
|
|
|
"/mark_stream_as_read": {
|
|
|
|
"post": {
|
|
|
|
"description": "Mark all the unread messages in a stream as read.",
|
|
|
|
"parameters": [
|
|
|
|
{
|
|
|
|
"name": "stream_id",
|
|
|
|
"in": "query",
|
|
|
|
"description": "The ID of the stream whose messages should be marked as read.",
|
|
|
|
"schema": {
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
"type": "integer",
|
2019-07-29 15:46:48 +02:00
|
|
|
},
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
"required": True,
|
2019-07-29 15:46:48 +02:00
|
|
|
},
|
|
|
|
{
|
|
|
|
"name": "bool_param",
|
|
|
|
"in": "query",
|
|
|
|
"description": "Just a boolean parameter.",
|
|
|
|
"schema": {
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
"type": "boolean",
|
2019-07-29 15:46:48 +02:00
|
|
|
},
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
"required": True,
|
|
|
|
},
|
2019-07-29 15:46:48 +02:00
|
|
|
],
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
2019-07-29 15:46:48 +02:00
|
|
|
}
|
|
|
|
|
python: Convert assignment type annotations to Python 3.6 style.
This commit was split by tabbott; this piece covers the vast majority
of files in Zulip, but excludes scripts/, tools/, and puppet/ to help
ensure we at least show the right error messages for Xenial systems.
We can likely further refine the remaining pieces with some testing.
Generated by com2ann, with whitespace fixes and various manual fixes
for runtime issues:
- invoiced_through: Optional[LicenseLedger] = models.ForeignKey(
+ invoiced_through: Optional["LicenseLedger"] = models.ForeignKey(
-_apns_client: Optional[APNsClient] = None
+_apns_client: Optional["APNsClient"] = None
- notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- signup_notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ signup_notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- author: Optional[UserProfile] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
+ author: Optional["UserProfile"] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
- bot_owner: Optional[UserProfile] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
+ bot_owner: Optional["UserProfile"] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
- default_sending_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
- default_events_register_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_sending_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_events_register_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
-descriptors_by_handler_id: Dict[int, ClientDescriptor] = {}
+descriptors_by_handler_id: Dict[int, "ClientDescriptor"] = {}
-worker_classes: Dict[str, Type[QueueProcessingWorker]] = {}
-queues: Dict[str, Dict[str, Type[QueueProcessingWorker]]] = {}
+worker_classes: Dict[str, Type["QueueProcessingWorker"]] = {}
+queues: Dict[str, Dict[str, Type["QueueProcessingWorker"]]] = {}
-AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional[LDAPSearch] = None
+AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional["LDAPSearch"] = None
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-22 01:09:50 +02:00
|
|
|
spec_mock_with_invalid_method: Dict[str, object] = {
|
2019-12-04 12:27:15 +01:00
|
|
|
"security": [{"basicAuth": []}],
|
2019-07-29 15:46:48 +02:00
|
|
|
"paths": {
|
|
|
|
"/endpoint": {
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
"brew": {}, # the data is irrelevant as is should be rejected.
|
|
|
|
},
|
|
|
|
},
|
python: Convert assignment type annotations to Python 3.6 style.
This commit was split by tabbott; this piece covers the vast majority
of files in Zulip, but excludes scripts/, tools/, and puppet/ to help
ensure we at least show the right error messages for Xenial systems.
We can likely further refine the remaining pieces with some testing.
Generated by com2ann, with whitespace fixes and various manual fixes
for runtime issues:
- invoiced_through: Optional[LicenseLedger] = models.ForeignKey(
+ invoiced_through: Optional["LicenseLedger"] = models.ForeignKey(
-_apns_client: Optional[APNsClient] = None
+_apns_client: Optional["APNsClient"] = None
- notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- signup_notifications_stream: Optional[Stream] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
+ signup_notifications_stream: Optional["Stream"] = models.ForeignKey('Stream', related_name='+', null=True, blank=True, on_delete=CASCADE)
- author: Optional[UserProfile] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
+ author: Optional["UserProfile"] = models.ForeignKey('UserProfile', blank=True, null=True, on_delete=CASCADE)
- bot_owner: Optional[UserProfile] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
+ bot_owner: Optional["UserProfile"] = models.ForeignKey('self', null=True, on_delete=models.SET_NULL)
- default_sending_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
- default_events_register_stream: Optional[Stream] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_sending_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
+ default_events_register_stream: Optional["Stream"] = models.ForeignKey('zerver.Stream', null=True, related_name='+', on_delete=CASCADE)
-descriptors_by_handler_id: Dict[int, ClientDescriptor] = {}
+descriptors_by_handler_id: Dict[int, "ClientDescriptor"] = {}
-worker_classes: Dict[str, Type[QueueProcessingWorker]] = {}
-queues: Dict[str, Dict[str, Type[QueueProcessingWorker]]] = {}
+worker_classes: Dict[str, Type["QueueProcessingWorker"]] = {}
+queues: Dict[str, Dict[str, Type["QueueProcessingWorker"]]] = {}
-AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional[LDAPSearch] = None
+AUTH_LDAP_REVERSE_EMAIL_SEARCH: Optional["LDAPSearch"] = None
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-22 01:09:50 +02:00
|
|
|
}
|
2019-07-29 15:46:48 +02:00
|
|
|
|
|
|
|
spec_mock_using_object = {
|
2019-12-04 12:27:15 +01:00
|
|
|
"security": [{"basicAuth": []}],
|
2019-07-29 15:46:48 +02:00
|
|
|
"paths": {
|
|
|
|
"/endpoint": {
|
2019-10-03 15:02:51 +02:00
|
|
|
"get": {
|
|
|
|
"description": "Get some info.",
|
|
|
|
"parameters": [
|
|
|
|
{
|
|
|
|
"name": "param1",
|
|
|
|
"in": "query",
|
|
|
|
"description": "An object",
|
|
|
|
"schema": {
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
"type": "object",
|
2019-10-03 15:02:51 +02:00
|
|
|
},
|
|
|
|
"example": {
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
"key": "value",
|
2019-10-03 15:02:51 +02:00
|
|
|
},
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
"required": True,
|
|
|
|
},
|
|
|
|
],
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
2019-10-03 15:02:51 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
spec_mock_using_param_in_path = {
|
2019-12-04 12:27:15 +01:00
|
|
|
"security": [{"basicAuth": []}],
|
2019-10-03 15:02:51 +02:00
|
|
|
"paths": {
|
|
|
|
"/endpoint/{param1}": {
|
2019-07-29 15:46:48 +02:00
|
|
|
"get": {
|
|
|
|
"description": "Get some info.",
|
|
|
|
"parameters": [
|
|
|
|
{
|
|
|
|
"name": "param1",
|
|
|
|
"in": "path",
|
2019-10-03 15:59:28 +02:00
|
|
|
"description": "Param in path",
|
|
|
|
"schema": {
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
"type": "integer",
|
2019-10-03 15:59:28 +02:00
|
|
|
},
|
|
|
|
"example": 35,
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
"required": True,
|
2019-10-03 15:59:28 +02:00
|
|
|
},
|
|
|
|
{
|
|
|
|
"name": "param2",
|
|
|
|
"in": "query",
|
2019-07-29 15:46:48 +02:00
|
|
|
"description": "An object",
|
|
|
|
"schema": {
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
"type": "object",
|
2019-07-29 15:46:48 +02:00
|
|
|
},
|
|
|
|
"example": {
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
"key": "value",
|
2019-07-29 15:46:48 +02:00
|
|
|
},
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
"required": True,
|
|
|
|
},
|
|
|
|
],
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
2019-07-29 15:46:48 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
spec_mock_using_object_without_example = {
|
2019-12-04 12:27:15 +01:00
|
|
|
"security": [{"basicAuth": []}],
|
2019-07-29 15:46:48 +02:00
|
|
|
"paths": {
|
|
|
|
"/endpoint": {
|
|
|
|
"get": {
|
|
|
|
"description": "Get some info.",
|
|
|
|
"parameters": [
|
|
|
|
{
|
|
|
|
"name": "param1",
|
2019-10-03 15:02:51 +02:00
|
|
|
"in": "query",
|
2019-07-29 15:46:48 +02:00
|
|
|
"description": "An object",
|
|
|
|
"schema": {
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
"type": "object",
|
2019-07-29 15:46:48 +02:00
|
|
|
},
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
"required": True,
|
|
|
|
},
|
|
|
|
],
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
2019-07-29 15:46:48 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
spec_mock_using_array_without_example = {
|
2019-12-04 12:27:15 +01:00
|
|
|
"security": [{"basicAuth": []}],
|
2019-07-29 15:46:48 +02:00
|
|
|
"paths": {
|
|
|
|
"/endpoint": {
|
|
|
|
"get": {
|
|
|
|
"description": "Get some info.",
|
|
|
|
"parameters": [
|
|
|
|
{
|
|
|
|
"name": "param1",
|
2019-10-03 15:02:51 +02:00
|
|
|
"in": "query",
|
2019-07-29 15:46:48 +02:00
|
|
|
"description": "An array",
|
|
|
|
"schema": {
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
"type": "array",
|
2019-07-29 15:46:48 +02:00
|
|
|
},
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
"required": True,
|
|
|
|
},
|
|
|
|
],
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
2019-07-29 15:46:48 +02:00
|
|
|
}
|
|
|
|
|
2019-08-16 21:17:01 +02:00
|
|
|
def curl_example(self, endpoint: str, method: str, *args: Any, **kwargs: Any) -> List[str]:
|
|
|
|
return generate_curl_example(endpoint, method,
|
|
|
|
"http://localhost:9991/api", *args, **kwargs)
|
|
|
|
|
2019-07-29 15:46:48 +02:00
|
|
|
def test_generate_and_render_curl_example(self) -> None:
|
2019-08-16 21:17:01 +02:00
|
|
|
generated_curl_example = self.curl_example("/get_stream_id", "GET")
|
2019-07-29 15:46:48 +02:00
|
|
|
expected_curl_example = [
|
|
|
|
"```curl",
|
2019-08-07 10:55:41 +02:00
|
|
|
"curl -sSX GET -G http://localhost:9991/api/v1/get_stream_id \\",
|
2019-07-29 15:46:48 +02:00
|
|
|
" -u BOT_EMAIL_ADDRESS:BOT_API_KEY \\",
|
|
|
|
" -d 'stream=Denmark'",
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
"```",
|
2019-07-29 15:46:48 +02:00
|
|
|
]
|
|
|
|
self.assertEqual(generated_curl_example, expected_curl_example)
|
|
|
|
|
|
|
|
def test_generate_and_render_curl_example_with_nonexistant_endpoints(self) -> None:
|
|
|
|
with self.assertRaises(KeyError):
|
2019-08-16 21:17:01 +02:00
|
|
|
self.curl_example("/mark_this_stream_as_read", "POST")
|
2019-07-29 15:46:48 +02:00
|
|
|
with self.assertRaises(KeyError):
|
2019-08-16 21:17:01 +02:00
|
|
|
self.curl_example("/mark_stream_as_read", "GET")
|
2019-07-29 15:46:48 +02:00
|
|
|
|
|
|
|
def test_generate_and_render_curl_without_auth(self) -> None:
|
2019-08-16 21:17:01 +02:00
|
|
|
generated_curl_example = self.curl_example("/dev_fetch_api_key", "POST")
|
2019-07-29 15:46:48 +02:00
|
|
|
expected_curl_example = [
|
|
|
|
"```curl",
|
2019-08-07 10:55:41 +02:00
|
|
|
"curl -sSX POST http://localhost:9991/api/v1/dev_fetch_api_key \\",
|
2019-07-29 15:46:48 +02:00
|
|
|
" -d 'username=iago@zulip.com'",
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
"```",
|
2019-07-29 15:46:48 +02:00
|
|
|
]
|
|
|
|
self.assertEqual(generated_curl_example, expected_curl_example)
|
|
|
|
|
2020-02-23 18:10:42 +01:00
|
|
|
@patch("zerver.openapi.openapi.OpenAPISpec.spec")
|
2019-07-29 15:46:48 +02:00
|
|
|
def test_generate_and_render_curl_with_default_examples(self, spec_mock: MagicMock) -> None:
|
|
|
|
spec_mock.return_value = self.spec_mock_without_examples
|
2019-08-16 21:17:01 +02:00
|
|
|
generated_curl_example = self.curl_example("/mark_stream_as_read", "POST")
|
2019-07-29 15:46:48 +02:00
|
|
|
expected_curl_example = [
|
|
|
|
"```curl",
|
2019-08-07 10:55:41 +02:00
|
|
|
"curl -sSX POST http://localhost:9991/api/v1/mark_stream_as_read \\",
|
2019-12-04 12:27:15 +01:00
|
|
|
" -u BOT_EMAIL_ADDRESS:BOT_API_KEY \\",
|
2019-07-29 15:46:48 +02:00
|
|
|
" -d 'stream_id=1' \\",
|
|
|
|
" -d 'bool_param=false'",
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
"```",
|
2019-07-29 15:46:48 +02:00
|
|
|
]
|
|
|
|
self.assertEqual(generated_curl_example, expected_curl_example)
|
|
|
|
|
2020-02-23 18:10:42 +01:00
|
|
|
@patch("zerver.openapi.openapi.OpenAPISpec.spec")
|
2019-07-29 15:46:48 +02:00
|
|
|
def test_generate_and_render_curl_with_invalid_method(self, spec_mock: MagicMock) -> None:
|
|
|
|
spec_mock.return_value = self.spec_mock_with_invalid_method
|
|
|
|
with self.assertRaises(ValueError):
|
2019-08-16 21:17:01 +02:00
|
|
|
self.curl_example("/endpoint", "BREW") # see: HTCPCP
|
2019-07-29 15:46:48 +02:00
|
|
|
|
|
|
|
def test_generate_and_render_curl_with_array_example(self) -> None:
|
2019-08-16 21:17:01 +02:00
|
|
|
generated_curl_example = self.curl_example("/messages", "GET")
|
2019-07-29 15:46:48 +02:00
|
|
|
expected_curl_example = [
|
|
|
|
'```curl',
|
2019-08-07 10:55:41 +02:00
|
|
|
'curl -sSX GET -G http://localhost:9991/api/v1/messages \\',
|
2019-07-29 15:46:48 +02:00
|
|
|
' -u BOT_EMAIL_ADDRESS:BOT_API_KEY \\',
|
|
|
|
" -d 'anchor=42' \\",
|
|
|
|
" -d 'num_before=4' \\",
|
|
|
|
" -d 'num_after=8' \\",
|
2019-08-02 14:47:18 +02:00
|
|
|
' --data-urlencode narrow=\'[{"operand": "Denmark", "operator": "stream"}]\' \\',
|
2019-07-29 15:46:48 +02:00
|
|
|
" -d 'client_gravatar=true' \\",
|
2020-03-27 04:53:11 +01:00
|
|
|
" -d 'apply_markdown=false' \\",
|
|
|
|
" -d 'use_first_unread_anchor=true'",
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
'```',
|
2019-07-29 15:46:48 +02:00
|
|
|
]
|
|
|
|
self.assertEqual(generated_curl_example, expected_curl_example)
|
|
|
|
|
2020-02-23 18:10:42 +01:00
|
|
|
@patch("zerver.openapi.openapi.OpenAPISpec.spec")
|
2019-07-29 15:46:48 +02:00
|
|
|
def test_generate_and_render_curl_with_object(self, spec_mock: MagicMock) -> None:
|
|
|
|
spec_mock.return_value = self.spec_mock_using_object
|
2019-08-16 21:17:01 +02:00
|
|
|
generated_curl_example = self.curl_example("/endpoint", "GET")
|
2019-07-29 15:46:48 +02:00
|
|
|
expected_curl_example = [
|
|
|
|
'```curl',
|
2019-08-07 10:55:41 +02:00
|
|
|
'curl -sSX GET -G http://localhost:9991/api/v1/endpoint \\',
|
2019-12-04 12:27:15 +01:00
|
|
|
' -u BOT_EMAIL_ADDRESS:BOT_API_KEY \\',
|
2019-07-29 15:46:48 +02:00
|
|
|
' --data-urlencode param1=\'{"key": "value"}\'',
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
'```',
|
2019-07-29 15:46:48 +02:00
|
|
|
]
|
|
|
|
self.assertEqual(generated_curl_example, expected_curl_example)
|
2019-10-03 15:02:51 +02:00
|
|
|
|
2020-02-23 18:10:42 +01:00
|
|
|
@patch("zerver.openapi.openapi.OpenAPISpec.spec")
|
2019-07-29 15:46:48 +02:00
|
|
|
def test_generate_and_render_curl_with_object_without_example(self, spec_mock: MagicMock) -> None:
|
|
|
|
spec_mock.return_value = self.spec_mock_using_object_without_example
|
|
|
|
with self.assertRaises(ValueError):
|
2019-08-16 21:17:01 +02:00
|
|
|
self.curl_example("/endpoint", "GET")
|
2019-07-29 15:46:48 +02:00
|
|
|
|
2020-02-23 18:10:42 +01:00
|
|
|
@patch("zerver.openapi.openapi.OpenAPISpec.spec")
|
2019-07-29 15:46:48 +02:00
|
|
|
def test_generate_and_render_curl_with_array_without_example(self, spec_mock: MagicMock) -> None:
|
|
|
|
spec_mock.return_value = self.spec_mock_using_array_without_example
|
|
|
|
with self.assertRaises(ValueError):
|
2019-08-16 21:17:01 +02:00
|
|
|
self.curl_example("/endpoint", "GET")
|
2019-07-29 15:46:48 +02:00
|
|
|
|
2020-02-23 18:10:42 +01:00
|
|
|
@patch("zerver.openapi.openapi.OpenAPISpec.spec")
|
2019-10-03 15:59:28 +02:00
|
|
|
def test_generate_and_render_curl_with_param_in_path(self, spec_mock: MagicMock) -> None:
|
|
|
|
spec_mock.return_value = self.spec_mock_using_param_in_path
|
|
|
|
generated_curl_example = self.curl_example("/endpoint/{param1}", "GET")
|
|
|
|
expected_curl_example = [
|
|
|
|
'```curl',
|
|
|
|
'curl -sSX GET -G http://localhost:9991/api/v1/endpoint/35 \\',
|
2019-12-04 12:27:15 +01:00
|
|
|
' -u BOT_EMAIL_ADDRESS:BOT_API_KEY \\',
|
2019-10-03 15:59:28 +02:00
|
|
|
' --data-urlencode param2=\'{"key": "value"}\'',
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
'```',
|
2019-10-03 15:59:28 +02:00
|
|
|
]
|
|
|
|
self.assertEqual(generated_curl_example, expected_curl_example)
|
|
|
|
|
2019-07-29 15:46:48 +02:00
|
|
|
def test_generate_and_render_curl_wrapper(self) -> None:
|
2019-08-16 21:17:01 +02:00
|
|
|
generated_curl_example = render_curl_example("/get_stream_id:GET:email:key",
|
|
|
|
api_url="https://zulip.example.com/api")
|
2019-07-29 15:46:48 +02:00
|
|
|
expected_curl_example = [
|
|
|
|
"```curl",
|
2019-08-07 10:55:41 +02:00
|
|
|
"curl -sSX GET -G https://zulip.example.com/api/v1/get_stream_id \\",
|
2019-07-29 15:46:48 +02:00
|
|
|
" -u email:key \\",
|
|
|
|
" -d 'stream=Denmark'",
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
"```",
|
2019-07-29 15:46:48 +02:00
|
|
|
]
|
|
|
|
self.assertEqual(generated_curl_example, expected_curl_example)
|
2019-08-04 08:14:08 +02:00
|
|
|
|
|
|
|
def test_generate_and_render_curl_example_with_excludes(self) -> None:
|
2019-08-16 21:17:01 +02:00
|
|
|
generated_curl_example = self.curl_example("/messages", "GET",
|
|
|
|
exclude=["client_gravatar", "apply_markdown"])
|
2019-08-04 08:14:08 +02:00
|
|
|
expected_curl_example = [
|
|
|
|
'```curl',
|
2019-08-07 10:55:41 +02:00
|
|
|
'curl -sSX GET -G http://localhost:9991/api/v1/messages \\',
|
2019-08-04 08:14:08 +02:00
|
|
|
' -u BOT_EMAIL_ADDRESS:BOT_API_KEY \\',
|
|
|
|
" -d 'anchor=42' \\",
|
|
|
|
" -d 'num_before=4' \\",
|
|
|
|
" -d 'num_after=8' \\",
|
2020-03-27 04:53:11 +01:00
|
|
|
' --data-urlencode narrow=\'[{"operand": "Denmark", "operator": "stream"}]\' \\',
|
|
|
|
" -d 'use_first_unread_anchor=true'",
|
python: Use trailing commas consistently.
Automatically generated by the following script, based on the output
of lint with flake8-comma:
import re
import sys
last_filename = None
last_row = None
lines = []
for msg in sys.stdin:
m = re.match(
r"\x1b\[35mflake8 \|\x1b\[0m \x1b\[1;31m(.+):(\d+):(\d+): (\w+)", msg
)
if m:
filename, row_str, col_str, err = m.groups()
row, col = int(row_str), int(col_str)
if filename == last_filename:
assert last_row != row
else:
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
with open(filename) as f:
lines = f.readlines()
last_filename = filename
last_row = row
line = lines[row - 1]
if err in ["C812", "C815"]:
lines[row - 1] = line[: col - 1] + "," + line[col - 1 :]
elif err in ["C819"]:
assert line[col - 2] == ","
lines[row - 1] = line[: col - 2] + line[col - 1 :].lstrip(" ")
if last_filename is not None:
with open(last_filename, "w") as f:
f.writelines(lines)
Signed-off-by: Anders Kaseorg <anders@zulipchat.com>
2020-04-10 05:23:40 +02:00
|
|
|
'```',
|
2019-08-04 08:14:08 +02:00
|
|
|
]
|
|
|
|
self.assertEqual(generated_curl_example, expected_curl_example)
|
2020-05-20 19:53:41 +02:00
|
|
|
|
|
|
|
class OpenAPIAttributesTest(ZulipTestCase):
|
|
|
|
def test_attributes(self) -> None:
|
|
|
|
EXCLUDE = ["/real-time"]
|
|
|
|
VALID_TAGS = ["users", "server_and_organizations", "authentication",
|
|
|
|
"real_time_events", "streams", "messages", "users",
|
|
|
|
"webhooks"]
|
|
|
|
openapi_spec = OpenAPISpec(OPENAPI_SPEC_PATH).spec()["paths"]
|
|
|
|
for path in openapi_spec:
|
|
|
|
if path in EXCLUDE:
|
|
|
|
continue
|
|
|
|
for method in openapi_spec[path]:
|
|
|
|
# Check if every file has an operationId
|
|
|
|
assert("operationId" in openapi_spec[path][method])
|
|
|
|
assert("tags" in openapi_spec[path][method])
|
|
|
|
tag = openapi_spec[path][method]["tags"][0]
|
|
|
|
assert(tag in VALID_TAGS)
|
2020-06-10 19:39:24 +02:00
|
|
|
for response in openapi_spec[path][method]['responses']:
|
|
|
|
response_schema = (openapi_spec[path][method]['responses'][response]
|
|
|
|
['content']['application/json']['schema'])
|
|
|
|
if 'oneOf' in response_schema:
|
|
|
|
cnt = 0
|
|
|
|
for entry in response_schema['oneOf']:
|
|
|
|
validate_against_openapi_schema(entry['example'], path,
|
|
|
|
method, response + '_' + str(cnt))
|
|
|
|
cnt += 1
|
|
|
|
continue
|
|
|
|
validate_against_openapi_schema(response_schema['example'], path,
|
|
|
|
method, response)
|