text: Fix some typos (most of them found and fixed by codespell).

Signed-off-by: Stefan Weil <sw@weilnetz.de>
This commit is contained in:
Stefan Weil 2020-03-28 01:25:56 +01:00 committed by GitHub
parent 6eb39c8668
commit d2fa058cc1
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
114 changed files with 151 additions and 151 deletions

View File

@ -214,7 +214,7 @@ def do_replace_payment_source(user: UserProfile, stripe_token: str,
for stripe_invoice in stripe.Invoice.list( for stripe_invoice in stripe.Invoice.list(
billing='charge_automatically', customer=stripe_customer.id, status='open'): billing='charge_automatically', customer=stripe_customer.id, status='open'):
# The user will get either a receipt or a "failed payment" email, but the in-app # The user will get either a receipt or a "failed payment" email, but the in-app
# messaging could be clearer here (e.g. it could explictly tell the user that there # messaging could be clearer here (e.g. it could explicitly tell the user that there
# were payment(s) and that they succeeded or failed). # were payment(s) and that they succeeded or failed).
# Worth fixing if we notice that a lot of cards end up failing at this step. # Worth fixing if we notice that a lot of cards end up failing at this step.
stripe.Invoice.pay(stripe_invoice) stripe.Invoice.pay(stripe_invoice)

View File

@ -1093,7 +1093,7 @@ class BillingHelpersTest(ZulipTestCase):
datetime(2021, 2, 28, 1, 2, 3).replace(tzinfo=timezone_utc)] datetime(2021, 2, 28, 1, 2, 3).replace(tzinfo=timezone_utc)]
with self.assertRaises(AssertionError): with self.assertRaises(AssertionError):
add_months(anchor, -1) add_months(anchor, -1)
# Explictly test add_months for each value of MAX_DAY_FOR_MONTH and # Explicitly test add_months for each value of MAX_DAY_FOR_MONTH and
# for crossing a year boundary # for crossing a year boundary
for i, boundary in enumerate(period_boundaries): for i, boundary in enumerate(period_boundaries):
self.assertEqual(add_months(anchor, i), boundary) self.assertEqual(add_months(anchor, i), boundary)

View File

@ -328,8 +328,8 @@ run_test('validate_stream_message', () => {
}); });
run_test('test_validate_stream_message_post_policy', () => { run_test('test_validate_stream_message_post_policy', () => {
// This test is in continuation with test_validate but it has been seperated out // This test is in continuation with test_validate but it has been separated out
// for better readabilty. Their relative position of execution should not be changed. // for better readability. Their relative position of execution should not be changed.
// Although the position with respect to test_validate_stream_message does not matter // Although the position with respect to test_validate_stream_message does not matter
// as `get_stream_post_policy` is reset at the end. // as `get_stream_post_policy` is reset at the end.
page_params.is_admin = false; page_params.is_admin = false;

View File

@ -766,7 +766,7 @@ with_overrides(function (override) {
}); });
with_overrides(function (override) { with_overrides(function (override) {
// attachements // attachments
const event = event_fixtures.attachment; const event = event_fixtures.attachment;
global.with_stub(function (stub) { global.with_stub(function (stub) {
override('attachments_ui.update_attachments', stub.f); override('attachments_ui.update_attachments', stub.f);
@ -1524,7 +1524,7 @@ with_overrides(function (override) {
}); });
with_overrides(function (override) { with_overrides(function (override) {
// attachements // attachments
let event = event_fixtures.user_status__set_away; let event = event_fixtures.user_status__set_away;
global.with_stub(function (stub) { global.with_stub(function (stub) {
override('activity.on_set_away', stub.f); override('activity.on_set_away', stub.f);

View File

@ -11,7 +11,7 @@ run_test('get_unused_fence', () => {
// ``` // ```
// ```` // ````
// ````` // `````
// ... upto N chars // ... up to N chars
// We insert a N + 1 character fence. // We insert a N + 1 character fence.
for (let i = 3; i <= 20; i += 1) { for (let i = 3; i <= 20; i += 1) {
large_testcase += '`'.repeat(i) + '\n'; large_testcase += '`'.repeat(i) + '\n';

View File

@ -303,7 +303,7 @@ run_test('narrow_to_compose_target', () => {
{operator: 'pm-with', operand: 'alice@example.com,ray@example.com'}, {operator: 'pm-with', operand: 'alice@example.com,ray@example.com'},
]); ]);
// Test with some inavlid persons // Test with some invalid persons
global.compose_state.private_message_recipient = () => 'alice@example.com,random,ray@example.com'; global.compose_state.private_message_recipient = () => 'alice@example.com,random,ray@example.com';
args.called = false; args.called = false;
narrow.to_compose_target(); narrow.to_compose_target();
@ -312,7 +312,7 @@ run_test('narrow_to_compose_target', () => {
{operator: 'is', operand: 'private'}, {operator: 'is', operand: 'private'},
]); ]);
// Test with all inavlid persons // Test with all invalid persons
global.compose_state.private_message_recipient = () => 'alice,random,ray'; global.compose_state.private_message_recipient = () => 'alice,random,ray';
args.called = false; args.called = false;
narrow.to_compose_target(); narrow.to_compose_target();

View File

@ -78,7 +78,7 @@ run_test('message_is_notifiable', () => {
// Not notifiable because it was sent by the current user // Not notifiable because it was sent by the current user
assert.equal(notifications.message_is_notifiable(message), false); assert.equal(notifications.message_is_notifiable(message), false);
// Case 2: If the user has already been sent a notificaton about this message, // Case 2: If the user has already been sent a notification about this message,
// DO NOT notify the user // DO NOT notify the user
// In this test, all other circumstances should trigger notification // In this test, all other circumstances should trigger notification
// EXCEPT notification_sent, which should trump them // EXCEPT notification_sent, which should trump them

View File

@ -763,7 +763,7 @@ run_test('get_people_for_search_bar', () => {
// As long as there are 5+ results among the user_ids // As long as there are 5+ results among the user_ids
// in message_store, we will get a small result and not // in message_store, we will get a small result and not
// seach all people. // search all people.
assert.equal(small_results.length, 6); assert.equal(small_results.length, 6);
}); });

View File

@ -794,7 +794,7 @@ run_test('set_up', () => {
const stub_render_notifications_stream_ui = settings_org.render_notifications_stream_ui; const stub_render_notifications_stream_ui = settings_org.render_notifications_stream_ui;
settings_org.render_notifications_stream_ui = noop; settings_org.render_notifications_stream_ui = noop;
$("#id_realm_message_content_edit_limit_minutes").set_parent($.create('<stub edit limit parent>')); $("#id_realm_message_content_edit_limit_minutes").set_parent($.create('<stub edit limit parent>'));
$("#id_realm_message_content_delete_limit_minutes").set_parent($.create('<stub delete limti parent>')); $("#id_realm_message_content_delete_limit_minutes").set_parent($.create('<stub delete limit parent>'));
$("#message_content_in_email_notifications_label").set_parent($.create('<stub in-content setting checkbox>')); $("#message_content_in_email_notifications_label").set_parent($.create('<stub in-content setting checkbox>'));
$("#enable_digest_emails_label").set_parent($.create('<stub digest setting checkbox>')); $("#enable_digest_emails_label").set_parent($.create('<stub digest setting checkbox>'));
$("#id_realm_digest_weekday").set_parent($.create('<stub digest weekday setting dropdown>')); $("#id_realm_digest_weekday").set_parent($.create('<stub digest weekday setting dropdown>'));

View File

@ -529,7 +529,7 @@ run_test('separators_only_pinned_and_dormant', () => {
subscribed: true, subscribed: true,
}; };
add_row(RomeSub); add_row(RomeSub);
// dorment stream // dormant stream
const DenmarkSub = { const DenmarkSub = {
name: 'Denmark', name: 'Denmark',
stream_id: 3000, stream_id: 3000,

View File

@ -250,7 +250,7 @@ run_test('sort_recipients', () => {
stream_data.update_calculated_fields(dev_sub); stream_data.update_calculated_fields(dev_sub);
stream_data.update_calculated_fields(linux_sub); stream_data.update_calculated_fields(linux_sub);
// For spliting based on whether a PM was sent // For splitting based on whether a PM was sent
global.pm_conversations.set_partner(5); global.pm_conversations.set_partner(5);
global.pm_conversations.set_partner(6); global.pm_conversations.set_partner(6);
global.pm_conversations.set_partner(2); global.pm_conversations.set_partner(2);
@ -502,7 +502,7 @@ run_test('highlight_with_escaping', () => {
}); });
run_test('render_person when emails hidden', () => { run_test('render_person when emails hidden', () => {
// Test render_person with regular person, under hidden email visiblity case // Test render_person with regular person, under hidden email visibility case
page_params.is_admin = false; page_params.is_admin = false;
let rendered = false; let rendered = false;
global.stub_templates(function (template_name, args) { global.stub_templates(function (template_name, args) {

View File

@ -274,7 +274,7 @@ run_test('basics', () => {
}); });
} }
// User ids of poeple in compose narrow doesn't change and is same as stat.current_recipent // User ids of people in compose narrow doesn't change and is same as stat.current_recipient
// so counts of function should increase except stop_last_notification // so counts of function should increase except stop_last_notification
typing_status.update(worker, typing.get_recipient()); typing_status.update(worker, typing.get_recipient());
assert.deepEqual(call_count.maybe_ping_server, 1); assert.deepEqual(call_count.maybe_ping_server, 1);

View File

@ -60,7 +60,7 @@ run_test('basics', () => {
blueslip.clear_test_data(); blueslip.clear_test_data();
// Let's repeat the above procedue with warnings. Unlike errors, // Let's repeat the above procedue with warnings. Unlike errors,
// warnings shoudln't stop the code execution, and thus, the // warnings shouldn't stop the code execution, and thus, the
// behaviour is slightly different. // behaviour is slightly different.
function throw_a_warning() { function throw_a_warning() {

View File

@ -167,7 +167,7 @@ run_test('events', () => {
stopPropagation: function () {}, stopPropagation: function () {},
}; };
// Now call the hander. // Now call the handler.
red_handler_func(stub_event); red_handler_func(stub_event);
// And verify it did what it was supposed to do. // And verify it did what it was supposed to do.

View File

@ -12,7 +12,7 @@
* assert.notEqual(). * assert.notEqual().
* *
* There is a default _output_formatter used to create the * There is a default _output_formatter used to create the
* AssertionError error message; this function can be overriden using * AssertionError error message; this function can be overridden using
* the exported setFormatter() function below. * the exported setFormatter() function below.
* *
* The HTML passed to the _output_formatter is not the original HTML, but * The HTML passed to the _output_formatter is not the original HTML, but

View File

@ -1,6 +1,6 @@
# This class includes all the modules you need to install/run a Zulip installation # This class includes all the modules you need to install/run a Zulip installation
# in a single container (without the database, memcached, redis services). # in a single container (without the database, memcached, redis services).
# The database, memcached, redis services need to be run in seperate containers. # The database, memcached, redis services need to be run in separate containers.
# Through this split of services, it is easier to scale the services to the needs. # Through this split of services, it is easier to scale the services to the needs.
class zulip::dockervoyager { class zulip::dockervoyager {
include zulip::base include zulip::base

View File

@ -26,7 +26,7 @@ Unattended-Upgrade::Package-Blacklist {
// is running is possible (with a small delay) // is running is possible (with a small delay)
//Unattended-Upgrade::MinimalSteps "true"; //Unattended-Upgrade::MinimalSteps "true";
// Install all unattended-upgrades when the machine is shuting down // Install all unattended-upgrades when the machine is shutting down
// instead of doing it in the background while the machine is running // instead of doing it in the background while the machine is running
// This will (obviously) make shutdown slower // This will (obviously) make shutdown slower
//Unattended-Upgrade::InstallOnShutdown "true"; //Unattended-Upgrade::InstallOnShutdown "true";

View File

@ -14,8 +14,8 @@
#%# family=auto #%# family=auto
#%# capabilities=autoconf #%# capabilities=autoconf
# If run with the "autoconf"-parameter, give our opinion on wether we # If run with the "autoconf"-parameter, give our opinion on whether we
# should be run on this system or not. This is optinal, and only used by # should be run on this system or not. This is optional, and only used by
# munin-config. In the case of this plugin, we should most probably # munin-config. In the case of this plugin, we should most probably
# always be included. # always be included.

View File

@ -14,8 +14,8 @@
#%# family=auto #%# family=auto
#%# capabilities=autoconf #%# capabilities=autoconf
# If run with the "autoconf"-parameter, give our opinion on wether we # If run with the "autoconf"-parameter, give our opinion on whether we
# should be run on this system or not. This is optinal, and only used by # should be run on this system or not. This is optional, and only used by
# munin-config. In the case of this plugin, we should most probably # munin-config. In the case of this plugin, we should most probably
# always be included. # always be included.

View File

@ -14,8 +14,8 @@
#%# family=auto #%# family=auto
#%# capabilities=autoconf #%# capabilities=autoconf
# If run with the "autoconf"-parameter, give our opinion on wether we # If run with the "autoconf"-parameter, give our opinion on whether we
# should be run on this system or not. This is optinal, and only used by # should be run on this system or not. This is optional, and only used by
# munin-config. In the case of this plugin, we should most probably # munin-config. In the case of this plugin, we should most probably
# always be included. # always be included.

View File

@ -14,8 +14,8 @@
#%# family=auto #%# family=auto
#%# capabilities=autoconf #%# capabilities=autoconf
# If run with the "autoconf"-parameter, give our opinion on wether we # If run with the "autoconf"-parameter, give our opinion on whether we
# should be run on this system or not. This is optinal, and only used by # should be run on this system or not. This is optional, and only used by
# munin-config. In the case of this plugin, we should most probably # munin-config. In the case of this plugin, we should most probably
# always be included. # always be included.

View File

@ -14,8 +14,8 @@
#%# family=auto #%# family=auto
#%# capabilities=autoconf #%# capabilities=autoconf
# If run with the "autoconf"-parameter, give our opinion on wether we # If run with the "autoconf"-parameter, give our opinion on whether we
# should be run on this system or not. This is optinal, and only used by # should be run on this system or not. This is optional, and only used by
# munin-config. In the case of this plugin, we should most probably # munin-config. In the case of this plugin, we should most probably
# always be included. # always be included.

View File

@ -14,8 +14,8 @@
#%# family=auto #%# family=auto
#%# capabilities=autoconf #%# capabilities=autoconf
# If run with the "autoconf"-parameter, give our opinion on wether we # If run with the "autoconf"-parameter, give our opinion on whether we
# should be run on this system or not. This is optinal, and only used by # should be run on this system or not. This is optional, and only used by
# munin-config. In the case of this plugin, we should most probably # munin-config. In the case of this plugin, we should most probably
# always be included. # always be included.

View File

@ -12,8 +12,8 @@
#%# family=auto #%# family=auto
#%# capabilities=autoconf #%# capabilities=autoconf
# If run with the "autoconf"-parameter, give our opinion on wether we # If run with the "autoconf"-parameter, give our opinion on whether we
# should be run on this system or not. This is optinal, and only used by # should be run on this system or not. This is optional, and only used by
# munin-config. In the case of this plugin, we should most probably # munin-config. In the case of this plugin, we should most probably
# always be included. # always be included.

View File

@ -877,7 +877,7 @@ process_performance_data=0
# performance data files are opened in write ("w") or append ("a") # performance data files are opened in write ("w") or append ("a")
# mode. If you want to use named pipes, you should use the special # mode. If you want to use named pipes, you should use the special
# pipe ("p") mode which avoid blocking at startup, otherwise you will # pipe ("p") mode which avoid blocking at startup, otherwise you will
# likely want the defult append ("a") mode. # likely want the default append ("a") mode.
#host_perfdata_file_mode=a #host_perfdata_file_mode=a
#service_perfdata_file_mode=a #service_perfdata_file_mode=a
@ -906,7 +906,7 @@ process_performance_data=0
# HOST AND SERVICE PERFORMANCE DATA PROCESS EMPTY RESULTS # HOST AND SERVICE PERFORMANCE DATA PROCESS EMPTY RESULTS
# THese options determine wether the core will process empty perfdata # THese options determine whether the core will process empty perfdata
# results or not. This is needed for distributed monitoring, and intentionally # results or not. This is needed for distributed monitoring, and intentionally
# turned on by default. # turned on by default.
# If you don't require empty perfdata - saving some cpu cycles # If you don't require empty perfdata - saving some cpu cycles

View File

@ -142,7 +142,7 @@ for device in macs.values():
ip, '-j', 'MARK', '--set-mark', str(device_number)]) ip, '-j', 'MARK', '--set-mark', str(device_number)])
for throwaway in range(2): for throwaway in range(2):
# Don't freak out if this doens't work. # Don't freak out if this doesn't work.
subprocess.call( subprocess.call(
['/sbin/ip', 'route', 'del', '10.0.0.0/8']) ['/sbin/ip', 'route', 'del', '10.0.0.0/8'])

View File

@ -88,7 +88,7 @@ end
# #
#rrdcached_socket /var/run/rrdcached.sock #rrdcached_socket /var/run/rrdcached.sock
# Drop somejuser@fnord.comm and anotheruser@blibb.comm an email everytime # Drop somejuser@fnord.comm and anotheruser@blibb.comm an email every time
# something changes (OK -> WARNING, CRITICAL -> OK, etc) # something changes (OK -> WARNING, CRITICAL -> OK, etc)
#contact.someuser.command mail -s "Munin notification" somejuser@fnord.comm #contact.someuser.command mail -s "Munin notification" somejuser@fnord.comm
#contact.anotheruser.command mail -s "Munin notification" anotheruser@blibb.comm #contact.anotheruser.command mail -s "Munin notification" anotheruser@blibb.comm

View File

@ -399,7 +399,7 @@ def file_or_package_hash_updated(paths, hash_name, is_force, package_versions=[]
with open(path, 'rb') as file_to_hash: with open(path, 'rb') as file_to_hash:
sha1sum.update(file_to_hash.read()) sha1sum.update(file_to_hash.read())
# The ouput of tools like build_pygments_data depends # The output of tools like build_pygments_data depends
# on the version of some pip packages as well. # on the version of some pip packages as well.
for package_version in package_versions: for package_version in package_versions:
sha1sum.update(package_version.encode("utf-8")) sha1sum.update(package_version.encode("utf-8"))

View File

@ -271,7 +271,7 @@ exports.get_item = function (user_id) {
}; };
function user_is_recently_active(user_id) { function user_is_recently_active(user_id) {
// return true if the user has a green/orange cirle // return true if the user has a green/orange circle
return exports.level(user_id) <= 2; return exports.level(user_id) <= 2;
} }

View File

@ -685,10 +685,10 @@ exports.handle_keydown = function (event, textarea) {
const position = textarea.caret(); const position = textarea.caret();
const txt = document.getElementById(textarea[0].id); const txt = document.getElementById(textarea[0].id);
// Include selected text in between [] parantheses and insert '(url)' // Include selected text in between [] parentheses and insert '(url)'
// where "url" should be automatically selected. // where "url" should be automatically selected.
// Position of cursor depends on whether browser supports exec // Position of cursor depends on whether browser supports exec
// command or not. So set cursor position accrodingly. // command or not. So set cursor position accordingly.
if (range.length > 0) { if (range.length > 0) {
if (document.queryCommandEnabled('insertText')) { if (document.queryCommandEnabled('insertText')) {
txt.selectionStart = position - 4; txt.selectionStart = position - 4;
@ -1039,7 +1039,7 @@ exports.initialize = function () {
e.preventDefault(); e.preventDefault();
let target_textarea; let target_textarea;
// The data-message-id atribute is only present in the video // The data-message-id attribute is only present in the video
// call icon present in the message edit form. If present, // call icon present in the message edit form. If present,
// the request is for the edit UI; otherwise, it's for the // the request is for the edit UI; otherwise, it's for the
// compose box. // compose box.

View File

@ -777,7 +777,7 @@ exports.content_typeahead_selected = function (item, event) {
beginning = beginning.substring(0, backticks) + item; beginning = beginning.substring(0, backticks) + item;
} }
} else if (this.completing === 'topic_jump') { } else if (this.completing === 'topic_jump') {
// Put the cursor at the end of immediately preceeding stream mention syntax, // Put the cursor at the end of immediately preceding stream mention syntax,
// just before where the `**` at the end of the syntax. This will delete that // just before where the `**` at the end of the syntax. This will delete that
// final ** and set things up for the topic_list typeahead. // final ** and set things up for the topic_list typeahead.
const index = beginning.lastIndexOf('**'); const index = beginning.lastIndexOf('**');

View File

@ -644,12 +644,12 @@ exports.register_click_handlers = function () {
$("body").on("click", "#emoji_map", function (e) { $("body").on("click", "#emoji_map", function (e) {
e.preventDefault(); e.preventDefault();
e.stopPropagation(); e.stopPropagation();
// The data-message-id atribute is only present in the emoji icon present in // The data-message-id attribute is only present in the emoji icon present in
// the message edit form. So the following check will return false if this // the message edit form. So the following check will return false if this
// event was not fired from message edit form. // event was not fired from message edit form.
if ($(this).attr("data-message-id") !== undefined) { if ($(this).attr("data-message-id") !== undefined) {
// Store data-message-id value in global variable edit_message_id so that // Store data-message-id value in global variable edit_message_id so that
// its value can be further used to correclty find the message textarea element. // its value can be further used to correctly find the message textarea element.
edit_message_id = $(this).attr("data-message-id"); edit_message_id = $(this).attr("data-message-id");
} else { } else {
edit_message_id = null; edit_message_id = null;

View File

@ -212,7 +212,7 @@ exports.create = function (opts) {
// this sends a flag that the operation wasn't completely successful, // this sends a flag that the operation wasn't completely successful,
// which in this case is defined as some of the pills not autofilling // which in this case is defined as some of the pills not autofilling
// correclty. // correctly.
if (drafts.length > 0) { if (drafts.length > 0) {
return false; return false;
} }

View File

@ -79,7 +79,7 @@ const funcs = {
canvas.addEventListener("wheel", function (e) { canvas.addEventListener("wheel", function (e) {
e.preventDefault(); e.preventDefault();
// this is to reverese scrolling directions for the image. // this is to reverse scrolling directions for the image.
let delta = meta.direction * e.deltaY; let delta = meta.direction * e.deltaY;
if (e.deltaMode === DELTA_MODE.LINE) { if (e.deltaMode === DELTA_MODE.LINE) {

View File

@ -86,7 +86,7 @@ exports.create = function ($container, list, opts) {
const prototype = { const prototype = {
// Reads the provided list (in the scope directly above) // Reads the provided list (in the scope directly above)
// and renders the next block of messages automatically // and renders the next block of messages automatically
// into the specified contianer. // into the specified container.
render: function (load_count) { render: function (load_count) {
load_count = load_count || opts.load_count || DEFAULTS.LOAD_COUNT; load_count = load_count || opts.load_count || DEFAULTS.LOAD_COUNT;

View File

@ -263,7 +263,7 @@ exports.get_frontfill_anchor = function (msg_list) {
exports.maybe_load_older_messages = function (opts) { exports.maybe_load_older_messages = function (opts) {
// This function gets called when you scroll to the top // This function gets called when you scroll to the top
// of your window, and you want to get messages older // of your window, and you want to get messages older
// than what the browers originally fetched. // than what the browsers originally fetched.
const msg_list = opts.msg_list; const msg_list = opts.msg_list;
if (!msg_list.fetch_status.can_load_older_messages()) { if (!msg_list.fetch_status.can_load_older_messages()) {
// We may already be loading old messages or already // We may already be loading old messages or already
@ -301,7 +301,7 @@ exports.do_backfill = function (opts) {
exports.maybe_load_newer_messages = function (opts) { exports.maybe_load_newer_messages = function (opts) {
// This function gets called when you scroll to the top // This function gets called when you scroll to the top
// of your window, and you want to get messages newer // of your window, and you want to get messages newer
// than what the browers originally fetched. // than what the browsers originally fetched.
const msg_list = opts.msg_list; const msg_list = opts.msg_list;
if (!msg_list.fetch_status.can_load_newer_messages()) { if (!msg_list.fetch_status.can_load_newer_messages()) {

View File

@ -872,7 +872,7 @@ function show_search_query() {
// if query contains stop words, it is enclosed by a <del> tag // if query contains stop words, it is enclosed by a <del> tag
if (page_params.stop_words.includes(query_word)) { if (page_params.stop_words.includes(query_word)) {
// stop_words do not need sanitization so this is unnecesary but it is fail-safe. // stop_words do not need sanitization so this is unnecessary but it is fail-safe.
search_string_display.append($('<del>').text(query_word)); search_string_display.append($('<del>').text(query_word));
query_contains_stop_words = true; query_contains_stop_words = true;
} else { } else {

View File

@ -90,7 +90,7 @@ exports.page_up_the_right_amount = function () {
// because we can't rely on the browser to account for certain // because we can't rely on the browser to account for certain
// page elements, like the compose box, that sit in fixed // page elements, like the compose box, that sit in fixed
// positions above the message pane. For other scrolling // positions above the message pane. For other scrolling
// related adjustements, try to make those happen in the // related adjustments, try to make those happen in the
// scroll handlers, not here. // scroll handlers, not here.
const delta = amount_to_paginate(); const delta = amount_to_paginate();
message_viewport.scrollTop(message_viewport.scrollTop() - delta); message_viewport.scrollTop(message_viewport.scrollTop() - delta);

View File

@ -267,7 +267,7 @@ exports.notify_above_composebox = function (note, link_class, link_msg_id, link_
if (window.electron_bridge !== undefined) { if (window.electron_bridge !== undefined) {
// The code below is for sending a message received from notification reply which // The code below is for sending a message received from notification reply which
// is often refered to as inline reply feature. This is done so desktop app doesn't // is often referred to as inline reply feature. This is done so desktop app doesn't
// have to depend on channel.post for setting crsf_token and narrow.by_topic // have to depend on channel.post for setting crsf_token and narrow.by_topic
// to narrow to the message being sent. // to narrow to the message being sent.
window.electron_bridge.send_notification_reply_message_supported = true; window.electron_bridge.send_notification_reply_message_supported = true;

View File

@ -100,7 +100,7 @@ exports.open_modal = function (name) {
$("#" + name).modal("show").attr("aria-hidden", false); $("#" + name).modal("show").attr("aria-hidden", false);
// Disable background mouse events when modal is active // Disable background mouse events when modal is active
$('.overlay.show').attr("style", "pointer-events: none"); $('.overlay.show').attr("style", "pointer-events: none");
// Remove previous alert messsages from modal, if exists. // Remove previous alert messages from modal, if exists.
$("#" + name).find(".alert").hide(); $("#" + name).find(".alert").hide();
$("#" + name).find(".alert-notification").html(""); $("#" + name).find(".alert-notification").html("");
}; };

View File

@ -143,7 +143,7 @@ function update_url() {
/* Construct the URL that the webhook should be targeting, using /* Construct the URL that the webhook should be targeting, using
the bot's API key and the integration name. The stream and topic the bot's API key and the integration name. The stream and topic
are both optional, and for the sake of completeness, it should be are both optional, and for the sake of completeness, it should be
noted that the topic is irrelavent without specifying the noted that the topic is irrelevant without specifying the
stream. */ stream. */
const url_field = $("#URL")[0]; const url_field = $("#URL")[0];

View File

@ -74,7 +74,7 @@ function update_ui_and_send_reaction_ajax(message_id, reaction_info) {
success: function () {}, success: function () {},
error: function (xhr) { error: function (xhr) {
const response = channel.xhr_error_message("Error sending reaction", xhr); const response = channel.xhr_error_message("Error sending reaction", xhr);
// Errors are somewhat commmon here, due to race conditions // Errors are somewhat common here, due to race conditions
// where the user tries to add/remove the reaction when there is already // where the user tries to add/remove the reaction when there is already
// an in-flight request. We eventually want to make this a blueslip // an in-flight request. We eventually want to make this a blueslip
// error, rather than a warning, but we need to implement either // error, rather than a warning, but we need to implement either

View File

@ -453,7 +453,7 @@ exports.on_load_success = function (realm_people_data) {
} else { } else {
const new_profile_data = []; const new_profile_data = [];
$("#user-info-form-modal .custom_user_field_value").each(function () { $("#user-info-form-modal .custom_user_field_value").each(function () {
// Remove duplicate datepicker input element genearted flatpicker library // Remove duplicate datepicker input element generated flatpicker library
if (!$(this).hasClass("form-control")) { if (!$(this).hasClass("form-control")) {
new_profile_data.push({ new_profile_data.push({
id: parseInt($(this).closest(".custom_user_field").attr("data-field-id"), 10), id: parseInt($(this).closest(".custom_user_field").attr("data-field-id"), 10),

View File

@ -240,7 +240,7 @@ exports.new_stream_clicked = function (stream_name) {
// focus the button on that page, the entire app view jumps over to // focus the button on that page, the entire app view jumps over to
// the other tab, and the animation breaks. // the other tab, and the animation breaks.
// it is unclear whether this is a browser bug or "feature", however what // it is unclear whether this is a browser bug or "feature", however what
// is clear is that this shoudn't be touched unless you're also changing // is clear is that this shouldn't be touched unless you're also changing
// the mobile @media query at 700px. // the mobile @media query at 700px.
if (window.innerWidth > 700) { if (window.innerWidth > 700) {
$('#create_stream_name').focus(); $('#create_stream_name').focus();

View File

@ -23,7 +23,7 @@ exports.build_widget = function (
upload_button, // jQuery button to open file dialog upload_button, // jQuery button to open file dialog
max_file_upload_size max_file_upload_size
) { ) {
// default value of max upladed file size // default value of max uploaded file size
max_file_upload_size = max_file_upload_size || default_max_file_size; max_file_upload_size = max_file_upload_size || default_max_file_size;
function accept(file) { function accept(file) {
@ -109,7 +109,7 @@ exports.build_direct_upload_widget = function (
upload_function, upload_function,
max_file_upload_size max_file_upload_size
) { ) {
// default value of max upladed file size // default value of max uploaded file size
max_file_upload_size = max_file_upload_size || default_max_file_size; max_file_upload_size = max_file_upload_size || default_max_file_size;
function accept() { function accept() {
input_error.hide(); input_error.hide();

View File

@ -378,7 +378,7 @@ on a dark background, and don't change the dark labels dark either. */
} }
#invite_user_form .modal-footer { #invite_user_form .modal-footer {
// no transperancy prevents overlap issues // no transparency prevents overlap issues
background-color: hsl(211, 28%, 14%); background-color: hsl(211, 28%, 14%);
} }

View File

@ -297,7 +297,7 @@ $ cat hello.html
The point of marked was to create a markdown compiler where it was possible to The point of marked was to create a markdown compiler where it was possible to
frequently parse huge chunks of markdown without having to worry about frequently parse huge chunks of markdown without having to worry about
caching the compiled output somehow...or blocking for an unnecesarily long time. caching the compiled output somehow...or blocking for an unnecessarily long time.
marked is very concise and still implements all markdown features. It is also marked is very concise and still implements all markdown features. It is also
now fully compatible with the client-side. now fully compatible with the client-side.

View File

@ -59,7 +59,7 @@ You may pass the `client_gravatar` query parameter as follows:
* `email`: The email address of the user or bot. * `email`: The email address of the user or bot.
* `is_bot`: A boolean specifying whether the user is a bot or not. * `is_bot`: A boolean specifying whether the user is a bot or not.
* `avatar_url`: URL to the user's gravatar. `None` if the `client_gravatar` * `avatar_url`: URL to the user's gravatar. `None` if the `client_gravatar`
query paramater was set to `True`. query parameter was set to `True`.
* `full_name`: Full name of the user or bot. * `full_name`: Full name of the user or bot.
* `is_admin`: A boolean specifying whether the user is an admin or not. * `is_admin`: A boolean specifying whether the user is an admin or not.
* `bot_type`: `None` if the user isn't a bot. `1` for a `Generic` bot. * `bot_type`: `None` if the user isn't a bot. `1` for a `Generic` bot.

View File

@ -81,7 +81,7 @@ endpoint and a queue would be registered in the absence of a `queue_id`.
#### Return values #### Return values
* `events`: An array (possibly zero-length if `dont_block` is set) of events * `events`: An array (possibly zero-length if `dont_block` is set) of events
with IDs newer than `last_event_id`. Event IDs are guaranted to be increasing, with IDs newer than `last_event_id`. Event IDs are guaranteed to be increasing,
but they are not guaranteed to be consecutive. but they are not guaranteed to be consecutive.
#### Example response #### Example response

View File

@ -57,11 +57,11 @@ You may pass the `include_subscribers` query parameter as follows:
Only people who have been invited can access a private stream. Only people who have been invited can access a private stream.
* `subscribers`: A list of email addresses of users who are also subscribed * `subscribers`: A list of email addresses of users who are also subscribed
to a given stream. Included only if `include_subscribers` is `true`. to a given stream. Included only if `include_subscribers` is `true`.
* `desktop_notifications`: A boolean specifiying whether desktop notifications * `desktop_notifications`: A boolean specifying whether desktop notifications
are enabled for the given stream. are enabled for the given stream.
* `push_notifications`: A boolean specifiying whether push notifications * `push_notifications`: A boolean specifying whether push notifications
are enabled for the given stream. are enabled for the given stream.
* `audible_notifications`: A boolean specifiying whether audible notifications * `audible_notifications`: A boolean specifying whether audible notifications
are enabled for the given stream. are enabled for the given stream.
* `pin_to_top`: A boolean specifying whether the given stream has been pinned * `pin_to_top`: A boolean specifying whether the given stream has been pinned
to the top. to the top.

View File

@ -39,7 +39,7 @@ You may pass the `client_gravatar` or `include_custom_profile_fields` query para
* `email`: The email address of the user or bot. * `email`: The email address of the user or bot.
* `is_bot`: A boolean specifying whether the user is a bot or not. * `is_bot`: A boolean specifying whether the user is a bot or not.
* `avatar_url`: URL to the user's gravatar. `None` if the `client_gravatar` * `avatar_url`: URL to the user's gravatar. `None` if the `client_gravatar`
query paramater was set to `True`. query parameter was set to `True`.
* `full_name`: Full name of the user or bot. * `full_name`: Full name of the user or bot.
* `is_admin`: A boolean specifying whether the user is an admin or not. * `is_admin`: A boolean specifying whether the user is an admin or not.
* `bot_type`: `None` if the user isn't a bot. `1` for a `Generic` bot. * `bot_type`: `None` if the user isn't a bot. `1` for a `Generic` bot.

View File

@ -90,7 +90,7 @@ below are for a webhook named `MyWebHook`.
platform/server/product you are integrating. Used on the documentation platform/server/product you are integrating. Used on the documentation
pages as well as the sender's avatar for messages sent by the integration. pages as well as the sender's avatar for messages sent by the integration.
* `static/images/integrations/mywebbook/001.svg`: A screenshot of a message * `static/images/integrations/mywebbook/001.svg`: A screenshot of a message
sent by the integration, used on the documenation page. sent by the integration, used on the documentation page.
### Files that need to be updated ### Files that need to be updated

View File

@ -131,7 +131,7 @@
notifications? Enjoy Zulip on your desktop.</p> notifications? Enjoy Zulip on your desktop.</p>
</a> </a>
<!--Hack: These two pseudo elements are here to ensure the flex <!--Hack: These two pseudo elements are here to ensure the flex
arrangment uses the proper cell size with 4 elements in 2 rows.--> arrangement uses the proper cell size with 4 elements in 2 rows.-->
<div class="feature-block"></div> <div class="feature-block"></div>
<div class="feature-block"></div> <div class="feature-block"></div>
</section> </section>
@ -301,7 +301,7 @@
your use case is missing, you can make it happen!</p> your use case is missing, you can make it happen!</p>
</a> </a>
<!--Hack: These two pseudo elements are here to ensure the flex <!--Hack: These two pseudo elements are here to ensure the flex
arrangment uses the proper cell size with 4 elements in 2 rows.--> arrangement uses the proper cell size with 4 elements in 2 rows.-->
<div class="feature-block"></div> <div class="feature-block"></div>
</section> </section>
</div> </div>

View File

@ -2,7 +2,7 @@
By default, anyone can change their email address at any time. By default, anyone can change their email address at any time.
Organization adminstrators can Organization administrators can
[restrict users to certain email domains](/help/change-a-users-name), or [restrict users to certain email domains](/help/change-a-users-name), or
[prevent users from changing their email](/help/restrict-name-and-email-changes). [prevent users from changing their email](/help/restrict-name-and-email-changes).

View File

@ -103,7 +103,7 @@ Get a peek at new features before they're released!
Start by finding the latest version marked "Pre-release" on the Start by finding the latest version marked "Pre-release" on the
[release list page][release-list]. There may or may not be a "Pre-release" [release list page][release-list]. There may or may not be a "Pre-release"
later than the latest release. If there is, download the approriate Zulip later than the latest release. If there is, download the appropriate Zulip
installer or app from there, and follow the instructions for your operating installer or app from there, and follow the instructions for your operating
system above. system above.

View File

@ -22,7 +22,7 @@ apt-mark hold initramfs-tools initramfs-tools-bin udev base-files linux-firmware
apt-mark hold accountsservice apparmor apport apt apt-transport-https apt-utils bash bash-completion bind9-host binutils binutils-doc bsdutils bzr cloud-guest-utils cloud-init coreutils cpio dbus dnsutils dosfstools dpkg dpkg-dev e2fslibs e2fsprogs eject gcc-4.9-base git-core grub-common grub-pc grub-pc-bin grub2-common icu-devtools ifupdown imagemagick imagemagick-common init-system-helpers initscripts irqbalance isc-dhcp-client isc-dhcp-common klibc-utils krb5-locales krb5-multidev libaccountsservice0 libapparmor-perl libapparmor1 libblkid1 libc-bin libc-dev-bin libc6 libc6-dev libcdt5 libcgmanager0 libcgraph6 libcups2 libcurl3-gnutls libdbus-1-3 libdpkg-perl libdrm-intel1 libdrm-nouveau2 libdrm-radeon1 libdrm2 libevent-2.0-5 libexpat1 libexpat1-dev libgc1c2 libgcc1 libgd3 libgl1-mesa-dri libgl1-mesa-glx libglapi-mesa libgnutls-openssl27 libgraphite2-3 libgraphviz-dev libgssapi-krb5-2 libgssrpc4 libgtk2.0-0 libgtk2.0-common libgvc6 libgvpr2 libicu-dev libjasper-dev libjasper1 libk5crypto3 libkadm5clnt-mit9 libkadm5srv-mit9 libklibc libkrb5-3 libkrb5-dev libkrb5support0 liblcms2-2 liblcms2-dev libmagickwand-dev libmount1 libmysqlclient-dev libnl-3-200 libnl-genl-3-200 libnspr4 libnss3 libnss3-nssdb libnuma1 libpam-modules libpam-modules-bin libpam-runtime libpam-systemd libpam0g libpam0g-dev libpathplan4 libpci3 libpcre3 libpcre3-dev libpcsclite1 libpixman-1-0 libpixman-1-dev libpng12-0 libpng12-dev libpolkit-agent-1-0 libpolkit-backend-1-0 libpolkit-gobject-1-0 libpython3.4 libsndfile1 libss2 libssl-dev libssl1.0.0 libtasn1-6 libtiff5 libtiff5-dev libtiffxx5 libuuid1 libxdot4 libxml2 libxml2-dev libxpm4 linux-libc-dev login lsb-base lshw makedev mongodb-org mongodb-org-mongos mongodb-org-server mongodb-org-shell mongodb-org-tools mount multiarch-support mysql-client-5.7 mysql-client-core-5.7 mysql-common mysql-server-5.7 mysql-server-core-5.7 ntpdate openssh-client openssh-server openssh-sftp-server os-prober overlayroot passwd pciutils perl perl-base pgdg-keyring policykit-1 pollinate postgresql-client postgresql-client-common postgresql-common python-apt python-apt-common python-bzrlib python-urllib3 python3-apport python3-apt python3-distupgrade python3-gdbm python3-problem-report python3-software-properties python3-update-manager python3.4 python3.4-minimal rsync software-properties-common sudo sysv-rc sysvinit-utils tar tcpdump tzdata ubuntu-release-upgrader-core unzip update-manager-core usbutils util-linux uuid-runtime w3m apt-mark hold accountsservice apparmor apport apt apt-transport-https apt-utils bash bash-completion bind9-host binutils binutils-doc bsdutils bzr cloud-guest-utils cloud-init coreutils cpio dbus dnsutils dosfstools dpkg dpkg-dev e2fslibs e2fsprogs eject gcc-4.9-base git-core grub-common grub-pc grub-pc-bin grub2-common icu-devtools ifupdown imagemagick imagemagick-common init-system-helpers initscripts irqbalance isc-dhcp-client isc-dhcp-common klibc-utils krb5-locales krb5-multidev libaccountsservice0 libapparmor-perl libapparmor1 libblkid1 libc-bin libc-dev-bin libc6 libc6-dev libcdt5 libcgmanager0 libcgraph6 libcups2 libcurl3-gnutls libdbus-1-3 libdpkg-perl libdrm-intel1 libdrm-nouveau2 libdrm-radeon1 libdrm2 libevent-2.0-5 libexpat1 libexpat1-dev libgc1c2 libgcc1 libgd3 libgl1-mesa-dri libgl1-mesa-glx libglapi-mesa libgnutls-openssl27 libgraphite2-3 libgraphviz-dev libgssapi-krb5-2 libgssrpc4 libgtk2.0-0 libgtk2.0-common libgvc6 libgvpr2 libicu-dev libjasper-dev libjasper1 libk5crypto3 libkadm5clnt-mit9 libkadm5srv-mit9 libklibc libkrb5-3 libkrb5-dev libkrb5support0 liblcms2-2 liblcms2-dev libmagickwand-dev libmount1 libmysqlclient-dev libnl-3-200 libnl-genl-3-200 libnspr4 libnss3 libnss3-nssdb libnuma1 libpam-modules libpam-modules-bin libpam-runtime libpam-systemd libpam0g libpam0g-dev libpathplan4 libpci3 libpcre3 libpcre3-dev libpcsclite1 libpixman-1-0 libpixman-1-dev libpng12-0 libpng12-dev libpolkit-agent-1-0 libpolkit-backend-1-0 libpolkit-gobject-1-0 libpython3.4 libsndfile1 libss2 libssl-dev libssl1.0.0 libtasn1-6 libtiff5 libtiff5-dev libtiffxx5 libuuid1 libxdot4 libxml2 libxml2-dev libxpm4 linux-libc-dev login lsb-base lshw makedev mongodb-org mongodb-org-mongos mongodb-org-server mongodb-org-shell mongodb-org-tools mount multiarch-support mysql-client-5.7 mysql-client-core-5.7 mysql-common mysql-server-5.7 mysql-server-core-5.7 ntpdate openssh-client openssh-server openssh-sftp-server os-prober overlayroot passwd pciutils perl perl-base pgdg-keyring policykit-1 pollinate postgresql-client postgresql-client-common postgresql-common python-apt python-apt-common python-bzrlib python-urllib3 python3-apport python3-apt python3-distupgrade python3-gdbm python3-problem-report python3-software-properties python3-update-manager python3.4 python3.4-minimal rsync software-properties-common sudo sysv-rc sysvinit-utils tar tcpdump tzdata ubuntu-release-upgrader-core unzip update-manager-core usbutils util-linux uuid-runtime w3m
if ! apt-get dist-upgrade -y "${APT_OPTIONS[@]}"; then if ! apt-get dist-upgrade -y "${APT_OPTIONS[@]}"; then
echo "\`apt-get dist-upgrade\`: Failure occured while trying to perform distribution upgrade, Retrying..." echo "\`apt-get dist-upgrade\`: Failure occurred while trying to perform distribution upgrade, Retrying..."
apt-get dist-upgrade -y "${APT_OPTIONS[@]}" apt-get dist-upgrade -y "${APT_OPTIONS[@]}"
fi fi

View File

@ -9,7 +9,7 @@ set -e
# usage: clean-branches --reviews # usage: clean-branches --reviews
# Deletes all the above mentioned branches as well as branches # Deletes all the above mentioned branches as well as branches
# created by the scripts like `fetch-rebase-pull-request`. Be careful # created by the scripts like `fetch-rebase-pull-request`. Be careful
# as this would also remove other branches woth names like review-* # as this would also remove other branches with names like review-*
review=0 review=0
if [ $# -ne 0 ] && [ "$1" == "--reviews" ]; then if [ $# -ne 0 ] && [ "$1" == "--reviews" ]; then

View File

@ -15,7 +15,7 @@ EMOJI_NAME_MAPS = {
# laughter_tears from https://beebom.com/emoji-meanings/ # laughter_tears from https://beebom.com/emoji-meanings/
'1f602': {'canonical_name': 'joy', 'aliases': ['tears', 'laughter_tears']}, '1f602': {'canonical_name': 'joy', 'aliases': ['tears', 'laughter_tears']},
'1f923': {'canonical_name': 'rolling_on_the_floor_laughing', 'aliases': ['rofl']}, '1f923': {'canonical_name': 'rolling_on_the_floor_laughing', 'aliases': ['rofl']},
# not sure how the glpyhs match relaxed, but both iamcal and gemoji have it # not sure how the glyphs match relaxed, but both iamcal and gemoji have it
'263a': {'canonical_name': 'smile', 'aliases': ['relaxed']}, '263a': {'canonical_name': 'smile', 'aliases': ['relaxed']},
'1f60a': {'canonical_name': 'blush', 'aliases': []}, '1f60a': {'canonical_name': 'blush', 'aliases': []},
# halo comes from gemoji/unicode # halo comes from gemoji/unicode
@ -894,7 +894,7 @@ EMOJI_NAME_MAPS = {
'26f5': {'canonical_name': 'boat', 'aliases': ['sailboat']}, '26f5': {'canonical_name': 'boat', 'aliases': ['sailboat']},
'1f6e5': {'canonical_name': 'motor_boat', 'aliases': []}, '1f6e5': {'canonical_name': 'motor_boat', 'aliases': []},
'1f6a4': {'canonical_name': 'speedboat', 'aliases': []}, '1f6a4': {'canonical_name': 'speedboat', 'aliases': []},
# yatch and cruise seem like reasonable additions # yacht and cruise seem like reasonable additions
'1f6f3': {'canonical_name': 'passenger_ship', 'aliases': ['yacht', 'cruise']}, '1f6f3': {'canonical_name': 'passenger_ship', 'aliases': ['yacht', 'cruise']},
'26f4': {'canonical_name': 'ferry', 'aliases': []}, '26f4': {'canonical_name': 'ferry', 'aliases': []},
'1f6a2': {'canonical_name': 'ship', 'aliases': []}, '1f6a2': {'canonical_name': 'ship', 'aliases': []},

View File

@ -92,7 +92,7 @@ def generate_emoji_catalog(emoji_data: List[Dict[str, Any]],
return dict(emoji_catalog) return dict(emoji_catalog)
# Use only those names for which images are present in all # Use only those names for which images are present in all
# the emoji sets so that we can switch emoji sets seemlessly. # the emoji sets so that we can switch emoji sets seamlessly.
def emoji_is_universal(emoji_dict: Dict[str, Any]) -> bool: def emoji_is_universal(emoji_dict: Dict[str, Any]) -> bool:
for emoji_set in EMOJISETS: for emoji_set in EMOJISETS:
if not emoji_dict['has_img_' + emoji_set]: if not emoji_dict['has_img_' + emoji_set]:

View File

@ -113,7 +113,7 @@ def run_production() -> None:
contribs_list[username] = contrib_data contribs_list[username] = contrib_data
# remove duplicate contributions count # remove duplicate contributions count
# find commits at the time of split and substract from zulip-server # find commits at the time of split and subtract from zulip-server
with open(duplicate_commits_file, 'r') as f: with open(duplicate_commits_file, 'r') as f:
duplicate_commits = json.load(f) duplicate_commits = json.load(f)
for committer in duplicate_commits: for committer in duplicate_commits:

View File

@ -528,7 +528,7 @@ def process_avatars(avatar_list: List[ZerverFieldsT], avatar_dir: str, realm_id:
avatar_original['s3_path'] = original_image_path avatar_original['s3_path'] = original_image_path
avatar_original_list.append(avatar_original) avatar_original_list.append(avatar_original)
# Run downloads parallely # Run downloads in parallel
output = [] output = []
for (status, job) in run_parallel_wrapper(get_avatar, avatar_upload_list, threads=threads): for (status, job) in run_parallel_wrapper(get_avatar, avatar_upload_list, threads=threads):
output.append(job) output.append(job)
@ -617,7 +617,7 @@ def process_uploads(upload_list: List[ZerverFieldsT], upload_dir: str,
upload_url_list.append([upload_url, upload_s3_path]) upload_url_list.append([upload_url, upload_s3_path])
upload['path'] = upload_s3_path upload['path'] = upload_s3_path
# Run downloads parallely # Run downloads in parallel
output = [] output = []
for (status, job) in run_parallel_wrapper(get_uploads, upload_url_list, threads=threads): for (status, job) in run_parallel_wrapper(get_uploads, upload_url_list, threads=threads):
output.append(job) output.append(job)
@ -678,7 +678,7 @@ def process_emojis(zerver_realmemoji: List[ZerverFieldsT], emoji_dir: str,
emoji_records.append(emoji_record) emoji_records.append(emoji_record)
# Run downloads parallely # Run downloads in parallel
output = [] output = []
for (status, job) in run_parallel_wrapper(get_emojis, upload_emoji_list, threads=threads): for (status, job) in run_parallel_wrapper(get_emojis, upload_emoji_list, threads=threads):
output.append(job) output.append(job)

View File

@ -651,7 +651,7 @@ def label_mirror_dummy_users(num_teams: int, team_name: str, mattermost_data: Di
username_to_user: Dict[str, Dict[str, Any]]) -> None: username_to_user: Dict[str, Dict[str, Any]]) -> None:
# This function might looks like a great place to label admin users. But # This function might looks like a great place to label admin users. But
# that won't be fully correct since we are iterating only though posts and # that won't be fully correct since we are iterating only though posts and
# it covers only users that has sent atleast one message. # it covers only users that have sent at least one message.
for post in mattermost_data["post"]["channel_post"]: for post in mattermost_data["post"]["channel_post"]:
post_team = post["team"] post_team = post["team"]
if post_team == team_name: if post_team == team_name:

View File

@ -493,7 +493,7 @@ def process_long_term_idle_users(slack_data_dir: str, users: List[ZerverFieldsT]
"""Algorithmically, we treat users who have sent at least 10 messages """Algorithmically, we treat users who have sent at least 10 messages
or have sent a message within the last 60 days as active. or have sent a message within the last 60 days as active.
Everyone else is treated as long-term idle, which means they will Everyone else is treated as long-term idle, which means they will
have a slighly slower first page load when coming back to have a slightly slower first page load when coming back to
Zulip. Zulip.
""" """
all_messages = get_messages_iterator(slack_data_dir, added_channels, added_mpims, dm_members) all_messages = get_messages_iterator(slack_data_dir, added_channels, added_mpims, dm_members)

View File

@ -826,7 +826,7 @@ def do_deactivate_stream(stream: Stream, log: bool=True) -> None:
new_name = ("!DEACTIVATED:" + old_name)[:Stream.MAX_NAME_LENGTH] new_name = ("!DEACTIVATED:" + old_name)[:Stream.MAX_NAME_LENGTH]
for i in range(20): for i in range(20):
if stream_name_in_use(new_name, stream.realm_id): if stream_name_in_use(new_name, stream.realm_id):
# This stream has alrady been deactivated, keep prepending !s until # This stream has already been deactivated, keep prepending !s until
# we have a unique stream name or you've hit a rename limit. # we have a unique stream name or you've hit a rename limit.
new_name = ("!" + new_name)[:Stream.MAX_NAME_LENGTH] new_name = ("!" + new_name)[:Stream.MAX_NAME_LENGTH]
else: else:
@ -1223,7 +1223,7 @@ def get_service_bot_events(sender: UserProfile, service_bot_tuples: List[Tuple[i
# Mention triggers, for stream messages # Mention triggers, for stream messages
if is_stream and user_profile_id in mentioned_user_ids: if is_stream and user_profile_id in mentioned_user_ids:
trigger = 'mention' trigger = 'mention'
# PM triggers for personal and huddle messsages # PM triggers for personal and huddle messages
elif (not is_stream) and (user_profile_id in active_user_ids): elif (not is_stream) and (user_profile_id in active_user_ids):
trigger = 'private_message' trigger = 'private_message'
else: else:
@ -5239,7 +5239,7 @@ def do_revoke_user_invite(prereg_user: PreregistrationUser) -> None:
email = prereg_user.email email = prereg_user.email
# Delete both the confirmation objects and the prereg_user object. # Delete both the confirmation objects and the prereg_user object.
# TODO: Probably we actaully want to set the confirmation objects # TODO: Probably we actually want to set the confirmation objects
# to a "revoked" status so that we can give the invited user a better # to a "revoked" status so that we can give the invited user a better
# error message. # error message.
content_type = ContentType.objects.get_for_model(PreregistrationUser) content_type = ContentType.objects.get_for_model(PreregistrationUser)

View File

@ -2091,7 +2091,7 @@ def privacy_clean_markdown(content: str) -> str:
def log_bugdown_error(msg: str) -> None: def log_bugdown_error(msg: str) -> None:
"""We use this unusual logging approach to log the bugdown error, in """We use this unusual logging approach to log the bugdown error, in
order to prevent AdminNotifyHandler from sending the santized order to prevent AdminNotifyHandler from sending the sanitized
original markdown formatting into another Zulip message, which original markdown formatting into another Zulip message, which
could cause an infinite exception loop.""" could cause an infinite exception loop."""
bugdown_logger.error(msg) bugdown_logger.error(msg)

View File

@ -99,7 +99,7 @@ def bulk_set_users_or_streams_recipient_fields(model: Model,
# we take adventage of this, instead of calling save individually. # we take adventage of this, instead of calling save individually.
result.save(update_fields=['recipient']) result.save(update_fields=['recipient'])
# This is only sed in populate_db, so doesn't realy need tests # This is only sed in populate_db, so doesn't really need tests
def bulk_create_streams(realm: Realm, def bulk_create_streams(realm: Realm,
stream_dict: Dict[str, Dict[str, Any]]) -> None: # nocoverage stream_dict: Dict[str, Dict[str, Any]]) -> None: # nocoverage
existing_streams = frozenset([name.lower() for name in existing_streams = frozenset([name.lower() for name in

View File

@ -656,7 +656,7 @@ def ignore_unhashable_lru_cache(maxsize: int=128, typed: bool=False) -> DECORATO
pass pass
# Deliberately calling this function from outside of exception # Deliberately calling this function from outside of exception
# handler to get a more descriptive traceback. Otherise traceback # handler to get a more descriptive traceback. Otherwise traceback
# can include the exception from cached_enabled_user_function as # can include the exception from cached_enabled_user_function as
# well. # well.
return user_function(*args, **kwargs) return user_function(*args, **kwargs)

View File

@ -8,7 +8,7 @@ from zerver.lib.validator import check_required_string, \
check_external_account_url_pattern, check_dict_only check_external_account_url_pattern, check_dict_only
from zerver.lib.types import ProfileFieldData from zerver.lib.types import ProfileFieldData
# Default external account fields are by default avaliable # Default external account fields are by default available
# to realm admins, where realm admin only need to select # to realm admins, where realm admin only need to select
# the default field and other values(i.e. name, url) will be # the default field and other values(i.e. name, url) will be
# fetch from this dictionary. # fetch from this dictionary.

View File

@ -48,7 +48,7 @@ realm_tables = [("zerver_defaultstream", DefaultStream, "defaultstream"),
# that map old ids to new ids. We use this in # that map old ids to new ids. We use this in
# re_map_foreign_keys and other places. # re_map_foreign_keys and other places.
# #
# We explicity initialize ID_MAP with the tables that support # We explicitly initialize ID_MAP with the tables that support
# id re-mapping. # id re-mapping.
# #
# Code reviewers: give these tables extra scrutiny, as we need to # Code reviewers: give these tables extra scrutiny, as we need to

View File

@ -170,7 +170,7 @@ class MessageDict:
@staticmethod @staticmethod
def wide_dict(message: Message) -> Dict[str, Any]: def wide_dict(message: Message) -> Dict[str, Any]:
''' '''
The next two lines get the cachable field related The next two lines get the cacheable field related
to our message object, with the side effect of to our message object, with the side effect of
populating the cache. populating the cache.
''' '''

View File

@ -238,7 +238,7 @@ def list_to_streams(streams_raw: Iterable[Mapping[str, Any]],
@param streams_raw The list of stream dictionaries to process; @param streams_raw The list of stream dictionaries to process;
names should already be stripped of whitespace by the caller. names should already be stripped of whitespace by the caller.
@param user_profile The user for whom we are retreiving the streams @param user_profile The user for whom we are retrieving the streams
@param autocreate Whether we should create streams if they don't already exist @param autocreate Whether we should create streams if they don't already exist
""" """
# Validate all streams, getting extant ones, then get-or-creating the rest. # Validate all streams, getting extant ones, then get-or-creating the rest.

View File

@ -382,7 +382,7 @@ class ParallelTestSuite(django_runner.ParallelTestSuite):
def check_import_error(test_name: str) -> None: def check_import_error(test_name: str) -> None:
try: try:
# Directly using __import__ is not recommeded, but here it gives # Directly using __import__ is not recommended, but here it gives
# clearer traceback as compared to importlib.import_module. # clearer traceback as compared to importlib.import_module.
__import__(test_name) __import__(test_name)
except ImportError as exc: except ImportError as exc:

View File

@ -18,7 +18,7 @@ class Command(ZulipBaseCommand):
def handle(self, *args: Any, **options: Any) -> None: def handle(self, *args: Any, **options: Any) -> None:
try: try:
# first check if the db has been initalized # first check if the db has been initialized
Realm.objects.first() Realm.objects.first()
except ProgrammingError: except ProgrammingError:
raise CommandError("The Zulip database does not appear to exist. " raise CommandError("The Zulip database does not appear to exist. "

View File

@ -20,7 +20,7 @@ Zulip's message retention and deletion features.
Examples: Examples:
To restore all recently deleted messages: To restore all recently deleted messages:
./manage.py restore_messages ./manage.py restore_messages
To restore a specfic ArchiveTransaction: To restore a specific ArchiveTransaction:
./manage.py restore_messages --transaction-id=1 ./manage.py restore_messages --transaction-id=1
""" """

View File

@ -23,7 +23,7 @@ from zerver.models import Realm, get_realm, get_stream
# See zerver/tests/fixtures/email/1.txt for a very simple example, # See zerver/tests/fixtures/email/1.txt for a very simple example,
# but anything that the message_from_binary_file function # but anything that the message_from_binary_file function
# from the email library can parse should work. # from the email library can parse should work.
# Value of the TO: header doesn't matter, as it is overriden # Value of the TO: header doesn't matter, as it is overridden
# by the command in order for the email to be sent to the correct stream. # by the command in order for the email to be sent to the correct stream.
class Command(ZulipBaseCommand): class Command(ZulipBaseCommand):

View File

@ -377,7 +377,7 @@ class Realm(models.Model):
max_length=1) # type: str max_length=1) # type: str
icon_version = models.PositiveSmallIntegerField(default=1) # type: int icon_version = models.PositiveSmallIntegerField(default=1) # type: int
# Logo is the horizonal logo we show in top-left of webapp navbar UI. # Logo is the horizontal logo we show in top-left of webapp navbar UI.
LOGO_DEFAULT = u'D' LOGO_DEFAULT = u'D'
LOGO_UPLOADED = u'U' LOGO_UPLOADED = u'U'
LOGO_SOURCES = ( LOGO_SOURCES = (
@ -795,7 +795,7 @@ class UserProfile(AbstractBaseUser, PermissionsMixin):
# with EMAIL_ADDRESS_VISIBILITY_EVERYONE. For other # with EMAIL_ADDRESS_VISIBILITY_EVERYONE. For other
# organizations, it will be a unique value of the form # organizations, it will be a unique value of the form
# user1234@example.com. This field exists for backwards # user1234@example.com. This field exists for backwards
# compatibility in Zulip APIs where users are refered to by their # compatibility in Zulip APIs where users are referred to by their
# email address, not their ID; it should be used in all API use cases. # email address, not their ID; it should be used in all API use cases.
# #
# Both fields are unique within a realm (in a case-insensitive fashion). # Both fields are unique within a realm (in a case-insensitive fashion).
@ -1668,7 +1668,7 @@ class Message(AbstractMessage):
Find out whether a message is a stream message by Find out whether a message is a stream message by
looking up its recipient.type. TODO: Make this looking up its recipient.type. TODO: Make this
an easier operation by denormalizing the message an easier operation by denormalizing the message
type onto Message, either explicity (message.type) type onto Message, either explicitly (message.type)
or implicitly (message.stream_id is not None). or implicitly (message.stream_id is not None).
''' '''
return self.recipient.type == Recipient.STREAM return self.recipient.type == Recipient.STREAM
@ -2061,7 +2061,7 @@ class Subscription(models.Model):
# Whether the user has since unsubscribed. We mark Subscription # Whether the user has since unsubscribed. We mark Subscription
# objects as inactive, rather than deleting them, when a user # objects as inactive, rather than deleting them, when a user
# unsubscribes, so we can preseve user customizations like # unsubscribes, so we can preserve user customizations like
# notification settings, stream color, etc., if the user later # notification settings, stream color, etc., if the user later
# resubscribes. # resubscribes.
active = models.BooleanField(default=True) # type: bool active = models.BooleanField(default=True) # type: bool

View File

@ -117,7 +117,7 @@ def validate_against_openapi_schema(content: Dict[str, Any], endpoint: str,
raise SchemaError('Expected to find the "{}" required key') raise SchemaError('Expected to find the "{}" required key')
def to_python_type(py_type: str) -> type: def to_python_type(py_type: str) -> type:
"""Transform an OpenAPI-like type to a Pyton one. """Transform an OpenAPI-like type to a Python one.
https://swagger.io/docs/specification/data-models/data-types https://swagger.io/docs/specification/data-models/data-types
""" """
TYPES = { TYPES = {

View File

@ -125,7 +125,7 @@ paths:
description: An array of `event` objects (possibly description: An array of `event` objects (possibly
zero-length if `dont_block` is set) of events with zero-length if `dont_block` is set) of events with
IDs newer than `last_event_id`. Event IDs are IDs newer than `last_event_id`. Event IDs are
guaranted to be increasing, but they are not guaranteed to be increasing, but they are not
guaranteed to be consecutive. guaranteed to be consecutive.
- example: - example:
{ {

View File

@ -8,7 +8,7 @@ from zerver.models import get_realm
class GlobalPublicStreamTest(ZulipTestCase): class GlobalPublicStreamTest(ZulipTestCase):
def test_non_existant_stream_id(self) -> None: def test_non_existant_stream_id(self) -> None:
# Here we use a relatively big number as stream id assumming such an id # Here we use a relatively big number as stream id assuming such an id
# won't exist in the test DB. # won't exist in the test DB.
result = self.client_get("/archive/streams/100000000/topics/TopicGlobal") result = self.client_get("/archive/streams/100000000/topics/TopicGlobal")
self.assert_in_success_response(["This stream does not exist."], result) self.assert_in_success_response(["This stream does not exist."], result)

View File

@ -646,7 +646,7 @@ class DesktopFlowTestingLib(ZulipTestCase):
class SocialAuthBase(DesktopFlowTestingLib, ZulipTestCase): class SocialAuthBase(DesktopFlowTestingLib, ZulipTestCase):
"""This is a base class for testing social-auth backends. These """This is a base class for testing social-auth backends. These
methods are often overriden by subclasses: methods are often overridden by subclasses:
register_extra_endpoints() - If the backend being tested calls some extra register_extra_endpoints() - If the backend being tested calls some extra
endpoints then they can be added here. endpoints then they can be added here.
@ -1576,7 +1576,7 @@ class SAMLAuthBackendTest(SocialAuthBase):
self.assert_in_success_response(["Log in with Second Test IdP"], result) self.assert_in_success_response(["Log in with Second Test IdP"], result)
self.assert_in_success_response(["/accounts/login/social/saml/test_idp2"], result) self.assert_in_success_response(["/accounts/login/social/saml/test_idp2"], result)
# Try succesful authentication with the regular idp from all previous tests: # Try successful authentication with the regular idp from all previous tests:
self.test_social_auth_success() self.test_social_auth_success()
# Now test with the second idp: # Now test with the second idp:

View File

@ -143,7 +143,7 @@ class CreateCustomProfileFieldTest(CustomProfileFieldTestCase):
name=invalid_field_name, name=invalid_field_name,
)) ))
self.assert_json_success(result) self.assert_json_success(result)
# Sliently overwrite name and hint with values set in default fields dict # Silently overwrite name and hint with values set in default fields dict
# for default custom external account fields. # for default custom external account fields.
with self.assertRaises(CustomProfileField.DoesNotExist): with self.assertRaises(CustomProfileField.DoesNotExist):
field = CustomProfileField.objects.get(name=invalid_field_name, realm=realm) field = CustomProfileField.objects.get(name=invalid_field_name, realm=realm)

View File

@ -365,7 +365,7 @@ class PlansPageTest(ZulipTestCase):
# Test root domain # Test root domain
result = self.client_get("/plans/", subdomain="") result = self.client_get("/plans/", subdomain="")
self.assert_in_success_response(["Sign up now"], result) self.assert_in_success_response(["Sign up now"], result)
# Test non-existant domain # Test non-existent domain
result = self.client_get("/plans/", subdomain="moo") result = self.client_get("/plans/", subdomain="moo")
self.assertEqual(result.status_code, 404) self.assertEqual(result.status_code, 404)
self.assert_in_response("does not exist", result) self.assert_in_response("does not exist", result)

View File

@ -56,7 +56,7 @@ class EmailTranslationTestCase(ZulipTestCase):
class TranslationTestCase(ZulipTestCase): class TranslationTestCase(ZulipTestCase):
""" """
Tranlations strings should change with locale. URLs should be locale Translations strings should change with locale. URLs should be locale
aware. aware.
""" """

View File

@ -233,7 +233,7 @@ class TestIntegrationsDevPanel(ZulipTestCase):
# We have to use this roundabout manner since the order may vary each time. This is not # We have to use this roundabout manner since the order may vary each time. This is not
# an issue. Basically, we're trying to compare 2 lists and since we're not resorting to # an issue. Basically, we're trying to compare 2 lists and since we're not resorting to
# using sets or a sorted order, we're sticking with O(n*m) time complexity for this # using sets or a sorted order, we're sticking with O(n*m) time complexity for this
# comparision (where n and m are the lengths of the two lists respectively). But since # comparison (where n and m are the lengths of the two lists respectively). But since
# this is just a unit test and more importantly n = m = some-low-number we don't really # this is just a unit test and more importantly n = m = some-low-number we don't really
# care about the time complexity being what it is. # care about the time complexity being what it is.
self.assertTrue(r in expected_responses) self.assertTrue(r in expected_responses)

View File

@ -4333,14 +4333,14 @@ class DeleteMessageTest(ZulipTestCase):
result = test_delete_message_by_admin(msg_id=msg_id_2) result = test_delete_message_by_admin(msg_id=msg_id_2)
self.assert_json_success(result) self.assert_json_success(result)
# Test mulitple delete requests with no latency issues # Test multiple delete requests with no latency issues
msg_id = self.send_stream_message(hamlet, "Scotland") msg_id = self.send_stream_message(hamlet, "Scotland")
result = test_delete_message_by_owner(msg_id=msg_id) result = test_delete_message_by_owner(msg_id=msg_id)
self.assert_json_success(result) self.assert_json_success(result)
result = test_delete_message_by_owner(msg_id=msg_id) result = test_delete_message_by_owner(msg_id=msg_id)
self.assert_json_error(result, "Invalid message(s)") self.assert_json_error(result, "Invalid message(s)")
# Test handling of 500 error caused by mulitple delete requests due to latency. # Test handling of 500 error caused by multiple delete requests due to latency.
# see issue #11219. # see issue #11219.
with mock.patch("zerver.views.messages.do_delete_messages") as m, \ with mock.patch("zerver.views.messages.do_delete_messages") as m, \
mock.patch("zerver.views.messages.validate_can_delete_message", return_value=None), \ mock.patch("zerver.views.messages.validate_can_delete_message", return_value=None), \

View File

@ -376,7 +376,7 @@ class PushBouncerNotificationTest(BouncerTestCase):
server=server)) server=server))
self.assertEqual(len(tokens), 2) self.assertEqual(len(tokens), 2)
# Now we succesfully remove them: # Now we successfully remove them:
do_regenerate_api_key(user, user) do_regenerate_api_key(user, user)
tokens = list(RemotePushDeviceToken.objects.filter(user_id=user.id, tokens = list(RemotePushDeviceToken.objects.filter(user_id=user.id,
server=server)) server=server))

View File

@ -309,7 +309,7 @@ class TestArchiveMessagesGeneral(ArchiveMessagesTestingBase):
expired_usermsg_ids = self._get_usermessage_ids(expired_msg_ids) expired_usermsg_ids = self._get_usermessage_ids(expired_msg_ids)
archive_messages(chunk_size=2) # Specify low chunk_size to test batching. archive_messages(chunk_size=2) # Specify low chunk_size to test batching.
# Make sure we archived what neeeded: # Make sure we archived what needed:
self._verify_archive_data(expired_msg_ids, expired_usermsg_ids) self._verify_archive_data(expired_msg_ids, expired_usermsg_ids)
restore_all_data_from_archive() restore_all_data_from_archive()

View File

@ -3186,7 +3186,7 @@ class UserSignUpTest(InviteUserBase):
self.assertEqual(result.url, "/accounts/login/?email=newuser%40zulip.com") self.assertEqual(result.url, "/accounts/login/?email=newuser%40zulip.com")
self.assertFalse(UserProfile.objects.filter(delivery_email=email).exists()) self.assertFalse(UserProfile.objects.filter(delivery_email=email).exists())
# If the email is outside of LDAP_APPEND_DOMAIN, we succesfully create a non-ldap account, # If the email is outside of LDAP_APPEND_DOMAIN, we successfully create a non-ldap account,
# with the password managed in the zulip database. # with the password managed in the zulip database.
with self.settings( with self.settings(
POPULATE_PROFILE_VIA_LDAP=True, POPULATE_PROFILE_VIA_LDAP=True,

View File

@ -438,7 +438,7 @@ class SlackImporter(ZulipTestCase):
self.assertDictEqual(test_dm_members, dm_members) self.assertDictEqual(test_dm_members, dm_members)
# We can't do an assertDictEqual since during the construction of Personal # We can't do an assertDictEqual since during the construction of Personal
# recipients, slack_user_id_to_zulip_user_id are iterated in diffrent order in Python 3.5 and 3.6. # recipients, slack_user_id_to_zulip_user_id are iterated in different order in Python 3.5 and 3.6.
self.assertEqual(set(slack_recipient_name_to_zulip_recipient_id.keys()), slack_recipient_names) self.assertEqual(set(slack_recipient_name_to_zulip_recipient_id.keys()), slack_recipient_names)
self.assertEqual(set(slack_recipient_name_to_zulip_recipient_id.values()), set(i for i in range(11))) self.assertEqual(set(slack_recipient_name_to_zulip_recipient_id.values()), set(i for i in range(11)))

View File

@ -87,7 +87,7 @@ class FileUploadTest(UploadSerializeMixin, ZulipTestCase):
data = b"".join(response.streaming_content) data = b"".join(response.streaming_content)
self.assertEqual(b"zulip!", data) self.assertEqual(b"zulip!", data)
# Files uploaded through the API should be accesible via the web client # Files uploaded through the API should be accessible via the web client
self.login('hamlet') self.login('hamlet')
self.assert_url_serves_contents_of_file(uri, b"zulip!") self.assert_url_serves_contents_of_file(uri, b"zulip!")
@ -529,7 +529,7 @@ class FileUploadTest(UploadSerializeMixin, ZulipTestCase):
self.logout() self.logout()
self.assertEqual(len(queries), 5) self.assertEqual(len(queries), 5)
# Subscribed user who recieved the message should be able to view file # Subscribed user who received the message should be able to view file
self.login_user(cordelia) self.login_user(cordelia)
with queries_captured() as queries: with queries_captured() as queries:
response = self.client_get(uri) response = self.client_get(uri)

View File

@ -500,7 +500,7 @@ class PermissionTest(ZulipTestCase):
{'profile_data': ujson.dumps(new_profile_data)}) {'profile_data': ujson.dumps(new_profile_data)})
self.assert_json_error(result, error_msg) self.assert_json_error(result, error_msg)
# non-existant field and no data # non-existent field and no data
invalid_profile_data = [{ invalid_profile_data = [{
'id': 9001, 'id': 9001,
'value': '' 'value': ''
@ -509,7 +509,7 @@ class PermissionTest(ZulipTestCase):
{'profile_data': ujson.dumps(invalid_profile_data)}) {'profile_data': ujson.dumps(invalid_profile_data)})
self.assert_json_error(result, 'Field id 9001 not found.') self.assert_json_error(result, 'Field id 9001 not found.')
# non-existant field and data # non-existent field and data
invalid_profile_data = [{ invalid_profile_data = [{
'id': 9001, 'id': 9001,
'value': 'some data' 'value': 'some data'
@ -776,7 +776,7 @@ class UserProfileTest(ZulipTestCase):
self.assertEqual(check_valid_user_ids(realm.id, [bot.id]), self.assertEqual(check_valid_user_ids(realm.id, [bot.id]),
"User with ID %d is a bot" % (bot.id,)) "User with ID %d is a bot" % (bot.id,))
# Succesfully get non-bot, active user belong to your realm # Successfully get non-bot, active user belong to your realm
self.assertEqual(check_valid_user_ids(realm.id, [othello.id]), None) self.assertEqual(check_valid_user_ids(realm.id, [othello.id]), None)
def test_cache_invalidation(self) -> None: def test_cache_invalidation(self) -> None:

View File

@ -50,7 +50,7 @@ def add_reaction(request: HttpRequest, user_profile: UserProfile, message_id: in
# If another user has already reacted to this message with # If another user has already reacted to this message with
# same emoji code, we treat the new reaction as a vote for the # same emoji code, we treat the new reaction as a vote for the
# existing reaction. So the emoji name used by that earlier # existing reaction. So the emoji name used by that earlier
# reaction takes precendence over whatever was passed in this # reaction takes precedence over whatever was passed in this
# request. This is necessary to avoid a message having 2 # request. This is necessary to avoid a message having 2
# "different" emoji reactions with the same emoji code (and # "different" emoji reactions with the same emoji code (and
# thus same image) on the same message, which looks ugly. # thus same image) on the same message, which looks ugly.

View File

@ -379,7 +379,7 @@ def get_bots_backend(request: HttpRequest, user_profile: UserProfile) -> HttpRes
default_events_register_stream = get_stream_name(bot_profile.default_events_register_stream) default_events_register_stream = get_stream_name(bot_profile.default_events_register_stream)
# Bots are supposed to have only one API key, at least for now. # Bots are supposed to have only one API key, at least for now.
# Therefore we can safely asume that one and only valid API key will be # Therefore we can safely assume that one and only valid API key will be
# the first one. # the first one.
api_key = get_api_key(bot_profile) api_key = get_api_key(bot_profile)

View File

@ -1,6 +1,6 @@
Front lets you manage all of your communication channels in one place, Front lets you manage all of your communication channels in one place,
and helps your team collaborate around every message. Follow these steps and helps your team collaborate around every message. Follow these steps
to recieve Front notifications without leaving Zulip! to receive Front notifications without leaving Zulip!
1. {!create-stream.md!} 1. {!create-stream.md!}

View File

@ -43,7 +43,7 @@
"created_at": "2015-05-05T23:40:28Z", "created_at": "2015-05-05T23:40:28Z",
"updated_at": "2015-05-05T23:40:28Z", "updated_at": "2015-05-05T23:40:28Z",
"closed_at": null, "closed_at": null,
"body": "It looks like you accidently spelled 'commit' with two 't's." "body": "It looks like you accidentally spelled 'commit' with two 't's."
}, },
"comment": { "comment": {
"url": "https://api.github.com/repos/baxterthehacker/public-repo/issues/comments/99262140", "url": "https://api.github.com/repos/baxterthehacker/public-repo/issues/comments/99262140",

View File

@ -43,7 +43,7 @@
"created_at": "2015-05-05T23:40:28Z", "created_at": "2015-05-05T23:40:28Z",
"updated_at": "2015-05-05T23:40:28Z", "updated_at": "2015-05-05T23:40:28Z",
"closed_at": null, "closed_at": null,
"body": "It looks like you accidently spelled 'commit' with two 't's." "body": "It looks like you accidentally spelled 'commit' with two 't's."
}, },
"repository": { "repository": {
"id": 35129377, "id": 35129377,

View File

@ -131,13 +131,13 @@ class GithubWebhookTest(WebhookTestCase):
self.send_and_test_stream_message('issue_comment', expected_topic, expected_message) self.send_and_test_stream_message('issue_comment', expected_topic, expected_message)
def test_issue_msg(self) -> None: def test_issue_msg(self) -> None:
expected_message = u"baxterthehacker opened [Issue #2](https://github.com/baxterthehacker/public-repo/issues/2):\n\n~~~ quote\nIt looks like you accidently spelled 'commit' with two 't's.\n~~~" expected_message = u"baxterthehacker opened [Issue #2](https://github.com/baxterthehacker/public-repo/issues/2):\n\n~~~ quote\nIt looks like you accidentally spelled 'commit' with two 't's.\n~~~"
self.send_and_test_stream_message('issues', self.EXPECTED_TOPIC_ISSUE_EVENTS, expected_message) self.send_and_test_stream_message('issues', self.EXPECTED_TOPIC_ISSUE_EVENTS, expected_message)
def test_issue_msg_with_custom_topic_in_url(self) -> None: def test_issue_msg_with_custom_topic_in_url(self) -> None:
self.url = self.build_webhook_url(topic='notifications') self.url = self.build_webhook_url(topic='notifications')
expected_topic = u"notifications" expected_topic = u"notifications"
expected_message = u"baxterthehacker opened [Issue #2 Spelling error in the README file](https://github.com/baxterthehacker/public-repo/issues/2):\n\n~~~ quote\nIt looks like you accidently spelled 'commit' with two 't's.\n~~~" expected_message = u"baxterthehacker opened [Issue #2 Spelling error in the README file](https://github.com/baxterthehacker/public-repo/issues/2):\n\n~~~ quote\nIt looks like you accidentally spelled 'commit' with two 't's.\n~~~"
self.send_and_test_stream_message('issues', expected_topic, expected_message) self.send_and_test_stream_message('issues', expected_topic, expected_message)
def test_membership_msg(self) -> None: def test_membership_msg(self) -> None:
@ -210,7 +210,7 @@ class GithubWebhookTest(WebhookTestCase):
self.send_and_test_stream_message('release', self.EXPECTED_TOPIC_REPO_EVENTS, expected_message) self.send_and_test_stream_message('release', self.EXPECTED_TOPIC_REPO_EVENTS, expected_message)
def test_page_build_msg(self) -> None: def test_page_build_msg(self) -> None:
expected_message = u"Github Pages build, trigerred by baxterthehacker, has finished building." expected_message = u"Github Pages build, triggered by baxterthehacker, has finished building."
self.send_and_test_stream_message('page_build', self.EXPECTED_TOPIC_REPO_EVENTS, expected_message) self.send_and_test_stream_message('page_build', self.EXPECTED_TOPIC_REPO_EVENTS, expected_message)
def test_status_msg(self) -> None: def test_status_msg(self) -> None:

View File

@ -257,7 +257,7 @@ def get_page_build_body(payload: Dict[str, Any]) -> str:
CONTENT_MESSAGE_TEMPLATE.format(message=build['error']['message']) CONTENT_MESSAGE_TEMPLATE.format(message=build['error']['message'])
) )
return u"Github Pages build, trigerred by {}, {}.".format( return u"Github Pages build, triggered by {}, {}.".format(
payload['build']['pusher']['login'], payload['build']['pusher']['login'],
action action
) )

View File

@ -73,7 +73,7 @@ def replied_body(payload: Dict[str, Any], actor: str, action: str) -> str:
return body return body
def get_event_handler(event: str) -> Callable[..., str]: def get_event_handler(event: str) -> Callable[..., str]:
# The main reason for this function existance is because of mypy # The main reason for this function existence is because of mypy
handler = EVENTS_FUNCTION_MAPPER.get(event) # type: Any handler = EVENTS_FUNCTION_MAPPER.get(event) # type: Any
if handler is None: if handler is None:
raise UnexpectedWebhookEventType("Groove", event) raise UnexpectedWebhookEventType("Groove", event)

View File

@ -4,7 +4,7 @@
"name": "ToHighTemeprature", "name": "ToHighTemeprature",
"runbook_url": "http://www.use.water.pl", "runbook_url": "http://www.use.water.pl",
"version": 2, "version": 2,
"description": "Measurment of temperature in your computer" "description": "Measurement of temperature in your computer"
}, },
"account": "lizonr@gmail.com", "account": "lizonr@gmail.com",
"trigger_time": 1460407214, "trigger_time": 1460407214,

Some files were not shown because too many files have changed in this diff Show More