mirror of https://github.com/zulip/zulip.git
475 lines
12 KiB
JavaScript
475 lines
12 KiB
JavaScript
"use strict";
|
|
|
|
const {strict: assert} = require("assert");
|
|
|
|
const _ = require("lodash");
|
|
|
|
const {mock_module, set_global, zrequire} = require("../zjsunit/namespace");
|
|
const {run_test} = require("../zjsunit/test");
|
|
const $ = require("../zjsunit/zjquery");
|
|
|
|
set_global("document", "document-stub");
|
|
|
|
const noop = () => {};
|
|
|
|
function MessageListView() {
|
|
return {};
|
|
}
|
|
mock_module("message_list_view", {
|
|
MessageListView,
|
|
});
|
|
|
|
mock_module("recent_topics", {
|
|
process_messages: noop,
|
|
});
|
|
// Still required for page_params.initial_pointer
|
|
set_global("page_params", {});
|
|
mock_module("ui_report", {
|
|
hide_error: noop,
|
|
});
|
|
|
|
const channel = mock_module("channel");
|
|
const message_store = mock_module("message_store");
|
|
const message_util = mock_module("message_util");
|
|
const pm_list = mock_module("pm_list");
|
|
const server_events = mock_module("server_events");
|
|
const stream_list = mock_module("stream_list", {
|
|
maybe_scroll_narrow_into_view: () => {},
|
|
});
|
|
mock_module("message_scroll", {
|
|
show_loading_older: noop,
|
|
hide_loading_older: noop,
|
|
show_loading_newer: noop,
|
|
hide_loading_newer: noop,
|
|
update_top_of_narrow_notices: () => {},
|
|
});
|
|
set_global("document", "document-stub");
|
|
|
|
const message_fetch = zrequire("message_fetch");
|
|
|
|
const {Filter} = zrequire("../js/filter");
|
|
const message_list = zrequire("message_list");
|
|
const people = zrequire("people");
|
|
|
|
const alice = {
|
|
email: "alice@example.com",
|
|
user_id: 7,
|
|
full_name: "Alice",
|
|
};
|
|
people.add_active_user(alice);
|
|
|
|
server_events.home_view_loaded = noop;
|
|
|
|
function stub_message_view(list) {
|
|
list.view.append = noop;
|
|
list.view.maybe_rerender = noop;
|
|
list.view.prepend = noop;
|
|
}
|
|
|
|
function make_home_msg_list() {
|
|
const table_name = "whatever";
|
|
const filter = new Filter();
|
|
|
|
const list = new message_list.MessageList({
|
|
table_name,
|
|
filter,
|
|
});
|
|
return list;
|
|
}
|
|
|
|
function make_all_list() {
|
|
return new message_list.MessageList({});
|
|
}
|
|
|
|
function reset_lists() {
|
|
set_global("home_msg_list", make_home_msg_list());
|
|
set_global("current_msg_list", home_msg_list);
|
|
message_list.__Rewire__("all", make_all_list());
|
|
stub_message_view(home_msg_list);
|
|
stub_message_view(message_list.all);
|
|
}
|
|
|
|
function config_fake_channel(conf) {
|
|
const self = {};
|
|
let called;
|
|
let called_with_newest_flag = false;
|
|
|
|
channel.get = (opts) => {
|
|
assert.equal(opts.url, "/json/messages");
|
|
// There's a separate call with anchor="newest" that happens
|
|
// unconditionally; do basic verfication of that call.
|
|
if (opts.data.anchor === "newest") {
|
|
if (!called_with_newest_flag) {
|
|
called_with_newest_flag = true;
|
|
assert.equal(opts.data.num_after, 0);
|
|
return;
|
|
}
|
|
throw new Error("Only one 'newest' call allowed");
|
|
}
|
|
|
|
if (called && !conf.can_call_again) {
|
|
throw new Error("only use this for one call");
|
|
}
|
|
if (!conf.can_call_again) {
|
|
assert(self.success === undefined);
|
|
}
|
|
assert.deepEqual(opts.data, conf.expected_opts_data);
|
|
self.success = opts.success;
|
|
called = true;
|
|
};
|
|
|
|
return self;
|
|
}
|
|
|
|
function config_process_results(messages) {
|
|
const self = {};
|
|
|
|
const messages_processed_for_bools = [];
|
|
|
|
message_store.set_message_booleans = (message) => {
|
|
messages_processed_for_bools.push(message);
|
|
};
|
|
|
|
message_store.add_message_metadata = (message) => message;
|
|
|
|
message_util.do_unread_count_updates = (arg) => {
|
|
assert.deepEqual(arg, messages);
|
|
};
|
|
|
|
message_util.add_old_messages = (new_messages, msg_list) => {
|
|
assert.deepEqual(new_messages, messages);
|
|
msg_list.add_messages(new_messages);
|
|
};
|
|
|
|
stream_list.update_streams_sidebar = noop;
|
|
|
|
pm_list.update_private_messages = noop;
|
|
|
|
self.verify = () => {
|
|
assert.deepEqual(messages_processed_for_bools, messages);
|
|
};
|
|
|
|
return self;
|
|
}
|
|
|
|
function message_range(start, end) {
|
|
return _.range(start, end).map((idx) => ({
|
|
id: idx,
|
|
}));
|
|
}
|
|
|
|
const initialize_data = {
|
|
initial_fetch: {
|
|
req: {
|
|
anchor: "first_unread",
|
|
num_before: 200,
|
|
num_after: 200,
|
|
client_gravatar: true,
|
|
},
|
|
resp: {
|
|
messages: message_range(201, 801),
|
|
found_newest: false,
|
|
anchor: 444,
|
|
},
|
|
},
|
|
|
|
forward_fill: {
|
|
req: {
|
|
anchor: "800",
|
|
num_before: 0,
|
|
num_after: 1000,
|
|
client_gravatar: true,
|
|
},
|
|
resp: {
|
|
messages: message_range(800, 1000),
|
|
found_newest: true,
|
|
},
|
|
},
|
|
|
|
back_fill: {
|
|
req: {
|
|
anchor: "201",
|
|
num_before: 1000,
|
|
num_after: 0,
|
|
client_gravatar: true,
|
|
},
|
|
resp: {
|
|
messages: message_range(100, 200),
|
|
found_oldest: true,
|
|
},
|
|
},
|
|
};
|
|
|
|
function test_fetch_success(opts) {
|
|
const response = opts.response;
|
|
const messages = response.messages;
|
|
|
|
const process_results = config_process_results(messages);
|
|
opts.fetch.success(response);
|
|
process_results.verify();
|
|
}
|
|
|
|
function initial_fetch_step() {
|
|
const self = {};
|
|
|
|
let fetch;
|
|
const response = initialize_data.initial_fetch.resp;
|
|
|
|
self.prep = () => {
|
|
fetch = config_fake_channel({
|
|
expected_opts_data: initialize_data.initial_fetch.req,
|
|
});
|
|
|
|
message_fetch.initialize();
|
|
};
|
|
|
|
self.finish = () => {
|
|
test_fetch_success({
|
|
fetch,
|
|
response,
|
|
});
|
|
};
|
|
|
|
return self;
|
|
}
|
|
|
|
function forward_fill_step() {
|
|
const self = {};
|
|
|
|
let fetch;
|
|
|
|
self.prep = () => {
|
|
fetch = config_fake_channel({
|
|
expected_opts_data: initialize_data.forward_fill.req,
|
|
});
|
|
};
|
|
|
|
self.finish = () => {
|
|
const response = initialize_data.forward_fill.resp;
|
|
|
|
let idle_config;
|
|
$("document-stub").idle = (config) => {
|
|
idle_config = config;
|
|
};
|
|
|
|
test_fetch_success({
|
|
fetch,
|
|
response,
|
|
});
|
|
|
|
assert.equal(idle_config.idle, 10000);
|
|
|
|
return idle_config;
|
|
};
|
|
|
|
return self;
|
|
}
|
|
|
|
function test_backfill_idle(idle_config) {
|
|
const fetch = config_fake_channel({
|
|
expected_opts_data: initialize_data.back_fill.req,
|
|
});
|
|
|
|
const response = initialize_data.back_fill.resp;
|
|
|
|
idle_config.onIdle();
|
|
|
|
test_fetch_success({
|
|
fetch,
|
|
response,
|
|
});
|
|
}
|
|
|
|
run_test("initialize", () => {
|
|
reset_lists();
|
|
|
|
const step1 = initial_fetch_step();
|
|
|
|
step1.prep();
|
|
|
|
const step2 = forward_fill_step();
|
|
|
|
step2.prep();
|
|
step1.finish();
|
|
|
|
const idle_config = step2.finish();
|
|
|
|
test_backfill_idle(idle_config);
|
|
});
|
|
|
|
function simulate_narrow() {
|
|
const filter = {
|
|
predicate: () => () => false,
|
|
public_operators: () => [{operator: "pm-with", operand: alice.email}],
|
|
};
|
|
|
|
const msg_list = new message_list.MessageList({
|
|
table_name: "zfilt",
|
|
filter,
|
|
});
|
|
set_global("current_msg_list", msg_list);
|
|
|
|
return msg_list;
|
|
}
|
|
|
|
run_test("loading_newer", () => {
|
|
function test_dup_new_fetch(msg_list) {
|
|
assert.equal(msg_list.data.fetch_status.can_load_newer_messages(), false);
|
|
message_fetch.maybe_load_newer_messages({
|
|
msg_list,
|
|
});
|
|
}
|
|
|
|
function test_happy_path(opts) {
|
|
const msg_list = opts.msg_list;
|
|
const data = opts.data;
|
|
|
|
const fetch = config_fake_channel({
|
|
expected_opts_data: data.req,
|
|
can_call_again: true,
|
|
});
|
|
|
|
// The msg_list is empty and we are calling frontfill, which should
|
|
// raise fatal error.
|
|
if (opts.empty_msg_list) {
|
|
assert.throws(
|
|
() => {
|
|
message_fetch.maybe_load_newer_messages({
|
|
msg_list,
|
|
show_loading: noop,
|
|
hide_loading: noop,
|
|
});
|
|
},
|
|
{
|
|
name: "Error",
|
|
message: "There are no message available to frontfill.",
|
|
},
|
|
);
|
|
} else {
|
|
message_fetch.maybe_load_newer_messages({
|
|
msg_list,
|
|
show_loading: noop,
|
|
hide_loading: noop,
|
|
});
|
|
|
|
test_dup_new_fetch(msg_list);
|
|
|
|
test_fetch_success({
|
|
fetch,
|
|
response: data.resp,
|
|
});
|
|
}
|
|
}
|
|
|
|
(function test_narrow() {
|
|
const msg_list = simulate_narrow();
|
|
|
|
const data = {
|
|
req: {
|
|
anchor: "444",
|
|
num_before: 0,
|
|
num_after: 100,
|
|
narrow: `[{"operator":"pm-with","operand":[${alice.user_id}]}]`,
|
|
client_gravatar: true,
|
|
},
|
|
resp: {
|
|
messages: message_range(500, 600),
|
|
found_newest: false,
|
|
},
|
|
};
|
|
|
|
test_happy_path({
|
|
msg_list,
|
|
data,
|
|
empty_msg_list: true,
|
|
});
|
|
|
|
msg_list.append_to_view = () => {};
|
|
// Instead of using 444 as page_param.pointer, we
|
|
// should have a message with that id in the message_list.
|
|
msg_list.append(message_range(444, 445), false);
|
|
|
|
test_happy_path({
|
|
msg_list,
|
|
data,
|
|
empty_msg_list: false,
|
|
});
|
|
|
|
assert.equal(msg_list.data.fetch_status.can_load_newer_messages(), true);
|
|
|
|
// The server successfully responded with messages having id's from 500-599.
|
|
// We test for the case that this was the last batch of messages for the narrow
|
|
// so no more fetching should occur.
|
|
// And also while fetching for the above condition the server received a new message
|
|
// event, updating the last message's id for that narrow to 600 from 599.
|
|
data.resp.found_newest = true;
|
|
msg_list.data.fetch_status.update_expected_max_message_id([{id: 600}]);
|
|
|
|
test_happy_path({
|
|
msg_list,
|
|
data,
|
|
});
|
|
|
|
// To handle this special case we should allow another fetch to occur,
|
|
// since the last message event's data had been discarded.
|
|
// This fetch goes on until the newest message has been found.
|
|
assert.equal(msg_list.data.fetch_status.can_load_newer_messages(), false);
|
|
})();
|
|
|
|
(function test_home() {
|
|
reset_lists();
|
|
const msg_list = home_msg_list;
|
|
|
|
const data = [
|
|
{
|
|
req: {
|
|
anchor: "444",
|
|
num_before: 0,
|
|
num_after: 100,
|
|
client_gravatar: true,
|
|
},
|
|
resp: {
|
|
messages: message_range(500, 600),
|
|
found_newest: false,
|
|
},
|
|
},
|
|
{
|
|
req: {
|
|
anchor: "599",
|
|
num_before: 0,
|
|
num_after: 100,
|
|
client_gravatar: true,
|
|
},
|
|
resp: {
|
|
messages: message_range(700, 800),
|
|
found_newest: true,
|
|
},
|
|
},
|
|
];
|
|
|
|
test_happy_path({
|
|
msg_list,
|
|
data: data[0],
|
|
empty_msg_list: true,
|
|
});
|
|
|
|
message_list.all.append_to_view = () => {};
|
|
message_list.all.append(message_range(444, 445), false);
|
|
|
|
test_happy_path({
|
|
msg_list,
|
|
data: data[0],
|
|
empty_msg_list: false,
|
|
});
|
|
|
|
assert.equal(msg_list.data.fetch_status.can_load_newer_messages(), true);
|
|
|
|
test_happy_path({
|
|
msg_list,
|
|
data: data[1],
|
|
});
|
|
|
|
assert.equal(msg_list.data.fetch_status.can_load_newer_messages(), false);
|
|
})();
|
|
});
|