py3: Switch almost all shebang lines to use `python3`.
This causes `upgrade-zulip-from-git`, as well as a no-option run of
`tools/build-release-tarball`, to produce a Zulip install running
Python 3, rather than Python 2. In particular this means that the
virtualenv we create, in which all application code runs, is Python 3.
One shebang line, on `zulip-ec2-configure-interfaces`, explicitly
keeps Python 2, and at least one external ops script, `wal-e`, also
still runs on Python 2. See discussion on the respective previous
commits that made those explicit. There may also be some other
third-party scripts we use, outside of this source tree and running
outside our virtualenv, that still run on Python 2.
2017-08-02 23:15:16 +02:00
|
|
|
#!/usr/bin/env python3
|
2016-03-10 17:15:34 +01:00
|
|
|
from __future__ import print_function
|
2013-05-21 23:09:08 +02:00
|
|
|
import optparse
|
2017-01-13 01:09:52 +01:00
|
|
|
import subprocess
|
2012-11-09 21:03:57 +01:00
|
|
|
import sys
|
|
|
|
import os
|
2013-03-05 22:50:06 +01:00
|
|
|
import glob
|
2012-11-09 21:03:57 +01:00
|
|
|
|
2013-05-21 23:09:08 +02:00
|
|
|
#
|
|
|
|
# In order to use remote casperjs debugging, pass the --remote-debug flag
|
|
|
|
# This will start a remote debugging session listening on port 7777
|
|
|
|
#
|
2013-07-24 23:29:47 +02:00
|
|
|
# See https://wiki.zulip.net/wiki/Testing_the_app for more information
|
2013-05-21 23:09:08 +02:00
|
|
|
# on how to use remote debugging
|
|
|
|
#
|
|
|
|
|
2016-12-15 07:02:42 +01:00
|
|
|
os.environ["CASPER_TESTS"] = "1"
|
Upgrade caspersjs to version 1.1.3. (w/acrefoot)
(Most of this work was done by acrefoot in an earlier branch.
I took over the branch to fix casper tests that were broken during
the upgrade (which were fixed in a different commit). I also
made most of the changes to run-casper.)
This also upgrades phantomjs to 2.1.7.
The huge structural change here is that we no longer vendor casperjs
or download phantomjs with our own script. Instead, we just use
casperjs and phantomjs from npm, via package.json.
Another thing that we do now is run casperjs tests individually, so
that we don't get strange test flakes from test interactions. (Tests
can still influence each other in terms of changing data, since we
don't yet have code to clear the test database in between tests.)
A lot of this diff is just removing files and obsolete configurations.
The main new piece is in package.json, which causes npm to install the
new version.
Also, run-casper now runs files individually, as mentioned above.
We had vendored casperjs in the past. I didn't bring over any of our
changes. Some of the changes were performance-related (primarily
5fd58cf24927359dce26588d59690c40c6ce6d4c), so the upgraded version may
be slower in some instances. (I didn't do much measurement of that,
since most of our slowness when running tests is about the setup
environment, not casper itself.) Any bug fixes that we may have
implemented in the past were either magically fixed by changes to
casper itself or by improvements we have made in the tests themselves
over the years.
Tim tested the Casper suite on his machine and running the full Casper
test suite is faster than it was before this change (1m30 vs. 1m50),
so we're at least not regressing overall performance.
2016-10-07 18:20:59 +02:00
|
|
|
os.environ["PHANTOMJS_EXECUTABLE"] = os.path.join(os.path.dirname(__file__), "../node_modules/.bin/phantomjs")
|
2017-03-24 12:48:56 +01:00
|
|
|
os.environ["http_proxy"] = ""
|
|
|
|
os.environ["https_proxy"] = ""
|
2013-01-10 19:05:53 +01:00
|
|
|
|
2016-06-27 23:15:20 +02:00
|
|
|
usage = """%prog [options]
|
|
|
|
test-js-with-casper # Run all test files
|
|
|
|
test-js-with-casper 09-navigation.js # Run a single test file
|
2016-11-02 17:40:06 +01:00
|
|
|
test-js-with-casper 09 # Run a single test file 09-navigation.js
|
|
|
|
test-js-with-casper 01-login.js 03-narrow.js # Run a few test files
|
|
|
|
test-js-with-casper 01 03 # Run a few test files, 01-login.js and 03-narrow.js here"""
|
2016-06-27 23:15:20 +02:00
|
|
|
parser = optparse.OptionParser(usage)
|
2013-05-21 23:09:08 +02:00
|
|
|
|
2017-01-13 02:02:49 +01:00
|
|
|
parser.add_option('--skip-flaky-tests', dest='skip_flaky',
|
|
|
|
action="store_true",
|
|
|
|
default=False, help='Skip flaky tests')
|
2016-10-15 17:39:27 +02:00
|
|
|
parser.add_option('--force', dest='force',
|
|
|
|
action="store_true",
|
|
|
|
default=False, help='Run tests despite possible problems.')
|
2013-05-21 23:09:08 +02:00
|
|
|
parser.add_option('--remote-debug',
|
|
|
|
help='Whether or not to enable remote debugging on port 7777',
|
|
|
|
action="store_true",
|
|
|
|
default=False)
|
|
|
|
(options, args) = parser.parse_args()
|
|
|
|
|
2016-09-13 22:40:13 +02:00
|
|
|
TOOLS_DIR = os.path.dirname(os.path.abspath(__file__))
|
|
|
|
sys.path.insert(0, os.path.dirname(TOOLS_DIR))
|
2016-10-15 17:23:28 +02:00
|
|
|
|
2017-02-05 21:24:28 +01:00
|
|
|
# check for the venv
|
|
|
|
from tools.lib import sanity_check
|
|
|
|
sanity_check.check_venv(__file__)
|
|
|
|
|
2017-01-13 01:09:52 +01:00
|
|
|
from tools.lib.test_script import get_provisioning_status
|
|
|
|
from tools.lib.test_server import test_server_running
|
2016-10-15 17:23:28 +02:00
|
|
|
|
2017-02-05 21:24:28 +01:00
|
|
|
from typing import Iterable
|
|
|
|
|
2016-10-15 17:39:27 +02:00
|
|
|
if not options.force:
|
|
|
|
ok, msg = get_provisioning_status()
|
|
|
|
if not ok:
|
|
|
|
print(msg)
|
|
|
|
print('If you really know what you are doing, use --force to run anyway.')
|
|
|
|
sys.exit(1)
|
2016-10-15 17:23:28 +02:00
|
|
|
|
2015-10-13 23:34:50 +02:00
|
|
|
os.chdir(os.path.join(os.path.dirname(os.path.realpath(__file__)), '..'))
|
2012-11-09 21:03:57 +01:00
|
|
|
|
2016-05-26 12:36:22 +02:00
|
|
|
subprocess.check_call('tools/setup/generate-test-credentials')
|
2015-08-20 02:57:30 +02:00
|
|
|
|
2016-07-24 07:02:22 +02:00
|
|
|
subprocess.check_call(['mkdir', '-p', 'var/casper'])
|
|
|
|
|
2016-09-27 10:12:40 +02:00
|
|
|
subprocess.check_call(['rm', '-f'] + glob.glob('var/casper/casper-failure*.png'))
|
|
|
|
|
2016-11-02 01:01:21 +01:00
|
|
|
LOG_FILE = 'var/casper/server.log'
|
2012-11-09 21:03:57 +01:00
|
|
|
|
2017-08-25 05:41:04 +02:00
|
|
|
def run_tests(files, external_host):
|
2017-01-13 01:09:52 +01:00
|
|
|
# type: (bool, Iterable[str], str) -> None
|
Upgrade caspersjs to version 1.1.3. (w/acrefoot)
(Most of this work was done by acrefoot in an earlier branch.
I took over the branch to fix casper tests that were broken during
the upgrade (which were fixed in a different commit). I also
made most of the changes to run-casper.)
This also upgrades phantomjs to 2.1.7.
The huge structural change here is that we no longer vendor casperjs
or download phantomjs with our own script. Instead, we just use
casperjs and phantomjs from npm, via package.json.
Another thing that we do now is run casperjs tests individually, so
that we don't get strange test flakes from test interactions. (Tests
can still influence each other in terms of changing data, since we
don't yet have code to clear the test database in between tests.)
A lot of this diff is just removing files and obsolete configurations.
The main new piece is in package.json, which causes npm to install the
new version.
Also, run-casper now runs files individually, as mentioned above.
We had vendored casperjs in the past. I didn't bring over any of our
changes. Some of the changes were performance-related (primarily
5fd58cf24927359dce26588d59690c40c6ce6d4c), so the upgraded version may
be slower in some instances. (I didn't do much measurement of that,
since most of our slowness when running tests is about the setup
environment, not casper itself.) Any bug fixes that we may have
implemented in the past were either magically fixed by changes to
casper itself or by improvements we have made in the tests themselves
over the years.
Tim tested the Casper suite on his machine and running the full Casper
test suite is faster than it was before this change (1m30 vs. 1m50),
so we're at least not regressing overall performance.
2016-10-07 18:20:59 +02:00
|
|
|
test_dir = os.path.join(os.path.dirname(__file__), '../frontend_tests/casper_tests')
|
2016-07-21 05:20:34 +02:00
|
|
|
test_files = []
|
2016-07-21 05:23:07 +02:00
|
|
|
for file in files:
|
2016-11-02 17:40:06 +01:00
|
|
|
for file_name in os.listdir(test_dir):
|
|
|
|
if file_name.startswith(file):
|
|
|
|
file = file_name
|
|
|
|
break
|
2016-07-21 05:20:34 +02:00
|
|
|
if not os.path.exists(file):
|
Upgrade caspersjs to version 1.1.3. (w/acrefoot)
(Most of this work was done by acrefoot in an earlier branch.
I took over the branch to fix casper tests that were broken during
the upgrade (which were fixed in a different commit). I also
made most of the changes to run-casper.)
This also upgrades phantomjs to 2.1.7.
The huge structural change here is that we no longer vendor casperjs
or download phantomjs with our own script. Instead, we just use
casperjs and phantomjs from npm, via package.json.
Another thing that we do now is run casperjs tests individually, so
that we don't get strange test flakes from test interactions. (Tests
can still influence each other in terms of changing data, since we
don't yet have code to clear the test database in between tests.)
A lot of this diff is just removing files and obsolete configurations.
The main new piece is in package.json, which causes npm to install the
new version.
Also, run-casper now runs files individually, as mentioned above.
We had vendored casperjs in the past. I didn't bring over any of our
changes. Some of the changes were performance-related (primarily
5fd58cf24927359dce26588d59690c40c6ce6d4c), so the upgraded version may
be slower in some instances. (I didn't do much measurement of that,
since most of our slowness when running tests is about the setup
environment, not casper itself.) Any bug fixes that we may have
implemented in the past were either magically fixed by changes to
casper itself or by improvements we have made in the tests themselves
over the years.
Tim tested the Casper suite on his machine and running the full Casper
test suite is faster than it was before this change (1m30 vs. 1m50),
so we're at least not regressing overall performance.
2016-10-07 18:20:59 +02:00
|
|
|
file = os.path.join(test_dir, file)
|
2016-07-21 05:20:34 +02:00
|
|
|
test_files.append(os.path.abspath(file))
|
2016-07-21 05:17:40 +02:00
|
|
|
|
Upgrade caspersjs to version 1.1.3. (w/acrefoot)
(Most of this work was done by acrefoot in an earlier branch.
I took over the branch to fix casper tests that were broken during
the upgrade (which were fixed in a different commit). I also
made most of the changes to run-casper.)
This also upgrades phantomjs to 2.1.7.
The huge structural change here is that we no longer vendor casperjs
or download phantomjs with our own script. Instead, we just use
casperjs and phantomjs from npm, via package.json.
Another thing that we do now is run casperjs tests individually, so
that we don't get strange test flakes from test interactions. (Tests
can still influence each other in terms of changing data, since we
don't yet have code to clear the test database in between tests.)
A lot of this diff is just removing files and obsolete configurations.
The main new piece is in package.json, which causes npm to install the
new version.
Also, run-casper now runs files individually, as mentioned above.
We had vendored casperjs in the past. I didn't bring over any of our
changes. Some of the changes were performance-related (primarily
5fd58cf24927359dce26588d59690c40c6ce6d4c), so the upgraded version may
be slower in some instances. (I didn't do much measurement of that,
since most of our slowness when running tests is about the setup
environment, not casper itself.) Any bug fixes that we may have
implemented in the past were either magically fixed by changes to
casper itself or by improvements we have made in the tests themselves
over the years.
Tim tested the Casper suite on his machine and running the full Casper
test suite is faster than it was before this change (1m30 vs. 1m50),
so we're at least not regressing overall performance.
2016-10-07 18:20:59 +02:00
|
|
|
if not test_files:
|
|
|
|
test_files = sorted(glob.glob(os.path.join(test_dir, '*.js')))
|
|
|
|
|
2017-01-13 02:02:49 +01:00
|
|
|
# 10-admin.js is too flaky!
|
|
|
|
if options.skip_flaky:
|
|
|
|
test_files = [fn for fn in test_files if '10-admin' not in fn]
|
|
|
|
|
2016-07-21 05:26:45 +02:00
|
|
|
remote_debug = ""
|
|
|
|
if options.remote_debug:
|
|
|
|
remote_debug = "--remote-debugger-port=7777 --remote-debugger-autorun=yes"
|
|
|
|
|
2017-01-25 18:31:51 +01:00
|
|
|
with test_server_running(options.force, external_host, log_file=LOG_FILE, dots=True):
|
2017-01-13 01:09:52 +01:00
|
|
|
ret = 1
|
Upgrade caspersjs to version 1.1.3. (w/acrefoot)
(Most of this work was done by acrefoot in an earlier branch.
I took over the branch to fix casper tests that were broken during
the upgrade (which were fixed in a different commit). I also
made most of the changes to run-casper.)
This also upgrades phantomjs to 2.1.7.
The huge structural change here is that we no longer vendor casperjs
or download phantomjs with our own script. Instead, we just use
casperjs and phantomjs from npm, via package.json.
Another thing that we do now is run casperjs tests individually, so
that we don't get strange test flakes from test interactions. (Tests
can still influence each other in terms of changing data, since we
don't yet have code to clear the test database in between tests.)
A lot of this diff is just removing files and obsolete configurations.
The main new piece is in package.json, which causes npm to install the
new version.
Also, run-casper now runs files individually, as mentioned above.
We had vendored casperjs in the past. I didn't bring over any of our
changes. Some of the changes were performance-related (primarily
5fd58cf24927359dce26588d59690c40c6ce6d4c), so the upgraded version may
be slower in some instances. (I didn't do much measurement of that,
since most of our slowness when running tests is about the setup
environment, not casper itself.) Any bug fixes that we may have
implemented in the past were either magically fixed by changes to
casper itself or by improvements we have made in the tests themselves
over the years.
Tim tested the Casper suite on his machine and running the full Casper
test suite is faster than it was before this change (1m30 vs. 1m50),
so we're at least not regressing overall performance.
2016-10-07 18:20:59 +02:00
|
|
|
for test_file in test_files:
|
2017-08-25 05:41:04 +02:00
|
|
|
cmd = "node_modules/.bin/casperjs %s test %s" % (remote_debug, test_file)
|
Upgrade caspersjs to version 1.1.3. (w/acrefoot)
(Most of this work was done by acrefoot in an earlier branch.
I took over the branch to fix casper tests that were broken during
the upgrade (which were fixed in a different commit). I also
made most of the changes to run-casper.)
This also upgrades phantomjs to 2.1.7.
The huge structural change here is that we no longer vendor casperjs
or download phantomjs with our own script. Instead, we just use
casperjs and phantomjs from npm, via package.json.
Another thing that we do now is run casperjs tests individually, so
that we don't get strange test flakes from test interactions. (Tests
can still influence each other in terms of changing data, since we
don't yet have code to clear the test database in between tests.)
A lot of this diff is just removing files and obsolete configurations.
The main new piece is in package.json, which causes npm to install the
new version.
Also, run-casper now runs files individually, as mentioned above.
We had vendored casperjs in the past. I didn't bring over any of our
changes. Some of the changes were performance-related (primarily
5fd58cf24927359dce26588d59690c40c6ce6d4c), so the upgraded version may
be slower in some instances. (I didn't do much measurement of that,
since most of our slowness when running tests is about the setup
environment, not casper itself.) Any bug fixes that we may have
implemented in the past were either magically fixed by changes to
casper itself or by improvements we have made in the tests themselves
over the years.
Tim tested the Casper suite on his machine and running the full Casper
test suite is faster than it was before this change (1m30 vs. 1m50),
so we're at least not regressing overall performance.
2016-10-07 18:20:59 +02:00
|
|
|
print("\n\nRunning %s" % (cmd,))
|
|
|
|
ret = subprocess.call(cmd, shell=True)
|
|
|
|
if ret != 0:
|
|
|
|
break
|
2016-07-21 05:20:34 +02:00
|
|
|
if ret != 0:
|
|
|
|
print("""
|
2015-08-21 22:48:57 +02:00
|
|
|
Oops, the frontend tests failed. Tips for debugging:
|
2016-11-02 01:01:21 +01:00
|
|
|
* Check the frontend test server logs at %s
|
2016-09-27 10:12:40 +02:00
|
|
|
* Check the screenshots of failed tests at var/casper/casper-failure*.png
|
2016-12-03 12:25:35 +01:00
|
|
|
* Try remote debugging the test web browser as described in docs/testing-with-casper.md
|
2016-11-02 01:01:21 +01:00
|
|
|
""" % (LOG_FILE,), file=sys.stderr)
|
2015-08-21 22:48:57 +02:00
|
|
|
|
2016-07-21 05:20:34 +02:00
|
|
|
sys.exit(ret)
|
|
|
|
|
2017-01-13 01:09:52 +01:00
|
|
|
external_host = "zulipdev.com:9981"
|
2017-08-25 05:41:04 +02:00
|
|
|
run_tests(args, external_host)
|
2016-07-21 05:20:34 +02:00
|
|
|
sys.exit(0)
|