py3: Switch almost all shebang lines to use `python3`.
This causes `upgrade-zulip-from-git`, as well as a no-option run of
`tools/build-release-tarball`, to produce a Zulip install running
Python 3, rather than Python 2. In particular this means that the
virtualenv we create, in which all application code runs, is Python 3.
One shebang line, on `zulip-ec2-configure-interfaces`, explicitly
keeps Python 2, and at least one external ops script, `wal-e`, also
still runs on Python 2. See discussion on the respective previous
commits that made those explicit. There may also be some other
third-party scripts we use, outside of this source tree and running
outside our virtualenv, that still run on Python 2.
2017-08-02 23:15:16 +02:00
|
|
|
#!/usr/bin/env python3
|
2017-01-06 18:56:36 +01:00
|
|
|
"""
|
|
|
|
Fetch contributors data from Github using their API, convert it to structured
|
2017-10-31 20:08:32 +01:00
|
|
|
JSON data for the /team page contributors section.
|
2017-01-06 18:56:36 +01:00
|
|
|
"""
|
|
|
|
|
2017-02-05 21:24:28 +01:00
|
|
|
# check for the venv
|
|
|
|
from lib import sanity_check
|
|
|
|
sanity_check.check_venv(__file__)
|
|
|
|
|
2018-03-03 13:21:55 +01:00
|
|
|
from typing import Any, Dict, List, Optional, Union, Text, cast
|
2017-11-16 14:05:26 +01:00
|
|
|
from mypy_extensions import TypedDict
|
2017-01-06 18:56:36 +01:00
|
|
|
|
|
|
|
import os
|
|
|
|
import sys
|
|
|
|
import argparse
|
2017-11-16 14:05:26 +01:00
|
|
|
from time import sleep
|
2017-01-06 18:56:36 +01:00
|
|
|
from datetime import date
|
|
|
|
import subprocess
|
|
|
|
|
|
|
|
import requests
|
2017-10-12 07:54:25 +02:00
|
|
|
import json
|
2017-01-06 18:56:36 +01:00
|
|
|
|
|
|
|
sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..'))
|
|
|
|
os.environ['DJANGO_SETTINGS_MODULE'] = 'zproject.settings'
|
|
|
|
from django.conf import settings
|
|
|
|
from zerver.lib.utils import split_by
|
|
|
|
|
|
|
|
FIXTURE_FILE = os.path.join(os.path.dirname(__file__), '../zerver/fixtures/authors.json')
|
2018-03-03 13:21:55 +01:00
|
|
|
duplicate_commits_file = os.path.join(os.path.dirname(__file__), '../zerver/fixtures/duplicate_commits.json')
|
2017-01-06 18:56:36 +01:00
|
|
|
|
|
|
|
parser = argparse.ArgumentParser()
|
|
|
|
parser.add_argument('--max-retries', type=int, default=3,
|
|
|
|
help='Number of times to retry fetching data from Github')
|
|
|
|
# In Travis CI and development environment, we use test fixture to avoid
|
|
|
|
# fetching from Github constantly.
|
|
|
|
parser.add_argument('--use-fixture', action='store_true', default=False,
|
|
|
|
help='Use fixture data instead of fetching from Github')
|
2017-01-24 07:19:25 +01:00
|
|
|
parser.add_argument('--not-required', action='store_true', default=False,
|
|
|
|
help='Consider failures to reach GitHub nonfatal')
|
2017-01-06 18:56:36 +01:00
|
|
|
args = parser.parse_args()
|
|
|
|
|
2017-11-20 21:49:03 +01:00
|
|
|
|
2017-11-16 14:05:26 +01:00
|
|
|
ContributorsJSON = TypedDict('ContributorsJSON', {
|
|
|
|
'date': str,
|
2017-11-20 21:49:03 +01:00
|
|
|
'contrib': List[Dict[str, Union[str, int]]],
|
2017-11-16 14:05:26 +01:00
|
|
|
})
|
|
|
|
|
2017-01-06 18:56:36 +01:00
|
|
|
|
2017-11-16 14:05:26 +01:00
|
|
|
def fetch_contributors(repo_link: str) -> Optional[List[Dict[str, Dict[str, Any]]]]:
|
|
|
|
r = requests.get(repo_link) # type: requests.Response
|
|
|
|
return r.json() if r.status_code == 200 else None
|
2017-01-06 18:56:36 +01:00
|
|
|
|
2017-11-16 14:05:26 +01:00
|
|
|
def write_to_disk(json_data: ContributorsJSON, out_file: str) -> None:
|
2017-01-06 18:56:36 +01:00
|
|
|
with open(out_file, 'w') as f:
|
|
|
|
try:
|
|
|
|
f.write("{}\n".format(json.dumps(json_data)))
|
|
|
|
except IOError as e:
|
|
|
|
print(e)
|
|
|
|
sys.exit(1)
|
|
|
|
|
2017-11-16 14:05:26 +01:00
|
|
|
|
|
|
|
def run_production() -> None:
|
2017-01-06 18:56:36 +01:00
|
|
|
"""
|
2017-11-20 21:49:03 +01:00
|
|
|
Get contributors data from Github and insert them into a temporary
|
|
|
|
dictionary. Retry fetching each repository if responded with non HTTP 200
|
|
|
|
status.
|
2017-01-06 18:56:36 +01:00
|
|
|
"""
|
2017-11-16 14:05:26 +01:00
|
|
|
repositories = {
|
|
|
|
'server': 'https://api.github.com/repos/zulip/zulip/stats/contributors',
|
|
|
|
'desktop': 'https://api.github.com/repos/zulip/zulip-electron/stats/contributors',
|
|
|
|
'mobile': 'https://api.github.com/repos/zulip/zulip-mobile/stats/contributors',
|
|
|
|
'python-zulip-api': 'https://api.github.com/repos/zulip/python-zulip-api/stats/contributors',
|
2018-01-15 15:54:22 +01:00
|
|
|
'zulip-js': 'https://api.github.com/repos/zulip/zulip-js/stats/contributors',
|
2017-11-16 14:05:26 +01:00
|
|
|
'zulipbot': 'https://api.github.com/repos/zulip/zulipbot/stats/contributors',
|
|
|
|
}
|
|
|
|
|
2017-11-20 21:49:03 +01:00
|
|
|
data = dict(date=str(date.today()), contrib=[]) # type: ContributorsJSON
|
|
|
|
contribs_list = {} # type: Dict[str, Dict[str, Union[str, int]]]
|
2017-11-16 14:05:26 +01:00
|
|
|
|
2017-11-20 21:49:03 +01:00
|
|
|
for _ in range(args.max_retries):
|
|
|
|
repos_done = []
|
|
|
|
for name, link in repositories.items():
|
|
|
|
contribs = fetch_contributors(link)
|
2017-11-16 14:05:26 +01:00
|
|
|
if contribs:
|
2017-11-20 21:49:03 +01:00
|
|
|
repos_done.append(name)
|
2017-11-16 14:05:26 +01:00
|
|
|
for contrib in contribs:
|
2018-02-08 21:16:14 +01:00
|
|
|
if contrib.get('author') is None:
|
|
|
|
# This happens for users who've deleted their GitHub account.
|
|
|
|
continue
|
2017-11-16 14:05:26 +01:00
|
|
|
username = contrib.get('author').get('login')
|
|
|
|
contrib_data = {
|
|
|
|
'avatar': contrib.get('author').get('avatar_url'),
|
2017-11-20 21:49:03 +01:00
|
|
|
name: contrib.get('total'),
|
2017-11-16 14:05:26 +01:00
|
|
|
}
|
2017-11-20 21:49:03 +01:00
|
|
|
if username in contribs_list:
|
|
|
|
contribs_list[username].update(contrib_data)
|
2017-11-16 14:05:26 +01:00
|
|
|
else:
|
2017-11-20 21:49:03 +01:00
|
|
|
contribs_list[username] = contrib_data
|
2018-03-03 13:21:55 +01:00
|
|
|
|
|
|
|
# remove duplicate contributions count
|
|
|
|
# find commits at the time of split and substract from zulip-server
|
|
|
|
with open(duplicate_commits_file, 'r') as f:
|
|
|
|
duplicate_commits = json.loads(f.read())
|
|
|
|
for committer in duplicate_commits:
|
|
|
|
if contribs_list[committer].get('server'):
|
|
|
|
total_commits = cast(int, contribs_list[committer]['server'])
|
|
|
|
duplicate_commits_count = duplicate_commits[committer]
|
|
|
|
original_commits = total_commits - duplicate_commits_count
|
|
|
|
contribs_list[committer]['server'] = original_commits
|
|
|
|
|
2017-11-20 21:49:03 +01:00
|
|
|
for repo in repos_done:
|
|
|
|
del repositories[repo]
|
2017-11-16 14:05:26 +01:00
|
|
|
|
|
|
|
if not repositories:
|
|
|
|
break
|
|
|
|
|
|
|
|
# Wait before retrying failed requests for Github to aggregate data.
|
|
|
|
sleep(2)
|
|
|
|
else:
|
|
|
|
print("ERROR: Failed fetching contributors data from Github.")
|
|
|
|
if not args.not_required:
|
|
|
|
sys.exit(1)
|
|
|
|
|
2017-11-20 21:49:03 +01:00
|
|
|
for contributor_name, contributor_data in contribs_list.items():
|
|
|
|
contributor_data['name'] = contributor_name
|
|
|
|
data['contrib'].append(contributor_data)
|
2017-11-16 14:05:26 +01:00
|
|
|
|
|
|
|
write_to_disk(data, settings.CONTRIBUTORS_DATA)
|
|
|
|
|
|
|
|
|
|
|
|
def copy_fixture() -> None:
|
2017-01-06 18:56:36 +01:00
|
|
|
"""
|
|
|
|
Copy test fixture file from zerver/fixtures. This is used to avoid
|
|
|
|
constantly fetching data from Github during testing.
|
|
|
|
"""
|
|
|
|
subprocess.check_call(['cp', FIXTURE_FILE, settings.CONTRIBUTORS_DATA])
|
|
|
|
|
2017-11-16 14:05:26 +01:00
|
|
|
|
2017-01-06 18:56:36 +01:00
|
|
|
if args.use_fixture:
|
|
|
|
copy_fixture()
|
|
|
|
else:
|
|
|
|
run_production()
|