kitty/publish.py

601 lines
22 KiB
Python
Raw Normal View History

#!/usr/bin/env python
2017-02-02 13:49:05 +03:00
# License: GPL v3 Copyright: 2017, Kovid Goyal <kovid at kovidgoyal.net>
import argparse
import base64
import contextlib
import datetime
import glob
2017-02-02 13:49:05 +03:00
import io
import json
import mimetypes
import os
import pprint
import re
import shlex
import shutil
2017-02-02 13:49:05 +03:00
import subprocess
import sys
import tempfile
2017-02-02 13:49:05 +03:00
import time
from contextlib import contextmanager, suppress
from http.client import HTTPResponse, HTTPSConnection
from typing import Any, Callable, Dict, Generator, Iterable, List, Optional, Tuple, Union
from urllib.parse import urlencode, urlparse
2017-02-02 13:49:05 +03:00
os.chdir(os.path.dirname(os.path.abspath(__file__)))
docs_dir = os.path.abspath('docs')
publish_dir = os.path.abspath(os.path.join('..', 'kovidgoyal.github.io', 'kitty'))
building_nightly = False
with open('kitty/constants.py') as f:
raw = f.read()
2020-03-15 11:06:09 +03:00
nv = re.search(r'^version: Version\s+=\s+Version\((\d+), (\d+), (\d+)\)', raw, flags=re.MULTILINE)
2020-03-05 15:53:45 +03:00
if nv is not None:
version = f'{nv.group(1)}.{nv.group(2)}.{nv.group(3)}'
2020-03-15 11:06:09 +03:00
ap = re.search(r"^appname: str\s+=\s+'([^']+)'", raw, flags=re.MULTILINE)
if ap is not None:
appname = ap.group(1)
2017-02-02 13:49:05 +03:00
2023-10-31 03:48:23 +03:00
ALL_ACTIONS = 'local_build man html build tag sdist upload website'.split()
NIGHTLY_ACTIONS = 'local_build man html build sdist upload_nightly'.split()
2017-02-02 13:49:05 +03:00
2021-09-21 05:47:53 +03:00
def echo_cmd(cmd: Iterable[str]) -> None:
isatty = sys.stdout.isatty()
end = '\n'
if isatty:
2022-01-29 15:14:36 +03:00
end = f'\x1b[m{end}'
2021-09-21 05:47:53 +03:00
print('\x1b[92m', end='')
print(shlex.join(cmd), end=end, flush=True)
def call(*cmd: str, cwd: Optional[str] = None, echo: bool = False) -> None:
2017-02-02 13:49:05 +03:00
if len(cmd) == 1:
2020-03-08 19:38:18 +03:00
q = shlex.split(cmd[0])
else:
q = list(cmd)
2021-09-21 05:47:53 +03:00
if echo:
echo_cmd(cmd)
2020-03-08 19:38:18 +03:00
ret = subprocess.Popen(q, cwd=cwd).wait()
2017-02-02 13:49:05 +03:00
if ret != 0:
raise SystemExit(ret)
2023-10-31 03:48:23 +03:00
def run_local_build(args: Any) -> None:
call('make debug')
2020-03-14 12:07:11 +03:00
def run_build(args: Any) -> None:
import runpy
m = runpy.run_path('./setup.py', run_name='__publish__')
vcs_rev: str = m['get_vcs_rev']()
def run_with_retry(cmd: str) -> None:
try:
call(cmd, echo=True)
2021-12-23 05:41:16 +03:00
except (SystemExit, Exception):
needs_retry = 'arm64' in cmd or building_nightly
if not needs_retry:
raise
print('Build failed, retrying in a few seconds...', file=sys.stderr)
2022-02-28 04:51:46 +03:00
if 'macos' in cmd:
call('python ../bypy macos shutdown')
time.sleep(25)
call(cmd, echo=True)
for x in ('64', 'arm64'):
2022-11-15 09:02:34 +03:00
prefix = f'python ../bypy linux --arch {x} '
run_with_retry(prefix + f'program --non-interactive --extra-program-data "{vcs_rev}"')
2022-11-15 09:02:34 +03:00
call(prefix + 'shutdown', echo=True)
run_with_retry(f'python ../bypy macos program --sign-installers --notarize --non-interactive --extra-program-data "{vcs_rev}"')
2022-11-15 09:02:34 +03:00
call('python ../bypy macos shutdown', echo=True)
call('make debug')
call('./setup.py build-static-binaries')
2017-02-02 13:49:05 +03:00
2020-03-14 12:07:11 +03:00
def run_tag(args: Any) -> None:
call('git push')
2017-02-02 13:49:05 +03:00
call('git tag -s v{0} -m version-{0}'.format(version))
call(f'git push origin v{version}')
2017-02-02 13:49:05 +03:00
2020-03-14 12:07:11 +03:00
def run_man(args: Any) -> None:
2022-01-23 07:42:11 +03:00
call('make FAIL_WARN=1 man', cwd=docs_dir)
2020-03-14 12:07:11 +03:00
def run_html(args: Any) -> None:
# Force a fresh build otherwise the search index is not correct
2023-07-16 10:46:52 +03:00
with suppress(FileNotFoundError):
shutil.rmtree(os.path.join(docs_dir, '_build', 'dirhtml'))
2022-06-30 14:30:14 +03:00
call('make FAIL_WARN=1 "OPTS=-D analytics_id=G-XTJK3R7GF2" dirhtml', cwd=docs_dir)
add_old_redirects('docs/_build/dirhtml')
2023-07-16 10:46:52 +03:00
with suppress(FileNotFoundError):
shutil.rmtree(os.path.join(docs_dir, '_build', 'html'))
call('make FAIL_WARN=1 "OPTS=-D analytics_id=G-XTJK3R7GF2" html', cwd=docs_dir)
2021-12-10 12:34:39 +03:00
def generate_redirect_html(link_name: str, bname: str) -> None:
with open(link_name, 'w') as f:
f.write(f'''
<html>
<head>
<title>Redirecting...</title>
<link rel="canonical" href="{bname}/" />
2021-07-20 11:21:50 +03:00
<noscript>
<meta http-equiv="refresh" content="0;url={bname}/" />
2021-07-20 11:21:50 +03:00
</noscript>
<script type="text/javascript">
window.location.replace('./{bname}/' + window.location.hash);
</script>
</head>
<body>
<p>Redirecting, please wait...</p>
</body>
</html>
''')
2021-12-10 12:34:39 +03:00
def add_old_redirects(loc: str) -> None:
for dirpath, dirnames, filenames in os.walk(loc):
if dirpath != loc:
for fname in filenames:
if fname == 'index.html':
bname = os.path.basename(dirpath)
base = os.path.dirname(dirpath)
link_name = os.path.join(base, f'{bname}.html') if base else f'{bname}.html'
generate_redirect_html(link_name, bname)
2021-12-10 15:29:09 +03:00
old_unicode_input_path = os.path.join(loc, 'kittens', 'unicode-input')
os.makedirs(old_unicode_input_path, exist_ok=True)
generate_redirect_html(os.path.join(old_unicode_input_path, 'index.html'), '../unicode_input')
generate_redirect_html(f'{old_unicode_input_path}.html', 'unicode_input')
2021-12-10 12:34:39 +03:00
def run_docs(args: Any) -> None:
subprocess.check_call(['make', 'docs'])
2020-03-14 12:07:11 +03:00
def run_website(args: Any) -> None:
if os.path.exists(publish_dir):
shutil.rmtree(publish_dir)
shutil.copytree(os.path.join(docs_dir, '_build', 'dirhtml'), publish_dir, symlinks=True)
with open(os.path.join(publish_dir, 'current-version.txt'), 'w') as f:
f.write(version)
shutil.copy2(os.path.join(docs_dir, 'installer.sh'), publish_dir)
os.chdir(os.path.dirname(publish_dir))
subprocess.check_call(['optipng', '-o7'] + glob.glob('kitty/_images/social_previews/*.png'))
subprocess.check_call(['git', 'add', 'kitty'])
subprocess.check_call(['git', 'commit', '-m', 'kitty website updates'])
subprocess.check_call(['git', 'push'])
2018-05-30 08:11:37 +03:00
def sign_file(path: str) -> None:
2022-01-29 15:14:36 +03:00
dest = f'{path}.sig'
with suppress(FileNotFoundError):
os.remove(dest)
subprocess.check_call([
2022-01-29 15:14:36 +03:00
os.environ['PENV'] + '/gpg-as-kovid', '--output', f'{path}.sig',
'--detach-sig', path
])
2020-03-14 12:07:11 +03:00
def run_sdist(args: Any) -> None:
with tempfile.TemporaryDirectory() as tdir:
base = os.path.join(tdir, f'kitty-{version}')
os.mkdir(base)
2022-01-29 15:14:36 +03:00
subprocess.check_call(f'git archive HEAD | tar -x -C {base}', shell=True)
dest = os.path.join(base, 'docs', '_build')
os.mkdir(dest)
for x in 'html man'.split():
shutil.copytree(os.path.join(docs_dir, '_build', x), os.path.join(dest, x))
dest = os.path.abspath(os.path.join('build', f'kitty-{version}.tar'))
subprocess.check_call(['tar', '-cf', dest, os.path.basename(base)], cwd=tdir)
with suppress(FileNotFoundError):
2022-01-29 15:14:36 +03:00
os.remove(f'{dest}.xz')
subprocess.check_call(['xz', '-9', dest])
2022-01-29 15:14:36 +03:00
sign_file(f'{dest}.xz')
2020-03-08 19:38:18 +03:00
class ReadFileWithProgressReporting(io.FileIO): # {{{
2020-03-14 12:07:11 +03:00
def __init__(self, path: str):
2023-02-24 17:59:51 +03:00
super().__init__(path, 'rb')
2017-02-02 13:49:05 +03:00
self.seek(0, os.SEEK_END)
self._total = self.tell()
self.seek(0)
2020-03-08 19:38:18 +03:00
self.start_time = time.monotonic()
2023-02-24 17:59:51 +03:00
print('Starting upload of:', os.path.basename(path), 'size:', self._total)
2017-02-02 13:49:05 +03:00
2020-03-08 19:38:18 +03:00
def __len__(self) -> int:
2017-02-02 13:49:05 +03:00
return self._total
def read(self, size: int = -1) -> bytes:
2020-03-08 19:38:18 +03:00
data = io.FileIO.read(self, size)
2017-02-02 13:49:05 +03:00
if data:
self.report_progress(len(data))
return data
2020-03-08 19:38:18 +03:00
def report_progress(self, size: int) -> None:
def write(*args: str) -> None:
print(*args, end='')
2023-01-22 06:11:38 +03:00
frac = int(self.tell() * 100 / self._total)
2017-02-02 13:49:05 +03:00
mb_pos = self.tell() / float(1024**2)
mb_tot = self._total / float(1024**2)
kb_pos = self.tell() / 1024.0
2020-03-08 19:38:18 +03:00
kb_rate = kb_pos / (time.monotonic() - self.start_time)
2017-02-02 13:49:05 +03:00
bit_rate = kb_rate * 1024
eta = int((self._total - self.tell()) / bit_rate) + 1
2023-01-22 06:11:38 +03:00
eta_m, eta_s = divmod(eta, 60)
if sys.stdout.isatty():
write(
2023-01-22 06:11:38 +03:00
f'\r\033[K\033[?7h {frac}% {mb_pos:.1f}/{mb_tot:.1f}MB {kb_rate:.1f} KB/sec {eta_m} minutes, {eta_s} seconds left\033[?7l')
2017-02-02 13:49:05 +03:00
if self.tell() >= self._total:
t = int(time.monotonic() - self.start_time) + 1
2021-10-01 05:27:21 +03:00
print(f'\nUpload took {t//60} minutes and {t%60} seconds at {kb_rate:.1f} KB/sec')
2017-02-02 13:49:05 +03:00
sys.stdout.flush()
# }}}
2023-01-21 10:36:16 +03:00
class GitHub: # {{{
2017-02-02 13:49:05 +03:00
2022-01-29 15:14:36 +03:00
API = 'https://api.github.com'
2017-02-02 13:49:05 +03:00
2020-03-14 12:07:11 +03:00
def __init__(
self,
files: Dict[str, str],
reponame: str,
version: str,
username: str,
password: str,
replace: bool = False
):
2017-02-02 13:49:05 +03:00
self.files, self.reponame, self.version, self.username, self.password, self.replace = (
files, reponame, version, username, password, replace)
2022-01-29 15:14:36 +03:00
self.current_tag_name = self.version if self.version == 'nightly' else f'v{self.version}'
self.is_nightly = self.current_tag_name == 'nightly'
2023-01-22 05:30:29 +03:00
self.auth = 'Basic ' + base64.standard_b64encode(f'{self.username}:{self.password}'.encode()).decode()
2022-01-29 15:14:36 +03:00
self.url_base = f'{self.API}/repos/{self.username}/{self.reponame}/releases'
2017-02-02 13:49:05 +03:00
2023-01-21 10:36:16 +03:00
def info(self, *args: Any) -> None:
print(*args, flush=True)
def error(self, *args: Any) -> None:
print(*args, flush=True, file=sys.stderr)
def make_request(
self, url: str, data: Optional[Dict[str, Any]] = None, method:str = 'GET',
upload_data: Optional[ReadFileWithProgressReporting] = None,
params: Optional[Dict[str, str]] = None,
) -> HTTPSConnection:
headers={
'Authorization': self.auth,
2023-01-22 05:30:29 +03:00
'Accept': 'application/vnd.github+json',
'User-Agent': 'kitty',
'X-GitHub-Api-Version': '2022-11-28',
}
if params:
url += '?' + urlencode(params)
rdata: Optional[Union[bytes, io.FileIO]] = None
if data is not None:
rdata = json.dumps(data).encode('utf-8')
headers['Content-Type'] = 'application/json'
headers['Content-Length'] = str(len(rdata))
elif upload_data is not None:
rdata = upload_data
mime_type = mimetypes.guess_type(os.path.basename(str(upload_data.name)))[0] or 'application/octet-stream'
headers['Content-Type'] = mime_type
headers['Content-Length'] = str(upload_data._total)
purl = urlparse(url)
conn = HTTPSConnection(purl.netloc, timeout=60)
conn.request(method, url, body=rdata, headers=headers)
return conn
def make_request_with_retries(
self, url: str, data: Optional[Dict[str, str]] = None, method:str = 'GET',
num_tries: int = 2, sleep_between_tries: float = 15,
success_codes: Tuple[int, ...] = (200,),
failure_msg: str = 'Request failed',
return_data: bool = False,
upload_path: str = '',
params: Optional[Dict[str, str]] = None,
failure_callback: Callable[[HTTPResponse], None] = lambda r: None,
) -> Any:
for i in range(num_tries):
2024-03-24 18:10:00 +03:00
is_last_try = i == num_tries - 1
try:
2023-01-22 05:30:29 +03:00
if upload_path:
conn = self.make_request(url, method='POST', upload_data=ReadFileWithProgressReporting(upload_path), params=params)
else:
conn = self.make_request(url, data, method, params=params)
with contextlib.closing(conn):
r = conn.getresponse()
if r.status in success_codes:
2023-01-21 10:36:16 +03:00
return json.loads(r.read()) if return_data else None
2024-03-24 18:10:00 +03:00
if is_last_try:
self.fail(r, failure_msg)
else:
self.print_failed_response_details(r, failure_msg)
failure_callback(r)
except Exception as e:
2023-01-21 10:36:16 +03:00
self.error(failure_msg, 'with error:', e)
self.error(f'Retrying after {sleep_between_tries} seconds')
2024-03-24 18:10:00 +03:00
if is_last_try:
break
time.sleep(sleep_between_tries)
raise SystemExit('All retries failed, giving up')
def patch(self, url: str, fail_msg: str, **data: str) -> None:
self.make_request_with_retries(url, data, method='PATCH', failure_msg=fail_msg)
def update_nightly_description(self, release_id: int) -> None:
2022-01-29 15:14:36 +03:00
url = f'{self.url_base}/{release_id}'
now = str(datetime.datetime.now(datetime.timezone.utc)).split('.')[0] + ' UTC'
commit = subprocess.check_output(['git', 'rev-parse', '--verify', '--end-of-options', 'master^{commit}']).decode('utf-8').strip()
self.patch(
url, 'Failed to update nightly release description',
body=f'Nightly release, generated on: {now} from commit: {commit}.'
' For how to install nightly builds, see: https://sw.kovidgoyal.net/kitty/binary/#customizing-the-installation'
)
2020-03-14 12:07:11 +03:00
def __call__(self) -> None:
# See https://docs.github.com/en/rest/releases/assets#upload-a-release-asset
release = self.create_release()
2017-02-02 13:49:05 +03:00
upload_url = release['upload_url'].partition('{')[0]
all_assest_for_release = self.existing_assets_for_release(release)
assets_by_fname = {a['name']:a for a in all_assest_for_release}
2022-11-19 08:53:31 +03:00
def delete_asset(asset: Dict[str, Any], allow_not_found: bool = True) -> None:
success_codes = [204]
if allow_not_found:
success_codes.append(404)
self.make_request_with_retries(
asset['url'], method='DELETE', num_tries=5, sleep_between_tries=2, success_codes=tuple(success_codes),
2024-03-21 07:16:02 +03:00
failure_msg='Failed to delete asset from GitHub')
2022-11-19 08:53:31 +03:00
def upload_with_retries(path: str, desc: str, num_tries: int = 8, sleep_time: float = 60.0) -> None:
2017-02-02 13:49:05 +03:00
fname = os.path.basename(path)
if self.is_nightly:
fname = fname.replace(version, 'nightly')
if fname in assets_by_fname:
self.info(f'Deleting {fname} from GitHub with id: {assets_by_fname[fname]["id"]}')
delete_asset(assets_by_fname.pop(fname))
params = {'name': fname, 'label': desc}
self.make_request_with_retries(
upload_url, upload_path=path, params=params, num_tries=num_tries, sleep_between_tries=sleep_time,
failure_msg=f'Failed to upload file: {fname}', success_codes=(201,),
)
2022-11-19 09:01:43 +03:00
if self.is_nightly:
for fname in tuple(assets_by_fname):
self.info(f'Deleting {fname} from GitHub with id: {assets_by_fname[fname]["id"]}')
delete_asset(assets_by_fname.pop(fname))
2022-11-19 09:01:43 +03:00
for path, desc in self.files.items():
self.info('')
upload_with_retries(path, desc)
if self.is_nightly:
self.update_nightly_description(release['id'])
def print_failed_response_details(self, r: HTTPResponse, msg: str) -> None:
2023-01-21 10:36:16 +03:00
self.error(msg, f'\nStatus Code: {r.status} {r.reason}')
try:
jr = json.loads(r.read())
except Exception:
pass
else:
2023-01-21 10:36:16 +03:00
self.error('JSON from response:')
pprint.pprint(jr, stream=sys.stderr)
def fail(self, r: HTTPResponse, msg: str) -> None:
self.print_failed_response_details(r, msg)
2017-02-02 13:49:05 +03:00
raise SystemExit(1)
def existing_assets_for_release(self, release: Dict[str, Any]) -> List[Dict[str, Any]]:
if 'assets' in release:
d: List[Dict[str, Any]] = release['assets']
else:
d = self.make_request_with_retries(
release['assets_url'], params={'per_page': '64'}, failure_msg='Failed to get assets for release', return_data=True)
return d
2017-02-02 13:49:05 +03:00
def create_release(self) -> Dict[str, Any]:
2017-02-02 13:49:05 +03:00
' Create a release on GitHub or if it already exists, return the existing release '
# Check for existing release
2022-01-29 15:14:36 +03:00
url = f'{self.url_base}/tags/{self.current_tag_name}'
with contextlib.closing(self.make_request(url)) as conn:
r = conn.getresponse()
if r.status == 200:
return {str(k): v for k, v in json.loads(r.read()).items()}
if self.is_nightly:
2023-01-22 05:30:29 +03:00
self.fail(r, 'No existing nightly release found on GitHub')
data = {
'tag_name': self.current_tag_name,
'target_commitish': 'master',
'name': f'version {self.version}',
'body': f'Release version {self.version}.'
' For changelog, see https://sw.kovidgoyal.net/kitty/changelog/#detailed-list-of-changes'
' GPG key used for signing tarballs is: https://calibre-ebook.com/signatures/kovid.gpg',
'draft': False,
'prerelease': False
}
with contextlib.closing(self.make_request(self.url_base, method='POST', data=data)) as conn:
r = conn.getresponse()
if r.status != 201:
self.fail(r, f'Failed to create release for version: {self.version}')
return {str(k): v for k, v in json.loads(r.read()).items()}
2017-02-02 13:49:05 +03:00
# }}}
2020-03-14 12:07:11 +03:00
def get_github_data() -> Dict[str, str]:
2019-12-12 07:35:11 +03:00
with open(os.environ['PENV'] + '/github-token') as f:
2017-02-02 13:49:05 +03:00
un, pw = f.read().strip().split(':')
return {'username': un, 'password': pw}
def files_for_upload() -> Dict[str, str]:
files = {}
signatures = {}
for f, desc in {
'macos/dist/kitty-{}.dmg': 'macOS dmg',
'linux/64/dist/kitty-{}-x86_64.txz': 'Linux amd64 binary bundle',
'linux/32/dist/kitty-{}-i686.txz': 'Linux x86 binary bundle',
2021-12-10 10:17:34 +03:00
'linux/arm64/dist/kitty-{}-arm64.txz': 'Linux arm64 binary bundle',
}.items():
path = os.path.join('bypy', 'b', f.format(version))
if not os.path.exists(path):
raise SystemExit(f'The installer {path} does not exist')
files[path] = desc
2021-05-07 03:44:40 +03:00
signatures[path] = f'GPG signature for {desc}'
b = len(files)
for path in glob.glob('build/static/kitten-*'):
2022-11-16 11:15:27 +03:00
if path.endswith('.sig'):
continue
path = os.path.abspath(path)
exe_name = os.path.basename(path)
files[path] = f'Static {exe_name} executable'
signatures[path] = f'GPG signature for static {exe_name} executable'
if len(files) == b:
raise SystemExit('No static binaries found')
files[f'build/kitty-{version}.tar.xz'] = 'Source code'
files[f'build/kitty-{version}.tar.xz.sig'] = 'Source code GPG signature'
for path, desc in signatures.items():
sign_file(path)
2022-01-29 15:14:36 +03:00
files[f'{path}.sig'] = desc
2017-02-02 13:49:05 +03:00
for f in files:
if not os.path.exists(f):
raise SystemExit(f'The release artifact {f} does not exist')
return files
def run_upload(args: Any) -> None:
2017-02-02 13:49:05 +03:00
gd = get_github_data()
files = files_for_upload()
2017-02-02 13:54:38 +03:00
gh = GitHub(files, appname, version, gd['username'], gd['password'])
2017-02-02 13:49:05 +03:00
gh()
def run_upload_nightly(args: Any) -> None:
subprocess.check_call(['git', 'tag', '-f', 'nightly'])
subprocess.check_call(['git', 'push', 'origin', 'nightly', '-f'])
gd = get_github_data()
files = files_for_upload()
gh = GitHub(files, appname, 'nightly', gd['username'], gd['password'])
gh()
def current_branch() -> str:
return subprocess.check_output(['git', 'symbolic-ref', '--short', 'HEAD']).decode('utf-8').strip()
2020-03-14 12:07:11 +03:00
def require_git_master(branch: str = 'master') -> None:
if current_branch() != branch:
raise SystemExit(f'You must be in the {branch} git branch')
def safe_read(path: str) -> str:
with suppress(FileNotFoundError):
with open(path) as f:
return f.read()
return ''
2024-05-07 06:22:49 +03:00
def remove_pycache_only_folders() -> None:
folders_to_remove = []
for dirpath, folders, files in os.walk('.'):
if not files and folders == ['__pycache__']:
folders_to_remove.append(dirpath)
for x in folders_to_remove:
shutil.rmtree(x)
@contextmanager
def change_to_git_master() -> Generator[None, None, None]:
stash_ref_before = safe_read('.git/refs/stash')
subprocess.check_call(['git', 'stash', '-u'])
try:
branch_before = current_branch()
if branch_before != 'master':
subprocess.check_call(['git', 'switch', 'master'])
remove_pycache_only_folders()
subprocess.check_call(['make', 'clean', 'debug'])
try:
yield
finally:
if branch_before != 'master':
subprocess.check_call(['git', 'switch', branch_before])
subprocess.check_call(['make', 'clean', 'debug'])
finally:
if stash_ref_before != safe_read('.git/refs/stash'):
subprocess.check_call(['git', 'stash', 'pop'])
2020-03-14 12:07:11 +03:00
def require_penv() -> None:
2018-09-01 08:02:08 +03:00
if 'PENV' not in os.environ:
raise SystemExit('The PENV env var is not present, required for uploading releases')
def exec_actions(actions: Iterable[str], args: Any) -> None:
for action in actions:
print('Running', action)
cwd = os.getcwd()
2022-01-29 15:14:36 +03:00
globals()[f'run_{action}'](args)
os.chdir(cwd)
2020-03-14 12:07:11 +03:00
def main() -> None:
global building_nightly
2017-02-02 13:49:05 +03:00
parser = argparse.ArgumentParser(description='Publish kitty')
parser.add_argument(
'--only',
default=False,
action='store_true',
help='Only run the specified action, by default the specified action and all sub-sequent actions are run')
parser.add_argument(
'--nightly',
default=False,
action='store_true',
help='Upload a nightly release, ignores all other arguments')
2017-02-02 13:49:05 +03:00
parser.add_argument(
'action',
default='all',
2017-02-02 13:49:05 +03:00
nargs='?',
choices=list(ALL_ACTIONS) + ['all', 'upload_nightly'],
2017-02-02 13:49:05 +03:00
help='The action to start with')
args = parser.parse_args()
require_penv()
if args.nightly:
with change_to_git_master():
building_nightly = True
exec_actions(NIGHTLY_ACTIONS, args)
subprocess.run(['make', 'clean', 'debug'])
return
require_git_master()
if args.action == 'all':
actions = list(ALL_ACTIONS)
elif args.action == 'upload_nightly':
actions = ['upload_nightly']
else:
idx = ALL_ACTIONS.index(args.action)
actions = ALL_ACTIONS[idx:]
2017-02-02 13:49:05 +03:00
if args.only:
del actions[1:]
else:
try:
ans = input(f'Publish version \033[91m{version}\033[m (y/n): ')
except KeyboardInterrupt:
ans = 'n'
if ans.lower() != 'y':
return
if actions == ['website']:
actions.insert(0, 'html')
exec_actions(actions, args)
2017-02-02 13:49:05 +03:00
if __name__ == '__main__':
main()