diff options
Diffstat (limited to 'pkgs/development/tools/electron/update.py')
-rwxr-xr-x | pkgs/development/tools/electron/update.py | 792 |
1 files changed, 641 insertions, 151 deletions
diff --git a/pkgs/development/tools/electron/update.py b/pkgs/development/tools/electron/update.py index 128b1dc050672..bf0f7a3a4758f 100755 --- a/pkgs/development/tools/electron/update.py +++ b/pkgs/development/tools/electron/update.py @@ -1,81 +1,105 @@ #! /usr/bin/env nix-shell #! nix-shell -i python -p python3.pkgs.joblib python3.pkgs.click python3.pkgs.click-log nix nix-prefetch-git nix-universal-prefetch prefetch-yarn-deps prefetch-npm-deps +""" +electron updater -import logging -import click_log -import click -import random -import traceback -import csv +A script for updating both binary and source hashes. + +It supports the following modes: + +| Mode | Description | +|------------- | ----------------------------------------------- | +| `update` | for updating a specific Electron release | +| `update-all` | for updating all electron releases at once | +| `eval` | just print the necessary sources to fetch | + +The `eval` and `update` commands accept an optional `--version` flag +to restrict the mechanism only to a given major release. + +The `update` and `update-all` commands accept an optional `--commit` +flag to automatically commit the changes for you. + +The `update` and `update-all` commands accept optional `--bin-only` +and `--source-only` flags to restict the update to binary or source +releases. +""" import base64 +import csv +import json +import logging import os +import random import re -import tempfile import subprocess -import json import sys -from joblib import Parallel, delayed, Memory +import tempfile +import traceback +import urllib.request + +from abc import ABC from codecs import iterdecode from datetime import datetime +from typing import Iterable, Optional, Tuple from urllib.request import urlopen -os.chdir(os.path.dirname(__file__)) +import click +import click_log + +from joblib import Parallel, delayed, Memory depot_tools_checkout = tempfile.TemporaryDirectory() -subprocess.check_call([ - "nix-prefetch-git", - "--builder", "--quiet", - "--url", "https://chromium.googlesource.com/chromium/tools/depot_tools", - "--out", depot_tools_checkout.name, - "--rev", "7a69b031d58081d51c9e8e89557b343bba8518b1"]) +subprocess.check_call( + [ + "nix-prefetch-git", + "--builder", + "--quiet", + "--url", + "https://chromium.googlesource.com/chromium/tools/depot_tools", + "--out", + depot_tools_checkout.name, + "--rev", + "7a69b031d58081d51c9e8e89557b343bba8518b1", + ] +) sys.path.append(depot_tools_checkout.name) import gclient_eval import gclient_utils -memory = Memory("cache", verbose=0) -@memory.cache -def get_repo_hash(fetcher, args): - cmd = ['nix-universal-prefetch', fetcher] - for arg_name, arg in args.items(): - cmd.append(f'--{arg_name}') - cmd.append(arg) +# Relative path to the electron-source info.json +SOURCE_INFO_JSON = "info.json" - print(" ".join(cmd), file=sys.stderr) - out = subprocess.check_output(cmd) - return out.decode('utf-8').strip() +# Relatice path to the electron-bin info.json +BINARY_INFO_JSON = "binary/info.json" -@memory.cache -def _get_yarn_hash(file): - print(f'prefetch-yarn-deps', file=sys.stderr) - with tempfile.TemporaryDirectory() as tmp_dir: - with open(tmp_dir + '/yarn.lock', 'w') as f: - f.write(file) - return subprocess.check_output(['prefetch-yarn-deps', tmp_dir + '/yarn.lock']).decode('utf-8').strip() -def get_yarn_hash(repo, yarn_lock_path = 'yarn.lock'): - return _get_yarn_hash(repo.get_file(yarn_lock_path)) +# Number of spaces used for each indentation level +JSON_INDENT = 4 + +os.chdir(os.path.dirname(__file__)) + +memory: Memory = Memory("cache", verbose=0) + +logger = logging.getLogger(__name__) +click_log.basic_config(logger) -@memory.cache -def _get_npm_hash(file): - print(f'prefetch-npm-deps', file=sys.stderr) - with tempfile.TemporaryDirectory() as tmp_dir: - with open(tmp_dir + '/package-lock.json', 'w') as f: - f.write(file) - return subprocess.check_output(['prefetch-npm-deps', tmp_dir + '/package-lock.json']).decode('utf-8').strip() -def get_npm_hash(repo, package_lock_path = 'package-lock.json'): - return _get_npm_hash(repo.get_file(package_lock_path)) class Repo: - def __init__(self): - self.deps = {} + fetcher: str + args: dict + + def __init__(self) -> None: + self.deps: dict = {} self.hash = "sha256-AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA=" - def get_deps(self, repo_vars, path): - print("evaluating " + json.dumps(self, default = vars), file=sys.stderr) + def get_deps(self, repo_vars: dict, path: str) -> None: + print( + "evaluating " + json.dumps(self, default=vars, sort_keys=True), + file=sys.stderr, + ) deps_file = self.get_file("DEPS") - evaluated = gclient_eval.Parse(deps_file, filename='DEPS') + evaluated = gclient_eval.Parse(deps_file, filename="DEPS") repo_vars = dict(evaluated["vars"]) | repo_vars @@ -84,7 +108,12 @@ class Repo: self.deps = { prefix + dep_name: repo_from_dep(dep) for dep_name, dep in evaluated["deps"].items() - if (gclient_eval.EvaluateCondition(dep["condition"], repo_vars) if "condition" in dep else True) and repo_from_dep(dep) != None + if ( + gclient_eval.EvaluateCondition(dep["condition"], repo_vars) + if "condition" in dep + else True + ) + and repo_from_dep(dep) != None } for key in evaluated.get("recursedeps", []): @@ -92,80 +121,244 @@ class Repo: if dep_path in self.deps and dep_path != "src/third_party/squirrel.mac": self.deps[dep_path].get_deps(repo_vars, dep_path) - def prefetch(self): + def prefetch(self) -> None: self.hash = get_repo_hash(self.fetcher, self.args) - def prefetch_all(self): - return sum([dep.prefetch_all() for [_, dep] in self.deps.items()], [delayed(self.prefetch)()]) + def prefetch_all(self) -> int: + return sum( + [dep.prefetch_all() for [_, dep] in self.deps.items()], + [delayed(self.prefetch)()], + ) - def flatten_repr(self): - return { - "fetcher": self.fetcher, - "hash": self.hash, - **self.args - } + def flatten_repr(self) -> dict: + return {"fetcher": self.fetcher, "hash": self.hash, **self.args} - def flatten(self, path): - out = { - path: self.flatten_repr() - } + def flatten(self, path: str) -> dict: + out = {path: self.flatten_repr()} for dep_path, dep in self.deps.items(): out |= dep.flatten(dep_path) return out + def get_file(self, filepath: str) -> str: + raise NotImplementedError + + class GitRepo(Repo): - def __init__(self, url, rev): + def __init__(self, url: str, rev: str) -> None: super().__init__() - self.fetcher = 'fetchgit' + self.fetcher = "fetchgit" self.args = { "url": url, "rev": rev, } + class GitHubRepo(Repo): - def __init__(self, owner, repo, rev): + def __init__(self, owner: str, repo: str, rev: str) -> None: super().__init__() - self.fetcher = 'fetchFromGitHub' + self.fetcher = "fetchFromGitHub" self.args = { "owner": owner, "repo": repo, "rev": rev, } - def get_file(self, filepath): - return urlopen(f"https://raw.githubusercontent.com/{self.args['owner']}/{self.args['repo']}/{self.args['rev']}/{filepath}").read().decode('utf-8') + def get_file(self, filepath: str) -> str: + return ( + urlopen( + f"https://raw.githubusercontent.com/{self.args['owner']}/{self.args['repo']}/{self.args['rev']}/{filepath}" + ) + .read() + .decode("utf-8") + ) + class GitilesRepo(Repo): - def __init__(self, url, rev): + def __init__(self, url: str, rev: str) -> None: super().__init__() - self.fetcher = 'fetchFromGitiles' - #self.fetcher = 'fetchgit' + self.fetcher = "fetchFromGitiles" + # self.fetcher = 'fetchgit' self.args = { "url": url, "rev": rev, - #"fetchSubmodules": "false", + # "fetchSubmodules": "false", } if url == "https://chromium.googlesource.com/chromium/src.git": - self.args['postFetch'] = "rm -r $out/third_party/blink/web_tests; " - self.args['postFetch'] += "rm -r $out/third_party/hunspell/tests; " - self.args['postFetch'] += "rm -r $out/content/test/data; " - self.args['postFetch'] += "rm -r $out/courgette/testdata; " - self.args['postFetch'] += "rm -r $out/extensions/test/data; " - self.args['postFetch'] += "rm -r $out/media/test/data; " + self.args["postFetch"] = "rm -r $out/third_party/blink/web_tests; " + self.args["postFetch"] += "rm -r $out/third_party/hunspell/tests; " + self.args["postFetch"] += "rm -r $out/content/test/data; " + self.args["postFetch"] += "rm -r $out/courgette/testdata; " + self.args["postFetch"] += "rm -r $out/extensions/test/data; " + self.args["postFetch"] += "rm -r $out/media/test/data; " + + def get_file(self, filepath: str) -> str: + return base64.b64decode( + urlopen( + f"{self.args['url']}/+/{self.args['rev']}/{filepath}?format=TEXT" + ).read() + ).decode("utf-8") + + +class ElectronBinRepo(GitHubRepo): + def __init__(self, owner: str, repo: str, rev: str) -> None: + super().__init__(owner, repo, rev) + self.systems = { + "i686-linux": "linux-ia32", + "x86_64-linux": "linux-x64", + "armv7l-linux": "linux-armv7l", + "aarch64-linux": "linux-arm64", + "x86_64-darwin": "darwin-x64", + "aarch64-darwin": "darwin-arm64", + } + + def get_shasums256(self, version: str) -> list: + """Returns the contents of SHASUMS256.txt""" + try: + called_process: subprocess.CompletedProcess = subprocess.run( + [ + "nix-prefetch-url", + "--print-path", + f"https://github.com/electron/electron/releases/download/v{version}/SHASUMS256.txt", + ], + capture_output=True, + check=True, + text=True, + ) + + hash_file_path = called_process.stdout.split("\n")[1] + + with open(hash_file_path, "r") as f: + return f.read().split("\n") + + except subprocess.CalledProcessError as err: + print(err.stderr) + sys.exit(1) + + def get_headers(self, version: str) -> str: + """Returns the hash of the release headers tarball""" + try: + called_process: subprocess.CompletedProcess = subprocess.run( + [ + "nix-prefetch-url", + f"https://artifacts.electronjs.org/headers/dist/v{version}/node-v{version}-headers.tar.gz", + ], + capture_output=True, + check=True, + text=True, + ) + return called_process.stdout.split("\n")[0] + except subprocess.CalledProcessError as err: + print(err.stderr) + sys.exit(1) + + def get_hashes(self, major_version: str) -> dict: + """Returns a dictionary of hashes for a given major version""" + m, _ = get_latest_version(major_version) + version: str = m["version"] + + out = {} + out[major_version] = { + "hashes": {}, + "version": version, + } + + hashes: list = self.get_shasums256(version) + + for nix_system, electron_system in self.systems.items(): + filename = f"*electron-v{version}-{electron_system}.zip" + if any([x.endswith(filename) for x in hashes]): + out[major_version]["hashes"][nix_system] = [ + x.split(" ")[0] for x in hashes if x.endswith(filename) + ][0] + out[major_version]["hashes"]["headers"] = self.get_headers(version) + + return out - def get_file(self, filepath): - return base64.b64decode(urlopen(f"{self.args['url']}/+/{self.args['rev']}/{filepath}?format=TEXT").read()).decode('utf-8') -def repo_from_dep(dep): +# Releases that have reached end-of-life no longer receive any updates +# and it is rather pointless trying to update those. +# +# https://endoflife.date/electron +@memory.cache +def supported_version_range() -> range: + """Returns a range of electron releases that have not reached end-of-life yet""" + releases_json = json.loads( + urlopen("https://endoflife.date/api/electron.json").read() + ) + supported_releases = [ + int(x["cycle"]) + for x in releases_json + if x["eol"] == False + or datetime.strptime(x["eol"], "%Y-%m-%d") > datetime.today() + ] + + return range( + min(supported_releases), # incl. + # We have also packaged the beta release in nixpkgs, + # but it is not tracked by endoflife.date + max(supported_releases) + 2, # excl. + 1, + ) + + +@memory.cache +def get_repo_hash(fetcher: str, args: dict) -> str: + cmd = ["nix-universal-prefetch", fetcher] + for arg_name, arg in args.items(): + cmd.append(f"--{arg_name}") + cmd.append(arg) + + print(" ".join(cmd), file=sys.stderr) + out = subprocess.check_output(cmd) + return out.decode("utf-8").strip() + + +@memory.cache +def _get_yarn_hash(path: str) -> str: + print(f"prefetch-yarn-deps", file=sys.stderr) + with tempfile.TemporaryDirectory() as tmp_dir: + with open(tmp_dir + "/yarn.lock", "w") as f: + f.write(path) + return ( + subprocess.check_output(["prefetch-yarn-deps", tmp_dir + "/yarn.lock"]) + .decode("utf-8") + .strip() + ) + + +def get_yarn_hash(repo: Repo, yarn_lock_path: str = "yarn.lock") -> str: + return _get_yarn_hash(repo.get_file(yarn_lock_path)) + + +@memory.cache +def _get_npm_hash(filename: str) -> str: + print(f"prefetch-npm-deps", file=sys.stderr) + with tempfile.TemporaryDirectory() as tmp_dir: + with open(tmp_dir + "/package-lock.json", "w") as f: + f.write(filename) + return ( + subprocess.check_output( + ["prefetch-npm-deps", tmp_dir + "/package-lock.json"] + ) + .decode("utf-8") + .strip() + ) + + +def get_npm_hash(repo: Repo, package_lock_path: str = "package-lock.json") -> str: + return _get_npm_hash(repo.get_file(package_lock_path)) + + +def repo_from_dep(dep: dict) -> Optional[Repo]: if "url" in dep: url, rev = gclient_utils.SplitUrlRevision(dep["url"]) - search_object = re.search(r'https://github.com/(.+)/(.+?)(\.git)?$', url) + search_object = re.search(r"https://github.com/(.+)/(.+?)(\.git)?$", url) if search_object: return GitHubRepo(search_object.group(1), search_object.group(2), rev) - if re.match(r'https://.+.googlesource.com', url): + if re.match(r"https://.+.googlesource.com", url): return GitilesRepo(url, rev) return GitRepo(url, rev) @@ -173,107 +366,404 @@ def repo_from_dep(dep): # Not a git dependency; skip return None -def get_gn_source(repo): + +def get_gn_source(repo: Repo) -> dict: gn_pattern = r"'gn_version': 'git_revision:([0-9a-f]{40})'" gn_commit = re.search(gn_pattern, repo.get_file("DEPS")).group(1) - gn = subprocess.check_output([ - "nix-prefetch-git", - "--quiet", - "https://gn.googlesource.com/gn", - "--rev", gn_commit - ]) - gn = json.loads(gn) + gn_prefetch: bytes = subprocess.check_output( + [ + "nix-prefetch-git", + "--quiet", + "https://gn.googlesource.com/gn", + "--rev", + gn_commit, + ] + ) + gn: dict = json.loads(gn_prefetch) return { "gn": { "version": datetime.fromisoformat(gn["date"]).date().isoformat(), "url": gn["url"], "rev": gn["rev"], - "hash": gn["hash"] + "hash": gn["hash"], } } -def get_electron_info(major_version): - electron_releases = json.loads(urlopen("https://releases.electronjs.org/releases.json").read()) - major_version_releases = filter(lambda item: item["version"].startswith(f"{major_version}."), electron_releases) + +def get_latest_version(major_version: str) -> Tuple[str, str]: + """Returns the latest version for a given major version""" + electron_releases: dict = json.loads( + urlopen("https://releases.electronjs.org/releases.json").read() + ) + major_version_releases = filter( + lambda item: item["version"].startswith(f"{major_version}."), electron_releases + ) m = max(major_version_releases, key=lambda item: item["date"]) - rev=f"v{m['version']}" + rev = f"v{m['version']}" + return (m, rev) - electron_repo = GitHubRepo("electron", "electron", rev) - electron_repo.recurse = True - electron_repo.get_deps({ - f"checkout_{platform}": platform == "linux" - for platform in ["ios", "chromeos", "android", "mac", "win", "linux"] - }, "src/electron") +def get_electron_bin_info(major_version: str) -> Tuple[str, str, ElectronBinRepo]: + m, rev = get_latest_version(major_version) + electron_repo: ElectronBinRepo = ElectronBinRepo("electron", "electron", rev) return (major_version, m, electron_repo) -logger = logging.getLogger(__name__) -click_log.basic_config(logger) -@click.group() -def cli(): - pass +def get_electron_info(major_version: str) -> Tuple[str, str, GitHubRepo]: + m, rev = get_latest_version(major_version) + + electron_repo: GitHubRepo = GitHubRepo("electron", "electron", rev) + electron_repo.get_deps( + { + f"checkout_{platform}": platform == "linux" + for platform in ["ios", "chromeos", "android", "mac", "win", "linux"] + }, + "src/electron", + ) + + return (major_version, m, electron_repo) -@cli.command("eval") -@click.option("--version", help="The major version, e.g. '23'") -def eval(version): - (_, _, repo) = electron_repo = get_electron_info(version) - tree = electron_repo.flatten("src/electron") - print(json.dumps(tree, indent=4, default = vars)) -def get_update(repo): +def get_update(repo: Tuple[str, str, Repo]) -> Tuple[str, dict]: (major_version, m, electron_repo) = repo tasks = electron_repo.prefetch_all() - a = lambda: ( - ("electron_yarn_hash", get_yarn_hash(electron_repo)) - ) + a = lambda: (("electron_yarn_hash", get_yarn_hash(electron_repo))) tasks.append(delayed(a)()) a = lambda: ( - ("chromium_npm_hash", get_npm_hash(electron_repo.deps["src"], "third_party/node/package-lock.json")) + ( + "chromium_npm_hash", + get_npm_hash( + electron_repo.deps["src"], "third_party/node/package-lock.json" + ), + ) ) tasks.append(delayed(a)()) random.shuffle(tasks) - task_results = {n[0]: n[1] for n in Parallel(n_jobs=3, require='sharedmem', return_as="generator")(tasks) if n != None} + task_results = { + n[0]: n[1] + for n in Parallel(n_jobs=3, require="sharedmem", return_as="generator")(tasks) + if n != None + } tree = electron_repo.flatten("src/electron") - return (f"{major_version}", { - "deps": tree, - **{key: m[key] for key in ["version", "modules", "chrome", "node"]}, - "chromium": { - "version": m['chrome'], - "deps": get_gn_source(electron_repo.deps["src"]) - }, - **task_results - }) - -@cli.command("update") -@click.option("--version", help="The major version, e.g. '23'") -def update(version): + return ( + f"{major_version}", + { + "deps": tree, + **{key: m[key] for key in ["version", "modules", "chrome", "node"]}, + "chromium": { + "version": m["chrome"], + "deps": get_gn_source(electron_repo.deps["src"]), + }, + **task_results, + }, + ) + + +def load_info_json(path: str) -> dict: + """Load the contents of a JSON file + + Args: + path: The path to the JSON file + + Returns: An empty dict if the path does not exist, otherwise the contents of the JSON file. + """ try: - with open('info.json', 'r') as f: - old_info = json.loads(f.read()) + with open(path, "r") as f: + return json.loads(f.read()) except: - old_info = {} - repo = get_electron_info(version) - update = get_update(repo) - out = old_info | { update[0]: update[1] } - with open('info.json', 'w') as f: - f.write(json.dumps(out, indent=4, default = vars)) - f.write('\n') - -@cli.command("update-all") -def update_all(): - repos = Parallel(n_jobs=2, require='sharedmem')(delayed(get_electron_info)(major_version) for major_version in range(28, 24, -1)) - out = {n[0]: n[1] for n in Parallel(n_jobs=2, require='sharedmem')(delayed(get_update)(repo) for repo in repos)} - - with open('info.json', 'w') as f: - f.write(json.dumps(out, indent=4, default = vars)) - f.write('\n') + return {} + + +def save_info_json(path: str, content: dict) -> None: + """Saves the given info to a JSON file + + Args: + path: The path where the info should be saved + content: The content to be saved as JSON. + """ + with open(path, "w") as f: + f.write(json.dumps(content, indent=JSON_INDENT, default=vars, sort_keys=True)) + f.write("\n") + + +def update_bin(major_version: str, commit: bool) -> None: + """Update a given electron-bin release + + Args: + major_version: The major version number, e.g. '27' + commit: Whether the updater should commit the result + """ + package_name = f"electron_{major_version}-bin" + print(f"Updating {package_name}") + + electron_bin_info = get_electron_bin_info(major_version) + (_major_version, _version, repo) = electron_bin_info + + old_info = load_info_json(BINARY_INFO_JSON) + new_info = repo.get_hashes(major_version) + + out = old_info | new_info + + save_info_json(BINARY_INFO_JSON, out) + + old_version = ( + old_info[major_version]["version"] if major_version in old_info else None + ) + new_version = new_info[major_version]["version"] + if old_version == new_version: + print(f"{package_name} is up-to-date") + elif commit: + commit_result(package_name, old_version, new_version, BINARY_INFO_JSON) + + +def update_source(major_version: str, commit: bool) -> None: + """Update a given electron-source release + + Args: + major_version: The major version number, e.g. '27' + commit: Whether the updater should commit the result + """ + package_name = f"electron-source.electron_{major_version}" + print(f"Updating electron-source.electron_{major_version}") + + old_info = load_info_json(SOURCE_INFO_JSON) + old_version = ( + old_info[str(major_version)]["version"] + if str(major_version) in old_info + else None + ) + + electron_source_info = get_electron_info(major_version) + new_info = get_update(electron_source_info) + out = old_info | {new_info[0]: new_info[1]} + + save_info_json(SOURCE_INFO_JSON, out) + + new_version = new_info[1]["version"] + if old_version == new_version: + print(f"{package_name} is up-to-date") + elif commit: + commit_result(package_name, old_version, new_version, SOURCE_INFO_JSON) + + +def non_eol_releases(releases: Iterable[int]) -> Iterable[int]: + """Returns a list of releases that have not reached end-of-life yet.""" + return tuple(filter(lambda x: x in supported_version_range(), releases)) + + +def update_all_source(commit: bool) -> None: + """Update all eletron-source releases at once + + Args: + commit: Whether to commit the result + """ + old_info = load_info_json(SOURCE_INFO_JSON) + + filtered_releases = non_eol_releases(tuple(map(lambda x: int(x), old_info.keys()))) + + # This might take some time + repos = Parallel(n_jobs=2, require="sharedmem")( + delayed(get_electron_info)(major_version) for major_version in filtered_releases + ) + new_info = { + n[0]: n[1] + for n in Parallel(n_jobs=2, require="sharedmem")( + delayed(get_update)(repo) for repo in repos + ) + } + + if commit: + for major_version in filtered_releases: + # Since the sources have been fetched at this point already, + # fetching them again will be much faster. + update_source(str(major_version), commit) + else: + out = old_info | {new_info[0]: new_info[1]} + save_info_json(SOURCE_INFO_JSON, out) + + +def parse_cve_numbers(tag_name: str) -> Iterable[str]: + """Returns mentioned CVE numbers from a given release tag""" + cve_pattern = r"CVE-\d{4}-\d+" + url = f"https://api.github.com/repos/electron/electron/releases/tags/{tag_name}" + headers = { + "Accept": "application/vnd.github+json", + "X-GitHub-Api-Version": "2022-11-28", + } + request = urllib.request.Request(url=url, headers=headers) + release_note = "" + try: + with urlopen(request) as response: + release_note = json.loads(response.read().decode("utf-8"))["body"] + except: + print( + f"WARN: Fetching release note for {tag_name} from GitHub failed!", + file=sys.stderr, + ) + + return sorted(re.findall(cve_pattern, release_note)) + + +def commit_result( + package_name: str, old_version: Optional[str], new_version: str, path: str +) -> None: + """Creates a git commit with a short description of the change + + Args: + package_name: The package name, e.g. `electron-source.electron-{major_version}` + or `electron_{major_version}-bin` + + old_version: Version number before the update. + Can be left empty when initializing a new release. + + new_version: Version number after the update. + + path: Path to the lockfile to be committed + """ + assert ( + isinstance(package_name, str) and len(package_name) > 0 + ), "Argument `package_name` cannot be empty" + assert ( + isinstance(new_version, str) and len(new_version) > 0 + ), "Argument `new_version` cannot be empty" + + if old_version != new_version: + major_version = new_version.split(".")[0] + cve_fixes_text = "\n".join( + list( + map(lambda cve: f"- Fixes {cve}", parse_cve_numbers(f"v{new_version}")) + ) + ) + init_msg = f"init at {new_version}" + update_msg = f"{old_version} -> {new_version}" + diff = f"- Diff: https://github.com/electron/electron/compare/refs/tags/v{old_version}...v{new_version}\n" if old_version != None else "" + commit_message = f"""{package_name}: {update_msg if old_version != None else init_msg} + +- Changelog: https://github.com/electron/electron/releases/tag/v{new_version} +{diff}{cve_fixes_text} +""" + subprocess.run( + [ + "git", + "add", + path, + ] + ) + subprocess.run( + [ + "git", + "commit", + "-m", + commit_message, + ] + ) + + +@click.group() +def cli() -> None: + """A script for updating electron-bin and electron-source hashes""" + pass + + +@cli.command( + "eval", help="Print the necessary sources to fetch for a given major release" +) +@click.option("--version", help="The major version, e.g. '23'") +def eval(version): + (_, _, repo) = electron_repo = get_electron_info(version) + tree = repo.flatten("src/electron") + print(json.dumps(tree, indent=JSON_INDENT, default=vars, sort_keys=True)) + + +@cli.command("update", help="Update a single major release") +@click.option("-v", "--version", help="The major version, e.g. '23'") +@click.option( + "-b", + "--bin-only", + is_flag=True, + default=False, + help="Only update electron-bin packages", +) +@click.option( + "-s", + "--source-only", + is_flag=True, + default=False, + help="Only update electron-source packages", +) +@click.option("-c", "--commit", is_flag=True, default=False, help="Commit the result") +def update(version: str, bin_only: bool, source_only: bool, commit: bool) -> None: + assert isinstance(version, str) and len(version) > 0, "version must be non-empty" + + if bin_only and source_only: + print( + "Error: Omit --bin-only and --source-only if you want to update both source and binary packages.", + file=sys.stderr, + ) + sys.exit(1) + + elif bin_only: + update_bin(version, commit) + + elif source_only: + update_source(version, commit) + + else: + update_bin(version, commit) + update_source(version, commit) + + +@cli.command("update-all", help="Update all releases at once") +@click.option( + "-b", + "--bin-only", + is_flag=True, + default=False, + help="Only update electron-bin packages", +) +@click.option( + "-s", + "--source-only", + is_flag=True, + default=False, + help="Only update electron-source packages", +) +@click.option("-c", "--commit", is_flag=True, default=False, help="Commit the result") +def update_all(bin_only: bool, source_only: bool, commit: bool) -> None: + # Filter out releases that have reached end-of-life + filtered_bin_info = dict( + filter( + lambda entry: int(entry[0]) in supported_version_range(), + load_info_json(BINARY_INFO_JSON).items(), + ) + ) + + if bin_only and source_only: + print( + "Error: omit --bin-only and --source-only if you want to update both source and binary packages.", + file=sys.stderr, + ) + sys.exit(1) + + elif bin_only: + for major_version, _ in filtered_bin_info.items(): + update_bin(major_version, commit) + + elif source_only: + update_all_source(commit) + + else: + for major_version, _ in filtered_bin_info.items(): + update_bin(major_version, commit) + + update_all_source(commit) + if __name__ == "__main__": cli() |