AnotherAURHelper/update.py
Stephen Seo ed1b49d28a Specify "CCACHE_NOHASHDIR" when using ccache
AUR packages tend to be built with the package's name and version as
part of the current-working-directory during building. Specifying
CCACHE_NOHASHDIR=1 will cause ccache to ignore the
current-working-directory when hashing for ccache-cached objects. This
should hopefully fix issues where new package versions tend to cause
cache-misses on build.
2024-11-11 15:55:47 +09:00

2755 lines
92 KiB
Python
Executable file

#!/usr/bin/env python3
import os
import stat
import sys
import argparse
import subprocess
import re
import atexit
import glob
import toml
import datetime
import time
import shutil
import getpass
import tempfile
import threading
from pathlib import Path
from typing import Any, Union
import signal
import pwd
# SCRIPT_DIR = os.path.dirname(os.path.realpath(__file__))
SUDO_PROC = False
AUR_GIT_REPO_PATH = "https://aur.archlinux.org"
AUR_GIT_REPO_PATH_TEMPLATE = AUR_GIT_REPO_PATH + "/{}.git"
GLOBAL_LOG_FILE = "log.txt"
DEFAULT_EDITOR = "/usr/bin/nano"
IS_DIGIT_REGEX = re.compile("^[0-9]+$")
STRFTIME_FORMAT = "%Y-%m-%dT%H:%M:%SZ"
STRFTIME_LOCAL_FORMAT = "%Y-%m-%dT%H:%M:%S"
PKG_STATE = None
OTHER_STATE = None
DUMMY_PKGBUILD = """pkgname=dummy_pkg
pkgver=1.0
pkgrel=1
pkgdesc="dummy pkg"
arch=(any)
url="https://example.com"
license=('MIT')
source=()
prepare() {
echo prepare
}
build() {
echo build
}
check() {
echo check
}
package() {
echo package
}
"""
class ArchPkgVersion:
"""Holds a version (typically of an ArchLinux package) for comparison."""
def __init__(self, version_str: str):
self.versions = []
self.pkgver = 0
end_dash_idx = version_str.rfind("-")
if end_dash_idx != -1 and end_dash_idx + 1 < len(version_str):
try:
self.pkgver = int(version_str[end_dash_idx + 1 :])
except ValueError:
self.pkgver = version_str[end_dash_idx + 1 :]
version_str = version_str[:end_dash_idx]
for sub in version_str.split("."):
if IS_DIGIT_REGEX.match(sub) is not None:
self.versions.append(int(sub))
else:
subversion = []
string = None
integer = None
for char in sub:
if IS_DIGIT_REGEX.match(char) is not None:
if string is not None:
subversion.append(string)
string = None
if integer is None:
integer = int(char)
else:
integer = integer * 10 + int(char)
else:
if integer is not None:
subversion.append(integer)
integer = None
if string is None:
string = char
else:
string = string + char
if string is not None:
subversion.append(string)
string = None
if integer is not None:
subversion.append(integer)
integer = None
self.versions.append(tuple(subversion))
self.versions = tuple(self.versions)
def compare_with(self, other_self: "ArchPkgVersion"):
"""Returns -1 if self is less than other_self, 0 if they are equal, and
1 if self is greater than other_self."""
self_count = len(self.versions)
other_count = len(other_self.versions)
if other_count < self_count:
count = other_count
else:
count = self_count
for i in range(count):
if type(self.versions[i]) is tuple:
if type(other_self.versions[i]) is tuple:
self_subcount = len(self.versions[i])
other_subcount = len(other_self.versions[i])
if other_subcount < self_subcount:
subcount = other_subcount
else:
subcount = self_subcount
for j in range(subcount):
try:
if self.versions[i][j] < other_self.versions[i][j]:
return -1
elif (
self.versions[i][j] > other_self.versions[i][j]
):
return 1
except TypeError:
if str(self.versions[i][j]) < str(
other_self.versions[i][j]
):
return -1
elif str(self.versions[i][j]) > str(
other_self.versions[i][j]
):
return 1
if self_subcount < other_subcount:
return -1
elif self_subcount > other_subcount:
return 1
else:
# self is tuple but other is not
return 1
elif type(other_self.versions[i]) is tuple:
# other is tuple but self is not
return -1
else:
try:
if self.versions[i] < other_self.versions[i]:
return -1
elif self.versions[i] > other_self.versions[i]:
return 1
except TypeError:
if str(self.versions[i]) < str(other_self.versions[i]):
return -1
elif str(self.versions[i]) > str(other_self.versions[i]):
return 1
if self_count < other_count:
return -1
elif self_count > other_count:
return 1
else:
try:
if self.pkgver < other_self.pkgver:
return -1
elif self.pkgver > other_self.pkgver:
return 1
else:
return 0
except TypeError:
if str(self.pkgver) < str(other_self.pkgver):
return -1
elif str(self.pkgver) > str(other_self.pkgver):
return 1
else:
return 0
def __eq__(self, other: Any):
if isinstance(other, ArchPkgVersion):
return self.compare_with(other) == 0
else:
return False
def __ne__(self, other: Any):
if isinstance(other, ArchPkgVersion):
return self.compare_with(other) != 0
else:
return False
def __lt__(self, other: Any):
if isinstance(other, ArchPkgVersion):
return self.compare_with(other) < 0
else:
return False
def __le__(self, other: Any):
if isinstance(other, ArchPkgVersion):
return self.compare_with(other) <= 0
else:
return False
def __gt__(self, other: Any):
if isinstance(other, ArchPkgVersion):
return self.compare_with(other) > 0
else:
return False
def __ge__(self, other: Any):
if isinstance(other, ArchPkgVersion):
return self.compare_with(other) >= 0
else:
return False
def __str__(self):
self_str = ""
for idx in range(len(self.versions)):
if type(self.versions[idx]) is tuple:
for sub in self.versions[idx]:
self_str += str(sub)
else:
self_str += str(self.versions[idx])
if idx + 1 < len(self.versions):
self_str += "."
self_str += "-" + str(self.pkgver)
return self_str
def timedelta_to_offset_string(timed: datetime.timedelta) -> str:
"""Returns a timedelta string in the format "+HH:MM" or "-HH:MM"."""
seconds = timed.days * 24 * 60 * 60 + timed.seconds
minutes_offset = int(seconds / 60)
hours_offset = int(minutes_offset / 60)
minutes_offset = abs(minutes_offset - hours_offset * 60)
return f"{hours_offset:+03d}:{minutes_offset:02d}"
def get_datetime_timezone_now(other_state) -> str:
"""Returns a datetime string compatible with RFC 3339 and ISO 8601.
If other_state["datetime_in_local_time"] is True, then the returned string
is in localtime."""
if other_state["datetime_in_local_time"]:
lt = datetime.datetime.now(datetime.timezone.utc).astimezone()
return lt.strftime(STRFTIME_LOCAL_FORMAT) + timedelta_to_offset_string(
lt.tzinfo.utcoffset(None)
)
else:
return datetime.datetime.now(datetime.timezone.utc).strftime(
STRFTIME_FORMAT
)
def log_print(*args, **kwargs):
"""Prints to stdout, then logs to GLOBAL_LOG_FILE."""
if (
"other_state" in kwargs
and "is_timed" in kwargs["other_state"]
and kwargs["other_state"]["is_timed"]
):
t = get_datetime_timezone_now(kwargs["other_state"])
print(t, end=" ")
with open(GLOBAL_LOG_FILE, "a", encoding="utf-8") as lf:
print(t, end=" ", file=lf)
if "other_state" in kwargs:
del kwargs["other_state"]
if "file" in kwargs:
kwargs["file"] = sys.stdout
print(*args, **kwargs)
with open(GLOBAL_LOG_FILE, "a", encoding="utf-8") as lf:
kwargs["file"] = lf
print(*args, **kwargs)
def ensure_pkg_dir_exists(
pkg: str,
pkg_state: dict[str, Any],
other_state: dict[str, Any],
):
"""Ensures that an AUR-pkg-dir exists, returning False on failure.
If no such directory exists, this script attempts to clone it from the AUR.
True is returned on successful cloning.
If a file exists with the same name, returns False.
If a directory exists with the same name, returns True.
If "repo_path" is specified and the directory doesn't exist, returns False.
If "NO_REPO" is specified as "repo_path" and the directory doesn't exist,
returns False.
"""
log_print(
'Checking that dir for "{}" exists...'.format(pkg),
other_state=other_state,
)
pkgdir = os.path.join(other_state["clones_dir"], pkg)
if os.path.isdir(pkgdir):
log_print('Dir for "{}" exists.'.format(pkg), other_state=other_state)
return True
elif os.path.exists(pkgdir):
log_print(
'"{}" exists but is not a dir'.format(pkgdir),
other_state=other_state,
)
return False
elif "repo_path" not in pkg_state[pkg]:
pkg_state[pkg]["repo_path"] = AUR_GIT_REPO_PATH_TEMPLATE.format(pkg)
try:
subprocess.run(
(
"/usr/bin/env",
"git",
"clone",
pkg_state[pkg]["repo_path"],
pkgdir,
),
check=True,
)
except subprocess.CalledProcessError:
log_print(
'ERROR: Failed to git clone "{}" (tried repo path "{}")'.format(
pkgdir, pkg_state[pkg]["repo_path"]
),
other_state=other_state,
)
return False
log_print('Created dir for "{}".'.format(pkg), other_state=other_state)
return True
elif pkg_state[pkg]["repo_path"] == "NO_REPO":
log_print(
'"{}" does not exist, but NO_REPO specified for repo_path',
other_state=other_state,
)
return False
return False
def update_pkg_dir(
pkg: str,
pkg_state: dict[str, Any],
other_state: dict[str, Any],
):
"""Updates the pkg by invoking "git pull".
If "git pull" failes, it is retried after invoking "git restore .".
If the local working directory fails to update via "git pull" (or if some
other handling of the local repository fails), then this function returns
(False, False).
(True, True) is returned if "skip_branch_up_to_date" is True for the package
and the local repository is up to date.
(True, False) is returned by default (successful "git pull"; regardless of
if an update was fetched).
"""
log_print(
'Making sure pkg dir for "{}" is up to date...'.format(pkg),
other_state=other_state,
)
pkgdir = os.path.join(other_state["clones_dir"], pkg)
# fetch all
try:
subprocess.run(
("/usr/bin/env", "git", "fetch", "-p", "--all"),
check=True,
cwd=pkgdir,
)
except subprocess.CalledProcessError:
log_print(
'ERROR: Failed to update pkg dir of "{}" (fetching).'.format(pkg),
other_state=other_state,
)
return False, False
# get remotes
remotes = []
try:
result = subprocess.run(
("/usr/bin/env", "git", "remote"),
check=True,
cwd=pkgdir,
capture_output=True,
encoding="UTF-8",
)
remotes = result.stdout.split(sep="\n")
except subprocess.CalledProcessError:
log_print(
'ERROR: Failed to update pkg dir of "{}" (getting remotes).'.format(
pkg
),
other_state=other_state,
)
return False, False
remotes = list(filter(lambda s: len(s) > 0, remotes))
if len(remotes) == 0:
log_print(
'ERROR: Failed to update pkg dir of "{}" (getting remotes).'.format(
pkg
),
other_state=other_state,
)
return False, False
# get remote that current branch is tracking
selected_remote = None
try:
result = subprocess.run(
("/usr/bin/env", "git", "status", "-sb", "--porcelain"),
check=True,
cwd=pkgdir,
capture_output=True,
encoding="UTF-8",
)
result_lines = result.stdout.split(sep="\n")
for matching_line in filter(lambda s: s.startswith("##"), result_lines):
for remote in map(lambda r: r.strip(), remotes):
if matching_line.find(remote) != -1:
selected_remote = remote
break
except subprocess.CalledProcessError:
log_print(
f'ERROR: Failed to update pkg dir of "{pkg}" (getting branch\'s remote).',
other_state=other_state,
)
return False, False
if selected_remote is None or not isinstance(selected_remote, str):
log_print(
f'ERROR: Failed to update pkg dir of "{pkg}" (getting branch\'s remote).',
other_state=other_state,
)
return False, False
# get hash of current branch
current_branch_hash = None
try:
result = subprocess.run(
("/usr/bin/env", "git", "log", "-1", "--format=format:%H"),
check=True,
cwd=pkgdir,
capture_output=True,
encoding="UTF-8",
)
current_branch_hash = result.stdout.strip()
except subprocess.CalledProcessError:
log_print(
f'ERROR: Failed to update pkg dir of "{pkg}" (getting current branch\'s hash).',
other_state=other_state,
)
return False, False
if current_branch_hash is None or not isinstance(current_branch_hash, str):
log_print(
f'ERROR: Failed to update pkg dir of "{pkg}" (getting current branch\'s hash).',
other_state=other_state,
)
return False, False
# get hash of remote branch
remote_branch_hash = None
try:
result = subprocess.run(
(
"/usr/bin/env",
"git",
"log",
"-1",
"--format=format:%H",
selected_remote,
),
check=True,
cwd=pkgdir,
capture_output=True,
encoding="UTF-8",
)
remote_branch_hash = result.stdout.strip()
except subprocess.CalledProcessError:
log_print(
f'ERROR: Failed to update pkg dir of "{pkg}" (getting remote branch\'s hash).',
other_state=other_state,
)
return False, False
if remote_branch_hash is None or not isinstance(remote_branch_hash, str):
log_print(
f'ERROR: Failed to update pkg dir of "{pkg}" (getting remote branch\'s hash).',
other_state=other_state,
)
return False, False
# update current branch if not same commit
if current_branch_hash != remote_branch_hash:
try:
subprocess.run(
("/usr/bin/env", "git", "pull"), check=True, cwd=pkgdir
)
except subprocess.CalledProcessError:
try:
subprocess.run(
("/usr/bin/env", "git", "restore", "."),
check=True,
cwd=pkgdir,
)
subprocess.run(
("/usr/bin/env", "git", "pull"),
check=True,
cwd=pkgdir,
)
except subprocess.CalledProcessError:
log_print(
'ERROR: Failed to update pkg dir of "{}".'.format(pkg),
other_state=other_state,
)
return False, False
elif pkg_state[pkg]["skip_branch_up_to_date"]:
log_print(f'"{pkg}" is up to date', other_state=other_state)
return True, True
log_print('Updated pkg dir for "{}"'.format(pkg), other_state=other_state)
return True, False
def check_pkg_build(
pkg: str,
pkg_state: dict[str, Any],
other_state: dict[str, Any],
editor: str,
):
"""Opens the PKGBUILD in the editor, then prompts the user for an action.
Returns "ok", "not_ok", "abort", or "force_build"."""
pkgdir = os.path.join(other_state["clones_dir"], pkg)
log_print(
'Checking PKGBUILD for "{}"...'.format(pkg), other_state=other_state
)
try:
subprocess.run(
("/usr/bin/env", editor, "PKGBUILD"), check=True, cwd=pkgdir
)
except subprocess.CalledProcessError:
log_print(
'ERROR: Failed checking PKGBUILD for "{}"'.format(pkg),
other_state=other_state,
)
return "abort"
while True:
log_print(
"PKGBUILD okay? [Y/n/c(heck again)/a(bort)/f(orce build)/b(ack)]",
other_state=other_state,
)
user_input = sys.stdin.buffer.readline().decode().strip().lower()
if user_input == "y" or len(user_input) == 0:
log_print("User decided PKGBUILD is ok", other_state=other_state)
return "ok"
elif user_input == "n":
log_print(
"User decided PKGBUILD is not ok", other_state=other_state
)
return "not_ok"
elif user_input == "c":
log_print("User will check PKGBUILD again", other_state=other_state)
return check_pkg_build(pkg, pkg_state, other_state, editor)
elif user_input == "a":
return "abort"
elif user_input == "f":
return "force_build"
elif user_input == "b":
return "back"
else:
log_print(
"ERROR: User gave invalid input...", other_state=other_state
)
continue
def check_pkg_version(
pkg: str,
pkg_state: dict[str, Any],
repo: str,
force_check_srcinfo: bool,
other_state: dict[str, Any],
):
"""Gets the installed version and pkg version and checks them.
Returns "fail" (on failure), "install" (pkg is newer), or "done"
(installed pkg is up to date)."""
status, current_epoch, current_version = get_pkg_current_version(
pkg, pkg_state, repo, other_state
)
if status != "fetched":
return status
elif current_version is None:
log_print(
'ERROR: Failed to get version from package "{}".'.format(
pkg_state[pkg]["pkg_name"]
),
other_state=other_state,
)
return "fail"
log_print(
'Got version "{}:{}" for installed pkg "{}"'.format(
current_epoch if current_epoch is not None else "0",
current_version,
pkg_state[pkg]["pkg_name"],
),
other_state=other_state,
)
return get_srcinfo_check_result(
current_epoch,
current_version,
pkg,
force_check_srcinfo,
pkg_state,
other_state,
)
def get_srcinfo_version(pkg: str, other_state: dict[str, Any]):
"""Parses .SRCINFO for verison information.
Returns (success_bool, pkgepoch, pkgver, pkgrel)
When "success_bool" is False, all other values are None.
Otherwise, all other values are str (or None if not found)."""
if not os.path.exists(
os.path.join(other_state["clones_dir"], pkg, ".SRCINFO")
):
log_print(
f'ERROR: .SRCINFO does not exist for pkg "{pkg}"',
other_state=other_state,
)
return False, None, None, None
pkgver_reprog = re.compile("^\\s*pkgver\\s*=\\s*([a-zA-Z0-9._+-]+)\\s*$")
pkgrel_reprog = re.compile("^\\s*pkgrel\\s*=\\s*([0-9.]+)\\s*$")
pkgepoch_reprog = re.compile("^\\s*epoch\\s*=\\s*([0-9]+)\\s*$")
pkgver = None
pkgrel = None
pkgepoch = None
with open(
os.path.join(other_state["clones_dir"], pkg, ".SRCINFO"),
encoding="UTF-8",
) as fo:
line = fo.readline()
while len(line) > 0:
pkgver_result = pkgver_reprog.match(line)
pkgrel_result = pkgrel_reprog.match(line)
pkgepoch_result = pkgepoch_reprog.match(line)
if pkgver_result:
pkgver = pkgver_result.group(1)
elif pkgrel_result:
pkgrel = pkgrel_result.group(1)
elif pkgepoch_result:
pkgepoch = pkgepoch_result.group(1)
line = fo.readline()
return True, pkgepoch, pkgver, pkgrel
def get_pkgbuild_version(
pkg: str,
force_check_srcinfo: bool,
pkg_state: dict[str, Any],
other_state: dict[str, Any],
):
"""Gets the version of the pkg from .SRCINFO or PKGBUILD.
Returns (success, epoch, version, release).
If "success" is False, then all other values are None.
Otherwise, version and release should be a str type, but epoch may be
None."""
pkgdir = os.path.join(other_state["clones_dir"], pkg)
log_print(f'Getting version of "{pkg}"...', other_state=other_state)
while True and not force_check_srcinfo:
log_print(
"Use .SRCINFO or directly parse PKGBUILD?", other_state=other_state
)
user_input = input("1 for .SRCINFO, 2 for PKGBUILD > ")
if user_input == "1" or user_input == "2":
break
# TODO support split packages
if force_check_srcinfo or user_input == "1":
srcinfo_fetch_success, pkgepoch, pkgver, pkgrel = get_srcinfo_version(
pkg, other_state
)
if not srcinfo_fetch_success:
log_print(
"ERROR: Failed to get pkg info from .SRCINFO",
other_state=other_state,
)
return False, None, None, None
elif user_input == "2":
try:
log_print(
'Running "makechrootpkg ... --nobuild" to ensure pkgver in PKGBUILD is updated...',
other_state=other_state,
)
# Ensure ccache isn't enabled for this check.
if other_state["tmpfs"]:
cleanup_ccache(other_state["tmpfs_chroot"])
else:
cleanup_ccache(other_state["chroot"])
command_list = [
"/usr/bin/env",
"makechrootpkg",
"-c",
"-r",
(
other_state["tmpfs_chroot"]
if other_state["tmpfs"]
else other_state["chroot"]
),
]
post_command_list = ["--", "-s", "-r", "-c", "--nobuild"]
if "link_cargo_registry" in pkg_state[pkg]:
command_list.insert(2, "-d")
command_list.insert(
3,
f'{os.environ["HOME"]}/.cargo/registry:/build/.cargo/registry',
)
command_list.insert(4, "-d")
command_list.insert(
5,
f'{os.environ["HOME"]}/.cargo/git:/build/.cargo/git',
)
if len(pkg_state[pkg]["other_deps"]) != 0:
prefetch_result = prefetch_dependencies(
pkg_state[pkg]["other_deps"], other_state
)
if prefetch_result != "fetched":
log_print(
"ERROR: Failed to prefetch deps {}".format(
pkg_state[pkg]["other_deps"]
),
other_state=other_state,
)
return False, None, None, None
for dep in pkg_state[pkg]["other_deps"]:
dep_fullpath = get_latest_pkg(dep, "/var/cache/pacman/pkg")
if not dep_fullpath:
log_print(
'ERROR: Failed to get dep "{}"'.format(dep),
other_state=other_state,
)
return False, None, None, None
command_list.insert(2, "-I")
command_list.insert(3, dep_fullpath)
for aur_dep in pkg_state[pkg]["aur_deps"]:
aur_dep_fullpath = get_latest_pkg(
aur_dep, other_state["pkg_out_dir"]
)
if not aur_dep_fullpath:
log_print(
'ERROR: Failed to get aur_dep "{}"'.format(aur_dep),
other_state=other_state,
)
return False, None, None, None
command_list.insert(2, "-I")
command_list.insert(3, aur_dep_fullpath)
subprocess.run(
command_list + post_command_list,
check=True,
cwd=pkgdir,
)
except subprocess.CalledProcessError:
log_print(
f'ERROR: Failed to run "makechrootpkg ... --nobuild" in "{pkg}".',
other_state=other_state,
)
if os.path.exists(os.path.join(pkgdir, "src")):
shutil.rmtree(os.path.join(pkgdir, "src"))
return False, None, None, None
if os.path.exists(os.path.join(pkgdir, "src")):
shutil.rmtree(os.path.join(pkgdir, "src"))
pkgepoch = None
pkgver = None
pkgrel = None
# Setup checking the PKGBUILD from within the chroot.
chroot_user_path = os.path.join(
(
other_state["tmpfs_chroot"]
if other_state["tmpfs"]
else other_state["chroot"]
),
other_state["USER"],
)
chroot_build_path = os.path.join(chroot_user_path, "build")
chroot_check_pkgbuild_path = os.path.join(chroot_build_path, "PKGBUILD")
chroot_check_sh_path = os.path.join(chroot_build_path, "check.sh")
try:
subprocess.run(
(
"/usr/bin/cp",
os.path.join(pkgdir, "PKGBUILD"),
chroot_check_pkgbuild_path,
),
check=True,
)
except subprocess.CalledProcessError:
log_print(
f'ERROR: Failed to check PKGBUILD (moving PKGBUILD to chroot) for "{pkg}"!',
other_state=other_state,
)
return False, None, None, None
check_pkgbuild_script = """#!/usr/bin/env bash
set -e
source "/build/PKGBUILD"
echo "pkgver=$pkgver"
echo "pkgrel=$pkgrel"
echo "epoch=$epoch"
"""
if not create_executable_script(
chroot_check_sh_path, check_pkgbuild_script
):
log_print(
f'ERROR: Failed to check PKGBUILD (check PKGBUILD setup) for "{pkg}"!',
other_state=other_state,
)
return False, None, None, None
pkgbuild_output = str()
try:
pkgbuild_output = subprocess.run(
(
"/usr/bin/env",
"sudo",
"arch-nspawn",
chroot_user_path,
"/build/check.sh",
),
check=True,
capture_output=True,
text=True,
)
except subprocess.CalledProcessError:
log_print(
f'ERROR: Failed to check PKGBUILD (checking PKGBUILD) for "{pkg}"!',
other_state=other_state,
)
return False, None, None, None
output_ver_re = re.compile(
"^pkgver=([a-zA-Z0-9._+-]+)\\s*$", flags=re.M
)
output_rel_re = re.compile("^pkgrel=([0-9.]+)\\s*$", flags=re.M)
output_epoch_re = re.compile("^epoch=([0-9]+)\\s*$", flags=re.M)
match = output_ver_re.search(pkgbuild_output.stdout)
if match:
pkgver = match.group(1)
match = output_rel_re.search(pkgbuild_output.stdout)
if match:
pkgrel = match.group(1)
match = output_epoch_re.search(pkgbuild_output.stdout)
if match:
pkgepoch = match.group(1)
else:
log_print("ERROR: Unreachable code", other_state=other_state)
return False, None, None, None
if pkgver is not None and pkgrel is not None:
return True, pkgepoch, pkgver, pkgrel
else:
log_print(
'ERROR: Failed to get PKGBUILD version of "{}".'.format(pkg),
other_state=other_state,
)
return False, None, None, None
def get_srcinfo_check_result(
current_epoch: Union[str, None],
current_version: str,
pkg: str,
force_check_srcinfo: bool,
pkg_state: dict[str, Any],
other_state: dict[str, Any],
):
"""Checks the version of the pkg against the currently installed version.
Returns "install" if the version of the pkg is newer.
Otherwise returns "done" if the version is not newer.
Returns "fail" on error."""
ver_success, pkgepoch, pkgver, pkgrel = get_pkgbuild_version(
pkg, force_check_srcinfo, pkg_state, other_state
)
if ver_success:
if current_epoch is None and pkgepoch is not None:
log_print(
'Current installed version of "{}" is out of date (missing epoch).'.format(
pkg_state[pkg]["pkg_name"]
),
other_state=other_state,
)
return "install"
elif current_epoch is not None and pkgepoch is None:
log_print(
'Current installed version of "{}" is up to date (has epoch).'.format(
pkg_state[pkg]["pkg_name"]
),
other_state=other_state,
)
return "done"
elif (
current_epoch is not None
and pkgepoch is not None
and int(current_epoch) < int(pkgepoch)
):
log_print(
'Current installed version of "{}" is out of date (older epoch).'.format(
pkg_state[pkg]["pkg_name"]
),
other_state=other_state,
)
return "install"
elif (
pkgver is not None
and pkgrel is not None
and ArchPkgVersion(current_version)
< ArchPkgVersion(pkgver + "-" + pkgrel)
):
log_print(
'Current installed version of "{}" is out of date (older version).'.format(
pkg_state[pkg]["pkg_name"]
),
other_state=other_state,
)
return "install"
else:
log_print(
'Current installed version of "{}" is up to date.'.format(
pkg_state[pkg]["pkg_name"]
),
other_state=other_state,
)
return "done"
else:
log_print(
'ERROR: Failed to get pkg_version of "{}"'.format(
pkg_state[pkg]["pkg_name"]
),
other_state=other_state,
)
return "fail"
def get_pkg_current_version(
pkg: str, pkg_state: dict[str, Any], repo: str, other_state: dict[str, Any]
):
"""Fetches the version info and returns status of fetching and the version.
Returns (status, epoch, version)
"status" may be one of:
"fail", "install", "fetched"
"epoch" may be None.
"version" must be a str if "status" is "fetched". Otherwise, it is None if
"status" is "fail" or "install".
"""
log_print(
'Checking version of installed pkg "{}"...'.format(
pkg_state[pkg]["pkg_name"]
),
other_state=other_state,
)
current_epoch = None
current_version = None
try:
result = subprocess.run(
(
"/usr/bin/env",
"bash",
"-c",
"tar -tf {} | grep '{}.*/$'".format(
repo, pkg_state[pkg]["pkg_name"]
),
),
check=True,
capture_output=True,
encoding="UTF-8",
)
reprog = re.compile(
"^{}-(?P<epoch>[0-9]+:)?(?P<version>[^-/: ]*-[0-9]+)/$".format(
pkg_state[pkg]["pkg_name"]
),
flags=re.MULTILINE,
)
reresult = reprog.search(result.stdout)
if reresult:
result_dict = reresult.groupdict()
if not result_dict["epoch"] is None:
current_epoch = result_dict["epoch"][:-1]
if not result_dict["version"] is None:
current_version = result_dict["version"]
else:
log_print(
"ERROR: Failed to get current version from repo for package {}".format(
pkg_state[pkg]["pkg_name"]
),
other_state=other_state,
)
return "fail", None, None
except subprocess.CalledProcessError:
log_print(
"Package not found, assuming building first time.",
other_state=other_state,
)
return "install", None, None
return "fetched", current_epoch, current_version
def get_sudo_privileges(other_state: dict[str, Any]):
"""Starts a bash loop that ensures sudo privileges are ready while this
script is active."""
global SUDO_PROC
if not SUDO_PROC:
log_print("sudo -v", other_state=other_state)
try:
subprocess.run(("/usr/bin/env", "sudo", "-v"), check=True)
except subprocess.CalledProcessError:
return False
SUDO_PROC = subprocess.Popen(
"while true; do sudo -v; sleep 2m; done", shell=True
)
atexit.register(cleanup_sudo, sudo_proc=SUDO_PROC)
return True
return True
def cleanup_sudo(sudo_proc):
"""Stops the bash loop keeping sudo privileges."""
sudo_proc.terminate()
def create_executable_script(dest_filename: str, script_contents: str):
"""Creates a script via use of sudo to be executed later.
This is currently used to set up sccache by placing custom commands in
"/usr/local/bin" for gcc and friends."""
tempf_name = "unknown"
with tempfile.NamedTemporaryFile(
mode="w", encoding="utf-8", delete=False
) as tempf:
print(script_contents, file=tempf)
tempf_name = tempf.name
try:
subprocess.run(
(
"/usr/bin/env",
"sudo",
"cp",
tempf_name,
dest_filename,
),
check=True,
)
subprocess.run(
(
"/usr/bin/env",
"sudo",
"chmod",
"a+rx",
dest_filename,
),
check=True,
)
subprocess.run(
(
"/usr/bin/env",
"rm",
"-f",
tempf_name,
),
check=True,
)
except subprocess.CalledProcessError:
log_print(
f'ERROR: Failed to create executable script "{dest_filename}"'
)
return False
return True
def setup_ccache(chroot: str):
"""Sets up the chroot for ccache."""
# set up ccache stuff
try:
subprocess.run(
(
"/usr/bin/env",
"sudo",
"sed",
"-i",
"/^BUILDENV=/s/!ccache/ccache/",
f"{chroot}/root/etc/makepkg.conf",
),
check=True,
)
except subprocess.CalledProcessError:
log_print(
"ERROR: Failed to enable ccache in makepkg.conf",
other_state=other_state,
)
sys.exit(1)
def cleanup_ccache(chroot: str):
"""Unsets up the chroot for ccache."""
# cleanup ccache stuff
try:
subprocess.run(
(
"/usr/bin/env",
"sudo",
"sed",
"-i",
"/^BUILDENV=/s/ ccache/ !ccache/",
f"{chroot}/root/etc/makepkg.conf",
),
check=True,
)
except subprocess.CalledProcessError:
log_print(
"ERROR: Failed to disable ccache in makepkg.conf",
other_state=other_state,
)
sys.exit(1)
def setup_sccache(chroot: str):
"""Sets up sccache for the chroot."""
sccache_script = """#!/usr/bin/env sh
export PATH=${PATH/:\\/usr\\/local\\/bin/}
/usr/bin/env sccache $(basename "$0") "$@"
"""
if (
not create_executable_script(
f"{chroot}/root/usr/local/bin/gcc", sccache_script
)
or not create_executable_script(
f"{chroot}/root/usr/local/bin/g++", sccache_script
)
or not create_executable_script(
f"{chroot}/root/usr/local/bin/cc", sccache_script
)
or not create_executable_script(
f"{chroot}/root/usr/local/bin/c++", sccache_script
)
or not create_executable_script(
f"{chroot}/root/usr/local/bin/cpp", sccache_script
)
or not create_executable_script(
f"{chroot}/root/usr/local/bin/clang", sccache_script
)
or not create_executable_script(
f"{chroot}/root/usr/local/bin/clang++", sccache_script
)
or not create_executable_script(
f"{chroot}/root/usr/local/bin/rustc", sccache_script
)
):
log_print(
"ERROR: Failed to set up sccache wrapper scripts",
other_state=other_state,
)
sys.exit(1)
def cleanup_sccache(chroot: str):
"""Unsets up sccache for the chroot."""
# cleanup sccache stuff
try:
subprocess.run(
(
"/usr/bin/env",
"sudo",
"rm",
"-f",
f"{chroot}/root/usr/local/bin/gcc",
f"{chroot}/root/usr/local/bin/g++",
f"{chroot}/root/usr/local/bin/cc",
f"{chroot}/root/usr/local/bin/c++",
f"{chroot}/root/usr/local/bin/cpp",
f"{chroot}/root/usr/local/bin/clang",
f"{chroot}/root/usr/local/bin/clang++",
f"{chroot}/root/usr/local/bin/rustc",
),
check=False,
)
except BaseException:
log_print(
"WARNING: Failed to cleanup sccache files", other_state=other_state
)
def handle_output_stream(
handle,
output_file,
other_state,
print_to_log=False,
ignore_output_file=False,
):
"""Reads lines from an input stream "handle" and writes them to
"output_file". Flags in "other_state" determine certain behaviors, such as
prepending a timestamp to each line, or the filesize-limit for the
"output_file"."""
log_count = 0
limit_reached = False
while True:
line = handle.readline()
if len(line) == 0:
break
if print_to_log:
if line[-1] == "\n":
log_print(line[0:-1], other_state=other_state)
else:
log_print(line, other_state=other_state)
if ignore_output_file:
continue
if not limit_reached:
if other_state["is_log_timed"]:
nowstring = get_datetime_timezone_now(other_state)
line = nowstring + " " + line
log_count += len(line)
if log_count > other_state["log_limit"]:
limit_reached = True
if other_state["error_on_limit"]:
output_file.write(
"\nERROR: Reached log_limit! No longer logging to file!\n"
)
output_file.flush()
log_print(
"ERROR: Reached log_limit! No longer logging to file!",
other_state=other_state,
)
handle.close()
break
else:
output_file.write(
"\nWARNING: Reached log_limit! No longer logging to file!\n"
)
output_file.flush()
log_print(
"WARNING: Reached log_limit! No longer logging to file!",
other_state=other_state,
)
else:
output_file.write(line)
output_file.flush()
def update_pkg_list(
pkgs: list[str],
pkg_state: dict[str, Any],
other_state: dict[str, Any],
signing_gpg_dir: str,
signing_gpg_key_fp: str,
signing_gpg_pass: str,
no_store: bool,
):
"""For each package to build: builds it, signs it, and moves it to
"pkg_out_dir"."""
atexit.register(build_print_pkg_info, pkgs, pkg_state, other_state)
if not get_sudo_privileges(other_state):
log_print(
"ERROR: Failed to get sudo privileges", other_state=other_state
)
pkg_state[pkg]["build_status"] = "get_sudo_fail"
sys.exit(1)
for pkg in pkgs:
if other_state["stop_building"]:
sys.exit(0)
pkgdir = os.path.join(other_state["clones_dir"], pkg)
if "ccache_dir" in pkg_state[pkg]:
cleanup_sccache(
other_state["tmpfs_chroot"]
if other_state["tmpfs"]
else other_state["chroot"]
)
setup_ccache(
other_state["tmpfs_chroot"]
if other_state["tmpfs"]
else other_state["chroot"]
)
else:
cleanup_ccache(
other_state["tmpfs_chroot"]
if other_state["tmpfs"]
else other_state["chroot"]
)
if (
"sccache_dir" in pkg_state[pkg]
and not pkg_state[pkg]["sccache_rust_only"]
):
setup_sccache(
other_state["tmpfs_chroot"]
if other_state["tmpfs"]
else other_state["chroot"]
)
else:
cleanup_sccache(
other_state["tmpfs_chroot"]
if other_state["tmpfs"]
else other_state["chroot"]
)
# check integrity
log_print(
f'Checking files of "{pkg}" before building it...',
other_state=other_state,
)
try:
subprocess.run(
("/usr/bin/env", "makepkg", "--verifysource"),
check=True,
cwd=pkgdir,
)
except:
log_print(
f'ERROR: Failed to verify pkg "{pkg}"', other_state=other_state
)
pkg_state[pkg]["build_status"] = "pkg_verify_fail"
continue
log_print(f'Building "{pkg}"...', other_state=other_state)
command_list = [
"/usr/bin/env",
"makechrootpkg",
"-c",
"-r",
(
other_state["tmpfs_chroot"]
if other_state["tmpfs"]
else other_state["chroot"]
),
]
post_command_list = [
"--",
"--syncdeps",
"--noconfirm",
"--log",
"--holdver",
]
failure = False
if "ccache_dir" in pkg_state[pkg]:
if not "ccache" in pkg_state[pkg]["other_deps"]:
pkg_state[pkg]["other_deps"].append("ccache")
elif "sccache_dir" in pkg_state[pkg]:
if not "sccache" in pkg_state[pkg]["other_deps"]:
pkg_state[pkg]["other_deps"].append("sccache")
if len(pkg_state[pkg]["other_deps"]) != 0:
prefetch_result = prefetch_dependencies(
pkg_state[pkg]["other_deps"], other_state
)
if prefetch_result != "fetched":
log_print(
"ERROR: Failed to prefetch deps {}".format(
pkg_state[pkg]["other_deps"]
),
other_state=other_state,
)
failure = True
pkg_state[pkg]["build_status"] = "get_dep_fail"
break
log_print(
"Successfully prefetched deps, continuing on to build...",
other_state=other_state,
)
for dep in pkg_state[pkg]["other_deps"]:
dep_fullpath = get_latest_pkg(dep, "/var/cache/pacman/pkg")
if not dep_fullpath:
log_print(
'ERROR: Failed to get dep "{}"'.format(dep),
other_state=other_state,
)
failure = True
pkg_state[pkg]["build_status"] = "get_dep_fail"
break
command_list.insert(2, "-I")
command_list.insert(3, dep_fullpath)
if failure:
continue
for aur_dep in pkg_state[pkg]["aur_deps"]:
aur_dep_fullpath = get_latest_pkg(
aur_dep, other_state["pkg_out_dir"]
)
if not aur_dep_fullpath:
log_print(
'ERROR: Failed to get aur_dep "{}"'.format(aur_dep),
other_state=other_state,
)
failure = True
pkg_state[pkg]["build_status"] = "get_aur_dep_fail"
break
command_list.insert(2, "-I")
command_list.insert(3, aur_dep_fullpath)
if failure:
continue
if "ccache_dir" in pkg_state[pkg]:
command_list.insert(2, "-d")
command_list.insert(3, f'{pkg_state[pkg]["ccache_dir"]}:/ccache')
post_command_list.insert(1, "CCACHE_DIR=/ccache")
post_command_list.insert(2, "CCACHE_NOHASHDIR=1")
elif "sccache_dir" in pkg_state[pkg]:
command_list.insert(2, "-d")
command_list.insert(3, f'{pkg_state[pkg]["sccache_dir"]}:/sccache')
post_command_list.insert(1, "SCCACHE_DIR=/sccache")
post_command_list.insert(
2, f'SCCACHE_CACHE_SIZE={pkg_state[pkg]["sccache_cache_size"]}'
)
post_command_list.insert(3, "RUSTC_WRAPPER=/usr/bin/sccache")
nowstring = get_datetime_timezone_now(other_state)
if "link_cargo_registry" in pkg_state[pkg]:
command_list.insert(2, "-d")
command_list.insert(
3,
f'{os.environ["HOME"]}/.cargo/registry:/build/.cargo/registry',
)
command_list.insert(4, "-d")
command_list.insert(
5,
f'{os.environ["HOME"]}/.cargo/git:/build/.cargo/git',
)
# log_print(f"Using command list: {command_list + post_command_list}", other_state=other_state) # DEBUG
with open(
os.path.join(
other_state["logs_dir"], "{}_stdout_{}".format(pkg, nowstring)
),
mode="w",
encoding="utf-8",
) as log_stdout, open(
os.path.join(
other_state["logs_dir"], "{}_stderr_{}".format(pkg, nowstring)
),
mode="w",
encoding="utf-8",
) as log_stderr:
try:
p1 = subprocess.Popen(
command_list + post_command_list,
cwd=pkgdir,
text=True,
stdout=subprocess.PIPE,
stderr=subprocess.PIPE,
)
tout = threading.Thread(
target=handle_output_stream,
args=[p1.stdout, log_stdout, other_state],
)
terr = threading.Thread(
target=handle_output_stream,
args=[p1.stderr, log_stderr, other_state],
)
tout.start()
terr.start()
p1.wait()
tout.join()
terr.join()
if p1.returncode is None:
raise RuntimeError("pOpen process didn't finish")
elif type(p1.returncode) is not int:
raise RuntimeError("pOpen process non-integer returncode")
elif p1.returncode != 0:
raise RuntimeError(
f"pOpen process non-zero return code {p1.returncode}"
)
except BaseException as e:
log_print(
'ERROR: Failed to build pkg "{}" in chroot: {}'.format(
pkg, e
),
other_state=other_state,
)
pkg_state[pkg]["build_status"] = "build_fail"
continue
if no_store:
pkg_state[pkg]["build_status"] = "success"
continue
pkg_list = glob.glob(
os.path.join(other_state["clones_dir"], pkg, "*.pkg.tar*")
)
log_print("Signing package...", other_state=other_state)
for gpkg in pkg_list:
try:
command_list = [
"gpg",
"--batch",
"--passphrase-fd",
"0",
"--pinentry-mode",
"loopback",
"--default-key",
signing_gpg_key_fp,
"--detach-sign",
gpkg,
]
subprocess.run(
command_list,
check=True,
cwd=os.path.join(other_state["clones_dir"], pkg),
input=signing_gpg_pass,
text=True,
env={"GNUPGHOME": signing_gpg_dir},
)
except subprocess.CalledProcessError:
log_print(
f'ERROR: Failed to sign pkg "{pkg}"',
other_state=other_state,
)
log_print("Adding built pkgs to repo...", other_state=other_state)
try:
command_list = ["repo-add", "--include-sigs", other_state["repo"]]
for gpkg in pkg_list:
command_list.append(gpkg)
p1 = subprocess.Popen(
command_list,
text=True,
stdout=subprocess.PIPE,
stderr=subprocess.PIPE,
)
tout = threading.Thread(
target=handle_output_stream,
args=[p1.stdout, None, other_state, True, True],
)
terr = threading.Thread(
target=handle_output_stream,
args=[p1.stderr, None, other_state, True, True],
)
tout.start()
terr.start()
p1.wait()
tout.join()
terr.join()
if p1.returncode is None:
raise RuntimeError("pOpen process didn't finish")
elif type(p1.returncode) is not int:
raise RuntimeError("pOpen process non-integer returncode")
elif p1.returncode != 0:
raise RuntimeError(
f"pOpen process non-zero return code {p1.returncode}"
)
except subprocess.CalledProcessError:
log_print(
'ERROR: Failed to add built pkg(s) "{}" to repo.'.format(pkg),
other_state=other_state,
)
pkg_state[pkg]["build_status"] = "add_fail"
continue
except RuntimeError as e:
log_print(
'ERROR: Failed to add built pkg(s) "{}" to repo ({}).'.format(
pkg, e
),
other_state=other_state,
)
pkg_state[pkg]["build_status"] = "add_fail"
continue
log_print(
f'Signing "{other_state["repo"]}"...', other_state=other_state
)
try:
subprocess.run(
(
"/usr/bin/rm",
"-f",
str(
os.path.join(
other_state["pkg_out_dir"],
f"{other_state['repo']}.sig",
)
),
)
)
subprocess.run(
(
"/usr/bin/env",
"gpg",
"--batch",
"--passphrase-fd",
"0",
"--pinentry-mode",
"loopback",
"--default-key",
signing_gpg_key_fp,
"--detach-sign",
str(
os.path.join(
other_state["pkg_out_dir"], other_state["repo"]
)
),
),
check=True,
input=signing_gpg_pass,
text=True,
env={"GNUPGHOME": signing_gpg_dir},
)
repo_sig_name = f"{other_state['repo']}.sig"
if repo_sig_name.rfind("/") != -1:
repo_sig_name = repo_sig_name.rsplit(sep="/", maxsplit=1)[1]
subprocess.run(
(
"/usr/bin/env",
"ln",
"-sf",
repo_sig_name,
str(
os.path.join(
other_state["pkg_out_dir"], other_state["repo"]
)
).removesuffix(".tar")
+ ".sig",
)
)
except subprocess.CalledProcessError:
log_print(
f'WARNING: Failed to sign "{other_state["repo"]}"',
other_state=other_state,
)
pkg_state[pkg]["build_status"] = "success"
log_print("Moving pkg to pkgs directory...", other_state=other_state)
for f in pkg_list:
log_print(f'Moving "{f}"...', other_state=other_state)
os.rename(
f, os.path.join(other_state["pkg_out_dir"], os.path.basename(f))
)
sig_name = f + ".sig"
if os.path.exists(sig_name):
log_print(f'Moving "{sig_name}"...', other_state=other_state)
os.rename(
sig_name,
os.path.join(
other_state["pkg_out_dir"], os.path.basename(sig_name)
),
)
def get_latest_pkg(pkg: str, cache_dir: str):
"""Gets the latest pkg from the specified "cache_dir" and return its
filename."""
globbed = glob.glob(os.path.join(cache_dir, pkg + "*"))
if len(globbed) > 0:
globbed.sort()
reprog = re.compile(
".*"
+ pkg
+ "-[0-9a-zA-Z.+_:]+-[0-9a-zA-Z.+_]+-(any|x86_64).pkg.tar.(xz|gz|zst)$"
)
result = list(filter(lambda x: reprog.match(x), globbed))
if len(result) == 0:
return None
else:
return result[-1]
else:
return None
def confirm_result(pkg: str, state_result: str, other_state: dict[str, Any]):
"""Prompts the user the action to take for a pkg after checking its
PKGBUILD.
Returns "continue", "recheck", "force_build", "skip", "back", or "abort"."""
while True:
log_print(
'Got "{}" for pkg "{}", action: [C(ontinue), r(echeck), f(orce build),\
s(kip), b(ack) a(abort)]'.format(
state_result, pkg
),
other_state=other_state,
)
user_input = sys.stdin.buffer.readline().decode().strip().lower()
if user_input == "c" or len(user_input) == 0:
return "continue"
elif user_input == "r":
return "recheck"
elif user_input == "f":
return "force_build"
elif user_input == "s":
return "skip"
elif user_input == "b":
return "back"
elif user_input == "a":
return "abort"
else:
log_print("Got invalid input", other_state=other_state)
continue
def print_state_info_and_get_update_list(
other_state: dict[str, Any], pkg_state: dict[str, Any]
):
"""Prints the current "checked" state of all pkgs in the config."""
to_update = []
log_print("package state:", other_state=other_state)
max_name_len = 1
for pkg_name in pkg_state.keys():
if len(pkg_name) + 1 > max_name_len:
max_name_len = len(pkg_name) + 1
for pkg_name, pkg_dict in pkg_state.items():
name_space = " " * (max_name_len - len(pkg_name))
if "state" in pkg_dict:
state_str = '"' + pkg_dict["state"] + '"'
if (
"print_state_SIGUSR1" in other_state
and type(other_state["print_state_SIGUSR1"]) is bool
and other_state["print_state_SIGUSR1"]
and "print_state_info_only_building_sigusr1" in other_state
and type(other_state["print_state_info_only_building_sigusr1"])
is bool
and other_state["print_state_info_only_building_sigusr1"]
):
if state_str == '"install"':
log_print(
f" {pkg_name}{name_space}: pre_state is {state_str: <13}, build_state is \"{pkg_dict['build_status']}\"",
other_state=other_state,
)
if pkg_dict["state"] == "install":
to_update.append(pkg_name)
else:
log_print(
f" {pkg_name}{name_space}: pre_state is {state_str: <13}, build_state is \"{pkg_dict['build_status']}\"",
other_state=other_state,
)
if pkg_dict["state"] == "install":
to_update.append(pkg_name)
else:
log_print(
f" {pkg_name}{name_space}: not reached",
other_state=other_state,
)
return to_update
def build_print_pkg_info(
pkgs: tuple[str, ...],
pkg_state: dict[str, Any],
other_state: dict[str, Any],
):
"""Prints the current "build" state of the given pkgs."""
max_name_len = 1
for pkg in pkgs:
if len(pkg) + 1 > max_name_len:
max_name_len = len(pkg) + 1
for pkg in pkgs:
name_space = " " * (max_name_len - len(pkg))
log_print(
f'"{pkg}"{name_space}status: {pkg_state[pkg]["build_status"]}',
other_state=other_state,
)
def test_gpg_passphrase(
signing_gpg_dir: str,
signing_key_fp: str,
passphrase: str,
other_state: dict[str, Any],
):
"""Checks if the given gpg passphrase works with the gpg signing key."""
local_share_dir = os.path.join(os.environ["HOME"], ".local", "share")
local_share_dir_path = Path(local_share_dir)
if not local_share_dir_path.exists():
local_share_dir_path.mkdir(parents=True)
with tempfile.NamedTemporaryFile(dir=local_share_dir) as tempnf:
tempnf.write(b"Test file content")
tempnf.flush()
try:
# Clear gpg password cache so that incorrect passwords don't pass.
subprocess.run(
("/usr/bin/env", "gpg-connect-agent", "reloadagent", "/bye"),
check=True,
stdout=subprocess.DEVNULL,
stderr=subprocess.DEVNULL,
env={"GNUPGHOME": signing_gpg_dir},
)
subprocess.run(
(
"/usr/bin/env",
"gpg",
"--batch",
"--passphrase-fd",
"0",
"--pinentry-mode",
"loopback",
"--default-key",
signing_key_fp,
"--detach-sign",
tempnf.name,
),
check=True,
input=passphrase,
text=True,
env={"GNUPGHOME": signing_gpg_dir},
)
os.remove(tempnf.name + ".sig")
except subprocess.CalledProcessError:
log_print(
"ERROR: Failed to sign test file with gpg",
other_state=other_state,
)
return False
log_print(
"Verified passphrase works by signing dummy test file",
other_state=other_state,
)
return True
def validate_and_verify_paths(other_state: dict[str, Any]):
"""Checks and validates/ensures that certain directories exist."""
if not os.path.exists(other_state["chroot"]):
log_print(
f"ERROR: chroot at \"{other_state['chroot']}\" does not exist",
other_state=other_state,
)
sys.exit(1)
log_print("Ensuring pkgs directory exists...", other_state=other_state)
if not os.path.exists(other_state["pkg_out_dir"]):
pkg_out_dir_path = Path(other_state["pkg_out_dir"])
pkg_out_dir_path.mkdir(parents=True)
if not os.path.exists(other_state["gpg_home"]):
log_print(
f"ERROR: checkingGPG at \"{other_state['gpg_home']}\" does not exist",
other_state=other_state,
)
sys.exit(1)
if "signing_gpg_dir" in other_state and not os.path.exists(
other_state["signing_gpg_dir"]
):
log_print(
f"ERROR: signingGPG at \"{other_state['signing_gpg_dir']}\" does not exist",
other_state=other_state,
)
sys.exit(1)
log_print("Ensuring logs directory exists...", other_state=other_state)
if other_state["logs_dir"] is None:
log_print(
'ERROR: "logs_dir" was not specified!', other_state=other_state
)
sys.exit(1)
if not os.path.exists(other_state["logs_dir"]):
logs_dir_path = Path(other_state["logs_dir"])
logs_dir_path.mkdir(parents=True)
log_print("Ensuring clones directory exists...", other_state=other_state)
if not os.path.exists(other_state["clones_dir"]):
clones_dir_path = Path(other_state["clones_dir"])
clones_dir_path.mkdir(parents=True)
def signal_handler(sig, frame):
"""Handle SIGINT and SIGUSR1."""
global OTHER_STATE, PKG_STATE
if OTHER_STATE is not None and PKG_STATE is not None:
OTHER_STATE["print_state_SIGUSR1"] = (
signal.Signals(sig) is signal.SIGUSR1
)
print_state_info_and_get_update_list(OTHER_STATE, PKG_STATE)
OTHER_STATE["print_state_SIGUSR1"] = False
if signal.Signals(sig) is not signal.SIGINT:
return
OTHER_STATE["stop_building"] = True
sys.exit(0)
if signal.Signals(sig) is not signal.SIGINT:
return
OTHER_STATE["stop_building"] = True
sys.exit(1)
def check_install_script(
pkg_state: dict[str, Any],
other_state: dict[str, Any],
pkg: str,
editor: str,
skip_prepare_chroot=False,
):
"""Returns "error", "does_not_exist", and "ok"."""
pkgdir = os.path.join(other_state["clones_dir"], pkg)
chroot_user_path = os.path.join(
(
other_state["tmpfs_chroot"]
if other_state["tmpfs"]
else other_state["chroot"]
),
other_state["USER"],
)
chroot_build_path = os.path.join(chroot_user_path, "build")
chroot_check_pkgbuild_path = os.path.join(chroot_build_path, "PKGBUILD")
chroot_check_sh_path = os.path.join(chroot_build_path, "install_check.sh")
if not skip_prepare_chroot and not prepare_user_chroot(other_state):
log_print(
f"ERROR: Failed to prepare user chroot with dummy PKGBUILD!",
other_state=other_state,
)
return "error"
try:
subprocess.run(
(
"/usr/bin/cp",
os.path.join(pkgdir, "PKGBUILD"),
chroot_check_pkgbuild_path,
),
check=True,
)
except subprocess.CalledProcessError:
log_print(
f'ERROR: Failed to check PKGBUILD install (moving PKGBUILD to chroot) for "{pkg}"!',
other_state=other_state,
)
return "error"
get_install_script = """#!/usr/bin/env bash
set -e
source "/build/PKGBUILD"
if [[ -n "$install" ]]; then
echo "$install"
else
echo "PKGBUILD_INSTALL_DOES_NOT_EXIST"
fi
"""
if not create_executable_script(chroot_check_sh_path, get_install_script):
log_print(
f'ERROR: Failed to check PKGBUILD install (check PKGBUILD setup) for "{pkg}"!',
other_state=other_state,
)
return "error"
install_output = None
try:
install_output = subprocess.run(
(
"/usr/bin/env",
"sudo",
"arch-nspawn",
chroot_user_path,
"/build/install_check.sh",
),
check=True,
capture_output=True,
text=True,
)
except subprocess.CalledProcessError:
log_print(
f'ERROR: Failed to check PKGBUILD install (checking PKGBUILD) for "{pkg}"!',
other_state=other_state,
)
return "error"
if (
len(install_output.stdout.strip()) == 0
or install_output.stdout.strip() == "PKGBUILD_INSTALL_DOES_NOT_EXIST"
):
return "does_not_exist"
elif not os.path.exists(
os.path.join(pkgdir, install_output.stdout.strip())
):
log_print(
f'ERROR: PKGBUILD install file specified but doesn\'t exist for "{pkg}"!',
other_state=other_state,
)
return "error"
try:
subprocess.run(
("/usr/bin/env", editor, install_output.stdout.strip()),
check=True,
cwd=pkgdir,
)
except subprocess.CalledProcessError:
log_print(
'ERROR: Failed checking install file for "{}"'.format(pkg),
other_state=other_state,
)
return "error"
return "ok"
def prepare_user_chroot(other_state: dict[str, Any]):
try:
log_print(
'Running "makechrootpkg ... --nobuild" with dummy package to ensure user chroot is ready...',
other_state=other_state,
)
# Ensure ccache isn't enabled for this check.
if other_state["tmpfs"]:
cleanup_ccache(other_state["tmpfs_chroot"])
else:
cleanup_ccache(other_state["chroot"])
command_list = [
"/usr/bin/env",
"makechrootpkg",
"-c",
"-r",
(
other_state["tmpfs_chroot"]
if other_state["tmpfs"]
else other_state["chroot"]
),
]
post_command_list = ["--", "-s", "-r", "-c", "--nobuild"]
dummy_package_dir = os.path.join(
os.environ["HOME"], ".local", "share", "dummy_pkg_TEMPORARY_DIR"
)
Path(dummy_package_dir).mkdir(mode=0o700, parents=True, exist_ok=True)
dummy_package_pkgbuild = os.path.join(dummy_package_dir, "PKGBUILD")
Path(dummy_package_pkgbuild).write_text(DUMMY_PKGBUILD)
subprocess.run(
command_list + post_command_list,
check=True,
cwd=dummy_package_dir,
)
shutil.rmtree(dummy_package_dir)
except subprocess.CalledProcessError:
return False
return True
def prefetch_dependencies(pkg_names: [str], other_state: dict[str, Any]):
"""Returns "fetched" on success."""
log_print(
f'Prefetching packages "{pkg_names}" with "pacman -Sw"...',
other_state=other_state,
)
command_list = ["/usr/bin/env", "sudo", "pacman", "--noconfirm", "-Sw"]
command_list.extend(pkg_names)
try:
subprocess.run(
command_list,
check=True,
)
except subprocess.CalledProcessError:
return "fail"
return "fetched"
def main():
"""The main function."""
signal.signal(signal.SIGINT, signal_handler)
signal.signal(signal.SIGUSR1, signal_handler)
editor = None
parser = argparse.ArgumentParser(description="Update AUR pkgs")
parser.add_argument(
"--config", help="Info and pkg(s) to update in a .toml config"
)
parser.add_argument(
"-p", "--pkg", action="append", help="Pkg(s) to update", metavar="pkg"
)
parser.add_argument(
"--no-skip",
action="append",
help="Pkg(s) to not skip if up to date",
metavar="noskip",
)
parser.add_argument(
"-e",
"--editor",
default=None,
help="editor to use when viewing PKGBUILDs",
metavar="editor",
)
parser.add_argument("--chroot", help="Chroot to build in")
parser.add_argument("--pkg-dir", help="Destination for built pkgs")
parser.add_argument("--repo", help="repository tar file")
parser.add_argument("--gpg-dir", help="gpg home for checking signatures")
parser.add_argument("--logs-dir", help="dir to put logs")
parser.add_argument(
"--no-update", help="Do not update chroot", action="store_true"
)
parser.add_argument("--signing-gpg-dir", help="gpg home for signing key")
parser.add_argument(
"--signing-gpg-key-fp", help="gpg fingerprint for signing key"
)
parser.add_argument(
"--no-store",
action="store_true",
help="Don't sign built package and add to repo",
)
parser.add_argument(
"--tmpfs",
action="store_true",
help="Build in tmpfs",
)
args = parser.parse_args()
if (
args.pkg
and not args.config
and (
not args.chroot
or not args.pkg_dir
or not args.repo
or not args.gpg_dir
or not args.logs_dir
)
):
log_print(
"ERROR: --pkg requires also --chroot, --pkg_dir, --repo, --gpg_dir, and --logs_dir",
other_state=other_state,
)
sys.exit(1)
pkg_state = {}
other_state = {}
global PKG_STATE, OTHER_STATE, GLOBAL_LOG_FILE
PKG_STATE = pkg_state
OTHER_STATE = other_state
other_state["USER"] = os.environ["USER"]
other_state["UID"] = pwd.getpwnam(other_state["USER"]).pw_uid
other_state["stop_building"] = False
other_state["logs_dir"] = None
other_state["log_limit"] = 1024 * 1024 * 1024
other_state["error_on_limit"] = False
other_state["print_state_SIGUSR1"] = False
other_state["print_state_info_only_building_sigusr1"] = True
if args.pkg and not args.config:
for pkg in args.pkg:
pkg_state[pkg] = {}
pkg_state[pkg]["aur_deps"] = []
other_state["chroot"] = args.chroot
other_state["pkg_out_dir"] = args.pkg_dir
other_state["repo"] = args.repo
other_state["gpg_home"] = args.gpg_dir
other_state["logs_dir"] = args.logs_dir
if args_logs_dir is not None:
GLOBAL_LOG_FILE = args_logs_dir + "/update.py_logs"
log_print(
get_datetime_timezone_now(other_state),
other_state=other_state,
)
log_print(
f"Set GLOBAL_LOG_FILE to {GLOBAL_LOG_FILE}",
other_state=other_state,
)
other_state["signing_gpg_dir"] = args.signing_gpg_dir
other_state["signing_gpg_key_fp"] = args.signing_gpg_key_fp
if args_signing_gpg_key_fp is None:
log_print(
'ERROR: Signing key fingerprint "signing_gpg_key_fp" not present in config',
other_state=other_state,
)
sys.exit(1)
if args_signing_gpg_dir is not None and not args.no_store:
other_state["signing_gpg_pass"] = getpass.getpass(
"gpg signing key pass: "
)
if not test_gpg_passphrase(
other_state["signing_gpg_dir"],
other_state["signing_gpg_key_fp"],
other_state["signing_gpg_pass"],
other_state,
):
sys.exit(1)
elif args.config:
d = toml.load(args.config)
for entry in d["entry"]:
pkg_state[entry["name"]] = {}
if "aur_deps" in entry:
pkg_state[entry["name"]]["aur_deps"] = entry["aur_deps"]
else:
pkg_state[entry["name"]]["aur_deps"] = []
if "repo_path" in entry:
pkg_state[entry["name"]]["repo_path"] = entry["repo_path"]
if "pkg_name" in entry:
pkg_state[entry["name"]]["pkg_name"] = entry["pkg_name"]
else:
pkg_state[entry["name"]]["pkg_name"] = entry["name"]
if "ccache_dir" in entry:
pkg_state[entry["name"]]["ccache_dir"] = entry["ccache_dir"]
elif "sccache_dir" in entry:
pkg_state[entry["name"]]["sccache_dir"] = entry["sccache_dir"]
if "sccache_cache_size" in entry:
pkg_state[entry["name"]]["sccache_cache_size"] = entry[
"sccache_cache_size"
]
else:
pkg_state[entry["name"]]["sccache_cache_size"] = "5G"
if (
"sccache_rust_only" in entry
and type(entry["sccache_rust_only"]) is bool
and entry["sccache_rust_only"]
):
pkg_state[entry["name"]]["sccache_rust_only"] = True
else:
pkg_state[entry["name"]]["sccache_rust_only"] = False
if "other_deps" in entry:
pkg_state[entry["name"]]["other_deps"] = entry["other_deps"]
else:
pkg_state[entry["name"]]["other_deps"] = []
if (
"skip_branch_up_to_date" in entry
and type(entry["skip_branch_up_to_date"]) is bool
and entry["skip_branch_up_to_date"]
and not (
not args.no_skip is None and entry["name"] in args.no_skip
)
):
pkg_state[entry["name"]]["skip_branch_up_to_date"] = True
else:
pkg_state[entry["name"]]["skip_branch_up_to_date"] = False
if (
"link_cargo_registry" in entry
and type(entry["link_cargo_registry"]) is bool
and entry["link_cargo_registry"]
):
pkg_state[entry["name"]]["link_cargo_registry"] = True
other_state["chroot"] = d["chroot"]
other_state["pkg_out_dir"] = d["pkg_out_dir"]
other_state["repo"] = d["repo"]
other_state["gpg_home"] = d["gpg_dir"]
other_state["logs_dir"] = d["logs_dir"]
other_state["clones_dir"] = d["clones_dir"]
if (
"datetime_in_local_time" in d
and type(d["datetime_in_local_time"]) is bool
and d["datetime_in_local_time"]
):
other_state["datetime_in_local_time"] = True
else:
other_state["datetime_in_local_time"] = False
if other_state["logs_dir"] is not None:
GLOBAL_LOG_FILE = other_state["logs_dir"] + "/update.py_logs"
log_print(
get_datetime_timezone_now(other_state),
other_state=other_state,
)
log_print(
f"Set GLOBAL_LOG_FILE to {GLOBAL_LOG_FILE}",
other_state=other_state,
)
if args.pkg:
to_keep = [args_pkg for args_pkg in args.pkg]
removal = []
for existing in pkg_state.keys():
if existing in to_keep:
pass
else:
removal.append(existing)
for to_remove in removal:
del pkg_state[to_remove]
if "signing_gpg_dir" in d and not args.no_store:
other_state["signing_gpg_dir"] = d["signing_gpg_dir"]
other_state["signing_gpg_key_fp"] = d["signing_gpg_key_fp"]
other_state["signing_gpg_pass"] = getpass.getpass(
"gpg signing key pass: "
)
if not test_gpg_passphrase(
other_state["signing_gpg_dir"],
other_state["signing_gpg_key_fp"],
other_state["signing_gpg_pass"],
other_state,
):
sys.exit(1)
if "editor" in d:
editor = d["editor"]
if "is_timed" in d and d["is_timed"] is True:
other_state["is_timed"] = True
else:
other_state["is_timed"] = False
if "is_log_timed" in d and d["is_log_timed"] is True:
other_state["is_log_timed"] = True
else:
other_state["is_log_timed"] = False
if (
"log_limit" in d
and type(d["log_limit"]) is int
and d["log_limit"] > 0
):
other_state["log_limit"] = d["log_limit"]
log_print('Set "log_limit" to {}'.format(d["log_limit"]))
else:
log_print(
'Using default "log_limit" of {}'.format(
other_state["log_limit"]
)
)
log_print(" {} KiB".format(other_state["log_limit"] / 1024))
log_print(" {} MiB".format(other_state["log_limit"] / 1024 / 1024))
if (
"error_on_limit" in d
and type(d["error_on_limit"]) is bool
and d["error_on_limit"]
):
other_state["error_on_limit"] = True
log_print(
'Notice: "error_on_limit" is set to "{}"'.format(
other_state["error_on_limit"]
)
)
if "tmpfs" in d and type(d["tmpfs"]) is bool and d["tmpfs"]:
other_state["tmpfs"] = True
else:
other_state["tmpfs"] = False
if (
"print_state_info_only_building_sigusr1" in d
and type(d["print_state_info_only_building_sigusr1"]) is bool
):
other_state["print_state_info_only_building_sigusr1"] = d[
"print_state_info_only_building_sigusr1"
]
print(
'State info print on SIGUSR1 is set to: "{}"'.format(
other_state["print_state_info_only_building_sigusr1"]
)
)
else:
log_print(
'ERROR: At least "--config" or "--pkg" must be specified',
other_state=other_state,
)
sys.exit(1)
while len(other_state["chroot"]) > 1 and other_state["chroot"][-1] == "/":
other_state["chroot"] = other_state["chroot"][:-1]
if args.tmpfs:
other_state["tmpfs"] = True
if other_state["tmpfs"]:
other_state["tmpfs_chroot"] = os.path.join(
os.path.dirname(os.path.realpath(other_state["chroot"])),
"tmpfs_chroot",
)
get_sudo_privileges(other_state)
try:
old_umask = os.umask(0o077)
log_print(
"Ensuring tmpfs_chroot dir exists...", other_state=other_state
)
subprocess.run(
(
"/usr/bin/env",
"mkdir",
"-p",
other_state["tmpfs_chroot"],
),
check=True,
)
log_print("Creating tmpfs dir...", other_state=other_state)
subprocess.run(
(
"/usr/bin/env",
"sudo",
"mount",
"-t",
"tmpfs",
"-o",
f"size=90%,mode=0700,uid={other_state['UID']}",
"tmpfs",
other_state["tmpfs_chroot"],
),
check=True,
)
atexit.register(
lambda tmpfs_path: subprocess.run(
(
"/usr/bin/env",
"sudo",
"bash",
"-c",
f"for ((i=0; i<5; ++i)); do if umount {tmpfs_path}; then break; fi; sleep 1; done",
)
),
other_state["tmpfs_chroot"],
)
os.umask(old_umask)
except subprocess.CalledProcessError:
log_print("ERROR: Failed to set up tmpfs!")
sys.exit(1)
validate_and_verify_paths(other_state)
if args.editor is not None:
editor = args.editor
if editor is None:
editor = DEFAULT_EDITOR
os.putenv("CHROOT", os.path.realpath(other_state["chroot"]))
os.putenv("GNUPGHOME", os.path.realpath(other_state["gpg_home"]))
if not os.path.exists(other_state["logs_dir"]):
os.makedirs(other_state["logs_dir"])
elif not os.path.isdir(other_state["logs_dir"]):
log_print(
'ERROR: logs_dir "{}" must be a directory'.format(
other_state["logs_dir"]
),
other_state=other_state,
)
sys.exit(1)
if not args.no_update:
log_print("Updating the chroot...", other_state=other_state)
try:
subprocess.run(
(
"/usr/bin/env",
"arch-nspawn",
"{}/root".format(other_state["chroot"]),
"pacman",
"-Syu",
),
check=True,
)
except subprocess.CalledProcessError:
log_print(
"ERROR: Failed to update the chroot", other_state=other_state
)
sys.exit(1)
if other_state["tmpfs"]:
try:
log_print(
'Copying "chroot"/root to tmpfs_chroot/root...',
other_state=other_state,
)
subprocess.run(
(
"/usr/bin/env",
"sudo",
"cp",
"-a",
f'{other_state["chroot"]}/root',
f'{other_state["tmpfs_chroot"]}/root',
),
check=True,
)
except subprocess.CalledProcessError:
log_print(
'ERROR: Failed to copy "chroot"/root to tmpfs_chroot/root!',
other_state=other_state,
)
sys.exit(1)
os.putenv("CHROOT", os.path.realpath(other_state["tmpfs_chroot"]))
pkg_list = [temp_pkg_name for temp_pkg_name in pkg_state.keys()]
# ensure build_status is populated.
for pkg_name in pkg_list:
pkg_state[pkg_name]["build_status"] = "unknown"
i = 0
furthest_checked = 0
going_back = False
check_install_script_ran_once = False
while i < len(pkg_list):
if i > furthest_checked:
furthest_checked = i
if not ensure_pkg_dir_exists(pkg_list[i], pkg_state, other_state):
print_state_info_and_get_update_list(other_state, pkg_state)
sys.exit(1)
if (
"repo_path" not in pkg_state[pkg_list[i]]
or pkg_state[pkg_list[i]]["repo_path"] != "NO_REPO"
):
update_pkg_dir_count = 0
update_pkg_dir_success = False
while update_pkg_dir_count < 5:
(success, skip_on_same_ver) = update_pkg_dir(
pkg_list[i], pkg_state, other_state
)
if success:
update_pkg_dir_success = True
break
else:
time.sleep(1)
update_pkg_dir_count += 1
if not update_pkg_dir_success:
log_print(
'Failed to update pkg dir for "{}"',
pkg_list[i],
other_state=other_state,
)
pkg_state[pkg_list[i]]["state"] = "error_fetch"
pkg_state[pkg_list[i]]["build_status"] = "not_building"
i += 1
continue
if skip_on_same_ver and i >= furthest_checked:
check_pkg_version_result = check_pkg_version(
pkg_list[i], pkg_state, other_state["repo"], True, other_state
)
if check_pkg_version_result != "install":
log_print(
f"Pkg {pkg_list[i]} is up to date, skipping...",
other_state=other_state,
)
pkg_state[pkg_list[i]]["state"] = "up_to_date"
pkg_state[pkg_list[i]]["build_status"] = "not_building"
i += 1
continue
check_pkg_build_result = check_pkg_build(
pkg_list[i], pkg_state, other_state, editor
)
if check_pkg_build_result == "ok":
pass
elif check_pkg_build_result == "not_ok":
pkg_state[pkg_list[i]]["state"] = "skip"
pkg_state[pkg_list[i]]["build_status"] = "not_building"
i += 1
continue
elif check_pkg_build_result == "force_build":
pkg_state[pkg_list[i]]["state"] = "install"
pkg_state[pkg_list[i]]["build_status"] = "will_build"
i += 1
log_print(
'WARNING: force_build will skip "install" script check!',
other_state=other_state,
)
continue
elif check_pkg_build_result == "invalid":
continue
elif check_pkg_build_result == "back":
if i > 0:
i -= 1
continue
else: # check_pkg_build_result == "abort":
print_state_info_and_get_update_list(other_state, pkg_state)
sys.exit(1)
install_check = check_install_script(
pkg_state,
other_state,
pkg_list[i],
editor,
skip_prepare_chroot=check_install_script_ran_once,
)
check_install_script_ran_once = True
if install_check == "does_not_exist":
log_print(
'NOTICE: pkg does not have "install" script.',
other_state=other_state,
)
elif install_check == "error":
log_print(
"WARNING: Failed to check PKGBUILD install script!",
other_state=other_state,
)
pkg_state[pkg_list[i]]["state"] = "error"
pkg_state[pkg_list[i]]["build_status"] = "not_building"
i += 1
continue
elif install_check == "ok":
continue_on_loop_exit = False
recheck_install_script = False
while True:
if recheck_install_script:
recheck_install_script = False
check_install_script(
pkg_state,
other_state,
pkg_list[i],
editor,
skip_prepare_chroot=True,
)
log_print(
"install script ok? [Y/n/r(echeck)/a(bort)/f(orce build)/b(ack)]",
other_state=other_state,
)
user_input = (
sys.stdin.buffer.readline().decode().strip().lower()
)
if user_input == "y" or len(user_input) == 0:
log_print(
"User decided install script is ok.",
other_state=other_state,
)
break
elif user_input == "n":
log_print(
"User decided install script is NOT ok.",
other_state=other_state,
)
pkg_state[pkg_list[i]]["state"] = "skip"
pkg_state[pkg_list[i]]["build_status"] = "not_building"
i += 1
continue_on_loop_exit = True
break
elif user_input == "r":
recheck_install_script = True
continue
elif user_input == "a":
print_state_info_and_get_update_list(other_state, pkg_state)
sys.exit(1)
elif user_input == "f":
pkg_state[pkg_list[i]]["state"] = "install"
pkg_state[pkg_list[i]]["build_status"] = "will_build"
i += 1
continue_on_loop_exit = True
break
elif user_input == "b":
if i > 0:
i -= 1
continue_on_loop_exit = True
break
else:
continue
if continue_on_loop_exit:
continue
else: # Should be unreachable.
log_print(
"WARNING: Check PKGBUILD install script: unreachable code!",
other_state=other_state,
)
while True:
if (
skip_on_same_ver
and check_pkg_version_result is not None
and i >= furthest_checked
):
state_result = check_pkg_version_result
else:
state_result = check_pkg_version(
pkg_list[i],
pkg_state,
other_state["repo"],
False,
other_state,
)
confirm_result_result = confirm_result(
pkg_list[i], state_result, other_state
)
if confirm_result_result == "continue":
pkg_state[pkg_list[i]]["state"] = state_result
pkg_state[pkg_list[i]]["build_status"] = (
"will_build"
if state_result == "install"
else "not_building"
)
break
elif confirm_result_result == "recheck":
check_pkg_version_result = None
continue
elif confirm_result_result == "force_build":
pkg_state[pkg_list[i]]["state"] = "install"
pkg_state[pkg_list[i]]["build_status"] = "will_build"
break
elif confirm_result_result == "skip":
pkg_state[pkg_list[i]]["state"] = "skip"
pkg_state[pkg_list[i]]["build_status"] = "not_building"
break
elif confirm_result_result == "back":
if i > 0:
i -= 1
going_back = True
break
else: # confirm_result_result == "abort"
print_state_info_and_get_update_list(other_state, pkg_state)
sys.exit(1)
if going_back:
going_back = False
else:
i += 1
log_print("Showing current actions:", other_state=other_state)
pkgs_to_update = print_state_info_and_get_update_list(
other_state, pkg_state
)
if len(pkgs_to_update) > 0:
log_print("Continue? [Y/n]", other_state=other_state)
user_input = sys.stdin.buffer.readline().decode().strip().lower()
if user_input == "y" or len(user_input) == 0:
if args.no_update:
log_print(
"Updating (without updating chroot)...",
other_state=other_state,
)
else:
log_print("Updating...", other_state=other_state)
update_pkg_list(
pkgs_to_update,
pkg_state,
other_state,
"" if args.no_store else other_state["signing_gpg_dir"],
"" if args.no_store else other_state["signing_gpg_key_fp"],
"" if args.no_store else other_state["signing_gpg_pass"],
args.no_store,
)
else:
log_print("Canceled.", other_state=other_state)
else:
log_print("No packages to update, done.", other_state=other_state)
if __name__ == "__main__":
main()