blob: 4d920ea3b16a6c1a9af524b3b0a47856b8ab2feb [file] [log] [blame]
#!/usr/bin/env python
# Copyright 2009-2021 Gentoo Authors
# Distributed under the terms of the GNU General Public License v2
import argparse
import platform
import signal
import stat
import sys
# This block ensures that ^C interrupts are handled quietly.
try:
def exithandler(signum, _frame):
signal.signal(signal.SIGINT, signal.SIG_IGN)
signal.signal(signal.SIGTERM, signal.SIG_IGN)
sys.exit(128 + signum)
signal.signal(signal.SIGINT, exithandler)
signal.signal(signal.SIGTERM, exithandler)
except KeyboardInterrupt:
sys.exit(128 + signal.SIGINT)
def debug_signal(_signum, _frame):
import pdb
pdb.set_trace()
if platform.python_implementation() == "Jython":
debug_signum = signal.SIGUSR2 # bug #424259
else:
debug_signum = signal.SIGUSR1
signal.signal(debug_signum, debug_signal)
import functools
import io
import logging
import subprocess
import time
import textwrap
import re
from os import path as osp
if osp.isfile(
osp.join(osp.dirname(osp.dirname(osp.realpath(__file__))), ".portage_not_installed")
):
sys.path.insert(
0, osp.join(osp.dirname(osp.dirname(osp.realpath(__file__))), "lib")
)
import portage
portage._internal_caller = True
from portage import os, _encodings, _unicode_encode, _unicode_decode
from portage.cache.cache_errors import CacheError, StatCollision
from portage.cache.index.pkg_desc_index import (
pkg_desc_index_line_format,
pkg_desc_index_line_read,
)
from portage.const import TIMESTAMP_FORMAT
from portage.dep import _repo_separator
from portage.output import colorize, EOutput
from portage.package.ebuild._parallel_manifest.ManifestScheduler import (
ManifestScheduler,
)
from portage.util import cmp_sort_key, writemsg_level
from portage.util._async.AsyncFunction import AsyncFunction
from portage.util._async.run_main_scheduler import run_main_scheduler
from portage.util._async.TaskScheduler import TaskScheduler
from portage.util._eventloop.global_event_loop import global_event_loop
from portage.util.changelog import ChangeLogTypeSort
from portage import cpv_getkey
from portage.dep import Atom, isjustname
from portage.versions import vercmp
from _emerge.MetadataRegen import MetadataRegen
try:
from xml.etree import ElementTree
except ImportError:
pass
else:
try:
from xml.parsers.expat import ExpatError
except ImportError:
pass
else:
from portage.xml.metadata import ( # pylint: disable=ungrouped-imports
parse_metadata_use,
)
def parse_args(args):
usage = "egencache [options] <action> ... [atom] ..."
parser = argparse.ArgumentParser(usage=usage)
actions = parser.add_argument_group("Actions")
actions.add_argument(
"--update",
action="store_true",
help="update metadata/md5-cache/ (generate as necessary)",
)
actions.add_argument(
"--update-use-local-desc",
action="store_true",
help="update the use.local.desc file from metadata.xml",
)
actions.add_argument(
"--update-changelogs",
action="store_true",
help="update the ChangeLog files from SCM logs",
)
actions.add_argument(
"--update-pkg-desc-index",
action="store_true",
help="update package description index",
)
actions.add_argument(
"--update-manifests", action="store_true", help="update manifests"
)
common = parser.add_argument_group("Common options")
common.add_argument("--repo", action="store", help="name of repo to operate on")
common.add_argument(
"--config-root",
help="location of portage config files",
dest="portage_configroot",
)
common.add_argument(
"--external-cache-only",
action="store_true",
help="Output only to the external cache (not the repository itself)",
)
common.add_argument(
"--gpg-dir", help="override the PORTAGE_GPG_DIR variable", dest="gpg_dir"
)
common.add_argument(
"--gpg-key", help="override the PORTAGE_GPG_KEY variable", dest="gpg_key"
)
common.add_argument(
"--repositories-configuration",
help="override configuration of repositories (in format of repos.conf)",
dest="repositories_configuration",
)
common.add_argument(
"--sign-manifests",
choices=("y", "n"),
metavar="<y|n>",
help="manually override layout.conf sign-manifests setting",
)
common.add_argument(
"--strict-manifests",
choices=("y", "n"),
metavar="<y|n>",
help='manually override "strict" FEATURES setting',
)
common.add_argument(
"--thin-manifests",
choices=("y", "n"),
metavar="<y|n>",
help="manually override layout.conf thin-manifests setting",
)
common.add_argument(
"--tolerant",
action="store_true",
help="exit successfully if only minor errors occurred",
)
common.add_argument(
"--ignore-default-opts",
action="store_true",
help="do not use the EGENCACHE_DEFAULT_OPTS environment variable",
)
common.add_argument(
"-v", "--verbose", action="count", default=0, help="increase verbosity"
)
common.add_argument(
"--write-timestamp",
action="store_true",
help="write metadata/timestamp.chk as required for rsync repositories",
)
update = parser.add_argument_group("--update options")
update.add_argument(
"--cache-dir", help="location of the metadata cache", dest="cache_dir"
)
update.add_argument(
"-j", "--jobs", type=int, action="store", help="max ebuild processes to spawn"
)
update.add_argument(
"--load-average",
type=float,
action="store",
help="max load allowed when spawning multiple jobs",
dest="load_average",
)
update.add_argument(
"--rsync",
action="store_true",
help="enable rsync stat collision workaround "
+ "for bug 139134 (use with --update)",
)
uld = parser.add_argument_group("--update-use-local-desc options")
uld.add_argument(
"--preserve-comments",
action="store_true",
help="preserve the comments from the existing use.local.desc file",
)
uld.add_argument(
"--use-local-desc-output",
help="output file for use.local.desc data (or '-' for stdout)",
dest="uld_output",
)
uc = parser.add_argument_group("--update-changelogs options")
uc.add_argument(
"--changelog-reversed",
action="store_true",
help="log commits in reverse order (oldest first)",
)
uc.add_argument(
"--changelog-output",
help="output filename for change logs",
dest="changelog_output",
default="ChangeLog",
)
options, args = parser.parse_known_args(args)
if options.jobs:
jobs = None
try:
jobs = int(options.jobs)
except ValueError:
jobs = -1
if jobs < 1:
parser.error("Invalid: --jobs='%s'" % (options.jobs,))
options.jobs = jobs
else:
options.jobs = None
if options.load_average:
try:
load_average = float(options.load_average)
except ValueError:
load_average = 0.0
if load_average <= 0.0:
parser.error("Invalid: --load-average='%s'" % (options.load_average,))
options.load_average = load_average
else:
options.load_average = None
options.config_root = options.portage_configroot
if options.config_root is not None and not os.path.isdir(options.config_root):
parser.error("Not a directory: --config-root='%s'" % (options.config_root,))
if options.cache_dir is not None:
if not os.path.isdir(options.cache_dir):
parser.error("Not a directory: --cache-dir='%s'" % (options.cache_dir,))
if not os.access(options.cache_dir, os.W_OK):
parser.error("Write access denied: --cache-dir='%s'" % (options.cache_dir,))
for atom in args:
try:
atom = portage.dep.Atom(atom)
except portage.exception.InvalidAtom:
parser.error("Invalid atom: %s" % (atom,))
if not isjustname(atom):
parser.error("Atom is too specific: %s" % (atom,))
if options.update_use_local_desc:
try:
ElementTree
ExpatError
except NameError:
parser.error("--update-use-local-desc requires python with USE=xml!")
if options.uld_output == "-" and options.preserve_comments:
parser.error("--preserve-comments can not be used when outputting to stdout")
return parser, options, args
class GenCache:
def __init__(
self,
portdb,
cp_iter=None,
max_jobs=None,
max_load=None,
rsync=False,
external_cache_only=False,
):
# The caller must set portdb.porttrees in order to constrain
# findname, cp_list, and cpv_list to the desired tree.
tree = portdb.porttrees[0]
self._portdb = portdb
self._eclass_db = portdb.repositories.get_repo_for_location(tree).eclass_db
self._auxdbkeys = portdb._known_keys
# We can globally cleanse stale cache only if we
# iterate over every single cp.
self._global_cleanse = cp_iter is None
if cp_iter is not None:
self._cp_set = set(cp_iter)
cp_iter = iter(self._cp_set)
self._cp_missing = self._cp_set.copy()
else:
self._cp_set = None
self._cp_missing = set()
write_auxdb = (
external_cache_only or "metadata-transfer" in portdb.settings.features
)
self._regen = MetadataRegen(
portdb,
cp_iter=cp_iter,
consumer=self._metadata_callback,
max_jobs=max_jobs,
max_load=max_load,
write_auxdb=write_auxdb,
main=True,
)
self.returncode = os.EX_OK
conf = portdb.repositories.get_repo_for_location(tree)
if external_cache_only:
self._trg_caches = ()
else:
self._trg_caches = tuple(
conf.iter_pregenerated_caches(
self._auxdbkeys, force=True, readonly=False
)
)
if not self._trg_caches:
raise Exception(
"cache formats '%s' aren't supported"
% (" ".join(conf.cache_formats),)
)
if rsync:
for trg_cache in self._trg_caches:
if hasattr(trg_cache, "raise_stat_collision"):
trg_cache.raise_stat_collision = True
# Make _metadata_callback write this cache first, in case
# it raises a StatCollision and triggers mtime
# modification.
self._trg_caches = tuple(
[trg_cache]
+ [x for x in self._trg_caches if x is not trg_cache]
)
self._existing_nodes = set()
def _metadata_callback(self, cpv, repo_path, metadata, ebuild_hash, eapi_supported):
self._existing_nodes.add(cpv)
self._cp_missing.discard(cpv_getkey(cpv))
# Since we're supposed to be able to efficiently obtain the
# EAPI from _parse_eapi_ebuild_head, we don't write cache
# entries for unsupported EAPIs.
if metadata is not None and eapi_supported:
for trg_cache in self._trg_caches:
self._write_cache(trg_cache, cpv, repo_path, metadata, ebuild_hash)
def _write_cache(self, trg_cache, cpv, repo_path, metadata, ebuild_hash):
if not hasattr(trg_cache, "raise_stat_collision"):
# This cache does not avoid redundant writes automatically,
# so check for an identical existing entry before writing.
# This prevents unnecessary disk writes and can also prevent
# unnecessary rsync transfers.
try:
dest = trg_cache[cpv]
except (KeyError, CacheError):
pass
else:
if trg_cache.validate_entry(dest, ebuild_hash, self._eclass_db):
identical = True
for k in self._auxdbkeys:
if dest.get(k, "") != metadata.get(k, ""):
identical = False
break
if identical:
return
try:
chf = trg_cache.validation_chf
metadata["_%s_" % chf] = getattr(ebuild_hash, chf)
try:
trg_cache[cpv] = metadata
except StatCollision as sc:
# If the content of a cache entry changes and neither the
# file mtime nor size changes, it will prevent rsync from
# detecting changes. Cache backends may raise this
# exception from _setitem() if they detect this type of stat
# collision. These exceptions are handled by bumping the
# mtime on the ebuild (and the corresponding cache entry).
# See bug #139134. It is convenient to include checks for
# redundant writes along with the internal StatCollision
# detection code, so for caches with the
# raise_stat_collision attribute, we do not need to
# explicitly check for redundant writes like we do for the
# other cache types above.
max_mtime = sc.mtime
for _ec, ec_hash in metadata["_eclasses_"].items():
if max_mtime < ec_hash.mtime:
max_mtime = ec_hash.mtime
if max_mtime == sc.mtime:
max_mtime += 1
max_mtime = int(max_mtime)
try:
os.utime(ebuild_hash.location, (max_mtime, max_mtime))
except OSError as e:
self.returncode |= 1
writemsg_level(
"%s writing target: %s\n" % (cpv, e),
level=logging.ERROR,
noiselevel=-1,
)
else:
ebuild_hash.mtime = max_mtime
metadata["_mtime_"] = max_mtime
trg_cache[cpv] = metadata
self._portdb.auxdb[repo_path][cpv] = metadata
except CacheError as ce:
self.returncode |= 1
writemsg_level(
"%s writing target: %s\n" % (cpv, ce),
level=logging.ERROR,
noiselevel=-1,
)
def run(self):
signum = run_main_scheduler(self._regen)
if signum is not None:
sys.exit(128 + signum)
self.returncode |= self._regen.returncode
for trg_cache in self._trg_caches:
self._cleanse_cache(trg_cache)
def _cleanse_cache(self, trg_cache):
cp_missing = self._cp_missing
dead_nodes = set()
if self._global_cleanse:
try:
for cpv in trg_cache:
cp = cpv_getkey(cpv)
if cp is None:
self.returncode |= 1
writemsg_level(
"Unable to parse cp for '%s'\n" % (cpv,),
level=logging.ERROR,
noiselevel=-1,
)
else:
dead_nodes.add(cpv)
except CacheError as ce:
self.returncode |= 1
writemsg_level(
"Error listing cache entries for "
+ "'%s': %s, continuing...\n" % (trg_cache.location, ce),
level=logging.ERROR,
noiselevel=-1,
)
else:
cp_set = self._cp_set
try:
for cpv in trg_cache:
cp = cpv_getkey(cpv)
if cp is None:
self.returncode |= 1
writemsg_level(
"Unable to parse cp for '%s'\n" % (cpv,),
level=logging.ERROR,
noiselevel=-1,
)
else:
cp_missing.discard(cp)
if cp in cp_set:
dead_nodes.add(cpv)
except CacheError as ce:
self.returncode |= 1
writemsg_level(
"Error listing cache entries for "
+ "'%s': %s, continuing...\n" % (trg_cache.location, ce),
level=logging.ERROR,
noiselevel=-1,
)
if cp_missing:
self.returncode |= 1
for cp in sorted(cp_missing):
writemsg_level(
"No ebuilds or cache entries found for '%s'\n" % (cp,),
level=logging.ERROR,
noiselevel=-1,
)
if dead_nodes:
dead_nodes.difference_update(self._existing_nodes)
for k in dead_nodes:
try:
del trg_cache[k]
except KeyError:
pass
except CacheError as ce:
self.returncode |= 1
writemsg_level(
"%s deleting stale cache: %s\n" % (k, ce),
level=logging.ERROR,
noiselevel=-1,
)
if not trg_cache.autocommits:
try:
trg_cache.commit()
except CacheError as ce:
self.returncode |= 1
writemsg_level(
"committing target: %s\n" % (ce,),
level=logging.ERROR,
noiselevel=-1,
)
if hasattr(trg_cache, "_prune_empty_dirs"):
trg_cache._prune_empty_dirs()
class GenPkgDescIndex:
def __init__(self, repo_config, portdb, output_file, verbose=False):
self.returncode = os.EX_OK
self._repo_config = repo_config
self._portdb = portdb
self._output_file = output_file
self._verbose = verbose
def run(self):
display_updates = self._verbose > 0
old = {}
new = {}
if display_updates:
try:
with open(
self._output_file, "rt", encoding=_encodings["repo.content"]
) as f:
for line in f:
pkg_desc = pkg_desc_index_line_read(line)
old[pkg_desc.cp] = pkg_desc
except FileNotFoundError:
pass
portage.util.ensure_dirs(os.path.dirname(self._output_file))
f = portage.util.atomic_ofstream(
self._output_file, encoding=_encodings["repo.content"]
)
portdb = self._portdb
for cp in portdb.cp_all():
pkgs = portdb.cp_list(cp)
if not pkgs:
continue
(desc,) = portdb.aux_get(pkgs[-1], ["DESCRIPTION"])
line = pkg_desc_index_line_format(cp, pkgs, desc)
f.write(line)
if display_updates:
new[cp] = pkg_desc_index_line_read(line)
f.close()
if display_updates:
out = EOutput()
out.einfo("Searching for changes")
print("")
items = sorted(new.values(), key=lambda pkg_desc: pkg_desc.cp)
haspkgs = False
for pkg_desc in items:
masked = False
version = self._portdb.xmatch(
"bestmatch-visible",
Atom(
"{}{}{}".format(
pkg_desc.cp, _repo_separator, self._repo_config.name
)
),
)
if not version:
version = pkg_desc.cpv_list[-1]
masked = True
old_versions = old.get(pkg_desc.cp)
if old_versions is None or version not in old_versions.cpv_list:
prefix0 = " "
prefix1 = " "
if old_versions is None:
color = functools.partial(colorize, "darkgreen")
prefix1 = "N"
else:
color = functools.partial(colorize, "turquoise")
prefix1 = "U"
if masked:
prefix0 = "M"
print(
" [%s%s] %s (%s): %s"
% (
colorize("red", prefix0),
color(prefix1),
colorize("bold", pkg_desc.cp),
color(version[len(pkg_desc.cp) + 1 :]),
pkg_desc.desc,
)
)
haspkgs = True
if not haspkgs:
out.einfo("No updates found")
class GenUseLocalDesc:
def __init__(self, portdb, output=None, preserve_comments=False):
self.returncode = os.EX_OK
self._portdb = portdb
self._output = output
self._preserve_comments = preserve_comments
def run(self):
repo_path = self._portdb.porttrees[0]
ops = {"<": 0, "<=": 1, "=": 2, ">=": 3, ">": 4}
prev_mtime = None
prev_md5 = None
if self._output is None or self._output != "-":
if self._output is None:
prof_path = os.path.join(repo_path, "profiles")
desc_path = os.path.join(prof_path, "use.local.desc")
try:
os.mkdir(prof_path)
except OSError:
pass
else:
desc_path = self._output
try:
prev_md5 = portage.checksum.perform_md5(desc_path)
prev_mtime = os.stat(desc_path)[stat.ST_MTIME]
except (portage.exception.FileNotFound, OSError):
pass
try:
if self._preserve_comments:
# Probe in binary mode, in order to avoid
# potential character encoding issues.
output = open(
_unicode_encode(
desc_path, encoding=_encodings["fs"], errors="strict"
),
"r+b",
)
else:
output = io.open(
_unicode_encode(
desc_path, encoding=_encodings["fs"], errors="strict"
),
mode="w",
encoding=_encodings["repo.content"],
errors="backslashreplace",
)
except IOError as e:
if not self._preserve_comments or os.path.isfile(desc_path):
writemsg_level(
"ERROR: failed to open output file %s: %s\n" % (desc_path, e),
level=logging.ERROR,
noiselevel=-1,
)
self.returncode |= 2
return
# Open in r+b mode failed because the file doesn't
# exist yet. We can probably recover if we disable
# preserve_comments mode now.
writemsg_level(
"WARNING: --preserve-comments enabled, but "
+ "output file not found: %s\n" % (desc_path,),
level=logging.WARNING,
noiselevel=-1,
)
self._preserve_comments = False
try:
output = io.open(
_unicode_encode(
desc_path, encoding=_encodings["fs"], errors="strict"
),
mode="w",
encoding=_encodings["repo.content"],
errors="backslashreplace",
)
except IOError as e:
writemsg_level(
"ERROR: failed to open output file %s: %s\n" % (desc_path, e),
level=logging.ERROR,
noiselevel=-1,
)
self.returncode |= 2
return
else:
output = sys.stdout
if self._preserve_comments:
while True:
pos = output.tell()
if not output.readline().startswith(b"#"):
break
output.seek(pos)
output.truncate()
output.close()
# Finished probing comments in binary mode, now append
# in text mode.
output = io.open(
_unicode_encode(desc_path, encoding=_encodings["fs"], errors="strict"),
mode="a",
encoding=_encodings["repo.content"],
errors="backslashreplace",
)
output.write("\n")
else:
output.write(
textwrap.dedent(
"""\
# This file is deprecated as per GLEP 56 in favor of metadata.xml. Please add
# your descriptions to your package's metadata.xml ONLY.
# * generated automatically using egencache *
"""
)
)
# The cmp function no longer exists in python3, so we'll
# implement our own here under a slightly different name
# since we don't want any confusion given that we never
# want to rely on the builtin cmp function.
def cmp_func(a, b):
if a is None or b is None:
# None can't be compared with other types in python3.
if a is None and b is None:
return 0
elif a is None:
return -1
else:
return 1
return (a > b) - (a < b)
class _MetadataTreeBuilder(ElementTree.TreeBuilder):
"""
Implements doctype() as required to avoid deprecation warnings
since Python >=2.7
"""
def doctype(self, name, pubid, system):
pass
for cp in self._portdb.cp_all():
metadata_path = os.path.join(repo_path, cp, "metadata.xml")
try:
metadata = ElementTree.parse(
_unicode_encode(
metadata_path, encoding=_encodings["fs"], errors="strict"
),
parser=ElementTree.XMLParser(target=_MetadataTreeBuilder()),
)
except IOError:
pass
except (ExpatError, EnvironmentError) as e:
writemsg_level(
"ERROR: failed parsing %s/metadata.xml: %s\n" % (cp, e),
level=logging.ERROR,
noiselevel=-1,
)
self.returncode |= 1
else:
try:
usedict = parse_metadata_use(metadata)
except portage.exception.ParseError as e:
writemsg_level(
"ERROR: failed parsing %s/metadata.xml: %s\n" % (cp, e),
level=logging.ERROR,
noiselevel=-1,
)
self.returncode |= 1
else:
for flag in sorted(usedict):
def atomcmp(atoma, atomb):
# None is better than an atom, that's why we reverse the args
if atoma is None or atomb is None:
return cmp_func(atomb, atoma)
# Same for plain PNs (.operator is None then)
elif atoma.operator is None or atomb.operator is None:
return cmp_func(atomb.operator, atoma.operator)
# Version matching
elif atoma.cpv != atomb.cpv:
return vercmp(atoma.version, atomb.version)
# Versions match, let's fallback to operator matching
else:
return cmp_func(
ops.get(atoma.operator, -1),
ops.get(atomb.operator, -1),
)
def _Atom(key):
if key is not None:
return Atom(key)
return None
resdict = usedict[flag]
if len(resdict) == 1:
resdesc = next(iter(resdict.items()))[1]
else:
try:
reskeys = dict((_Atom(k), k) for k in resdict)
except portage.exception.InvalidAtom as e:
writemsg_level(
"ERROR: failed parsing %s/metadata.xml: %s\n"
% (cp, e),
level=logging.ERROR,
noiselevel=-1,
)
self.returncode |= 1
resdesc = next(iter(resdict.items()))[1]
else:
resatoms = sorted(reskeys, key=cmp_sort_key(atomcmp))
resdesc = resdict[reskeys[resatoms[-1]]]
output.write("%s:%s - %s\n" % (cp, flag, resdesc))
output.close()
if prev_mtime is not None and prev_md5 == portage.checksum.perform_md5(
desc_path
):
# Preserve mtime for rsync.
mtime = prev_mtime
else:
# For portability, and consistency with the mtime preservation
# code, set mtime to an exact integer value.
mtime = int(time.time())
os.utime(desc_path, (mtime, mtime))
class GenChangeLogs:
def __init__(
self, portdb, changelog_output, changelog_reversed, max_jobs=None, max_load=None
):
self.returncode = os.EX_OK
self._portdb = portdb
self._wrapper = textwrap.TextWrapper(
width=78, initial_indent=" ", subsequent_indent=" "
)
self._changelog_output = changelog_output
self._changelog_reversed = changelog_reversed
self._max_jobs = max_jobs
self._max_load = max_load
self._repo_path = self._portdb.porttrees[0]
# --work-tree=... must be passed to Git if GIT_DIR is used
# and GIT_DIR is not a child of the root of the checkout
# eg:
# GIT_DIR=${parent}/work/.git/
# work-tree=${parent}/staging/
# If work-tree is not passed, Git tries to use the shared
# parent of the current directory and the ${GIT_DIR}, which can
# be outside the root of the checkout.
self._work_tree = "--work-tree=%s" % self._repo_path
@staticmethod
def grab(cmd):
p = subprocess.Popen(cmd, stdout=subprocess.PIPE)
return _unicode_decode(
p.communicate()[0], encoding=_encodings["stdio"], errors="strict"
)
def generate_changelog(self, cp):
os.chdir(os.path.join(self._repo_path, cp))
# Determine whether ChangeLog is up-to-date by comparing
# the newest commit timestamp with the ChangeLog timestamp.
lmod = self.grab(["git", self._work_tree, "log", "--format=%ct", "-1", "."])
if not lmod:
# This cp has not been added to the repo.
return
lmod = int(lmod)
try:
cmod = os.stat("ChangeLog")[stat.ST_MTIME]
except OSError:
cmod = 0
# Use exact comparison, since commit times are
# not necessarily ordered.
if cmod == lmod:
return
try:
output = io.open(
self._changelog_output,
mode="w",
encoding=_encodings["repo.content"],
errors="backslashreplace",
)
except IOError as e:
writemsg_level(
"ERROR: failed to open ChangeLog for %s: %s\n"
% (
cp,
e,
),
level=logging.ERROR,
noiselevel=-1,
)
self.returncode |= 2
return
output.write(
textwrap.dedent(
"""\
# ChangeLog for %s
# Copyright 1999-%s Gentoo Foundation; Distributed under the GPL v2
# (auto-generated from git log)
"""
% (cp, time.strftime("%Y"))
)
)
# now grab all the commits
revlist_cmd = ["git", self._work_tree, "rev-list"]
if self._changelog_reversed:
revlist_cmd.append("--reverse")
revlist_cmd.extend(["HEAD", "--", "."])
commits = self.grab(revlist_cmd).split()
for c in commits:
# Explaining the arguments:
# --name-status to get a list of added/removed files
# --no-renames to avoid getting more complex records on the list
# --format to get the timestamp, author and commit description
# --root to make it work fine even with the initial commit
# --relative=${cp} to get paths relative to ebuilddir
# -r (recursive) to get per-file changes
# then the commit-id and path.
cinfo = (
self.grab(
[
"git",
self._work_tree,
"diff-tree",
"--name-status",
"--no-renames",
"--format=%ct %cN <%cE>%n%B",
"--root",
"--relative=%s" % (cp,),
"-r",
c,
"--",
".",
]
)
.rstrip("\n")
.split("\n")
)
# Expected output:
# timestamp Author Name <author@email>
# commit message l1
# ...
# commit message ln
#
# status1 filename1
# ...
# statusn filenamen
changed = []
for n, l in enumerate(reversed(cinfo)):
if not l:
body = cinfo[1 : -n - 1]
break
else:
f = l.split()
if f[1] == "Manifest":
pass # XXX: remanifest commits?
elif f[1].startswith("ChangeLog"):
pass
elif f[0].startswith("A"):
changed.append(ChangeLogTypeSort("+", f[1]))
elif f[0].startswith("D"):
changed.append(ChangeLogTypeSort("-", f[1]))
elif f[0].startswith("M"):
changed.append(ChangeLogTypeSort("", f[1]))
else:
writemsg_level(
"ERROR: unexpected git file status for %s: %s\n"
% (
cp,
f,
),
level=logging.ERROR,
noiselevel=-1,
)
self.returncode |= 1
if not changed:
continue
(ts, author) = cinfo[0].split(" ", 1)
date = time.strftime("%d %b %Y", time.gmtime(float(ts)))
changed = [str(x) for x in sorted(changed)]
wroteheader = False
# Reverse the sort order for headers.
for c in reversed(changed):
if c.startswith("+") and c.endswith(".ebuild"):
output.write("*%s (%s)\n" % (c[1:-7], date))
wroteheader = True
if wroteheader:
output.write("\n")
# strip '<cp>: ', '[<cp>] ', and similar
body[0] = re.sub(r"^\W*" + re.escape(cp) + r"\W+", "", body[0])
# strip trailing newline
if not body[-1]:
body = body[:-1]
# strip git-svn id
if body[-1].startswith("git-svn-id:") and not body[-2]:
body = body[:-2]
# strip the repoman version/manifest note
if (
body[-1] == " (Signed Manifest commit)"
or body[-1] == " (Unsigned Manifest commit)"
):
body = body[:-1]
if body[-1].startswith("(Portage version:") and body[-1].endswith(")"):
body = body[:-1]
if not body[-1]:
body = body[:-1]
# don't break filenames on hyphens
self._wrapper.break_on_hyphens = False
output.write(
self._wrapper.fill("%s; %s %s:" % (date, author, ", ".join(changed)))
)
# but feel free to break commit messages there
self._wrapper.break_on_hyphens = True
output.write("\n%s\n\n" % "\n".join(self._wrapper.fill(x) for x in body))
output.close()
os.utime(self._changelog_output, (lmod, lmod))
def _task_iter(self):
if not os.path.isdir(
os.environ.get("GIT_DIR", os.path.join(self._repo_path, ".git"))
):
writemsg_level(
"ERROR: --update-changelogs supported only in git repos\n",
level=logging.ERROR,
noiselevel=-1,
)
self.returncode = 127
return
for cp in self._portdb.cp_all():
yield AsyncFunction(target=self.generate_changelog, args=[cp])
def run(self):
return run_main_scheduler(
TaskScheduler(
self._task_iter(),
event_loop=global_event_loop(),
max_jobs=self._max_jobs,
max_load=self._max_load,
)
)
def egencache_main(args):
# The calling environment is ignored, so the program is
# completely controlled by commandline arguments.
env = {}
if not sys.stdout.isatty() or os.environ.get("NOCOLOR", "").lower() in (
"yes",
"true",
):
portage.output.nocolor()
env["NOCOLOR"] = "true"
parser, options, atoms = parse_args(args)
config_root = options.config_root
if options.repositories_configuration is not None:
env["PORTAGE_REPOSITORIES"] = options.repositories_configuration
if options.cache_dir is not None:
env["PORTAGE_DEPCACHEDIR"] = options.cache_dir
settings = portage.config(config_root=config_root, local_config=False, env=env)
default_opts = None
if not options.ignore_default_opts:
default_opts = portage.util.shlex_split(
settings.get("EGENCACHE_DEFAULT_OPTS", "")
)
if default_opts:
parser, options, args = parse_args(default_opts + args)
if options.cache_dir is not None:
env["PORTAGE_DEPCACHEDIR"] = options.cache_dir
settings = portage.config(config_root=config_root, local_config=False, env=env)
if not (
options.update
or options.update_use_local_desc
or options.update_changelogs
or options.update_manifests
or options.update_pkg_desc_index
):
parser.error("No action specified")
return 1
if options.repo is None:
if len(settings.repositories.prepos) == 2:
for repo in settings.repositories:
if repo.name != "DEFAULT":
options.repo = repo.name
break
if options.repo is None:
parser.error("--repo option is required")
repo_path = settings.repositories.treemap.get(options.repo)
if repo_path is None:
parser.error("Unable to locate repository named '%s'" % (options.repo,))
return 1
repo_config = settings.repositories.get_repo_for_location(repo_path)
if options.strict_manifests is not None:
if options.strict_manifests == "y":
settings.features.add("strict")
else:
settings.features.discard("strict")
if options.update and "metadata-transfer" not in settings.features:
# Forcibly enable metadata-transfer if portdbapi has a pregenerated
# cache that does not support eclass validation.
cache = repo_config.get_pregenerated_cache(
portage.dbapi.dbapi._known_keys, readonly=True
)
if cache is not None and not cache.complete_eclass_entries:
settings.features.add("metadata-transfer")
cache = None
settings.lock()
portdb = portage.portdbapi(mysettings=settings)
# Limit ebuilds to the specified repo.
portdb.porttrees = [repo_path]
if options.update:
if options.cache_dir is not None:
# already validated earlier
pass
else:
# We check write access after the portdbapi constructor
# has had an opportunity to create it. This ensures that
# we don't use the cache in the "volatile" mode which is
# undesirable for egencache.
if not os.access(settings["PORTAGE_DEPCACHEDIR"], os.W_OK):
writemsg_level(
"ecachegen: error: "
+ "write access denied: %s\n" % (settings["PORTAGE_DEPCACHEDIR"],),
level=logging.ERROR,
noiselevel=-1,
)
return 1
if options.sign_manifests is not None:
repo_config.sign_manifest = options.sign_manifests == "y"
if options.thin_manifests is not None:
repo_config.thin_manifest = options.thin_manifests == "y"
gpg_cmd = None
gpg_vars = None
force_sign_key = None
if options.update_manifests:
if repo_config.sign_manifest:
sign_problem = False
gpg_dir = None
gpg_cmd = settings.get("PORTAGE_GPG_SIGNING_COMMAND")
if gpg_cmd is None:
writemsg_level(
"egencache: error: "
"PORTAGE_GPG_SIGNING_COMMAND is unset! "
"Is make.globals missing?\n",
level=logging.ERROR,
noiselevel=-1,
)
sign_problem = True
elif (
"${PORTAGE_GPG_KEY}" in gpg_cmd
and options.gpg_key is None
and "PORTAGE_GPG_KEY" not in settings
):
writemsg_level(
"egencache: error: " "PORTAGE_GPG_KEY is unset!\n",
level=logging.ERROR,
noiselevel=-1,
)
sign_problem = True
elif "${PORTAGE_GPG_DIR}" in gpg_cmd:
if options.gpg_dir is not None:
gpg_dir = options.gpg_dir
elif "PORTAGE_GPG_DIR" not in settings:
gpg_dir = os.path.expanduser("~/.gnupg")
else:
gpg_dir = os.path.expanduser(settings["PORTAGE_GPG_DIR"])
if not os.access(gpg_dir, os.X_OK):
writemsg_level(
(
"egencache: error: "
"Unable to access directory: "
"PORTAGE_GPG_DIR='%s'\n"
)
% gpg_dir,
level=logging.ERROR,
noiselevel=-1,
)
sign_problem = True
if sign_problem:
writemsg_level(
"egencache: You may disable manifest "
"signatures with --sign-manifests=n or by setting "
'"sign-manifests = false" in metadata/layout.conf\n',
level=logging.ERROR,
noiselevel=-1,
)
return 1
gpg_vars = {}
if gpg_dir is not None:
gpg_vars["PORTAGE_GPG_DIR"] = gpg_dir
gpg_var_names = []
if options.gpg_key is None:
gpg_var_names.append("PORTAGE_GPG_KEY")
else:
gpg_vars["PORTAGE_GPG_KEY"] = options.gpg_key
for k in gpg_var_names:
v = settings.get(k)
if v is not None:
gpg_vars[k] = v
force_sign_key = gpg_vars.get("PORTAGE_GPG_KEY")
ret = [os.EX_OK]
if options.update:
cp_iter = None
if atoms:
cp_iter = iter(atoms)
gen_cache = GenCache(
portdb,
cp_iter=cp_iter,
max_jobs=options.jobs,
max_load=options.load_average,
rsync=options.rsync,
external_cache_only=options.external_cache_only,
)
gen_cache.run()
if options.tolerant:
ret.append(os.EX_OK)
else:
ret.append(gen_cache.returncode)
if options.update_pkg_desc_index:
if not options.external_cache_only and repo_config.writable:
writable_location = repo_config.location
else:
writable_location = os.path.join(
portdb.depcachedir, repo_config.location.lstrip(os.sep)
)
if not options.external_cache_only:
msg = [
"WARNING: Repository is not writable: %s" % (repo_config.location,),
" Using cache directory instead: %s" % (writable_location,),
]
msg = "".join(line + "\n" for line in msg)
writemsg_level(msg, level=logging.WARNING, noiselevel=-1)
gen_index = GenPkgDescIndex(
repo_config,
portdb,
os.path.join(writable_location, "metadata", "pkg_desc_index"),
verbose=options.verbose,
)
gen_index.run()
ret.append(gen_index.returncode)
if options.update_use_local_desc:
gen_desc = GenUseLocalDesc(
portdb,
output=options.uld_output,
preserve_comments=options.preserve_comments,
)
gen_desc.run()
ret.append(gen_desc.returncode)
if options.update_changelogs:
gen_clogs = GenChangeLogs(
portdb,
changelog_output=options.changelog_output,
changelog_reversed=options.changelog_reversed,
max_jobs=options.jobs,
max_load=options.load_average,
)
signum = gen_clogs.run()
if signum is not None:
sys.exit(128 + signum)
ret.append(gen_clogs.returncode)
if options.update_manifests:
cp_iter = None
if atoms:
cp_iter = iter(atoms)
event_loop = global_event_loop()
scheduler = ManifestScheduler(
portdb,
cp_iter=cp_iter,
gpg_cmd=gpg_cmd,
gpg_vars=gpg_vars,
force_sign_key=force_sign_key,
max_jobs=options.jobs,
max_load=options.load_average,
event_loop=event_loop,
)
signum = run_main_scheduler(scheduler)
if signum is not None:
sys.exit(128 + signum)
if options.tolerant:
ret.append(os.EX_OK)
else:
ret.append(scheduler.returncode)
if options.write_timestamp:
timestamp_path = os.path.join(repo_path, "metadata", "timestamp.chk")
try:
portage.util.write_atomic(
timestamp_path, time.strftime("%s\n" % TIMESTAMP_FORMAT, time.gmtime())
)
except (EnvironmentError, portage.exception.PortageException):
ret.append(os.EX_IOERR)
else:
ret.append(os.EX_OK)
return max(ret)
if __name__ == "__main__":
portage._disable_legacy_globals()
portage.util.noiselimit = -1
try:
sys.exit(egencache_main(sys.argv[1:]))
finally:
global_event_loop().close()