| # -*- coding: utf-8 -*- |
| # Copyright 2015 The Chromium OS Authors. All rights reserved. |
| # Use of this source code is governed by a BSD-style license that can be |
| # found in the LICENSE file. |
| |
| """Deploy packages onto a target device. |
| |
| Integration tests for this file can be found at cli/cros/tests/cros_vm_tests.py. |
| See that file for more information. |
| """ |
| |
| from __future__ import division |
| from __future__ import print_function |
| |
| import bz2 |
| import fnmatch |
| import functools |
| import json |
| import os |
| import sys |
| import tempfile |
| |
| from chromite.cli import command |
| from chromite.lib import cros_build_lib |
| from chromite.lib import cros_logging as logging |
| from chromite.lib import operation |
| from chromite.lib import osutils |
| from chromite.lib import portage_util |
| from chromite.lib import remote_access |
| from chromite.lib import dlc_lib |
| from chromite.lib import workon_helper |
| try: |
| import portage |
| except ImportError: |
| if cros_build_lib.IsInsideChroot(): |
| raise |
| |
| |
| assert sys.version_info >= (3, 6), 'This module requires Python 3.6+' |
| |
| |
| _DEVICE_BASE_DIR = '/usr/local/tmp/cros-deploy' |
| # This is defined in src/platform/dev/builder.py |
| _STRIPPED_PACKAGES_DIR = 'stripped-packages' |
| |
| _MAX_UPDATES_NUM = 10 |
| _MAX_UPDATES_WARNING = ( |
| 'You are about to update a large number of installed packages, which ' |
| 'might take a long time, fail midway, or leave the target in an ' |
| 'inconsistent state. It is highly recommended that you flash a new image ' |
| 'instead.') |
| |
| _DLC_ID = 'DLC_ID' |
| _DLC_PACKAGE = 'DLC_PACKAGE' |
| _DLC_ENABLED = 'DLC_ENABLED' |
| _ENVIRONMENT_FILENAME = 'environment.bz2' |
| _DLC_INSTALL_ROOT = '/var/cache/dlc' |
| |
| |
| class DeployError(Exception): |
| """Thrown when an unrecoverable error is encountered during deploy.""" |
| |
| |
| class BrilloDeployOperation(operation.ProgressBarOperation): |
| """ProgressBarOperation specific for brillo deploy.""" |
| # These two variables are used to validate the output in the VM integration |
| # tests. Changes to the output must be reflected here. |
| MERGE_EVENTS = ['NOTICE: Copying', 'NOTICE: Installing', 'WARNING: Ignoring', |
| 'emerge --usepkg', 'has been installed.'] |
| UNMERGE_EVENTS = ['NOTICE: Unmerging', 'has been uninstalled.'] |
| |
| def __init__(self, pkg_count, emerge): |
| """Construct BrilloDeployOperation object. |
| |
| Args: |
| pkg_count: number of packages being built. |
| emerge: True if emerge, False is unmerge. |
| """ |
| super(BrilloDeployOperation, self).__init__() |
| if emerge: |
| self._events = self.MERGE_EVENTS |
| else: |
| self._events = self.UNMERGE_EVENTS |
| self._total = pkg_count * len(self._events) |
| self._completed = 0 |
| |
| def ParseOutput(self, output=None): |
| """Parse the output of brillo deploy to update a progress bar.""" |
| stdout = self._stdout.read() |
| stderr = self._stderr.read() |
| output = stdout + stderr |
| for event in self._events: |
| self._completed += output.count(event) |
| self.ProgressBar(self._completed / self._total) |
| |
| |
| class _InstallPackageScanner(object): |
| """Finds packages that need to be installed on a target device. |
| |
| Scans the sysroot bintree, beginning with a user-provided list of packages, |
| to find all packages that need to be installed. If so instructed, |
| transitively scans forward (mandatory) and backward (optional) dependencies |
| as well. A package will be installed if missing on the target (mandatory |
| packages only), or it will be updated if its sysroot version and build time |
| are different from the target. Common usage: |
| |
| pkg_scanner = _InstallPackageScanner(sysroot) |
| pkgs = pkg_scanner.Run(...) |
| """ |
| |
| class VartreeError(Exception): |
| """An error in the processing of the installed packages tree.""" |
| |
| class BintreeError(Exception): |
| """An error in the processing of the source binpkgs tree.""" |
| |
| class PkgInfo(object): |
| """A record containing package information.""" |
| |
| __slots__ = ('cpv', 'build_time', 'rdeps_raw', 'rdeps', 'rev_rdeps') |
| |
| def __init__(self, cpv, build_time, rdeps_raw, rdeps=None, rev_rdeps=None): |
| self.cpv = cpv |
| self.build_time = build_time |
| self.rdeps_raw = rdeps_raw |
| self.rdeps = set() if rdeps is None else rdeps |
| self.rev_rdeps = set() if rev_rdeps is None else rev_rdeps |
| |
| # Python snippet for dumping vartree info on the target. Instantiate using |
| # _GetVartreeSnippet(). |
| _GET_VARTREE = """ |
| import json |
| import os |
| import portage |
| |
| # Normalize the path to match what portage will index. |
| target_root = os.path.normpath('%(root)s') |
| if not target_root.endswith('/'): |
| target_root += '/' |
| trees = portage.create_trees(target_root=target_root, config_root='/') |
| vartree = trees[target_root]['vartree'] |
| pkg_info = [] |
| for cpv in vartree.dbapi.cpv_all(): |
| slot, rdep_raw, build_time = vartree.dbapi.aux_get( |
| cpv, ('SLOT', 'RDEPEND', 'BUILD_TIME')) |
| pkg_info.append((cpv, slot, rdep_raw, build_time)) |
| |
| print(json.dumps(pkg_info)) |
| """ |
| |
| def __init__(self, sysroot): |
| self.sysroot = sysroot |
| # Members containing the sysroot (binpkg) and target (installed) package DB. |
| self.target_db = None |
| self.binpkgs_db = None |
| # Members for managing the dependency resolution work queue. |
| self.queue = None |
| self.seen = None |
| self.listed = None |
| |
| @staticmethod |
| def _GetCP(cpv): |
| """Returns the CP value for a given CPV string.""" |
| attrs = portage_util.SplitCPV(cpv, strict=False) |
| if not attrs.cp: |
| raise ValueError('Cannot get CP value for %s' % cpv) |
| return attrs.cp |
| |
| @staticmethod |
| def _InDB(cp, slot, db): |
| """Returns whether CP and slot are found in a database (if provided).""" |
| cp_slots = db.get(cp) if db else None |
| return cp_slots is not None and (not slot or slot in cp_slots) |
| |
| @staticmethod |
| def _AtomStr(cp, slot): |
| """Returns 'CP:slot' if slot is non-empty, else just 'CP'.""" |
| return '%s:%s' % (cp, slot) if slot else cp |
| |
| @classmethod |
| def _GetVartreeSnippet(cls, root='/'): |
| """Returns a code snippet for dumping the vartree on the target. |
| |
| Args: |
| root: The installation root. |
| |
| Returns: |
| The said code snippet (string) with parameters filled in. |
| """ |
| return cls._GET_VARTREE % {'root': root} |
| |
| @classmethod |
| def _StripDepAtom(cls, dep_atom, installed_db=None): |
| """Strips a dependency atom and returns a (CP, slot) pair.""" |
| # TODO(garnold) This is a gross simplification of ebuild dependency |
| # semantics, stripping and ignoring various qualifiers (versions, slots, |
| # USE flag, negation) and will likely need to be fixed. chromium:447366. |
| |
| # Ignore unversioned blockers, leaving them for the user to resolve. |
| if dep_atom[0] == '!' and dep_atom[1] not in '<=>~': |
| return None, None |
| |
| cp = dep_atom |
| slot = None |
| require_installed = False |
| |
| # Versioned blockers should be updated, but only if already installed. |
| # These are often used for forcing cascaded updates of multiple packages, |
| # so we're treating them as ordinary constraints with hopes that it'll lead |
| # to the desired result. |
| if cp.startswith('!'): |
| cp = cp.lstrip('!') |
| require_installed = True |
| |
| # Remove USE flags. |
| if '[' in cp: |
| cp = cp[:cp.index('[')] + cp[cp.index(']') + 1:] |
| |
| # Separate the slot qualifier and strip off subslots. |
| if ':' in cp: |
| cp, slot = cp.split(':') |
| for delim in ('/', '='): |
| slot = slot.split(delim, 1)[0] |
| |
| # Strip version wildcards (right), comparators (left). |
| cp = cp.rstrip('*') |
| cp = cp.lstrip('<=>~') |
| |
| # Turn into CP form. |
| cp = cls._GetCP(cp) |
| |
| if require_installed and not cls._InDB(cp, None, installed_db): |
| return None, None |
| |
| return cp, slot |
| |
| @classmethod |
| def _ProcessDepStr(cls, dep_str, installed_db, avail_db): |
| """Resolves and returns a list of dependencies from a dependency string. |
| |
| This parses a dependency string and returns a list of package names and |
| slots. Other atom qualifiers (version, sub-slot, block) are ignored. When |
| resolving disjunctive deps, we include all choices that are fully present |
| in |installed_db|. If none is present, we choose an arbitrary one that is |
| available. |
| |
| Args: |
| dep_str: A raw dependency string. |
| installed_db: A database of installed packages. |
| avail_db: A database of packages available for installation. |
| |
| Returns: |
| A list of pairs (CP, slot). |
| |
| Raises: |
| ValueError: the dependencies string is malformed. |
| """ |
| def ProcessSubDeps(dep_exp, disjunct): |
| """Parses and processes a dependency (sub)expression.""" |
| deps = set() |
| default_deps = set() |
| sub_disjunct = False |
| for dep_sub_exp in dep_exp: |
| sub_deps = set() |
| |
| if isinstance(dep_sub_exp, (list, tuple)): |
| sub_deps = ProcessSubDeps(dep_sub_exp, sub_disjunct) |
| sub_disjunct = False |
| elif sub_disjunct: |
| raise ValueError('Malformed disjunctive operation in deps') |
| elif dep_sub_exp == '||': |
| sub_disjunct = True |
| elif dep_sub_exp.endswith('?'): |
| raise ValueError('Dependencies contain a conditional') |
| else: |
| cp, slot = cls._StripDepAtom(dep_sub_exp, installed_db) |
| if cp: |
| sub_deps = set([(cp, slot)]) |
| elif disjunct: |
| raise ValueError('Atom in disjunct ignored') |
| |
| # Handle sub-deps of a disjunctive expression. |
| if disjunct: |
| # Make the first available choice the default, for use in case that |
| # no option is installed. |
| if (not default_deps and avail_db is not None and |
| all([cls._InDB(cp, slot, avail_db) for cp, slot in sub_deps])): |
| default_deps = sub_deps |
| |
| # If not all sub-deps are installed, then don't consider them. |
| if not all([cls._InDB(cp, slot, installed_db) |
| for cp, slot in sub_deps]): |
| sub_deps = set() |
| |
| deps.update(sub_deps) |
| |
| return deps or default_deps |
| |
| try: |
| return ProcessSubDeps(portage.dep.paren_reduce(dep_str), False) |
| except portage.exception.InvalidDependString as e: |
| raise ValueError('Invalid dep string: %s' % e) |
| except ValueError as e: |
| raise ValueError('%s: %s' % (e, dep_str)) |
| |
| def _BuildDB(self, cpv_info, process_rdeps, process_rev_rdeps, |
| installed_db=None): |
| """Returns a database of packages given a list of CPV info. |
| |
| Args: |
| cpv_info: A list of tuples containing package CPV and attributes. |
| process_rdeps: Whether to populate forward dependencies. |
| process_rev_rdeps: Whether to populate reverse dependencies. |
| installed_db: A database of installed packages for filtering disjunctive |
| choices against; if None, using own built database. |
| |
| Returns: |
| A map from CP values to another dictionary that maps slots to package |
| attribute tuples. Tuples contain a CPV value (string), build time |
| (string), runtime dependencies (set), and reverse dependencies (set, |
| empty if not populated). |
| |
| Raises: |
| ValueError: If more than one CPV occupies a single slot. |
| """ |
| db = {} |
| logging.debug('Populating package DB...') |
| for cpv, slot, rdeps_raw, build_time in cpv_info: |
| cp = self._GetCP(cpv) |
| cp_slots = db.setdefault(cp, dict()) |
| if slot in cp_slots: |
| raise ValueError('More than one package found for %s' % |
| self._AtomStr(cp, slot)) |
| logging.debug(' %s -> %s, built %s, raw rdeps: %s', |
| self._AtomStr(cp, slot), cpv, build_time, rdeps_raw) |
| cp_slots[slot] = self.PkgInfo(cpv, build_time, rdeps_raw) |
| |
| avail_db = db |
| if installed_db is None: |
| installed_db = db |
| avail_db = None |
| |
| # Add approximate forward dependencies. |
| if process_rdeps: |
| logging.debug('Populating forward dependencies...') |
| for cp, cp_slots in db.items(): |
| for slot, pkg_info in cp_slots.items(): |
| pkg_info.rdeps.update(self._ProcessDepStr(pkg_info.rdeps_raw, |
| installed_db, avail_db)) |
| logging.debug(' %s (%s) processed rdeps: %s', |
| self._AtomStr(cp, slot), pkg_info.cpv, |
| ' '.join([self._AtomStr(rdep_cp, rdep_slot) |
| for rdep_cp, rdep_slot in pkg_info.rdeps])) |
| |
| # Add approximate reverse dependencies (optional). |
| if process_rev_rdeps: |
| logging.debug('Populating reverse dependencies...') |
| for cp, cp_slots in db.items(): |
| for slot, pkg_info in cp_slots.items(): |
| for rdep_cp, rdep_slot in pkg_info.rdeps: |
| to_slots = db.get(rdep_cp) |
| if not to_slots: |
| continue |
| |
| for to_slot, to_pkg_info in to_slots.items(): |
| if rdep_slot and to_slot != rdep_slot: |
| continue |
| logging.debug(' %s (%s) added as rev rdep for %s (%s)', |
| self._AtomStr(cp, slot), pkg_info.cpv, |
| self._AtomStr(rdep_cp, to_slot), to_pkg_info.cpv) |
| to_pkg_info.rev_rdeps.add((cp, slot)) |
| |
| return db |
| |
| def _InitTargetVarDB(self, device, root, process_rdeps, process_rev_rdeps): |
| """Initializes a dictionary of packages installed on |device|.""" |
| get_vartree_script = self._GetVartreeSnippet(root) |
| try: |
| result = device.GetAgent().RemoteSh(['python'], remote_sudo=True, |
| input=get_vartree_script) |
| except cros_build_lib.RunCommandError as e: |
| logging.error('Cannot get target vartree:\n%s', e.result.error) |
| raise |
| |
| try: |
| self.target_db = self._BuildDB(json.loads(result.output), |
| process_rdeps, process_rev_rdeps) |
| except ValueError as e: |
| raise self.VartreeError(str(e)) |
| |
| def _InitBinpkgDB(self, process_rdeps): |
| """Initializes a dictionary of binary packages for updating the target.""" |
| # Get build root trees; portage indexes require a trailing '/'. |
| build_root = os.path.join(self.sysroot, '') |
| trees = portage.create_trees(target_root=build_root, config_root=build_root) |
| bintree = trees[build_root]['bintree'] |
| binpkgs_info = [] |
| for cpv in bintree.dbapi.cpv_all(): |
| slot, rdep_raw, build_time = bintree.dbapi.aux_get( |
| cpv, ['SLOT', 'RDEPEND', 'BUILD_TIME']) |
| binpkgs_info.append((cpv, slot, rdep_raw, build_time)) |
| |
| try: |
| self.binpkgs_db = self._BuildDB(binpkgs_info, process_rdeps, False, |
| installed_db=self.target_db) |
| except ValueError as e: |
| raise self.BintreeError(str(e)) |
| |
| def _InitDepQueue(self): |
| """Initializes the dependency work queue.""" |
| self.queue = set() |
| self.seen = {} |
| self.listed = set() |
| |
| def _EnqDep(self, dep, listed, optional): |
| """Enqueues a dependency if not seen before or if turned non-optional.""" |
| if dep in self.seen and (optional or not self.seen[dep]): |
| return False |
| |
| self.queue.add(dep) |
| self.seen[dep] = optional |
| if listed: |
| self.listed.add(dep) |
| return True |
| |
| def _DeqDep(self): |
| """Dequeues and returns a dependency, its listed and optional flags. |
| |
| This returns listed packages first, if any are present, to ensure that we |
| correctly mark them as such when they are first being processed. |
| """ |
| if self.listed: |
| dep = self.listed.pop() |
| self.queue.remove(dep) |
| listed = True |
| else: |
| dep = self.queue.pop() |
| listed = False |
| |
| return dep, listed, self.seen[dep] |
| |
| def _FindPackageMatches(self, cpv_pattern): |
| """Returns list of binpkg (CP, slot) pairs that match |cpv_pattern|. |
| |
| This is breaking |cpv_pattern| into its C, P and V components, each of |
| which may or may not be present or contain wildcards. It then scans the |
| binpkgs database to find all atoms that match these components, returning a |
| list of CP and slot qualifier. When the pattern does not specify a version, |
| or when a CP has only one slot in the binpkgs database, we omit the slot |
| qualifier in the result. |
| |
| Args: |
| cpv_pattern: A CPV pattern, potentially partial and/or having wildcards. |
| |
| Returns: |
| A list of (CPV, slot) pairs of packages in the binpkgs database that |
| match the pattern. |
| """ |
| attrs = portage_util.SplitCPV(cpv_pattern, strict=False) |
| cp_pattern = os.path.join(attrs.category or '*', attrs.package or '*') |
| matches = [] |
| for cp, cp_slots in self.binpkgs_db.items(): |
| if not fnmatch.fnmatchcase(cp, cp_pattern): |
| continue |
| |
| # If no version attribute was given or there's only one slot, omit the |
| # slot qualifier. |
| if not attrs.version or len(cp_slots) == 1: |
| matches.append((cp, None)) |
| else: |
| cpv_pattern = '%s-%s' % (cp, attrs.version) |
| for slot, pkg_info in cp_slots.items(): |
| if fnmatch.fnmatchcase(pkg_info.cpv, cpv_pattern): |
| matches.append((cp, slot)) |
| |
| return matches |
| |
| def _FindPackage(self, pkg): |
| """Returns the (CP, slot) pair for a package matching |pkg|. |
| |
| Args: |
| pkg: Path to a binary package or a (partial) package CPV specifier. |
| |
| Returns: |
| A (CP, slot) pair for the given package; slot may be None (unspecified). |
| |
| Raises: |
| ValueError: if |pkg| is not a binpkg file nor does it match something |
| that's in the bintree. |
| """ |
| if pkg.endswith('.tbz2') and os.path.isfile(pkg): |
| package = os.path.basename(os.path.splitext(pkg)[0]) |
| category = os.path.basename(os.path.dirname(pkg)) |
| return self._GetCP(os.path.join(category, package)), None |
| |
| matches = self._FindPackageMatches(pkg) |
| if not matches: |
| raise ValueError('No package found for %s' % pkg) |
| |
| idx = 0 |
| if len(matches) > 1: |
| # Ask user to pick among multiple matches. |
| idx = cros_build_lib.GetChoice('Multiple matches found for %s: ' % pkg, |
| ['%s:%s' % (cp, slot) if slot else cp |
| for cp, slot in matches]) |
| |
| return matches[idx] |
| |
| def _NeedsInstall(self, cpv, slot, build_time, optional): |
| """Returns whether a package needs to be installed on the target. |
| |
| Args: |
| cpv: Fully qualified CPV (string) of the package. |
| slot: Slot identifier (string). |
| build_time: The BUILT_TIME value (string) of the binpkg. |
| optional: Whether package is optional on the target. |
| |
| Returns: |
| A tuple (install, update) indicating whether to |install| the package and |
| whether it is an |update| to an existing package. |
| |
| Raises: |
| ValueError: if slot is not provided. |
| """ |
| # If not checking installed packages, always install. |
| if not self.target_db: |
| return True, False |
| |
| cp = self._GetCP(cpv) |
| target_pkg_info = self.target_db.get(cp, dict()).get(slot) |
| if target_pkg_info is not None: |
| if cpv != target_pkg_info.cpv: |
| attrs = portage_util.SplitCPV(cpv) |
| target_attrs = portage_util.SplitCPV(target_pkg_info.cpv) |
| logging.debug('Updating %s: version (%s) different on target (%s)', |
| cp, attrs.version, target_attrs.version) |
| return True, True |
| |
| if build_time != target_pkg_info.build_time: |
| logging.debug('Updating %s: build time (%s) different on target (%s)', |
| cpv, build_time, target_pkg_info.build_time) |
| return True, True |
| |
| logging.debug('Not updating %s: already up-to-date (%s, built %s)', |
| cp, target_pkg_info.cpv, target_pkg_info.build_time) |
| return False, False |
| |
| if optional: |
| logging.debug('Not installing %s: missing on target but optional', cp) |
| return False, False |
| |
| logging.debug('Installing %s: missing on target and non-optional (%s)', |
| cp, cpv) |
| return True, False |
| |
| def _ProcessDeps(self, deps, reverse): |
| """Enqueues dependencies for processing. |
| |
| Args: |
| deps: List of dependencies to enqueue. |
| reverse: Whether these are reverse dependencies. |
| """ |
| if not deps: |
| return |
| |
| logging.debug('Processing %d %s dep(s)...', len(deps), |
| 'reverse' if reverse else 'forward') |
| num_already_seen = 0 |
| for dep in deps: |
| if self._EnqDep(dep, False, reverse): |
| logging.debug(' Queued dep %s', dep) |
| else: |
| num_already_seen += 1 |
| |
| if num_already_seen: |
| logging.debug('%d dep(s) already seen', num_already_seen) |
| |
| def _ComputeInstalls(self, process_rdeps, process_rev_rdeps): |
| """Returns a dictionary of packages that need to be installed on the target. |
| |
| Args: |
| process_rdeps: Whether to trace forward dependencies. |
| process_rev_rdeps: Whether to trace backward dependencies as well. |
| |
| Returns: |
| A dictionary mapping CP values (string) to tuples containing a CPV |
| (string), a slot (string), a boolean indicating whether the package |
| was initially listed in the queue, and a boolean indicating whether this |
| is an update to an existing package. |
| """ |
| installs = {} |
| while self.queue: |
| dep, listed, optional = self._DeqDep() |
| cp, required_slot = dep |
| if cp in installs: |
| logging.debug('Already updating %s', cp) |
| continue |
| |
| cp_slots = self.binpkgs_db.get(cp, dict()) |
| logging.debug('Checking packages matching %s%s%s...', cp, |
| ' (slot: %s)' % required_slot if required_slot else '', |
| ' (optional)' if optional else '') |
| num_processed = 0 |
| for slot, pkg_info in cp_slots.items(): |
| if required_slot and slot != required_slot: |
| continue |
| |
| num_processed += 1 |
| logging.debug(' Checking %s...', pkg_info.cpv) |
| |
| install, update = self._NeedsInstall(pkg_info.cpv, slot, |
| pkg_info.build_time, optional) |
| if not install: |
| continue |
| |
| installs[cp] = (pkg_info.cpv, slot, listed, update) |
| |
| # Add forward and backward runtime dependencies to queue. |
| if process_rdeps: |
| self._ProcessDeps(pkg_info.rdeps, False) |
| if process_rev_rdeps: |
| target_pkg_info = self.target_db.get(cp, dict()).get(slot) |
| if target_pkg_info: |
| self._ProcessDeps(target_pkg_info.rev_rdeps, True) |
| |
| if num_processed == 0: |
| logging.warning('No qualified bintree package corresponding to %s', cp) |
| |
| return installs |
| |
| def _SortInstalls(self, installs): |
| """Returns a sorted list of packages to install. |
| |
| Performs a topological sort based on dependencies found in the binary |
| package database. |
| |
| Args: |
| installs: Dictionary of packages to install indexed by CP. |
| |
| Returns: |
| A list of package CPVs (string). |
| |
| Raises: |
| ValueError: If dependency graph contains a cycle. |
| """ |
| not_visited = set(installs.keys()) |
| curr_path = [] |
| sorted_installs = [] |
| |
| def SortFrom(cp): |
| """Traverses dependencies recursively, emitting nodes in reverse order.""" |
| cpv, slot, _, _ = installs[cp] |
| if cpv in curr_path: |
| raise ValueError('Dependencies contain a cycle: %s -> %s' % |
| (' -> '.join(curr_path[curr_path.index(cpv):]), cpv)) |
| curr_path.append(cpv) |
| for rdep_cp, _ in self.binpkgs_db[cp][slot].rdeps: |
| if rdep_cp in not_visited: |
| not_visited.remove(rdep_cp) |
| SortFrom(rdep_cp) |
| |
| sorted_installs.append(cpv) |
| curr_path.pop() |
| |
| # So long as there's more packages, keep expanding dependency paths. |
| while not_visited: |
| SortFrom(not_visited.pop()) |
| |
| return sorted_installs |
| |
| def _EnqListedPkg(self, pkg): |
| """Finds and enqueues a listed package.""" |
| cp, slot = self._FindPackage(pkg) |
| if cp not in self.binpkgs_db: |
| raise self.BintreeError('Package %s not found in binpkgs tree' % pkg) |
| self._EnqDep((cp, slot), True, False) |
| |
| def _EnqInstalledPkgs(self): |
| """Enqueues all available binary packages that are already installed.""" |
| for cp, cp_slots in self.binpkgs_db.items(): |
| target_cp_slots = self.target_db.get(cp) |
| if target_cp_slots: |
| for slot in cp_slots.keys(): |
| if slot in target_cp_slots: |
| self._EnqDep((cp, slot), True, False) |
| |
| def Run(self, device, root, listed_pkgs, update, process_rdeps, |
| process_rev_rdeps): |
| """Computes the list of packages that need to be installed on a target. |
| |
| Args: |
| device: Target handler object. |
| root: Package installation root. |
| listed_pkgs: Package names/files listed by the user. |
| update: Whether to read the target's installed package database. |
| process_rdeps: Whether to trace forward dependencies. |
| process_rev_rdeps: Whether to trace backward dependencies as well. |
| |
| Returns: |
| A tuple (sorted, listed, num_updates, install_attrs) where |sorted| is a |
| list of package CPVs (string) to install on the target in an order that |
| satisfies their inter-dependencies, |listed| the subset that was |
| requested by the user, and |num_updates| the number of packages being |
| installed over preexisting versions. Note that installation order should |
| be reversed for removal, |install_attrs| is a dictionary mapping a package |
| CPV (string) to some of its extracted environment attributes. |
| """ |
| if process_rev_rdeps and not process_rdeps: |
| raise ValueError('Must processing forward deps when processing rev deps') |
| if process_rdeps and not update: |
| raise ValueError('Must check installed packages when processing deps') |
| |
| if update: |
| logging.info('Initializing target intalled packages database...') |
| self._InitTargetVarDB(device, root, process_rdeps, process_rev_rdeps) |
| |
| logging.info('Initializing binary packages database...') |
| self._InitBinpkgDB(process_rdeps) |
| |
| logging.info('Finding listed package(s)...') |
| self._InitDepQueue() |
| for pkg in listed_pkgs: |
| if pkg == '@installed': |
| if not update: |
| raise ValueError( |
| 'Must check installed packages when updating all of them.') |
| self._EnqInstalledPkgs() |
| else: |
| self._EnqListedPkg(pkg) |
| |
| logging.info('Computing set of packages to install...') |
| installs = self._ComputeInstalls(process_rdeps, process_rev_rdeps) |
| |
| num_updates = 0 |
| listed_installs = [] |
| for cpv, _, listed, isupdate in installs.values(): |
| if listed: |
| listed_installs.append(cpv) |
| if isupdate: |
| num_updates += 1 |
| |
| logging.info('Processed %d package(s), %d will be installed, %d are ' |
| 'updating existing packages', |
| len(self.seen), len(installs), num_updates) |
| |
| sorted_installs = self._SortInstalls(installs) |
| |
| install_attrs = {} |
| for pkg in sorted_installs: |
| pkg_path = os.path.join(root, portage_util.VDB_PATH, pkg) |
| dlc_id, dlc_package = _GetDLCInfo(device, pkg_path, from_dut=True) |
| install_attrs[pkg] = {} |
| if dlc_id and dlc_package: |
| install_attrs[pkg][_DLC_ID] = dlc_id |
| |
| return sorted_installs, listed_installs, num_updates, install_attrs |
| |
| |
| def _Emerge(device, pkg_path, root, extra_args=None): |
| """Copies |pkg| to |device| and emerges it. |
| |
| Args: |
| device: A ChromiumOSDevice object. |
| pkg_path: A path to a binary package. |
| root: Package installation root path. |
| extra_args: Extra arguments to pass to emerge. |
| |
| Raises: |
| DeployError: Unrecoverable error during emerge. |
| """ |
| pkgroot = os.path.join(device.work_dir, 'packages') |
| pkg_name = os.path.basename(pkg_path) |
| pkg_dirname = os.path.basename(os.path.dirname(pkg_path)) |
| pkg_dir = os.path.join(pkgroot, pkg_dirname) |
| portage_tmpdir = os.path.join(device.work_dir, 'portage-tmp') |
| # Clean out the dirs first if we had a previous emerge on the device so as to |
| # free up space for this emerge. The last emerge gets implicitly cleaned up |
| # when the device connection deletes its work_dir. |
| device.run( |
| ['rm', '-rf', pkg_dir, portage_tmpdir, '&&', |
| 'mkdir', '-p', pkg_dir, portage_tmpdir], remote_sudo=True) |
| |
| # This message is read by BrilloDeployOperation. |
| logging.notice('Copying %s to device.', pkg_name) |
| device.CopyToDevice(pkg_path, pkg_dir, mode='rsync', remote_sudo=True) |
| |
| logging.info('Use portage temp dir %s', portage_tmpdir) |
| |
| # This message is read by BrilloDeployOperation. |
| logging.notice('Installing %s.', pkg_name) |
| pkg_path = os.path.join(pkg_dir, pkg_name) |
| |
| # We set PORTAGE_CONFIGROOT to '/usr/local' because by default all |
| # chromeos-base packages will be skipped due to the configuration |
| # in /etc/protage/make.profile/package.provided. However, there is |
| # a known bug that /usr/local/etc/portage is not setup properly |
| # (crbug.com/312041). This does not affect `cros deploy` because |
| # we do not use the preset PKGDIR. |
| extra_env = { |
| 'FEATURES': '-sandbox', |
| 'PKGDIR': pkgroot, |
| 'PORTAGE_CONFIGROOT': '/usr/local', |
| 'PORTAGE_TMPDIR': portage_tmpdir, |
| 'PORTDIR': device.work_dir, |
| 'CONFIG_PROTECT': '-*', |
| } |
| # --ignore-built-slot-operator-deps because we don't rebuild everything. |
| # It can cause errors, but that's expected with cros deploy since it's just a |
| # best effort to prevent developers avoid rebuilding an image every time. |
| cmd = ['emerge', '--usepkg', '--ignore-built-slot-operator-deps=y', pkg_path, |
| '--root=%s' % root] |
| if extra_args: |
| cmd.append(extra_args) |
| |
| logging.warning('Ignoring slot dependencies! This may break things! e.g. ' |
| 'packages built against the old version may not be able to ' |
| 'load the new .so. This is expected, and you will just need ' |
| 'to build and flash a new image if you have problems.') |
| try: |
| result = device.run(cmd, extra_env=extra_env, remote_sudo=True, |
| capture_output=True, debug_level=logging.INFO) |
| |
| pattern = ('A requested package will not be merged because ' |
| 'it is listed in package.provided') |
| output = result.error.replace('\n', ' ').replace('\r', '') |
| if pattern in output: |
| error = ('Package failed to emerge: %s\n' |
| 'Remove %s from /etc/portage/make.profile/' |
| 'package.provided/chromeos-base.packages\n' |
| '(also see crbug.com/920140 for more context)\n' |
| % (pattern, pkg_name)) |
| cros_build_lib.Die(error) |
| except Exception: |
| logging.error('Failed to emerge package %s', pkg_name) |
| raise |
| else: |
| logging.notice('%s has been installed.', pkg_name) |
| |
| |
| def _RestoreSELinuxContext(device, pkgpath, root): |
| """Restore SELinux context for files in a given package. |
| |
| This reads the tarball from pkgpath, and calls restorecon on device to |
| restore SELinux context for files listed in the tarball, assuming those files |
| are installed to / |
| |
| Args: |
| device: a ChromiumOSDevice object |
| pkgpath: path to tarball |
| root: Package installation root path. |
| """ |
| enforced = device.IsSELinuxEnforced() |
| if enforced: |
| device.run(['setenforce', '0']) |
| pkgroot = os.path.join(device.work_dir, 'packages') |
| pkg_dirname = os.path.basename(os.path.dirname(pkgpath)) |
| pkgpath_device = os.path.join(pkgroot, pkg_dirname, os.path.basename(pkgpath)) |
| # Testing shows restorecon splits on newlines instead of spaces. |
| device.run( |
| ['cd', root, '&&', |
| 'tar', 'tf', pkgpath_device, '|', |
| 'restorecon', '-i', '-f', '-'], |
| remote_sudo=True) |
| if enforced: |
| device.run(['setenforce', '1']) |
| |
| |
| def _GetPackagesByCPV(cpvs, strip, sysroot): |
| """Returns paths to binary packages corresponding to |cpvs|. |
| |
| Args: |
| cpvs: List of CPV components given by portage_util.SplitCPV(). |
| strip: True to run strip_package. |
| sysroot: Sysroot path. |
| |
| Returns: |
| List of paths corresponding to |cpvs|. |
| |
| Raises: |
| DeployError: If a package is missing. |
| """ |
| packages_dir = None |
| if strip: |
| try: |
| cros_build_lib.run( |
| ['strip_package', '--sysroot', sysroot] + |
| [cpv.cpf for cpv in cpvs]) |
| packages_dir = _STRIPPED_PACKAGES_DIR |
| except cros_build_lib.RunCommandError: |
| logging.error('Cannot strip packages %s', |
| ' '.join([str(cpv) for cpv in cpvs])) |
| raise |
| |
| paths = [] |
| for cpv in cpvs: |
| path = portage_util.GetBinaryPackagePath( |
| cpv.category, cpv.package, cpv.version, sysroot=sysroot, |
| packages_dir=packages_dir) |
| if not path: |
| raise DeployError('Missing package %s.' % cpv) |
| paths.append(path) |
| |
| return paths |
| |
| |
| def _GetPackagesPaths(pkgs, strip, sysroot): |
| """Returns paths to binary |pkgs|. |
| |
| Args: |
| pkgs: List of package CPVs string. |
| strip: Whether or not to run strip_package for CPV packages. |
| sysroot: The sysroot path. |
| |
| Returns: |
| List of paths corresponding to |pkgs|. |
| """ |
| cpvs = [portage_util.SplitCPV(p) for p in pkgs] |
| return _GetPackagesByCPV(cpvs, strip, sysroot) |
| |
| |
| def _Unmerge(device, pkg, root): |
| """Unmerges |pkg| on |device|. |
| |
| Args: |
| device: A RemoteDevice object. |
| pkg: A package name. |
| root: Package installation root path. |
| """ |
| pkg_name = os.path.basename(pkg) |
| # This message is read by BrilloDeployOperation. |
| logging.notice('Unmerging %s.', pkg_name) |
| cmd = ['qmerge', '--yes'] |
| # Check if qmerge is available on the device. If not, use emerge. |
| if device.run(['qmerge', '--version'], check=False).returncode != 0: |
| cmd = ['emerge'] |
| |
| cmd.extend(['--unmerge', pkg, '--root=%s' % root]) |
| try: |
| # Always showing the emerge output for clarity. |
| device.run(cmd, capture_output=False, remote_sudo=True, |
| debug_level=logging.INFO) |
| except Exception: |
| logging.error('Failed to unmerge package %s', pkg_name) |
| raise |
| else: |
| logging.notice('%s has been uninstalled.', pkg_name) |
| |
| |
| def _ConfirmDeploy(num_updates): |
| """Returns whether we can continue deployment.""" |
| if num_updates > _MAX_UPDATES_NUM: |
| logging.warning(_MAX_UPDATES_WARNING) |
| return cros_build_lib.BooleanPrompt(default=False) |
| |
| return True |
| |
| |
| def _EmergePackages(pkgs, device, strip, sysroot, root, board, emerge_args): |
| """Call _Emerge for each package in pkgs.""" |
| dlc_deployed = False |
| for pkg_path in _GetPackagesPaths(pkgs, strip, sysroot): |
| _Emerge(device, pkg_path, root, extra_args=emerge_args) |
| if device.IsSELinuxAvailable(): |
| _RestoreSELinuxContext(device, pkg_path, root) |
| |
| dlc_id, dlc_package = _GetDLCInfo(device, pkg_path, from_dut=False) |
| if dlc_id and dlc_package: |
| _DeployDLCImage(device, sysroot, board, dlc_id, dlc_package) |
| dlc_deployed = True |
| # Clean up empty directories created by emerging DLCs. |
| device.run(['test', '-d', '/build/rootfs', '&&', 'rmdir', |
| '--ignore-fail-on-non-empty', '/build/rootfs', '/build'], |
| check=False) |
| |
| # Restart dlcservice so it picks up the newly installed DLC modules (in case |
| # we installed new DLC images). |
| if dlc_deployed: |
| device.run(['restart', 'dlcservice']) |
| |
| |
| def _UnmergePackages(pkgs, device, root, pkgs_attrs): |
| """Call _Unmege for each package in pkgs.""" |
| dlc_uninstalled = False |
| for pkg in pkgs: |
| _Unmerge(device, pkg, root) |
| if _UninstallDLCImage(device, pkgs_attrs[pkg]): |
| dlc_uninstalled = True |
| |
| # Restart dlcservice so it picks up the uninstalled DLC modules (in case we |
| # uninstalled DLC images). |
| if dlc_uninstalled: |
| device.run(['restart', 'dlcservice']) |
| |
| |
| def _UninstallDLCImage(device, pkg_attrs): |
| """Uninstall a DLC image.""" |
| if _DLC_ID in pkg_attrs: |
| dlc_id = pkg_attrs[_DLC_ID] |
| logging.notice('Uninstalling DLC image for %s', dlc_id) |
| |
| device.run(['dlcservice_util', '--uninstall', '--id=%s' % dlc_id]) |
| return True |
| else: |
| logging.debug('DLC_ID not found in package') |
| return False |
| |
| |
| def _DeployDLCImage(device, sysroot, board, dlc_id, dlc_package): |
| """Deploy (install and mount) a DLC image.""" |
| # Build the DLC image if the image is outdated or doesn't exist. |
| dlc_lib.InstallDlcImages(sysroot=sysroot, dlc_id=dlc_id, board=board) |
| |
| logging.debug('Uninstall DLC %s if it is installed.', dlc_id) |
| try: |
| device.run(['dlcservice_util', '--uninstall', '--id=%s' % dlc_id]) |
| except cros_build_lib.RunCommandError as e: |
| logging.info('Failed to uninstall DLC:%s. Continue anyway.', |
| e.result.error) |
| except Exception: |
| logging.error('Failed to uninstall DLC.') |
| raise |
| |
| # TODO(andrewlassalle): Copy the DLC image to the preload location instead |
| # of to dlc_a and dlc_b, and let dlcserive install the images to their final |
| # location. |
| logging.notice('Deploy the DLC image for %s', dlc_id) |
| dlc_img_path_src = os.path.join(sysroot, dlc_lib.DLC_BUILD_DIR, dlc_id, |
| dlc_package, dlc_lib.DLC_IMAGE) |
| dlc_img_path = os.path.join(_DLC_INSTALL_ROOT, dlc_id, dlc_package) |
| dlc_img_path_a = os.path.join(dlc_img_path, 'dlc_a') |
| dlc_img_path_b = os.path.join(dlc_img_path, 'dlc_b') |
| # Create directories for DLC images. |
| device.run(['mkdir', '-p', dlc_img_path_a, dlc_img_path_b]) |
| # Copy images to the destination directories. |
| device.CopyToDevice(dlc_img_path_src, os.path.join(dlc_img_path_a, |
| dlc_lib.DLC_IMAGE), |
| mode='rsync') |
| device.run(['cp', os.path.join(dlc_img_path_a, dlc_lib.DLC_IMAGE), |
| os.path.join(dlc_img_path_b, dlc_lib.DLC_IMAGE)]) |
| |
| # Set the proper perms and ownership so dlcservice can access the image. |
| device.run(['chmod', '-R', 'u+rwX,go+rX,go-w', _DLC_INSTALL_ROOT]) |
| device.run(['chown', '-R', 'dlcservice:dlcservice', _DLC_INSTALL_ROOT]) |
| |
| # Copy metadata to device. |
| dest_mata_dir = os.path.join('/', dlc_lib.DLC_META_DIR, dlc_id, |
| dlc_package) |
| device.run(['mkdir', '-p', dest_mata_dir]) |
| src_meta_dir = os.path.join(sysroot, dlc_lib.DLC_BUILD_DIR, dlc_id, |
| dlc_package, dlc_lib.DLC_TMP_META_DIR) |
| device.CopyToDevice(src_meta_dir + '/', |
| dest_mata_dir, |
| mode='rsync', |
| recursive=True, |
| remote_sudo=True) |
| |
| |
| def _GetDLCInfo(device, pkg_path, from_dut): |
| """Returns information of a DLC given its package path. |
| |
| Args: |
| device: commandline.Device object; None to use the default device. |
| pkg_path: path to the package. |
| from_dut: True if extracting DLC info from DUT, False if extracting DLC |
| info from host. |
| |
| Returns: |
| A tuple (dlc_id, dlc_package). |
| """ |
| environment_content = '' |
| if from_dut: |
| # On DUT, |pkg_path| is the directory which contains environment file. |
| environment_path = os.path.join(pkg_path, _ENVIRONMENT_FILENAME) |
| result = device.run(['test', '-f', environment_path], |
| check=False, encoding=None) |
| if result.returncode == 1: |
| # The package is not installed on DUT yet. Skip extracting info. |
| return None, None |
| result = device.run(['bzip2', '-d', '-c', environment_path], encoding=None) |
| environment_content = result.output |
| else: |
| # On host, pkg_path is tbz2 file which contains environment file. |
| # Extract the metadata of the package file. |
| data = portage.xpak.tbz2(pkg_path).get_data() |
| # Extract the environment metadata. |
| environment_content = bz2.decompress( |
| data[_ENVIRONMENT_FILENAME.encode('utf-8')]) |
| |
| with tempfile.NamedTemporaryFile() as f: |
| # Dumps content into a file so we can use osutils.SourceEnvironment. |
| path = os.path.realpath(f.name) |
| osutils.WriteFile(path, environment_content, mode='wb') |
| content = osutils.SourceEnvironment(path, (_DLC_ID, _DLC_PACKAGE, |
| _DLC_ENABLED)) |
| |
| dlc_enabled = content.get(_DLC_ENABLED) |
| if dlc_enabled is not None and (dlc_enabled is False or |
| str(dlc_enabled) == 'false'): |
| logging.info('Installing DLC in rootfs.') |
| return None, None |
| return content.get(_DLC_ID), content.get(_DLC_PACKAGE) |
| |
| |
| def Deploy(device, packages, board=None, emerge=True, update=False, deep=False, |
| deep_rev=False, clean_binpkg=True, root='/', strip=True, |
| emerge_args=None, ssh_private_key=None, ping=True, force=False, |
| dry_run=False): |
| """Deploys packages to a device. |
| |
| Args: |
| device: commandline.Device object; None to use the default device. |
| packages: List of packages (strings) to deploy to device. |
| board: Board to use; None to automatically detect. |
| emerge: True to emerge package, False to unmerge. |
| update: Check installed version on device. |
| deep: Install dependencies also. Implies |update|. |
| deep_rev: Install reverse dependencies. Implies |deep|. |
| clean_binpkg: Clean outdated binary packages. |
| root: Package installation root path. |
| strip: Run strip_package to filter out preset paths in the package. |
| emerge_args: Extra arguments to pass to emerge. |
| ssh_private_key: Path to an SSH private key file; None to use test keys. |
| ping: True to ping the device before trying to connect. |
| force: Ignore sanity checks and prompts. |
| dry_run: Print deployment plan but do not deploy anything. |
| |
| Raises: |
| ValueError: Invalid parameter or parameter combination. |
| DeployError: Unrecoverable failure during deploy. |
| """ |
| if deep_rev: |
| deep = True |
| if deep: |
| update = True |
| |
| if not packages: |
| raise DeployError('No packages provided, nothing to deploy.') |
| |
| if update and not emerge: |
| raise ValueError('Cannot update and unmerge.') |
| |
| if device: |
| hostname, username, port = device.hostname, device.username, device.port |
| else: |
| hostname, username, port = None, None, None |
| |
| lsb_release = None |
| sysroot = None |
| try: |
| # Somewhat confusing to clobber, but here we are. |
| # pylint: disable=redefined-argument-from-local |
| with remote_access.ChromiumOSDeviceHandler( |
| hostname, port=port, username=username, private_key=ssh_private_key, |
| base_dir=_DEVICE_BASE_DIR, ping=ping) as device: |
| lsb_release = device.lsb_release |
| |
| board = cros_build_lib.GetBoard(device_board=device.board, |
| override_board=board) |
| if not force and board != device.board: |
| raise DeployError('Device (%s) is incompatible with board %s. Use ' |
| '--force to deploy anyway.' % (device.board, board)) |
| |
| sysroot = cros_build_lib.GetSysroot(board=board) |
| |
| if clean_binpkg: |
| logging.notice('Cleaning outdated binary packages from %s', sysroot) |
| portage_util.CleanOutdatedBinaryPackages(sysroot) |
| |
| # Remount rootfs as writable if necessary. |
| if not device.MountRootfsReadWrite(): |
| raise DeployError('Cannot remount rootfs as read-write. Exiting.') |
| |
| # Obtain list of packages to upgrade/remove. |
| pkg_scanner = _InstallPackageScanner(sysroot) |
| pkgs, listed, num_updates, pkgs_attrs = pkg_scanner.Run( |
| device, root, packages, update, deep, deep_rev) |
| if emerge: |
| action_str = 'emerge' |
| else: |
| pkgs.reverse() |
| action_str = 'unmerge' |
| |
| if not pkgs: |
| logging.notice('No packages to %s', action_str) |
| return |
| |
| # Warn when the user seems to forget `cros workon start`. |
| worked_on_cps = workon_helper.WorkonHelper(sysroot).ListAtoms() |
| for package in listed: |
| cp = portage_util.SplitCPV(package).cp |
| if cp not in worked_on_cps: |
| logging.warning( |
| 'Are you intentionally deploying unmodified packages, or did ' |
| 'you forget to run `cros workon --board=$BOARD start %s`?', cp) |
| |
| logging.notice('These are the packages to %s:', action_str) |
| for i, pkg in enumerate(pkgs): |
| logging.notice('%s %d) %s', '*' if pkg in listed else ' ', i + 1, pkg) |
| |
| if dry_run or not _ConfirmDeploy(num_updates): |
| return |
| |
| # Select function (emerge or unmerge) and bind args. |
| if emerge: |
| func = functools.partial(_EmergePackages, pkgs, device, strip, |
| sysroot, root, board, emerge_args) |
| else: |
| func = functools.partial(_UnmergePackages, pkgs, device, root, |
| pkgs_attrs) |
| |
| # Call the function with the progress bar or with normal output. |
| if command.UseProgressBar(): |
| op = BrilloDeployOperation(len(pkgs), emerge) |
| op.Run(func, log_level=logging.DEBUG) |
| else: |
| func() |
| |
| if device.IsSELinuxAvailable(): |
| if sum(x.count('selinux-policy') for x in pkgs): |
| logging.warning( |
| 'Deploying SELinux policy will not take effect until reboot. ' |
| 'SELinux policy is loaded by init. Also, changing the security ' |
| 'contexts (labels) of a file will require building a new image ' |
| 'and flashing the image onto the device.') |
| |
| logging.warning('Please restart any updated services on the device, ' |
| 'or just reboot it.') |
| except Exception: |
| if lsb_release: |
| lsb_entries = sorted(lsb_release.items()) |
| logging.info('Following are the LSB version details of the device:\n%s', |
| '\n'.join('%s=%s' % (k, v) for k, v in lsb_entries)) |
| raise |