blob: 77e0861022f266bc236a36368c7b50822e917125 [file] [log] [blame]
# Lint as: python2, python3
# Copyright (c) 2019 The Chromium OS Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
#
# Expects to be run in an environment with sudo and no interactive password
# prompt, such as within the Chromium OS development chroot.
"""This is a base host class for servohost and labstation."""
import six.moves.http_client
import logging
import socket
import six.moves.xmlrpc_client
import time
import os
from autotest_lib.client.bin import utils
from autotest_lib.client.common_lib import autotest_enum
from autotest_lib.client.common_lib import error
from autotest_lib.client.common_lib import hosts
from autotest_lib.client.common_lib import lsbrelease_utils
from autotest_lib.client.common_lib.cros import dev_server
from autotest_lib.client.common_lib.cros import kernel_utils
from autotest_lib.client.cros import constants as client_constants
from autotest_lib.server import autotest
from autotest_lib.server import site_utils as server_utils
from autotest_lib.server.cros import provisioner
from autotest_lib.server.hosts import ssh_host
from autotest_lib.site_utils.rpm_control_system import rpm_client
class BaseServoHost(ssh_host.SSHHost):
"""Base host class for a host that manage servo(s).
E.g. beaglebone, labstation.
"""
REBOOT_CMD = 'sleep 5; reboot & sleep 10; reboot -f'
TEMP_FILE_DIR = '/var/lib/servod/'
LOCK_FILE_POSTFIX = '_in_use'
REBOOT_FILE_POSTFIX = '_reboot'
# Time to wait a rebooting servohost, in seconds.
REBOOT_TIMEOUT = 240
# Timeout value to power cycle a servohost, in seconds.
BOOT_TIMEOUT = 240
# Constants that reflect current host update state.
UPDATE_STATE = autotest_enum.AutotestEnum('IDLE', 'RUNNING',
'PENDING_REBOOT')
def _initialize(self, hostname, is_in_lab=None, *args, **dargs):
"""Construct a BaseServoHost object.
@param is_in_lab: True if the servo host is in Cros Lab. Default is set
to None, for which utils.host_is_in_lab_zone will be
called to check if the servo host is in Cros lab.
"""
super(BaseServoHost, self)._initialize(hostname=hostname,
*args, **dargs)
self._is_localhost = (self.hostname == 'localhost')
if self._is_localhost:
self._is_in_lab = False
elif is_in_lab is None:
self._is_in_lab = utils.host_is_in_lab_zone(self.hostname)
else:
self._is_in_lab = is_in_lab
# Commands on the servo host must be run by the superuser.
# Our account on a remote host is root, but if our target is
# localhost then we might be running unprivileged. If so,
# `sudo` will have to be added to the commands.
if self._is_localhost:
self._sudo_required = utils.system_output('id -u') != '0'
else:
self._sudo_required = False
self._is_labstation = None
self._dut_host_info = None
self._dut_hostname = None
def get_board(self):
"""Determine the board for this servo host. E.g. fizz-labstation
@returns a string representing this labstation's board or None if
target host is not using a ChromeOS image(e.g. test in chroot).
"""
output = self.run('cat /etc/lsb-release', ignore_status=True).stdout
return lsbrelease_utils.get_current_board(lsb_release_content=output)
def set_dut_host_info(self, dut_host_info):
"""
@param dut_host_info: A HostInfo object.
"""
logging.info('setting dut_host_info field to (%s)', dut_host_info)
self._dut_host_info = dut_host_info
def get_dut_host_info(self):
"""
@return A HostInfo object.
"""
return self._dut_host_info
def set_dut_hostname(self, dut_hostname):
"""
@param dut_hostname: hostname of the DUT that connected to this servo.
"""
logging.info('setting dut_hostname as (%s)', dut_hostname)
self._dut_hostname = dut_hostname
def get_dut_hostname(self):
"""
@returns hostname of the DUT that connected to this servo.
"""
return self._dut_hostname
def is_labstation(self):
"""Determine if the host is a labstation
@returns True if ths host is a labstation otherwise False.
"""
if self._is_labstation is None:
board = self.get_board()
self._is_labstation = board is not None and 'labstation' in board
return self._is_labstation
def _get_lsb_release_content(self):
"""Return the content of lsb-release file of host."""
return self.run(
'cat "%s"' % client_constants.LSB_RELEASE).stdout.strip()
def get_release_version(self):
"""Get the value of attribute CHROMEOS_RELEASE_VERSION from lsb-release.
@returns The version string in lsb-release, under attribute
CHROMEOS_RELEASE_VERSION(e.g. 12900.0.0). None on fail.
"""
return lsbrelease_utils.get_chromeos_release_version(
lsb_release_content=self._get_lsb_release_content()
)
def get_full_release_path(self):
"""Get full release path from servohost as string.
@returns full release path as a string
(e.g. fizz-labstation-release/R82.12900.0.0). None on fail.
"""
return lsbrelease_utils.get_chromeos_release_builder_path(
lsb_release_content=self._get_lsb_release_content()
)
def _check_update_status(self):
""" Check servohost's current update state.
@returns: one of below state of from self.UPDATE_STATE
IDLE -- if the target host is not currently updating and not
pending on a reboot.
RUNNING -- if there is another updating process that running on
target host(note: we don't expect to hit this scenario).
PENDING_REBOOT -- if the target host had an update and pending
on reboot.
"""
result = self.run('pgrep -f quick-provision | grep -v $$',
ignore_status=True)
# We don't expect any output unless there are another quick
# provision process is running.
if result.exit_status == 0:
return self.UPDATE_STATE.RUNNING
# Determine if we have an update that pending on reboot by check if
# the current inactive kernel has priority for the next boot.
try:
inactive_kernel = kernel_utils.get_kernel_state(self)[1]
next_kernel = kernel_utils.get_next_kernel(self)
if inactive_kernel == next_kernel:
return self.UPDATE_STATE.PENDING_REBOOT
except Exception as e:
logging.error('Unexpected error while checking kernel info; %s', e)
return self.UPDATE_STATE.IDLE
def is_in_lab(self):
"""Check whether the servo host is a lab device.
@returns: True if the servo host is in Cros Lab, otherwise False.
"""
return self._is_in_lab
def is_localhost(self):
"""Checks whether the servo host points to localhost.
@returns: True if it points to localhost, otherwise False.
"""
return self._is_localhost
def is_cros_host(self):
"""Check if a servo host is running chromeos.
@return: True if the servo host is running chromeos.
False if it isn't, or we don't have enough information.
"""
try:
result = self.run('grep -q CHROMEOS /etc/lsb-release',
ignore_status=True, timeout=10)
except (error.AutoservRunError, error.AutoservSSHTimeout):
return False
return result.exit_status == 0
def prepare_for_update(self):
"""Prepares the DUT for an update.
Subclasses may override this to perform any special actions
required before updating.
"""
pass
def reboot(self, *args, **dargs):
"""Reboot using special servo host reboot command."""
super(BaseServoHost, self).reboot(reboot_cmd=self.REBOOT_CMD,
*args, **dargs)
def update_image(self, stable_version=None):
"""Update the image on the servo host, if needed.
This method recognizes the following cases:
* If the Host is not running Chrome OS, do nothing.
* If a previously triggered update is now complete, reboot
to the new version.
* If the host is processing an update do nothing.
* If the host has an update that pending on reboot, do nothing.
* If the host is running a version of Chrome OS different
from the default for servo Hosts, start an update.
@stable_version the target build number.(e.g. R82-12900.0.0)
@raises dev_server.DevServerException: If all the devservers are down.
@raises site_utils.ParseBuildNameException: If the devserver returns
an invalid build name.
"""
# servod could be running in a Ubuntu workstation.
if not self.is_cros_host():
logging.info('Not attempting an update, either %s is not running '
'chromeos or we cannot find enough information about '
'the host.', self.hostname)
return
if lsbrelease_utils.is_moblab():
logging.info('Not attempting an update, %s is running moblab.',
self.hostname)
return
if not stable_version:
logging.debug("BaseServoHost::update_image attempting to get"
" servo cros stable version")
try:
stable_version = (self.get_dut_host_info().
servo_cros_stable_version)
except AttributeError:
logging.error("BaseServoHost::update_image failed to get"
" servo cros stable version.")
target_build = "%s-release/%s" % (self.get_board(), stable_version)
target_build_number = server_utils.ParseBuildName(
target_build)[3]
current_build_number = self.get_release_version()
if current_build_number == target_build_number:
logging.info('servo host %s does not require an update.',
self.hostname)
return
status = self._check_update_status()
if status == self.UPDATE_STATE.RUNNING:
logging.info('servo host %s already processing an update',
self.hostname)
return
if status == self.UPDATE_STATE.PENDING_REBOOT:
# Labstation reboot is handled separately here as it require
# synchronized reboot among all managed DUTs. For servo_v3, we'll
# reboot when initialize Servohost, if there is a update pending.
logging.info('An update has been completed and pending reboot.')
return
ds = dev_server.ImageServer.resolve(self.hostname,
hostname=self.hostname)
url = ds.get_update_url(target_build)
cros_provisioner = provisioner.ChromiumOSProvisioner(update_url=url,
host=self,
is_servohost=True)
logging.info('Using devserver url: %s to trigger update on '
'servo host %s, from %s to %s', url, self.hostname,
current_build_number, target_build_number)
cros_provisioner.run_provision()
def has_power(self):
"""Return whether or not the servo host is powered by PoE or RPM."""
# TODO(fdeng): See crbug.com/302791
# For now, assume all servo hosts in the lab have power.
return self.is_in_lab()
def _post_update_reboot(self):
""" Reboot servohost after an quick provision.
We need to do some specifal cleanup before and after reboot
when there is an update pending.
"""
# Regarding the 'crossystem' command below: In some cases,
# the update flow puts the TPM into a state such that it
# fails verification. We don't know why. However, this
# call papers over the problem by clearing the TPM during
# the reboot.
#
# We ignore failures from 'crossystem'. Although failure
# here is unexpected, and could signal a bug, the point of
# the exercise is to paper over problems; allowing this to
# fail would defeat the purpose.
# Preserve critical files before reboot since post-provision
# clobbering will wipe the stateful partition.
# TODO(xianuowang@) Remove this logic once we have updated to
# a image with https://crrev.com/c/2485908.
path_to_preserve = [
'/var/lib/servod',
'/var/lib/device_health_profile',
]
safe_location = '/mnt/stateful_partition/unencrypted/preserve/'
for item in path_to_preserve:
dest = os.path.join(safe_location, item.split('/')[-1])
self.run('rm -rf %s' % dest, ignore_status=True)
self.run('mv %s %s' % (item, safe_location), ignore_status=True)
self.run('crossystem clear_tpm_owner_request=1', ignore_status=True)
self._servo_host_reboot()
logging.debug('Cleaning up autotest directories if exist.')
try:
installed_autodir = autotest.Autotest.get_installed_autodir(self)
self.run('rm -rf ' + installed_autodir)
except autotest.AutodirNotFoundError:
logging.debug('No autotest installed directory found.')
# Recover preserved files to original location.
# TODO(xianuowang@) Remove this logic once we have updated to
# a image with https://crrev.com/c/2485908.
for item in path_to_preserve:
src = os.path.join(safe_location, item.split('/')[-1])
dest = '/'.join(item.split('/')[:-1])
self.run('mv %s %s' % (src, dest), ignore_status=True)
def power_cycle(self):
"""Cycle power to this host via PoE(servo v3) or RPM(labstation)
if it is a lab device.
@raises AutoservRepairError if it fails to power cycle the
servo host.
"""
if self.has_power():
try:
rpm_client.set_power(self, 'CYCLE')
except (socket.error, six.moves.xmlrpc_client.Error,
six.moves.http_client.BadStatusLine,
rpm_client.RemotePowerException) as e:
raise hosts.AutoservRepairError(
'Power cycling %s failed: %s' % (self.hostname, e),
'power_cycle_via_rpm_failed'
)
else:
logging.info('Skipping power cycling, not a lab device.')
def _servo_host_reboot(self):
"""Reboot this servo host because a reboot is requested."""
logging.info('Rebooting servo host %s from build %s', self.hostname,
self.get_release_version())
# Tell the reboot() call not to wait for completion.
# Otherwise, the call will log reboot failure if servo does
# not come back. The logged reboot failure will lead to
# test job failure. If the test does not require servo, we
# don't want servo failure to fail the test with error:
# `Host did not return from reboot` in status.log.
self.reboot(fastsync=True, wait=False)
# We told the reboot() call not to wait, but we need to wait
# for the reboot before we continue. Alas. The code from
# here below is basically a copy of Host.wait_for_restart(),
# with the logging bits ripped out, so that they can't cause
# the failure logging problem described above.
#
# The black stain that this has left on my soul can never be
# erased.
old_boot_id = self.get_boot_id()
if not self.wait_down(timeout=self.WAIT_DOWN_REBOOT_TIMEOUT,
warning_timer=self.WAIT_DOWN_REBOOT_WARNING,
old_boot_id=old_boot_id):
raise error.AutoservHostError(
'servo host %s failed to shut down.' %
self.hostname)
if self.wait_up(timeout=self.REBOOT_TIMEOUT):
logging.info('servo host %s back from reboot, with build %s',
self.hostname, self.get_release_version())
else:
raise error.AutoservHostError(
'servo host %s failed to come back from reboot.' %
self.hostname)
def make_ssh_command(self, user='root', port=22, opts='', hosts_file=None,
connect_timeout=None, alive_interval=None, alive_count_max=None,
connection_attempts=None):
"""Override default make_ssh_command to use tuned options.
Tuning changes:
- ConnectTimeout=30; maximum of 30 seconds allowed for an SSH
connection failure. Consistency with remote_access.py.
- ServerAliveInterval=180; which causes SSH to ping connection every
180 seconds. In conjunction with ServerAliveCountMax ensures
that if the connection dies, Autotest will bail out quickly.
- ServerAliveCountMax=3; consistency with remote_access.py.
- ConnectAttempts=4; reduce flakiness in connection errors;
consistency with remote_access.py.
- UserKnownHostsFile=/dev/null; we don't care about the keys.
- SSH protocol forced to 2; needed for ServerAliveInterval.
@param user User name to use for the ssh connection.
@param port Port on the target host to use for ssh connection.
@param opts Additional options to the ssh command.
@param hosts_file Ignored.
@param connect_timeout Ignored.
@param alive_interval Ignored.
@param alive_count_max Ignored.
@param connection_attempts Ignored.
@returns: An ssh command with the requested settings.
"""
options = ' '.join([opts, '-o Protocol=2'])
return super(BaseServoHost, self).make_ssh_command(
user=user, port=port, opts=options, hosts_file='/dev/null',
connect_timeout=30, alive_interval=180, alive_count_max=3,
connection_attempts=4)
def _make_scp_cmd(self, sources, dest):
"""Format scp command.
Given a list of source paths and a destination path, produces the
appropriate scp command for encoding it. Remote paths must be
pre-encoded. Overrides _make_scp_cmd in AbstractSSHHost
to allow additional ssh options.
@param sources: A list of source paths to copy from.
@param dest: Destination path to copy to.
@returns: An scp command that copies |sources| on local machine to
|dest| on the remote servo host.
"""
command = ('scp -rq %s -o BatchMode=yes -o StrictHostKeyChecking=no '
'-o UserKnownHostsFile=/dev/null -P %d %s "%s"')
return command % (self._master_ssh.ssh_option,
self.port, sources, dest)
def run(self, command, timeout=3600, ignore_status=False,
stdout_tee=utils.TEE_TO_LOGS, stderr_tee=utils.TEE_TO_LOGS,
connect_timeout=30, ssh_failure_retry_ok=False,
options='', stdin=None, verbose=True, args=()):
"""Run a command on the servo host.
Extends method `run` in SSHHost. If the servo host is a remote device,
it will call `run` in SSHost without changing anything.
If the servo host is 'localhost', it will call utils.system_output.
@param command: The command line string.
@param timeout: Time limit in seconds before attempting to
kill the running process. The run() function
will take a few seconds longer than 'timeout'
to complete if it has to kill the process.
@param ignore_status: Do not raise an exception, no matter
what the exit code of the command is.
@param stdout_tee/stderr_tee: Where to tee the stdout/stderr.
@param connect_timeout: SSH connection timeout (in seconds)
Ignored if host is 'localhost'.
@param options: String with additional ssh command options
Ignored if host is 'localhost'.
@param ssh_failure_retry_ok: when True and ssh connection failure is
suspected, OK to retry command (but not
compulsory, and likely not needed here)
@param stdin: Stdin to pass (a string) to the executed command.
@param verbose: Log the commands.
@param args: Sequence of strings to pass as arguments to command by
quoting them in " and escaping their contents if necessary.
@returns: A utils.CmdResult object.
@raises AutoservRunError if the command failed.
@raises AutoservSSHTimeout SSH connection has timed out. Only applies
when servo host is not 'localhost'.
"""
run_args = {
'command' : command,
'timeout' : timeout,
'ignore_status' : ignore_status,
'stdout_tee' : stdout_tee,
'stderr_tee' : stderr_tee,
# connect_timeout n/a for localhost
# options n/a for localhost
# ssh_failure_retry_ok n/a for localhost
'stdin' : stdin,
'verbose' : verbose,
'args' : args,
}
if self.is_localhost():
if self._sudo_required:
run_args['command'] = 'sudo -n sh -c "%s"' % utils.sh_escape(
command)
try:
return utils.run(**run_args)
except error.CmdError as e:
logging.error(e)
raise error.AutoservRunError('command execution error',
e.result_obj)
else:
run_args['connect_timeout'] = connect_timeout
run_args['options'] = options
run_args['ssh_failure_retry_ok'] = ssh_failure_retry_ok
return super(BaseServoHost, self).run(**run_args)
def _mount_drive(self, src_path, dst_path):
"""Mount an external drive on servohost.
@param: src_path the drive path to mount(e.g. /dev/sda3).
@param: dst_path the destination directory on servohost to mount
the drive.
@returns: True if mount success otherwise False.
"""
# Make sure the dst dir exists.
self.run('mkdir -p %s' % dst_path)
result = self.run('mount -o ro %s %s' % (src_path, dst_path),
ignore_status=True)
return result.exit_status == 0
def _unmount_drive(self, mount_path):
"""Unmount a drive from servohost.
@param: mount_path the path on servohost to unmount.
@returns: True if unmount success otherwise False.
"""
result = self.run('umount %s' % mount_path, ignore_status=True)
return result.exit_status == 0
def wait_ready(self, required_uptime=300):
"""Wait ready for a servohost if it has been rebooted recently.
It may take a few minutes until all servos and their componments
re-enumerated and become ready after a servohost(especially labstation
as it supports multiple servos) reboot, so we need to make sure the
servohost has been up for a given a mount of time before trying to
start any actions.
@param required_uptime: Minimum uptime in seconds that we can
consdier a servohost be ready.
"""
uptime = float(self.check_uptime())
# To prevent unexpected output from check_uptime() that causes long
# sleep, make sure the maximum wait time <= required_uptime.
diff = min(required_uptime - uptime, required_uptime)
if diff > 0:
logging.info(
'The servohost was just rebooted, wait %s'
' seconds for it to become ready', diff)
time.sleep(diff)