blob: 94454f0a4f631917e8df00d06954cf6225737ab4 [file] [log] [blame]
# Lint as: python2, python3
# Copyright (c) 2011 The Chromium OS Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
import logging
import os
import stat
from autotest_lib.client.common_lib import log
from autotest_lib.client.common_lib import error, utils, global_config
from autotest_lib.client.bin import base_sysinfo, utils
from autotest_lib.client.cros import constants
get_value = global_config.global_config.get_config_value
collect_corefiles = get_value('CLIENT', 'collect_corefiles',
type=bool, default=False)
logfile = base_sysinfo.logfile
command = base_sysinfo.command
class logdir(base_sysinfo.loggable):
"""Represents a log directory."""
DEFAULT_EXCLUDES = ("**autoserv*", "**.journal",)
def __init__(self, directory, excludes=DEFAULT_EXCLUDES):
super(logdir, self).__init__(directory, log_in_keyval=False)
self.dir = directory
self._excludes = excludes
self._infer_old_attributes()
def __setstate__(self, state):
"""Unpickle handler
When client tests are run without SSP, we pickle this object on the
server-side (using the version of the class deployed in the lab) and
unpickle it on the DUT (using the version of the class from the build).
This means that when adding a new attribute to this class, for a while
the server-side code does not populate that attribute. So, deal with
missing attributes in a sane way.
"""
self.__dict__ = state
if '_excludes' not in state:
self._excludes = self.DEFAULT_EXCLUDES
if self.additional_exclude:
self._excludes += tuple(self.additional_exclude)
def __repr__(self):
return "site_sysinfo.logdir(%r, %s)" % (self.dir,
self._excludes)
def __eq__(self, other):
if isinstance(other, logdir):
return (self.dir == other.dir and self._excludes == other._excludes)
elif isinstance(other, base_sysinfo.loggable):
return False
return NotImplemented
def __ne__(self, other):
result = self.__eq__(other)
if result is NotImplemented:
return result
return not result
def __hash__(self):
return hash(self.dir) + hash(self._excludes)
def run(self, log_dir):
"""Copies this log directory to the specified directory.
@param log_dir: The destination log directory.
"""
from_dir = os.path.realpath(self.dir)
if os.path.exists(from_dir):
parent_dir = os.path.dirname(from_dir)
utils.system("mkdir -p %s%s" % (log_dir, parent_dir))
excludes = [
"--exclude=%s" % self._anchored_exclude_pattern(from_dir, x)
for x in self._excludes]
# Take source permissions and add ugo+r so files are accessible via
# archive server.
utils.system(
"rsync --no-perms --chmod=ugo+r -a --safe-links %s %s %s%s"
% (" ".join(excludes), from_dir, log_dir, parent_dir))
def _anchored_exclude_pattern(self, from_dir, pattern):
return '/%s/%s' % (os.path.basename(from_dir), pattern)
def _infer_old_attributes(self):
""" Backwards compatibility attributes.
YOU MUST NEVER DROP / REINTERPRET THESE.
A logdir object is pickled on server-side and unpickled on
client-side. This means that, when running aginst client-side code
from an older build, we need to be able to unpickle an instance of
logdir pickled from a newer version of the class.
Some old attributes are not sanely handled via __setstate__, so we can't
drop them without breaking compatibility.
"""
additional_excludes = list(set(self._excludes) -
set(self.DEFAULT_EXCLUDES))
if additional_excludes:
# Old API only allowed a single additional exclude.
# Best effort, keep the first one, throw the rest.
self.additional_exclude = additional_excludes[0]
else:
self.additional_exclude = None
class file_stat(object):
"""Store the file size and inode, used for retrieving new data in file."""
def __init__(self, file_path):
"""Collect the size and inode information of a file.
@param file_path: full path to the file.
"""
stat = os.stat(file_path)
# Start size of the file, skip that amount of bytes when do diff.
self.st_size = stat.st_size
# inode of the file. If inode is changed, treat this as a new file and
# copy the whole file.
self.st_ino = stat.st_ino
class diffable_logdir(logdir):
"""Represents a log directory that only new content will be copied.
An instance of this class should be added in both
before_iteration_loggables and after_iteration_loggables. This is to
guarantee the file status information is collected when run method is
called in before_iteration_loggables, and diff is executed when run
method is called in after_iteration_loggables.
"""
def __init__(self, directory, excludes=logdir.DEFAULT_EXCLUDES,
keep_file_hierarchy=True, append_diff_in_name=True):
"""
Constructor of a diffable_logdir instance.
@param directory: directory to be diffed after an iteration finished.
@param excludes: path patterns to exclude for rsync.
@param keep_file_hierarchy: True if need to preserve full path, e.g.,
sysinfo/var/log/sysstat, v.s. sysinfo/sysstat if it's False.
@param append_diff_in_name: True if you want to append '_diff' to the
folder name to indicate it's a diff, e.g., var/log_diff. Option
keep_file_hierarchy must be True for this to take effect.
"""
super(diffable_logdir, self).__init__(directory, excludes)
self.keep_file_hierarchy = keep_file_hierarchy
self.append_diff_in_name = append_diff_in_name
# Init dictionary to store all file status for files in the directory.
self._log_stats = {}
def _get_init_status_of_src_dir(self, src_dir):
"""Get initial status of files in src_dir folder.
@param src_dir: directory to be diff-ed.
"""
# Dictionary used to store the initial status of files in src_dir.
for file_path in self._get_all_files(src_dir):
self._log_stats[file_path] = file_stat(file_path)
self.file_stats_collected = True
def _get_all_files(self, path):
"""Iterate through files in given path including subdirectories.
@param path: root directory.
@return: an iterator that iterates through all files in given path
including subdirectories.
"""
if not os.path.exists(path):
yield []
for root, dirs, files in os.walk(path):
for f in files:
if f.startswith('autoserv'):
continue
if f.endswith('.journal') or f.endswith('.journal~'):
continue
full_path = os.path.join(root, f)
# Only list regular files or symlinks to those (os.stat follows
# symlinks)
if stat.S_ISREG(os.stat(full_path).st_mode):
yield full_path
def _copy_new_data_in_file(self, file_path, src_dir, dest_dir):
"""Copy all new data in a file to target directory.
@param file_path: full path to the file to be copied.
@param src_dir: source directory to do the diff.
@param dest_dir: target directory to store new data of src_dir.
"""
bytes_to_skip = 0
if file_path in self._log_stats:
prev_stat = self._log_stats[file_path]
new_stat = os.stat(file_path)
if new_stat.st_ino == prev_stat.st_ino:
bytes_to_skip = prev_stat.st_size
if new_stat.st_size == bytes_to_skip:
return
elif new_stat.st_size < prev_stat.st_size:
# File is modified to a smaller size, copy whole file.
bytes_to_skip = 0
try:
with open(file_path, 'r') as in_log:
if bytes_to_skip > 0:
in_log.seek(bytes_to_skip)
# Skip src_dir in path, e.g., src_dir/[sub_dir]/file_name.
target_path = os.path.join(dest_dir,
os.path.relpath(file_path, src_dir))
target_dir = os.path.dirname(target_path)
if not os.path.exists(target_dir):
os.makedirs(target_dir)
with open(target_path, "w") as out_log:
out_log.write(in_log.read())
except IOError as e:
logging.error('Diff %s failed with error: %s', file_path, e)
def _log_diff(self, src_dir, dest_dir):
"""Log all of the new data in src_dir to dest_dir.
@param src_dir: source directory to do the diff.
@param dest_dir: target directory to store new data of src_dir.
"""
if self.keep_file_hierarchy:
dir = src_dir.lstrip('/')
if self.append_diff_in_name:
dir = dir.rstrip('/') + '_diff'
dest_dir = os.path.join(dest_dir, dir)
if not os.path.exists(dest_dir):
os.makedirs(dest_dir)
for src_file in self._get_all_files(src_dir):
self._copy_new_data_in_file(src_file, src_dir, dest_dir)
def run(self, log_dir, collect_init_status=True, collect_all=False):
"""Copies new content from self.dir to the destination log_dir.
@param log_dir: The destination log directory.
@param collect_init_status: Set to True if run method is called to
collect the initial status of files.
@param collect_all: Set to True to force to collect all files.
"""
if collect_init_status:
self._get_init_status_of_src_dir(self.dir)
elif os.path.exists(self.dir):
# Always create a copy of the new logs to help debugging.
self._log_diff(self.dir, log_dir)
if collect_all:
logdir_temp = logdir(self.dir)
logdir_temp.run(log_dir)
class purgeable_logdir(logdir):
"""Represents a log directory that will be purged."""
def __init__(self, directory, excludes=logdir.DEFAULT_EXCLUDES):
super(purgeable_logdir, self).__init__(directory, excludes)
def run(self, log_dir):
"""Copies this log dir to the destination dir, then purges the source.
@param log_dir: The destination log directory.
"""
super(purgeable_logdir, self).run(log_dir)
if os.path.exists(self.dir):
utils.system("rm -rf %s/*" % (self.dir))
class site_sysinfo(base_sysinfo.base_sysinfo):
"""Represents site system info."""
def __init__(self, job_resultsdir):
super(site_sysinfo, self).__init__(job_resultsdir)
crash_exclude_string = None
if not collect_corefiles:
crash_exclude_string = "*.core"
# This is added in before and after_iteration_loggables. When run is
# called in before_iteration_loggables, it collects file status in
# the directory. When run is called in after_iteration_loggables, diff
# is executed.
# self.diffable_loggables is only initialized if the instance does not
# have this attribute yet. The sysinfo instance could be loaded
# from an earlier pickle dump, which has already initialized attribute
# self.diffable_loggables.
if not hasattr(self, 'diffable_loggables'):
diffable_log = diffable_logdir(constants.LOG_DIR)
self.diffable_loggables = set()
self.diffable_loggables.add(diffable_log)
# add in some extra command logging
self.boot_loggables.add(command("ls -l /boot",
"boot_file_list"))
self.before_iteration_loggables.add(
command(constants.CHROME_VERSION_COMMAND, "chrome_version"))
self.boot_loggables.add(command("crossystem", "crossystem"))
self.test_loggables.add(
purgeable_logdir(
os.path.join(constants.CRYPTOHOME_MOUNT_PT, "log")))
# We only want to gather and purge crash reports after the client test
# runs in case a client test is checking that a crash found at boot
# (such as a kernel crash) is handled.
self.after_iteration_loggables.add(
purgeable_logdir(
os.path.join(constants.CRYPTOHOME_MOUNT_PT, "crash"),
excludes=logdir.DEFAULT_EXCLUDES + (crash_exclude_string,)))
self.after_iteration_loggables.add(
purgeable_logdir(
constants.CRASH_DIR,
excludes=logdir.DEFAULT_EXCLUDES + (crash_exclude_string,)))
self.test_loggables.add(
logfile(os.path.join(constants.USER_DATA_DIR,
".Google/Google Talk Plugin/gtbplugin.log")))
self.test_loggables.add(purgeable_logdir(
constants.CRASH_DIR,
excludes=logdir.DEFAULT_EXCLUDES + (crash_exclude_string,)))
# Collect files under /tmp/crash_reporter, which contain the procfs
# copy of those crashed processes whose core file didn't get converted
# into minidump. We need these additional files for post-mortem analysis
# of the conversion failure.
self.test_loggables.add(
purgeable_logdir(constants.CRASH_REPORTER_RESIDUE_DIR))
@log.log_and_ignore_errors("pre-test sysinfo error:")
def log_before_each_test(self, test):
"""Logging hook called before a test starts.
@param test: A test object.
"""
super(site_sysinfo, self).log_before_each_test(test)
for log in self.diffable_loggables:
log.run(log_dir=None, collect_init_status=True)
@log.log_and_ignore_errors("post-test sysinfo error:")
def log_after_each_test(self, test):
"""Logging hook called after a test finishs.
@param test: A test object.
"""
super(site_sysinfo, self).log_after_each_test(test)
test_sysinfodir = self._get_sysinfodir(test.outputdir)
for log in self.diffable_loggables:
log.run(log_dir=test_sysinfodir, collect_init_status=False,
collect_all=not test.success)
def _get_chrome_version(self):
"""Gets the Chrome version number and milestone as strings.
Invokes "chrome --version" to get the version number and milestone.
@return A tuple (chrome_ver, milestone) where "chrome_ver" is the
current Chrome version number as a string (in the form "W.X.Y.Z")
and "milestone" is the first component of the version number
(the "W" from "W.X.Y.Z"). If the version number cannot be parsed
in the "W.X.Y.Z" format, the "chrome_ver" will be the full output
of "chrome --version" and the milestone will be the empty string.
"""
version_string = utils.system_output(constants.CHROME_VERSION_COMMAND,
ignore_status=True)
return utils.parse_chrome_version(version_string)
def log_test_keyvals(self, test_sysinfodir):
"""Generate keyval for the sysinfo.
Collects keyval entries to be written in the test keyval.
@param test_sysinfodir: The test's system info directory.
"""
keyval = super(site_sysinfo, self).log_test_keyvals(test_sysinfodir)
lsb_lines = utils.system_output(
"cat /etc/lsb-release",
ignore_status=True).splitlines()
lsb_dict = dict(item.split("=") for item in lsb_lines)
for lsb_key in lsb_dict.keys():
# Special handling for build number
if lsb_key == "CHROMEOS_RELEASE_DESCRIPTION":
keyval["CHROMEOS_BUILD"] = (
lsb_dict[lsb_key].rstrip(")").split(" ")[3])
keyval[lsb_key] = lsb_dict[lsb_key]
# Get the hwid (hardware ID), if applicable.
try:
keyval["hwid"] = utils.system_output('crossystem hwid')
except error.CmdError:
# The hwid may not be available (e.g, when running on a VM).
# If the output of 'crossystem mainfw_type' is 'nonchrome', then
# we expect the hwid to not be avilable, and we can proceed in this
# case. Otherwise, the hwid is missing unexpectedly.
mainfw_type = utils.system_output('crossystem mainfw_type')
if mainfw_type == 'nonchrome':
logging.info(
'HWID not available; not logging it as a test keyval.')
else:
logging.exception('HWID expected but could not be identified; '
'output of "crossystem mainfw_type" is "%s"',
mainfw_type)
raise
# Get the chrome version and milestone numbers.
keyval["CHROME_VERSION"], keyval["MILESTONE"] = (
self._get_chrome_version())
# TODO(kinaba): crbug.com/707448 Import at the head of this file.
# Currently a server-side script server/server_job.py is indirectly
# importing this file, so we cannot globaly import cryptohome that
# has dependency to a client-only library.
from autotest_lib.client.cros import cryptohome
# Get the dictionary attack counter.
keyval["TPM_DICTIONARY_ATTACK_COUNTER"] = (
cryptohome.get_tpm_more_status().get(
'dictionary_attack_counter', 'Failed to query cryptohome'))
# Return the updated keyvals.
return keyval
def add_logdir(self, loggable):
"""Collect files in log_path to sysinfo folder.
This method can be called from a control file for test to collect files
in a specified folder. autotest creates a folder [test result
dir]/sysinfo folder with the full path of log_path and copy all files in
log_path to that folder.
@param loggable: A logdir instance corresponding to the logs to collect.
"""
self.test_loggables.add(loggable)