blob: f394ece14f12b2bf9156b3f999016b8c90172c9c [file] [log] [blame]
# Copyright (c) 2009 The Chromium OS Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
import logging
import os
import re
import shutil
import traceback
from autotest_lib.client.common_lib import error
from autotest_lib.client.bin import utils
from autotest_lib.server import test, autotest
LOWER_IS_BETTER_METRICS = set(['rdbytes', 'seconds'])
class platform_BootPerfServer(test.test):
"""A test that reboots the client and collect boot perf data."""
version = 1
def upload_perf_keyvals(self, keyvals):
"""Upload perf keyvals in dictionary |keyvals| to Chrome perf dashboard.
This method assumes that the key of a perf keyval is in the format
of "units_description". The text before the first underscore represents
the units and the rest of the text represents
a description of the measured perf value. For instance,
'seconds_kernel_to_login', 'rdbytes_kernel_to_startup'.
@param keyvals: A dictionary that maps a perf metric to its value.
"""
for key, val in keyvals.items():
match = re.match(r'^(.+?)_.+$', key)
if match:
units = match.group(1)
higher_is_better = units not in LOWER_IS_BETTER_METRICS
self.output_perf_value(
description=key, value=val,
units=units, higher_is_better=higher_is_better)
def run_once(self, host=None, upload_perf=False):
self.client = host
self.client_test = 'platform_BootPerf'
# Run a login test to complete the OOBE flow, if we haven't already.
# This is so that we measure boot times for the stable state.
try:
self.client.run('ls /home/chronos/.oobe_completed')
except error.AutoservRunError:
logging.info('Taking client through OOBE.')
client_at = autotest.Autotest(self.client)
client_at.run_test('login_LoginSuccess', disable_sysinfo=True)
# Reboot the client
logging.info('BootPerfServer: reboot %s', self.client.hostname)
try:
self.client.reboot(reboot_timeout=90)
except error.AutoservRebootError as e:
raise error.TestFail('%s.\nTest failed with error %s' % (
traceback.format_exc(), str(e)))
# Collect the performance metrics by running a client side test
logging.info('BootPerfServer: start client test')
client_at = autotest.Autotest(self.client)
client_at.run_test(
self.client_test, last_boot_was_reboot=True, disable_sysinfo=True)
# In the client results directory are a 'keyval' file, and
# various raw bootstat data files. First promote the client
# test 'keyval' as our own.
logging.info('BootPerfServer: gather client results')
client_results_dir = os.path.join(
self.outputdir, self.client_test, "results")
src = os.path.join(client_results_dir, "keyval")
dst = os.path.join(self.resultsdir, "keyval")
if os.path.exists(src):
client_results = open(src, "r")
server_results = open(dst, "a")
shutil.copyfileobj(client_results, server_results)
server_results.close()
client_results.close()
else:
logging.warning('Unable to locate %s', src)
# Upload perf keyvals in the client keyval file to perf dashboard.
if upload_perf:
logging.info('Output perf data for iteration %03d', self.iteration)
perf_keyvals = utils.read_keyval(src, type_tag='perf')
self.upload_perf_keyvals(perf_keyvals)
# Everything that isn't the client 'keyval' file is raw data
# from the client test: move it to a per-iteration
# subdirectory. We move instead of copying so we can be sure
# we don't have any stale results in the next iteration
if self.iteration is not None:
rawdata_dir = "rawdata.%03d" % self.iteration
else:
rawdata_dir = "rawdata"
rawdata_dir = os.path.join(self.resultsdir, rawdata_dir)
shutil.move(client_results_dir, rawdata_dir)
try:
os.remove(os.path.join(rawdata_dir, "keyval"))
except Exception:
pass