blob: aea4eeef5d8c95c338e8b3c739f3c2bf23f4db63 [file] [log] [blame]
#!/usr/bin/python
# Copyright (c) 2012 The Chromium OS Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
# Disable relative import warning from pylint.
# pylint: disable=W0403
import constants
import copy
import json
import optparse
import sys
import urllib
GS_PATH_DEFAULT = 'default' # Means gs://chromeos-image-archive/ + bot_id
# Contains the valid build config suffixes in the order that they are dumped.
CONFIG_TYPE_RELEASE = 'release'
CONFIG_TYPE_FULL = 'full'
CONFIG_TYPE_FIRMWARE = 'firmware'
CONFIG_TYPE_DUMP_ORDER = (
'paladin',
constants.PRE_CQ_BUILDER_NAME,
'pre-cq',
'pre-cq-launcher',
'incremental',
'telemetry',
CONFIG_TYPE_FULL,
CONFIG_TYPE_RELEASE,
'release-group',
'release-pgo',
'release-pgo-generate',
'release-pgo-use',
'sdk',
'chromium-pfq',
'chromium-pfq-informational',
'chrome-perf',
'chrome-pgo',
'chrome-pfq',
'chrome-pfq-informational',
'pre-flight-branch',
'factory',
CONFIG_TYPE_FIRMWARE,
'toolchain-major',
'toolchain-minor',
'asan',
'asan-informational',
'refresh-packages',
'platform2',
'test-ap',
constants.BRANCH_UTIL_CONFIG,
)
def OverrideConfigForTrybot(build_config, options):
"""Apply trybot-specific configuration settings.
Args:
build_config: The build configuration dictionary to override.
The dictionary is not modified.
options: The options passed on the commandline.
Returns:
A build configuration dictionary with the overrides applied.
"""
copy_config = copy.deepcopy(build_config)
for my_config in [copy_config] + copy_config['child_configs']:
my_config['uprev'] = True
if my_config['internal']:
my_config['overlays'] = constants.BOTH_OVERLAYS
# Most users don't have access to the pdf repository so disable pdf along
# with all other official_chrome useflags, so that we use the external
# chromium prebuilts.
useflags = my_config['useflags']
if not options.remote_trybot and useflags:
for chrome_use in official_chrome['useflags']:
if chrome_use in useflags:
useflags.remove(chrome_use)
# Use the local manifest which only requires elevated access if it's really
# needed to build.
if not options.remote_trybot:
my_config['manifest'] = my_config['dev_manifest']
my_config['push_image'] = False
if options.hwtest:
my_config['upload_hw_test_artifacts'] = True
if not my_config['hw_tests']:
my_config['hw_tests'] = HWTestConfig.DefaultList(
num=constants.HWTEST_TRYBOT_NUM, pool=constants.HWTEST_TRYBOT_POOL,
fatal_timeouts=True, file_bugs=False)
else:
for hw_config in my_config['hw_tests']:
hw_config.num = constants.HWTEST_TRYBOT_NUM
hw_config.pool = constants.HWTEST_TRYBOT_POOL
hw_config.fatal_timeouts = True
hw_config.file_bugs = False
hw_config.priority = constants.HWTEST_DEFAULT_PRIORITY
# Default to starting with a fresh chroot on remote trybot runs.
if options.remote_trybot:
my_config['chroot_replace'] = True
# In trybots, we want to always run VM tests and all unit tests, so that
# developers will get better testing for their changes.
if (my_config['build_type'] == constants.PALADIN_TYPE
and not my_config['arm']):
my_config['vm_tests'] = constants.SIMPLE_AU_TEST_TYPE
my_config['quick_unit'] = False
return copy_config
def GetManifestVersionsRepoUrl(internal_build, read_only=False, test=False):
"""Returns the url to the manifest versions repository.
Args:
internal_build: Whether to use the internal repo.
read_only: Whether the URL may be read only. If read_only is True,
pushing changes (even with dryrun option) may not work.
test: Whether we should use the corresponding test repositories. These
should be used when staging experimental features.
"""
# pylint: disable=W0613
if internal_build:
url = constants.INTERNAL_GOB_URL + constants.MANIFEST_VERSIONS_INT_SUFFIX
else:
url = constants.EXTERNAL_GOB_URL + constants.MANIFEST_VERSIONS_SUFFIX
if test:
url += '-test'
return url
def IsPFQType(b_type):
"""Returns True if this build type is a PFQ."""
return b_type in (constants.PFQ_TYPE, constants.PALADIN_TYPE,
constants.CHROME_PFQ_TYPE)
def IsCQType(b_type):
"""Returns True if this build type is a Commit Queue."""
return b_type == constants.PALADIN_TYPE
# List of usable cbuildbot configs; see add_config method.
# TODO(mtennant): This is seriously buried in this file. Move to top
# and rename something that stands out in a file where the word "config"
# is used everywhere.
config = {}
# pylint: disable=W0102
def GetCanariesForChromeLKGM(configs=config):
"""Grabs a list of builders that are important for the Chrome LKGM."""
builders = []
for build_name, conf in configs.iteritems():
if (conf['build_type'] == constants.CANARY_TYPE and
conf['critical_for_chrome']):
builders.append(build_name)
return builders
def FindFullConfigsForBoard(board=None):
"""Returns full builder configs for a board.
Args:
board: The board to match. By default, match all boards.
Returns:
A tuple containing a list of matching external configs and a list of
matching internal release configs for a board.
"""
ext_cfgs = []
int_cfgs = []
for name, c in config.iteritems():
if c['boards'] and (board is None or board in c['boards']):
if name.endswith('-%s' % CONFIG_TYPE_RELEASE) and c['internal']:
int_cfgs.append(copy.deepcopy(c))
elif name.endswith('-%s' % CONFIG_TYPE_FULL) and not c['internal']:
ext_cfgs.append(copy.deepcopy(c))
return ext_cfgs, int_cfgs
def FindCanonicalConfigForBoard(board):
"""Get the canonical cbuildbot builder config for a board."""
ext_cfgs, int_cfgs = FindFullConfigsForBoard(board)
# If both external and internal builds exist for this board, prefer the
# internal one.
both = int_cfgs + ext_cfgs
if not both:
raise ValueError('Invalid board specified: %s.' % board)
return both[0]
# Enumeration of valid settings; any/all config settings must be in this.
# All settings must be documented.
_settings = dict(
# name -- The name of the config.
name=None,
# boards -- A list of boards to build.
boards=None,
# TODO(mtennant): The description sounds independent of anything to do with a
# paladin. See if this should just be: "builder_waterfall_name".
# paladin_builder_name -- Used by paladin logic. The name of the builder on the
# buildbot waterfall if it differs from the config name.
# If None is used, defaults to config name.
paladin_builder_name=None,
# profile -- The profile of the variant to set up and build.
profile=None,
# master -- This bot pushes changes to the overlays.
master=False,
# important -- If False, this flag indicates that the CQ should not check
# whether this bot passed or failed. Set this to False if you are
# setting up a new bot. Once the bot is on the waterfall and is
# consistently green, mark the builder as important=True.
important=False,
# health_threshold -- An integer. If this builder fails this many times
# consecutively, send an alert email to the recipients
# health_alert_recipients. Applies only to builders that
# run CommitQueueSyncCompletion.
# This feature is similar to the ERROR_WATERMARK feature of
# upload_symbols, and it may make sense to merge the
# features at some point.
health_threshold=0,
# health_alert_recipients -- List of email addresses to send health alerts to
# for this builder.
health_alert_recipients=[],
# internal -- Whether this is an internal build config.
internal=False,
# branch -- Whether this is a branched build config. Used for pfq logic.
branch=False,
# manifest -- The name of the manifest to use. E.g., to use the buildtools
# manifest, specify 'buildtools'.
manifest=constants.DEFAULT_MANIFEST,
# dev_manifest -- The name of the manifest to use if we're building on a local
# trybot. This should only require elevated access if it's
# really needed to build this config.
dev_manifest=constants.DEFAULT_MANIFEST,
# do_not_apply_cq_patches -- Applies only to paladin builders. If True, Sync to
# the master manifest without applying any of the
# test patches, rather than running CommitQueueSync.
# This is basically ToT immediately prior to the
# current commit queue run.
do_not_apply_cq_patches=False,
# useflags -- emerge use flags to use while setting up the board, building
# packages, making images, etc.
useflags=[],
# chromeos_official -- Set the variable CHROMEOS_OFFICIAL for the build.
# Known to affect parallel_emerge, cros_set_lsb_release,
# and chromeos_version.sh. See bug chromium-os:14649
chromeos_official=False,
# usepkg_setup_board -- Use binary packages for setup_board. (emerge --usepkg)
usepkg_setup_board=True,
# usepkg_build_packages -- Use binary packages for build_packages.
usepkg_build_packages=True,
# chrome_binhost_only -- Only use binaries in build_packages for Chrome itself.
chrome_binhost_only=False,
# sync_chrome -- Does this profile need to sync chrome? If None, we guess based
# on other factors. If True/False, we always do that.
sync_chrome=None,
# latest_toolchain -- Use the newest ebuilds for all the toolchain packages.
latest_toolchain=False,
# gcc_githash -- This is only valid when latest_toolchain is True.
# If you set this to a commit-ish, the gcc ebuild will use it to build the
# toolchain compiler.
gcc_githash=None,
# board_replace -- wipe and replace the board inside the chroot.
board_replace=False,
# chroot_replace -- wipe and replace chroot, but not source.
chroot_replace=False,
# uprev -- Uprevs the local ebuilds to build new changes since last stable.
# build. If master then also pushes these changes on success.
# Note that we uprev on just about every bot config because it gives us
# a more deterministic build system (the tradeoff being that some bots
# build from source more frequently than if they never did an uprev).
# This way the release/factory/etc... builders will pick up changes
# that devs pushed before it runs, but after the correspoding PFQ bot
# ran (which is what creates+uploads binpkgs). The incremental bots
# are about the only ones that don't uprev because they mimic the flow
# a developer goes through on their own local systems.
uprev=True,
# overlays -- Select what overlays to look at for revving and prebuilts. This
# can be any constants.VALID_OVERLAYS.
overlays=constants.PUBLIC_OVERLAYS,
# push_overlays -- Select what overlays to push at. This should be a subset of
# overlays for the particular builder. Must be None if
# not a master. There should only be one master bot pushing
# changes to each overlay per branch.
push_overlays=None,
# chrome_rev -- Uprev Chrome, values of 'tot', 'stable_release', or None.
chrome_rev=None,
# compilecheck -- Exit the builder right after checking compilation.
# TODO(mtennant): Should be something like "compile_check_only".
compilecheck=False,
# pre_cq -- Test CLs to verify they're ready for the commit queue.
pre_cq=False,
# signer_tests -- Runs the tests that the signer would run.
signer_tests=False,
# unittests -- Runs unittests for packages.
unittests=True,
# quick_unit -- If unittests is true, only run the unit tests for packages which
# have changed since the previous build.
quick_unit=True,
# unittest_blacklist -- A list of the packages to blacklist from unittests.
unittest_blacklist=[],
# build_tests -- Builds autotest tests. Must be True if vm_tests is set.
build_tests=True,
# pgo_generate -- Generates PGO data.
pgo_generate=False,
# pgo_use -- Uses PGO data.
pgo_use=False,
# vm_tests -- Run vm test type defined in constants.
vm_tests=constants.SIMPLE_AU_TEST_TYPE,
# A list of HWTestConfig objects to run.
hw_tests=[],
# upload_hw_test_artifacts -- If true, uploads artifacts for hw testing.
# Upload payloads for test image if the image is
# built. If not, dev image is used and then base
# image.
upload_hw_test_artifacts=False,
# upload_standalone_images -- If true, uploads individual image tarballs.
upload_standalone_images=True,
# gs_path -- Google Storage path to offload files to.
# None - No upload
# GS_PATH_DEFAULT - 'gs://chromeos-image-archive/' + bot_id
# value - Upload to explicit path
gs_path=GS_PATH_DEFAULT,
# TODO(sosa): Deprecate binary.
# build_type -- Type of builder. Check constants.VALID_BUILD_TYPES.
build_type=constants.PFQ_TYPE,
# arm -- Whether the board we are building is arm-based.
arm=False,
archive_build_debug=False,
# images -- List of images we want to build -- see build_image for more details.
images=['test'],
factory_install_netboot=True,
# packages -- Tuple of specific packages we want to build. Most configs won't
# specify anything here and instead let build_packages calculate.
packages=(),
# push_image -- Do we push a final release image to chromeos-images.
push_image=False,
# upload_symbols -- Do we upload debug symbols.
upload_symbols=False,
# hwqual -- Whether we upload a hwqual tarball.
hwqual=False,
# manifest_repo_url -- git repository URL for our manifests.
# External: https://chromium.googlesource.com/chromiumos/manifest
# Internal: https://chrome-internal.googlesource.com/chromeos/manifest-internal
manifest_repo_url=constants.MANIFEST_URL,
# manifest_version -- Whether we are using the manifest_version repo that stores
# per-build manifests.
manifest_version=False,
# use_lkgm -- Use the Last Known Good Manifest blessed by Paladin.
use_lkgm=False,
# use_chrome_lkgm -- LKGM for Chrome OS generated for Chrome builds that are
# blessed from canary runs.
use_chrome_lkgm=False,
# True if this build config is critical for the chrome_lkgm decision.
critical_for_chrome=False,
# prebuilts -- Upload prebuilts for this build. Valid values are PUBLIC,
# PRIVATE, or False.
prebuilts=False,
# use_sdk -- Use SDK as opposed to building the chroot from source.
use_sdk=True,
# trybot_list -- List this config when user runs cbuildbot with --list option
# without the --all flag.
trybot_list=False,
# description -- The description string to print out for config when user runs
# --list.
description=None,
# git_sync -- Boolean that enables parameter --git-sync for upload_prebuilts.
git_sync=False,
# child_configs -- A list of the child config groups, if applicable. See the
# add_group method.
child_configs=[],
# shared_user_password -- Set shared user password for "chronos" user in built
# images. Use "None" (default) to remove the shared
# user password. Note that test images will always set
# the password to "test0000".
shared_user_password=None,
# grouped -- Whether this config belongs to a config group.
grouped=False,
# disk_layout -- layout of build_image resulting image.
# See scripts/build_library/legacy_disk_layout.json or
# overlay-<board>/scripts/disk_layout.json for possible values.
disk_layout=None,
# disk_vm_layout -- layout of image_to_vm.sh resulting image. See
# disk_layout for more info.
disk_vm_layout='2gb-rootfs-updatable',
# postsync_patch -- If enabled, run the PatchChanges stage. Enabled by default.
# Can be overridden by the --nopatch flag.
postsync_patch=True,
# postsync_rexec -- Reexec into the buildroot after syncing. Enabled by
# default.
postsync_reexec=True,
# create_delta_sysroot -- Create delta sysroot during ArchiveStage. Disabled by
# default.
create_delta_sysroot=False,
# TODO(sosa): Collapse to one option.
# ====================== Dev installer prebuilts options =======================
# binhost_bucket -- Upload prebuilts for this build to this bucket. If it equals
# None the default buckets are used.
binhost_bucket=None,
# binhost_key -- Parameter --key for upload_prebuilts. If it equals None, the
# default values are used, which depend on the build type.
binhost_key=None,
# binhost_base_url -- Parameter --binhost-base-url for upload_prebuilts. If it
# equals None, the default value is used.
binhost_base_url=None,
# Upload dev installer prebuilts.
dev_installer_prebuilts=False,
# Enable rootfs verification on the image.
rootfs_verification=True,
# Runs through the chrome-sdk workflow.
chrome_sdk=False,
# If chrome_sdk is set to True, this determines whether we use goma to build
# chrome.
chrome_sdk_goma=False,
# =============================================================================
)
class _JSONEncoder(json.JSONEncoder):
"""Json Encoder that encodes objects as their dictionaries."""
# pylint: disable=E0202
def default(self, obj):
return self.encode(obj.__dict__)
class HWTestConfig(object):
"""Config object for hardware tests suites.
Members:
copy_perf_results: If set to True, copy test results back from GS and send
them to the perf dashboard.
timeout: Number of seconds to wait before timing out waiting for results.
Defaults to 2 hours and ten minutes. This must be less than
lib.parallel._BackgroundTask.MINIMUM_SILENT_TIMEOUT.
pool: Pool to use for hw testing.
async: Fire-and-forget suite.
critical: Usually we consider structural failures here as OK.
num: Maximum number of devices to use when scheduling tests in the hw lab.
file_bugs: Should we file bugs if a test fails in a suite run.
"""
DEFAULT_HW_TEST = 'bvt'
DEFAULT_HW_TEST_TIMEOUT = 60 * 130
# Number of tests running in parallel in the AU suite.
AU_TESTS_NUM = 2
# Number of tests running in parallel in the QAV suite
QAV_TEST_NUM = 2
@classmethod
def DefaultList(cls, **dargs):
"""Returns a default list of HWTestConfig's for a canary build, with
overrides for optional args."""
# Set the number of machines for the au and qav suites. If we are
# constrained in the number of duts in the lab, only give 1 dut to each.
if (dargs.get('num', constants.HWTEST_DEFAULT_NUM) >=
constants.HWTEST_DEFAULT_NUM):
au_dict = dict(num=cls.AU_TESTS_NUM)
qav_dict = dict(num=cls.QAV_TEST_NUM)
else:
au_dict = dict(num=1)
qav_dict = dict(num=1)
au_dargs = dargs.copy()
au_dargs.update(au_dict)
qav_dargs = dargs.copy()
qav_dargs.update(qav_dict)
# BVT + AU suite.
return [cls(cls.DEFAULT_HW_TEST, **dargs),
cls(constants.HWTEST_AU_SUITE, **au_dargs),
cls(constants.HWTEST_QAV_SUITE, **qav_dargs)]
@classmethod
def PGOList(cls, **dargs):
"""Returns a default list of HWTestConfig's for a PGO build, with overrides
for optional args."""
pgo_dict = dict(pool=constants.HWTEST_CHROME_PERF_POOL,
timeout=90 * 60, num=1, async=True)
pgo_dict.update(dargs)
return [cls('pyauto_perf', **pgo_dict),
cls('perf_v2', **pgo_dict)]
@classmethod
def DefaultListCQ(cls, **dargs):
"""Returns a default list of HWTestConfig's for a CQ build,
with overrides for optional args."""
default_dict = dict(pool=constants.HWTEST_PALADIN_POOL, timeout=120 * 60,
file_bugs=False, priority=constants.HWTEST_CQ_PRIORITY)
# Allows dargs overrides to default_dict for cq.
default_dict.update(dargs)
return [cls(cls.DEFAULT_HW_TEST, **default_dict)]
@classmethod
def DefaultListPFQ(cls, **dargs):
"""Returns a default list of HWTestConfig's for a PFQ build,
with overrides for optional args."""
default_dict = dict(pool=constants.HWTEST_MACH_POOL, timeout=120 * 60,
file_bugs=True, priority=constants.HWTEST_PFQ_PRIORITY)
# Allows dargs overrides to default_dict for pfq.
default_dict.update(dargs)
return [cls(cls.DEFAULT_HW_TEST, **default_dict)]
def __init__(self, suite, num=constants.HWTEST_DEFAULT_NUM,
pool=constants.HWTEST_MACH_POOL, copy_perf_results=False,
timeout=DEFAULT_HW_TEST_TIMEOUT, async=False, critical=False,
fatal_timeouts=True, file_bugs=False,
priority=constants.HWTEST_BUILD_PRIORITY):
"""Constructor -- see members above."""
self.suite = suite
self.num = num
self.pool = pool
self.copy_perf_results = copy_perf_results
self.timeout = timeout
self.async = async
self.critical = critical
self.fatal_timeouts = fatal_timeouts
self.file_bugs = file_bugs
self.priority = priority
@property
def timeout_mins(self):
return int(self.timeout/60)
def PGORecordTest(**dargs):
default_dict = dict(pool=constants.HWTEST_CHROME_PFQ_POOL,
critical=True, num=1, file_bugs=False)
# Allows dargs overrides to default_dict for cq.
default_dict.update(dargs)
return HWTestConfig('PGO_record', **default_dict)
# TODO(mtennant): Rename this BuildConfig?
class _config(dict):
"""Dictionary of explicit configuration settings for a cbuildbot config
Each dictionary entry is in turn a dictionary of config_param->value.
See _settings for details on known configurations, and their documentation.
"""
_URLQUOTED_PARAMS = ('paladin_builder_name',)
def __getattr__(self, name):
"""Support attribute-like access to each dict entry."""
if name in self:
return self[name]
return super(_config, self).__getattr__(name)
def derive(self, *inherits, **overrides):
"""Create a new config derived from this one.
Args:
inherits: Mapping instances to mixin.
overrides: Settings to inject; see _settings for valid values.
Returns:
A new _config instance.
"""
new_config = copy.deepcopy(self)
for update_config in inherits:
new_config.update(update_config)
new_config.update(overrides)
return copy.deepcopy(new_config)
def add_config(self, name, *inherits, **overrides):
"""Derive and add the config to cbuildbots usable config targets
Args:
name: The name to label this configuration; this is what cbuildbot
would see.
inherits: See the docstring of derive.
overrides: See the docstring of derive.
Returns:
See the docstring of derive.
"""
overrides['name'] = name
new_config = self.derive(*inherits, **overrides)
# Derive directly from defaults so missing values are added.
# Store a dictionary, rather than our derivative- this is
# to ensure any far flung consumers of the config dictionary
# aren't affected by recent refactorings.
config_dict = _default.derive(self, *inherits, **overrides)
config_dict.update(
(key, urllib.quote(config_dict[key]))
for key in self._URLQUOTED_PARAMS if config_dict.get(key))
# TODO(mtennant): This is just confusing. Some random _config object
# (self) can add a new _config object to the global config dict. Even if
# self is itself not a part of the global config dict.
config[name] = config_dict
return new_config
@classmethod
def add_raw_config(cls, name, *inherits, **overrides):
return cls().add_config(name, *inherits, **overrides)
@classmethod
def add_group(cls, name, *configs, **group_overrides):
"""Create a new group of build configurations.
Args:
name: The name to label this configuration; this is what cbuildbot
would see.
configs: Configurations to build in this group. The first config in
the group is considered the primary configuration and is used
for syncing and creating the chroot.
group_overrides: See the docstring of derive. Applies to entire group.
"""
child_configs = [_default.derive(x, grouped=True) for x in configs]
group_overrides['child_configs'] = child_configs
return configs[0].add_config(name, **group_overrides)
_default = _config(**_settings)
# It is only safe to inherit prebuilts from generic boards, or from the
# same board without the variant. This rule helps keep inheritance trees
# sane.
generic_boards = set([
'x86-generic', 'amd64-generic', 'daisy'
])
# Arch-specific mixins.
arm = _config(
arm=True,
# VM/tests are broken on arm.
unittests=False,
vm_tests=None,
)
amd64 = _config()
# Builder-specific mixins
binary = _config(
# Full builds that build fully from binaries.
quick_unit=False,
build_type=constants.BUILD_FROM_SOURCE_TYPE,
archive_build_debug=True,
images=['test', 'factory_test', 'factory_install'],
git_sync=True,
)
full = _config(
# Full builds are test builds to show that we can build from scratch,
# so use settings to build from scratch, and archive the results.
usepkg_setup_board=False,
usepkg_build_packages=False,
chrome_sdk=True,
chroot_replace=True,
quick_unit=False,
build_type=constants.BUILD_FROM_SOURCE_TYPE,
archive_build_debug=True,
images=['base', 'test', 'factory_test', 'factory_install'],
git_sync=True,
trybot_list=True,
description='Full Builds',
)
# Full builders with prebuilts.
full_prebuilts = full.derive(
prebuilts=constants.PUBLIC,
)
pfq = _config(
build_type=constants.PFQ_TYPE,
important=True,
uprev=True,
overlays=constants.PUBLIC_OVERLAYS,
manifest_version=True,
trybot_list=True,
)
paladin = _config(
important=True,
build_type=constants.PALADIN_TYPE,
overlays=constants.PUBLIC_OVERLAYS,
prebuilts=constants.PUBLIC,
manifest_version=True,
trybot_list=True,
description='Commit Queue',
)
# Used for paladin builders that build from source.
full_paladin = _config(
board_replace=True,
chrome_binhost_only=True,
)
# Incremental builders are intended to test the developer workflow.
# For that reason, they don't uprev.
incremental = _config(
build_type=constants.INCREMENTAL_TYPE,
uprev=False,
overlays=constants.PUBLIC_OVERLAYS,
description='Incremental Builds',
)
# This builds with more source available.
internal = _config(
internal=True,
overlays=constants.BOTH_OVERLAYS,
manifest_repo_url=constants.MANIFEST_INT_URL,
)
# This adds Chrome branding.
official_chrome = _config(
useflags=[constants.USE_CHROME_INTERNAL, constants.USE_CHROME_PDF],
)
# This sets chromeos_official.
official = official_chrome.derive(
chromeos_official=True,
)
_cros_sdk = full_prebuilts.add_config('chromiumos-sdk',
# The amd64-host has to be last as that is when the toolchains
# are bundled up for inclusion in the sdk.
boards=('x86-generic', 'arm-generic', 'amd64-generic'),
build_type=constants.CHROOT_BUILDER_TYPE,
use_sdk=False,
trybot_list=True,
description='Build the SDK and all the cross-compilers',
)
asan = _config(
chroot_replace=True,
profile='asan',
useflags=['asan'], # see profile for more
disk_layout='2gb-rootfs',
disk_vm_layout='2gb-rootfs-updatable',
)
_config.add_raw_config('refresh-packages',
boards=['x86-generic', 'arm-generic'],
build_type=constants.REFRESH_PACKAGES_TYPE,
description='Check upstream Gentoo for package updates',
)
incremental.add_config('x86-generic-incremental',
boards=['x86-generic'],
)
incremental.add_config('daisy-incremental',
arm,
boards=['daisy'],
)
incremental.add_config('amd64-generic-incremental',
amd64,
boards=['amd64-generic'],
# This builder runs on a VM, so it can't run VM tests.
vm_tests=None,
)
incremental.add_config('x32-generic-incremental',
amd64,
boards=['x32-generic'],
# This builder runs on a VM, so it can't run VM tests.
vm_tests=None,
)
paladin.add_config('x86-generic-paladin',
boards=['x86-generic'],
paladin_builder_name='x86-generic paladin',
)
paladin.add_config('amd64-generic-paladin',
amd64,
boards=['amd64-generic'],
paladin_builder_name='amd64-generic paladin',
)
paladin.add_config('x32-generic-paladin',
amd64,
boards=['x32-generic'],
paladin_builder_name='x32-generic paladin',
important=False,
)
telemetry = _config(
build_type=constants.INCREMENTAL_TYPE,
uprev=False,
overlays=constants.PUBLIC_OVERLAYS,
vm_tests=constants.TELEMETRY_SUITE_TEST_TYPE,
description='Telemetry Builds',
)
telemetry.add_config('amd64-generic-telemetry',
amd64,
boards=['amd64-generic'],
)
telemetry.add_config('arm-generic-telemetry',
arm,
boards=['arm-generic'],
)
telemetry.add_config('x86-generic-telemetry',
boards=['x86-generic'],
)
chromium_pfq = _config(
build_type=constants.CHROME_PFQ_TYPE,
important=True,
uprev=False,
overlays=constants.PUBLIC_OVERLAYS,
manifest_version=True,
chrome_rev=constants.CHROME_REV_LATEST,
upload_hw_test_artifacts=True,
)
# TODO(davidjames): Convert this to an external config once the unified master
# logic is ready.
internal_chromium_pfq = internal.derive(
chromium_pfq,
description='Preflight Chromium Build (internal)',
overlays=constants.BOTH_OVERLAYS,
prebuilts=constants.PUBLIC,
)
internal_chromium_pfq.add_config('x86-generic-chromium-pfq',
boards=['x86-generic'],
master=True,
push_overlays=constants.BOTH_OVERLAYS,
)
internal_chromium_pfq.add_config('daisy-chromium-pfq',
arm,
boards=['daisy'],
)
internal_chromium_pfq.add_config('amd64-generic-chromium-pfq',
amd64,
disk_layout='2gb-rootfs',
boards=['amd64-generic'],
)
chrome_pfq = internal_chromium_pfq.derive(
official,
important=True,
overlays=constants.BOTH_OVERLAYS,
description='Preflight Chrome build (internal)',
upload_hw_test_artifacts=True,
prebuilts=constants.PRIVATE,
)
chrome_pfq.add_config('alex-chrome-pfq',
boards=['x86-alex'],
)
chrome_pfq.add_config('lumpy-chrome-pfq',
boards=['lumpy'],
hw_tests=HWTestConfig.DefaultListPFQ(),
upload_hw_test_artifacts=True,
)
chrome_pfq.add_config('daisy_spring-chrome-pfq',
arm,
boards=['daisy_spring'],
)
chromium_info = chromium_pfq.derive(
chrome_rev=constants.CHROME_REV_TOT,
use_lkgm=True,
important=False,
manifest_version=False,
vm_tests=constants.SMOKE_SUITE_TEST_TYPE,
disk_vm_layout='usb',
)
chrome_info = chromium_info.derive(
internal, official,
)
# Config with the official flags except pdf.
chrome_info_no_pdf = chrome_info.derive(
useflags=[y for y in official['useflags'] if y != constants.USE_CHROME_PDF],
)
chrome_perf = chrome_info.derive(
vm_tests=None,
upload_hw_test_artifacts=True,
hw_tests=[HWTestConfig('perf_v2', pool=constants.HWTEST_CHROME_PERF_POOL,
timeout=90 * 60, critical=True, num=1,
copy_perf_results=True)],
use_chrome_lkgm=True,
use_lkgm=False,
useflags=official['useflags'] + ['-cros-debug'],
)
chrome_perf.add_config('daisy-chrome-perf',
boards=['daisy'],
trybot_list=True,
)
chrome_perf.add_config('lumpy-chrome-perf',
boards=['lumpy'],
trybot_list=True,
)
chrome_perf.add_config('parrot-chrome-perf',
boards=['parrot'],
trybot_list=True,
)
chromium_info_x86 = \
chromium_info.add_config('x86-generic-tot-chrome-pfq-informational',
boards=['x86-generic'],
)
chromium_info_daisy = \
chromium_info.add_config('daisy-tot-chrome-pfq-informational',
arm,
boards=['daisy'],
)
chromium_info_amd64 = \
chromium_info.add_config('amd64-generic-tot-chrome-pfq-informational',
amd64,
boards=['amd64-generic'],
)
chromium_info.add_config('x32-generic-tot-chrome-pfq-informational',
amd64,
boards=['x32-generic'],
)
chrome_info.add_config('alex-tot-chrome-pfq-informational',
boards=['x86-alex'],
)
chrome_info.add_config('lumpy-tot-chrome-pfq-informational',
boards=['lumpy'],
)
# WebRTC configurations.
chrome_info_no_pdf.add_config('alex-webrtc-chrome-pfq-informational',
boards=['x86-alex'],
)
chrome_info_no_pdf.add_config('lumpy-webrtc-chrome-pfq-informational',
boards=['lumpy'],
)
chrome_info_no_pdf.add_config('daisy-webrtc-chrome-pfq-informational',
arm,
boards=['daisy'],
)
chromium_info_x86.add_config('x86-webrtc-chromium-pfq-informational',
archive_build_debug=True,
)
chromium_info_amd64.add_config('amd64-webrtc-chromium-pfq-informational',
archive_build_debug=True,
)
chromium_info_daisy.add_config('daisy-webrtc-chromium-pfq-informational',
archive_build_debug=True,
)
_arm_full_boards = frozenset([
'arm-generic',
'daisy',
'daisy_spring',
])
_x86_full_boards = frozenset([
'amd64-generic',
'butterfly',
'parrot',
'link',
'lumpy',
'stout',
'stout32',
'stumpy',
'x32-generic',
'x86-alex',
'x86-generic',
'x86-mario',
'x86-pineview',
'x86-zgb',
])
def _AddFullConfigs():
"""Add x86 and arm full configs."""
for board in _x86_full_boards:
full_prebuilts.add_config('%s-%s' % (board, CONFIG_TYPE_FULL),
boards=[board],
upload_hw_test_artifacts=True,
)
for board in _arm_full_boards:
full_prebuilts.add_config('%s-%s' % (board, CONFIG_TYPE_FULL),
arm,
boards=[board],
upload_hw_test_artifacts=True,
)
_AddFullConfigs()
_toolchain_major = _cros_sdk.add_config('toolchain-major',
latest_toolchain=True,
prebuilts=False,
gcc_githash='gcc.gnu.org/branches/google/main',
description='Test next major toolchain revision',
)
_toolchain_minor = _cros_sdk.add_config('toolchain-minor',
latest_toolchain=True,
prebuilts=False,
gcc_githash='gcc.gnu.org/branches/google/gcc-4_8-mobile',
description='Test next minor toolchain revision',
)
incremental.add_config('x86-generic-asan',
asan,
boards=['x86-generic'],
description='Build with Address Sanitizer (Clang)',
)
chromium_info.add_config('x86-generic-tot-asan-informational',
asan,
boards=['x86-generic'],
description='Build with Address Sanitizer (Clang) on TOT',
)
incremental.add_config('amd64-generic-asan',
amd64,
asan,
boards=['amd64-generic'],
description='Build with Address Sanitizer (Clang)',
)
chromium_info.add_config('amd64-generic-tot-asan-informational',
amd64,
asan,
boards=['amd64-generic'],
description='Build with Address Sanitizer (Clang) on TOT',
)
incremental_arm = incremental.derive(arm)
incremental_arm.add_config('beaglebone-incremental',
boards=['beaglebone'],
images=['base'],
packages=['chromeos-base/chromeos'],
trybot_list=True,
build_tests=False,
rootfs_verification=False,
description='Incremental Beaglebone Builder',
)
platform2 = _config(
useflags=['platform2'],
trybot_list=True,
description='Platform2 Builder',
)
platform2_incremental = platform2.derive(
incremental,
description='Incremental Platform2 Builder',
)
platform2_incremental.add_config('x86-incremental-platform2',
boards=['x86-generic'],
)
platform2_incremental.add_config('daisy-incremental-platform2',
arm,
boards=['daisy'],
)
#
# Internal Builds
#
internal_pfq = internal.derive(official_chrome, pfq,
overlays=constants.BOTH_OVERLAYS,
prebuilts=constants.PRIVATE,
)
# Because branch directories may be shared amongst builders on multiple
# branches, they must delete the chroot every time they run.
internal_pfq_branch = internal_pfq.derive(branch=True, chroot_replace=True,
trybot_list=False)
internal_paladin = internal.derive(official_chrome, paladin,
overlays=constants.BOTH_OVERLAYS,
prebuilts=constants.PRIVATE,
vm_tests=None,
description=paladin['description'] + ' (internal)',
)
# Used for paladin builders with nowithdebug flag (a.k.a -cros-debug)
internal_nowithdebug_paladin = internal_paladin.derive(
useflags=official['useflags'] + ['-cros-debug'],
description=paladin['description'] + ' (internal, nowithdebug)',
)
internal_nowithdebug_paladin.add_config('x86-generic-nowithdebug-paladin',
boards=['x86-generic'],
paladin_builder_name='x86-generic nowithdebug-paladin',
prebuilts=False,
important=False,
)
internal_nowithdebug_paladin.add_config('amd64-generic-nowithdebug-paladin',
boards=['amd64-generic'],
paladin_builder_name='amd64-generic nowithdebug-paladin',
prebuilts=False,
important=False,
)
internal_nowithdebug_paladin.add_config('x86-mario-nowithdebug-paladin',
boards=['x86-mario'],
paladin_builder_name='x86-mario nowithdebug-paladin',
prebuilts=False,
important=False,
)
internal_pre_cq = internal_paladin.derive(
build_type=constants.INCREMENTAL_TYPE,
compilecheck=True,
pre_cq=True,
description='Verifies compilation and unit tests',
)
internal_pre_cq.add_group(constants.PRE_CQ_BUILDER_NAME,
internal_pre_cq.add_config('parrot-pre-cq', boards=['parrot']),
internal_pre_cq.add_config('lumpy-pre-cq', boards=['lumpy']),
internal_pre_cq.add_config('daisy_spring-pre-cq',
arm, boards=['daisy_spring']),
internal_pre_cq.add_config('peach_pit-pre-cq',
arm, boards=['peach_pit']),
)
internal_paladin.add_config('pre-cq-launcher',
boards=[],
build_type=constants.PRE_CQ_LAUNCHER_TYPE,
description='Launcher for Pre-CQ builders.',
)
internal_paladin.add_config(constants.BRANCH_UTIL_CONFIG,
boards=[],
# Disable postsync_patch to prevent conflicting patches from being applied -
# e.g., patches from 'master' branch being applied to a branch.
postsync_patch=False,
# Disable postsync_reexec to continue running the 'master' branch chromite
# for all stages, rather than the chromite in the branch buildroot.
postsync_reexec=False,
build_type=constants.CREATE_BRANCH_TYPE,
description='Used for creating/deleting branches.',
)
# Internal incremental builders don't use official chrome because we want
# to test the developer workflow.
internal_incremental = internal.derive(
incremental,
overlays=constants.BOTH_OVERLAYS,
description='Incremental Builds (internal)',
)
sonic = _config(
manifest='sonic.xml',
boards=['sonic'],
# Until these are configured and ready, disable them.
images=['base', 'dev',],
unittests=False,
upload_hw_test_artifacts=False,
build_tests=False,
vm_tests=None,
signer_tests=False,
hw_tests=[],
)
internal_pfq_branch.add_config('x86-alex-pre-flight-branch',
master=True,
push_overlays=constants.BOTH_OVERLAYS,
boards=['x86-alex'],
)
internal.add_config('test-ap',
vm_tests=None,
description='stumpy image used for WiFi testing',
boards=['stumpy'],
images=['test'], # We need Python and utilities found only in test.
profile='testbed-ap',
)
### Master paladin (CQ builder).
internal_paladin.add_config('x86-mario-paladin',
master=True,
push_overlays=constants.BOTH_OVERLAYS,
boards=['x86-mario'],
paladin_builder_name='x86-mario paladin',
vm_tests=constants.SIMPLE_AU_TEST_TYPE,
health_threshold=3,
health_alert_recipients=['chromeos-build-alerts@google.com']
)
### Other paladins (CQ builders).
# These are slaves of the master paladin by virtue of matching
# in a few config values (e.g. 'build_type', 'branch', etc). If
# they are not 'important' then they are ignored slaves.
# TODO(mtennant): This master-slave relationship should be specified
# here in the configuration, rather than BuilderStage._GetSlavesForMaster.
# Something like the following:
# master_paladin = internal_paladin.add_config(...)
# master_paladin.AddSlave(internal_paladin.add_config(...))
# Sanity check builder, part of the CQ but builds without the patches
# under test.
paladin.add_config('link-tot-paladin',
boards=['link'],
paladin_builder_name='link ToT paladin',
do_not_apply_cq_patches=True,
important=False,
vm_tests=constants.SIMPLE_AU_TEST_TYPE,
prebuilts=False,
)
internal_paladin.add_config('x86-alex-paladin',
boards=['x86-alex'],
paladin_builder_name='x86-alex paladin',
hw_tests=HWTestConfig.DefaultListCQ(),
upload_hw_test_artifacts=True,
)
internal_paladin.add_config('bayleybay-paladin',
boards=['bayleybay'],
paladin_builder_name='bayleybay paladin',
important=False,
)
internal_paladin.add_config('beltino-paladin',
boards=['beltino'],
paladin_builder_name='beltino paladin',
)
# x86 full compile
internal_paladin.add_config('butterfly-paladin',
full_paladin,
boards=['butterfly'],
paladin_builder_name='butterfly paladin',
)
# amd64 full compile
internal_paladin.add_config('falco-paladin',
full_paladin,
boards=['falco'],
paladin_builder_name='falco paladin',
)
internal_paladin.add_config('fox_wtm2-paladin',
boards=['fox_wtm2'],
paladin_builder_name='fox_wtm2 paladin',
vm_tests=None,
important=False,
)
internal_paladin.add_config('leon-paladin',
boards=['leon'],
paladin_builder_name='leon paladin',
)
internal_paladin.add_config('link-paladin',
boards=['link'],
paladin_builder_name='link paladin',
upload_hw_test_artifacts=True,
)
internal_paladin.add_config('lumpy-paladin',
boards=['lumpy'],
paladin_builder_name='lumpy paladin',
hw_tests=HWTestConfig.DefaultListCQ(),
upload_hw_test_artifacts=True,
)
internal_paladin.add_config('parrot-paladin',
boards=['parrot'],
paladin_builder_name='parrot paladin',
hw_tests=HWTestConfig.DefaultListCQ(),
upload_hw_test_artifacts=True,
)
incompatible_instruction_set = _config(
# If a board has a newer instruction set, then the unit tests and VM tests
# cannot run on the builders, at least until we've implemented an emulator.
# Disable the VM tests and unit tests to be safe.
vm_tests=None,
unittests=False,
)
internal_paladin.add_config('rambi-paladin',
incompatible_instruction_set,
boards=['rambi'],
paladin_builder_name='rambi paladin',
)
internal_paladin.add_config('samus-paladin',
boards=['samus'],
paladin_builder_name='samus paladin',
vm_tests=None,
important=False,
)
# x86 full unit tests
internal_paladin.add_config('peppy-paladin',
boards=['peppy'],
paladin_builder_name='peppy paladin',
)
internal_paladin.add_config('slippy-paladin',
boards=['slippy'],
paladin_builder_name='slippy paladin',
)
internal_paladin.add_config('sonic-paladin',
sonic,
paladin_builder_name='sonic paladin',
important=False,
)
# amd64 full unit tests
internal_paladin.add_config('monroe-paladin',
boards=['monroe'],
hw_tests=[],
important=False,
quick_unit=False,
vm_tests=None,
paladin_builder_name='monroe paladin',
)
internal_paladin.add_config('panther-paladin',
boards=['panther'],
hw_tests=[],
important=False,
quick_unit=False,
vm_tests=None,
paladin_builder_name='panther paladin',
)
internal_paladin.add_config('stout-paladin',
boards=['stout'],
quick_unit=False,
paladin_builder_name='stout paladin',
)
internal_paladin.add_config('stout32-paladin',
boards=['stout32'],
paladin_builder_name='stout32 paladin',
important=False,
)
internal_paladin.add_config('stumpy-paladin',
boards=['stumpy'],
paladin_builder_name='stumpy paladin',
hw_tests=HWTestConfig.DefaultListCQ(),
upload_hw_test_artifacts=True,
)
internal_paladin.add_config('wolf-paladin',
boards=['wolf'],
paladin_builder_name='wolf paladin',
)
internal_paladin.add_config('x86-zgb-paladin',
boards=['x86-zgb'],
important=False,
paladin_builder_name='x86-zgb paladin',
)
### Arm paladins (CQ builders).
internal_arm_paladin = internal_paladin.derive(arm)
internal_arm_paladin.add_config('daisy-paladin',
boards=['daisy'],
paladin_builder_name='daisy paladin',
hw_tests=HWTestConfig.DefaultListCQ(),
upload_hw_test_artifacts=True,
)
# arm full compile
internal_arm_paladin.add_config('beaglebone-paladin',
boards=['beaglebone'],
packages=['chromeos-base/chromeos'],
images=['base'],
rootfs_verification=False,
important=False,
paladin_builder_name='beaglebone paladin',
)
internal_arm_paladin.add_config('daisy_spring-paladin',
full_paladin,
boards=['daisy_spring'],
paladin_builder_name='daisy_spring paladin',
)
internal_arm_paladin.add_config('peach_pit-paladin',
boards=['peach_pit'],
paladin_builder_name='peach_pit paladin',
)
internal_incremental.add_config('mario-incremental',
boards=['x86-mario'],
)
_toolchain_major.add_config('internal-toolchain-major', internal, official,
boards=('x86-alex', 'stumpy', 'daisy'),
use_lkgm=True,
useflags=[constants.USE_CHROME_INTERNAL],
build_tests=True,
description=_toolchain_major['description'] + ' (internal)',
)
_toolchain_minor.add_config('internal-toolchain-minor', internal, official,
boards=('x86-alex', 'stumpy', 'daisy'),
use_lkgm=True,
useflags=[constants.USE_CHROME_INTERNAL],
build_tests=True,
description=_toolchain_minor['description'] + ' (internal)',
)
_release = full.derive(official, internal,
build_type=constants.CANARY_TYPE,
useflags=official['useflags'] + ['-cros-debug', '-highdpi'],
build_tests=True,
manifest_version=True,
images=['base', 'test', 'factory_test', 'factory_install'],
push_image=True,
upload_symbols=True,
binhost_bucket='gs://chromeos-dev-installer',
binhost_key='RELEASE_BINHOST',
binhost_base_url=
'https://commondatastorage.googleapis.com/chromeos-dev-installer',
dev_installer_prebuilts=True,
git_sync=False,
vm_tests=constants.SMOKE_SUITE_TEST_TYPE,
disk_vm_layout='usb',
hw_tests=HWTestConfig.DefaultList(file_bugs=True),
upload_hw_test_artifacts=True,
signer_tests=True,
trybot_list=True,
hwqual=True,
description="Release Builds (canary) (internal)",
chrome_sdk=True,
)
### Master release config.
_release.add_config('x86-mario-release',
boards=['x86-mario'],
master=True,
)
### Release config groups.
_config.add_group('x86-alex-release-group',
_release.add_config('x86-alex-release',
boards=['x86-alex'],
),
_release.add_config('x86-alex_he-release',
boards=['x86-alex_he'],
vm_tests=None,
unittests=None,
hw_tests=[],
upload_hw_test_artifacts=False,
),
critical_for_chrome=True,
)
_config.add_group('x86-zgb-release-group',
_release.add_config('x86-zgb-release',
boards=['x86-zgb'],
),
_release.add_config('x86-zgb_he-release',
boards=['x86-zgb_he'],
vm_tests=None,
unittests=None,
hw_tests=[],
upload_hw_test_artifacts=False,
),
)
### Release PGO configs.
release_pgo = _release.derive(
hw_tests=HWTestConfig.DefaultList(pool=constants.HWTEST_CHROME_PERF_POOL,
num=4) +
HWTestConfig.PGOList(),
push_image=False,
dev_installer_prebuilts=False,
)
release_pgo.add_group('x86-alex-release-pgo',
release_pgo.add_config('x86-alex-release-pgo-generate',
boards=['x86-alex'],
pgo_generate=True,
),
release_pgo.add_config('x86-alex-release-pgo-use',
boards=['x86-alex'],
pgo_use=True,
),
)
release_pgo.add_group('lumpy-release-pgo',
release_pgo.add_config('lumpy-release-pgo-generate',
boards=['lumpy'],
pgo_generate=True,
),
release_pgo.add_config('lumpy-release-pgo-use',
boards=['lumpy'],
pgo_use=True,
),
)
release_pgo.add_group('parrot-release-pgo',
release_pgo.add_config('parrot-release-pgo-generate',
boards=['parrot'],
pgo_generate=True,
),
release_pgo.add_config('parrot-release-pgo-use',
boards=['parrot'],
pgo_use=True,
),
)
release_pgo.add_group('daisy-release-pgo',
release_pgo.add_config('daisy-release-pgo-generate',
boards=['daisy'],
pgo_generate=True,
),
release_pgo.add_config('daisy-release-pgo-use',
boards=['daisy'],
pgo_use=True,
),
)
### Release configs.
# bayleybay-release does not enable vm_tests or unittests due to the compiler
# flags enabled for baytrail.
_release.add_config('bayleybay-release',
boards=['bayleybay'],
hw_tests=[],
vm_tests=None,
unittests=False,
)
_release.add_config('beltino-release',
boards=['beltino'],
hw_tests=[],
vm_tests=None,
)
_release.add_config('butterfly-release',
boards=['butterfly'],
)
_release.add_config('falco-release',
boards=['falco'],
)
_release.add_config('fox_wtm2-release',
boards=['fox_wtm2'],
# Until these are configured and ready, disable them.
signer_tests=False,
vm_tests=None,
hw_tests=[],
)
_release.add_config('leon-release',
boards=['leon'],
)
_release.add_config('link-release',
boards=['link'],
useflags=_release['useflags'] + ['highdpi'],
)
_release.add_config('lumpy-release',
boards=['lumpy'],
critical_for_chrome=True,
)
_release.add_config('monroe-release',
boards=['monroe'],
hw_tests=[],
vm_tests=None,
)
_release.add_config('panther-release',
boards=['panther'],
hw_tests=[],
vm_tests=None,
)
_release.add_group('parrot-release-group',
_release.add_config('parrot-release',
boards=['parrot'],
),
_release.add_config('parrot_ivb-release',
boards=['parrot_ivb'],
)
)
_release.add_config('peppy-release',
boards=['peppy'],
)
# rambi-release does not enable vm_tests or unittests due to the compiler
# flags enabled for baytrail.
_release.add_config('rambi-release',
incompatible_instruction_set,
boards=['rambi'],
hw_tests=[],
)
_release.add_config('samus-release',
boards=['samus'],
hw_tests=[],
)
_release.add_config('slippy-release',
boards=['slippy'],
hw_tests=[],
)
_release.add_config('sonic-release',
sonic,
)
_release.add_config('stout-release',
boards=['stout'],
)
_release.add_config('stout32-release',
boards=['stout32'],
)
_release.add_config('stumpy-release',
boards=['stumpy'],
)
_release.add_config('wolf-release',
boards=['wolf'],
)
### Arm release configs.
_arm_release = _release.derive(arm)
_arm_release.add_config('beaglebone-release',
boards=['beaglebone'],
chrome_sdk=False,
build_tests=False,
dev_installer_prebuilts=False,
packages=['chromeos-base/chromeos'],
hw_tests=[],
images=['base'],
rootfs_verification=False,
signer_tests=False,
upload_hw_test_artifacts=True,
)
_arm_release.add_config('daisy-release',
boards=['daisy'],
critical_for_chrome=True,
)
_arm_release.add_config('daisy_spring-release',
boards=['daisy_spring'],
)
_arm_release.add_config('peach_pit-release',
boards=['peach_pit'],
)
# Factory and Firmware releases much inherit from these classes. Modifications
# for these release builders should go here.
# Naming conventions also must be followed. Factory and firmware branches must
# end in -factory or -firmware suffixes.
_factory_release = _release.derive(
upload_hw_test_artifacts=False,
upload_symbols=False,
hw_tests=[],
description='Factory Builds',
)
_firmware = _config(
images=[],
packages=('virtual/chromeos-firmware',),
usepkg_setup_board=True,
usepkg_build_packages=True,
sync_chrome=False,
build_tests=False,
chrome_sdk=False,
unittests=False,
vm_tests=None,
hw_tests=[],
dev_installer_prebuilts=False,
upload_hw_test_artifacts=False,
upload_symbols=False,
signer_tests=False,
trybot_list=False,
)
_firmware_release = _release.derive(_firmware,
description='Firmware Canary',
)
_depthcharge_release = _firmware_release.derive(useflags=['depthcharge'])
_depthcharge_full_internal = full.derive(
internal,
_firmware,
useflags=['depthcharge'],
description='Firmware Informational',
)
_x86_firmware_boards = (
'bayleybay',
'beltino',
'butterfly',
'falco',
'leon',
'link',
'lumpy',
'monroe',
'panther',
'parrot',
'peppy',
'rambi',
'samus',
'stout',
'stout32',
'slippy',
'stumpy',
'wolf',
'x86-mario',
)
_x86_depthcharge_firmware_boards = (
'bayleybay',
'leon',
'link',
'rambi',
'samus',
)
_arm_firmware_boards = (
'daisy',
'daisy_spring',
)
def _AddFirmwareConfigs():
"""Add x86 and arm firmware configs."""
for board in _x86_firmware_boards:
_firmware_release.add_config('%s-%s' % (board, CONFIG_TYPE_FIRMWARE),
boards=[board],
)
for board in _x86_depthcharge_firmware_boards:
_depthcharge_release.add_config(
'%s-%s-%s' % (board, 'depthcharge', CONFIG_TYPE_FIRMWARE),
boards=[board],
)
_depthcharge_full_internal.add_config(
'%s-%s-%s-%s' % (board, 'depthcharge', CONFIG_TYPE_FULL,
CONFIG_TYPE_FIRMWARE),
boards=[board],
)
for board in _arm_firmware_boards:
_firmware_release.add_config('%s-%s' % (board, CONFIG_TYPE_FIRMWARE),
arm,
boards=[board],
)
_AddFirmwareConfigs()
# This is an example factory branch configuration for x86.
# Modify it to match your factory branch.
_factory_release.add_config('x86-mario-factory',
boards=['x86-mario'],
)
# This is an example factory branch configuration for arm.
# Modify it to match your factory branch.
_factory_release.add_config('daisy-factory',
arm,
boards=['daisy'],
)
def _GetDisplayPosition(config_name, type_order=CONFIG_TYPE_DUMP_ORDER):
"""Given a config_name, return display position specified by suffix_order.
Args:
config_name: Name of config to look up.
type_order: A tuple/list of config types in the order they are to be
displayed.
If config name does not contain any of the suffixes, returns the index
position after the last element of suffix_order.
"""
for index, config_type in enumerate(type_order):
if config_name.endswith('-' + config_type) or config_name == config_type:
return index
return len(type_order)
def _InjectDisplayPosition(config_source):
"""Add field to help buildbot masters order builders on the waterfall."""
def _GetSortKey(items):
my_config = items[1]
# Allow configs to override the display_position.
return (my_config.get('display_position', 1000000),
_GetDisplayPosition(my_config['name']),
my_config['internal'], my_config['vm_tests'])
source = sorted(config_source.iteritems(), key=_GetSortKey)
return dict((name, dict(value.items() + [('display_position', idx)]))
for idx, (name, value) in enumerate(source))
def main(argv=None):
if not argv:
argv = sys.argv[1:]
usage = "usage: %prog [options] [config1 ... configN]"
parser = optparse.OptionParser(usage=usage)
parser.add_option('-c', '--compare', action='store', type='string',
default=None, metavar='file_name',
help=('Compare current config against a saved on disk '
'serialized version of a config.'))
parser.add_option('-d', '--dump', action='store_true', default=False,
help=('Dump the configs in JSON format.'))
parser.add_option('--for-buildbot', action='store_true', default=False,
help="Include the display position in json data.")
options, configs = parser.parse_args(argv)
if options.compare and options.dump:
parser.error('Cannot run with --load and --dump at the same time!')
elif not options.compare and not options.dump:
parser.print_help()
sys.exit(0)
convert = lambda x:x
if options.for_buildbot:
convert = _InjectDisplayPosition
my_config = convert(config)
# If configs specified, only dump/load those.
if configs:
temp_config = dict()
for c in configs:
temp_config[c] = my_config[c]
my_config = temp_config
if options.dump:
print json.dumps(my_config, cls=_JSONEncoder)
elif options.compare:
with open(options.compare, 'rb') as f:
original = convert(json.load(f))
my_config = json.loads(json.dumps(my_config, cls=_JSONEncoder))
for key in sorted(set(my_config.keys() + original.keys())):
obj1, obj2 = original.get(key), my_config.get(key)
if obj1 == obj2:
continue
elif obj1 is None:
print '%s: added to config\n' % (key,)
continue
elif obj2 is None:
print '%s: removed from config\n' % (key,)
continue
print '%s:' % (key,)
for subkey in sorted(set(obj1.keys() + obj2.keys())):
sobj1, sobj2 = obj1.get(subkey), obj2.get(subkey)
if sobj1 != sobj2:
print ' %s: %r, %r' % (subkey, sobj1, sobj2)
print
if __name__ == '__main__':
main()