blob: 954bc0aa112f8507e85e5239061c1b530923b147 [file] [log] [blame] [edit]
#!/usr/bin/env python2
# Copyright 2015 The Chromium Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
"""Cleanup orphaned containers.
If an autoserv process dies without being able to call handler of SIGTERM, the
container used to run the test will be orphaned. This adds overhead to the
drone. This script is used to clean up such containers.
This module also checks if the test job associated with a container has
finished. If so, kill the autoserv process for the test job and destroy the
container. To avoid racing condition, this only applies to job finished at least
1 hour ago.
"""
import argparse
import datetime
import logging
import os
import signal
import common
from autotest_lib.client.common_lib import logging_config
from autotest_lib.client.common_lib import time_utils
from autotest_lib.client.common_lib import utils
from autotest_lib.server.cros.dynamic_suite import frontend_wrappers
from autotest_lib.site_utils import lxc
AFE = frontend_wrappers.RetryingAFE(timeout_min=0.1, delay_sec=10)
# The cutoff time to declare a test job is completed and container is orphaned.
# This is to avoid a race condition that scheduler aborts a job and autoserv
# is still in the process of destroying the container it used.
FINISHED_JOB_CUTOFF_TIME = datetime.datetime.now() - datetime.timedelta(hours=1)
def is_container_orphaned(container):
"""Check if a container is orphaned.
A container is orphaned if any of these condition is True:
1. The autoserv process created the container is no longer running.
2. The test job is finished at least 1 hour ago.
@param container: A Container object.
@return: True if the container is orphaned.
"""
logging.debug('Checking if container is orphaned: %s', container.name)
if container.id is None:
logging.debug('Container %s is not created for test.', container.name)
return False
job_id = container.id.job_id
pid = container.id.pid
if pid and not utils.pid_is_alive(pid):
logging.debug('Process with PID %s is not alive, container %s is '
'orphaned.', pid, container.name)
return True
try:
hqes = AFE.get_host_queue_entries(job_id=job_id)
except Exception as e:
logging.error('Failed to get hqe for job %s. Error: %s.', job_id, e)
return False
if not hqes:
# The job has not run yet.
return False
for hqe in hqes:
if hqe.active or not hqe.complete:
logging.debug('Test job %s is not completed yet, container %s is '
'not orphaned.', job_id, container.name)
return False
if (hqe.finished_on and
(time_utils.time_string_to_datetime(hqe.finished_on) >
FINISHED_JOB_CUTOFF_TIME)):
logging.debug('Test job %s was completed less than an hour ago.',
job_id)
return False
logging.debug('Test job %s was completed, container %s is orphaned.',
job_id, container.name)
return True
def cleanup(container, options):
"""Cleanup orphaned container.
@param container: A Container object to be cleaned up.
@param options: Options to do cleanup.
@return: True if cleanup is successful. False otherwise.
"""
if not options.execute:
logging.info('dryrun: Cleanup container %s', container.name)
return False
try:
# cleanup is protected by is_container_orphaned. At this point the
# container may be assumed to have a valid ID.
pid = container.id.pid
# Kill autoserv process
if pid and utils.pid_is_alive(pid):
logging.info('Stopping process %s...', pid)
utils.nuke_pid(int(pid), (signal.SIGKILL,))
# Destroy container
logging.info('Destroying container %s...', container.name)
container.destroy()
return True
except Exception as e:
logging.error('Failed to cleanup container %s. Error: %s',
container.name, e)
return False
def parse_options():
"""Parse command line inputs.
@return: Options to run the script.
"""
parser = argparse.ArgumentParser()
parser.add_argument('-v', '--verbose', action='store_true',
default=False,
help='Print out ALL entries.')
parser.add_argument('-x', '--execute', action='store_true',
default=False,
help=('Execute the actions to kill autoserv processes '
'and destroy containers. Default is False to do '
'dry run'))
# TODO(dshi): Consider to adopt the scheduler log model:
# 1. Create one log per run.
# 2. Create a symlink to the latest log.
parser.add_argument('-l', '--logfile', type=str,
default=None,
help='Path to the log file to save logs.')
return parser.parse_args()
def main(options):
"""Main script.
@param options: Options to run the script.
"""
config = logging_config.LoggingConfig()
if options.logfile:
config.add_file_handler(
file_path=os.path.abspath(options.logfile),
level=logging.DEBUG if options.verbose else logging.INFO)
bucket = lxc.ContainerBucket()
logging.info('')
logging.info('Cleaning container bucket %s', bucket.container_path)
success_count = 0
failure_count = 0
for container in bucket.get_all().values():
if is_container_orphaned(container):
if cleanup(container, options):
success_count += 1
else:
failure_count += 1
logging.info('Cleanup finished.')
if __name__ == '__main__':
options = parse_options()
main(options)