| # Please keep this code python 2.4 compatible and stand alone. |
| |
| import logging, os, shutil, sys, tempfile, time, urllib2 |
| import subprocess, re |
| from distutils.version import LooseVersion |
| |
| from autotest_lib.client.common_lib import autotemp, revision_control, utils |
| |
| _READ_SIZE = 64*1024 |
| _MAX_PACKAGE_SIZE = 100*1024*1024 |
| _CHROMEOS_MIRROR = ('http://commondatastorage.googleapis.com/' |
| 'chromeos-mirror/gentoo/distfiles/') |
| |
| |
| class Error(Exception): |
| """Local exception to be raised by code in this file.""" |
| |
| class FetchError(Error): |
| """Failed to fetch a package from any of its listed URLs.""" |
| |
| |
| def _checksum_file(full_path): |
| """@returns The hex checksum of a file given its pathname.""" |
| inputfile = open(full_path, 'rb') |
| try: |
| hex_sum = utils.hash('sha1', inputfile.read()).hexdigest() |
| finally: |
| inputfile.close() |
| return hex_sum |
| |
| |
| def system(commandline): |
| """Same as os.system(commandline) but logs the command first. |
| |
| @param commandline: commandline to be called. |
| """ |
| logging.info(commandline) |
| return os.system(commandline) |
| |
| |
| def find_top_of_autotest_tree(): |
| """@returns The full path to the top of the autotest directory tree.""" |
| dirname = os.path.dirname(__file__) |
| autotest_dir = os.path.abspath(os.path.join(dirname, '..')) |
| return autotest_dir |
| |
| |
| class ExternalPackage(object): |
| """ |
| Defines an external package with URLs to fetch its sources from and |
| a build_and_install() method to unpack it, build it and install it |
| beneath our own autotest/site-packages directory. |
| |
| Base Class. Subclass this to define packages. |
| Note: Unless your subclass has a specific reason to, it should not |
| re-install the package every time build_externals is invoked, as this |
| happens periodically through the scheduler. To avoid doing so the is_needed |
| method needs to return an appropriate value. |
| |
| Attributes: |
| @attribute urls - A tuple of URLs to try fetching the package from. |
| @attribute local_filename - A local filename to use when saving the |
| fetched package. |
| @attribute dist_name - The name of the Python distribution. For example, |
| the package MySQLdb is included in the distribution named |
| MySQL-python. This is generally the PyPI name. Defaults to the |
| name part of the local_filename. |
| @attribute hex_sum - The hex digest (currently SHA1) of this package |
| to be used to verify its contents. |
| @attribute module_name - The installed python module name to be used for |
| for a version check. Defaults to the lower case class name with |
| the word Package stripped off. |
| @attribute extracted_package_path - The path to package directory after |
| extracting. |
| @attribute version - The desired minimum package version. |
| @attribute os_requirements - A dictionary mapping pathname tuples on the |
| the OS distribution to a likely name of a package the user |
| needs to install on their system in order to get this file. |
| One of the files in the tuple must exist. |
| @attribute name - Read only, the printable name of the package. |
| @attribute subclasses - This class attribute holds a list of all defined |
| subclasses. It is constructed dynamically using the metaclass. |
| """ |
| # Modules that are meant to be installed in system directory, rather than |
| # autotest/site-packages. These modules should be skipped if the module |
| # is already installed in system directory. This prevents an older version |
| # of the module from being installed in system directory. |
| SYSTEM_MODULES = ['setuptools'] |
| |
| subclasses = [] |
| urls = () |
| local_filename = None |
| dist_name = None |
| hex_sum = None |
| module_name = None |
| version = None |
| os_requirements = None |
| |
| |
| class __metaclass__(type): |
| """Any time a subclass is defined, add it to our list.""" |
| def __init__(mcs, name, bases, dict): |
| if name != 'ExternalPackage' and not name.startswith('_'): |
| mcs.subclasses.append(mcs) |
| |
| |
| def __init__(self): |
| self.verified_package = '' |
| if not self.module_name: |
| self.module_name = self.name.lower() |
| if not self.dist_name and self.local_filename: |
| self.dist_name = self.local_filename[:self.local_filename.rindex('-')] |
| self.installed_version = '' |
| |
| |
| @property |
| def extracted_package_path(self): |
| """Return the package path after extracting. |
| |
| If the package has assigned its own extracted_package_path, use it. |
| Or use part of its local_filename as the extracting path. |
| """ |
| return self.local_filename[:-len(self._get_extension( |
| self.local_filename))] |
| |
| |
| @property |
| def name(self): |
| """Return the class name with any trailing 'Package' stripped off.""" |
| class_name = self.__class__.__name__ |
| if class_name.endswith('Package'): |
| return class_name[:-len('Package')] |
| return class_name |
| |
| |
| def is_needed(self, install_dir): |
| """ |
| Check to see if we need to reinstall a package. This is contingent on: |
| 1. Module name: If the name of the module is different from the package, |
| the class that installs it needs to specify a module_name string, |
| so we can try importing the module. |
| |
| 2. Installed version: If the module doesn't contain a __version__ the |
| class that installs it needs to override the |
| _get_installed_version_from_module method to return an appropriate |
| version string. |
| |
| 3. Version/Minimum version: The class that installs the package should |
| contain a version string, and an optional minimum version string. |
| |
| 4. install_dir: If the module exists in a different directory, e.g., |
| /usr/lib/python2.7/dist-packages/, the module will be forced to be |
| installed in install_dir. |
| |
| @param install_dir: install directory. |
| @returns True if self.module_name needs to be built and installed. |
| """ |
| if not self.module_name or not self.version: |
| logging.warning('version and module_name required for ' |
| 'is_needed() check to work.') |
| return True |
| try: |
| module = __import__(self.module_name) |
| except ImportError, e: |
| logging.info("%s isn't present. Will install.", self.module_name) |
| return True |
| # Check if we're getting a module installed somewhere else, |
| # e.g. on the system. |
| if self.module_name not in self.SYSTEM_MODULES: |
| if (hasattr(module, '__file__') |
| and not module.__file__.startswith(install_dir)): |
| path = module.__file__ |
| elif (hasattr(module, '__path__') |
| and module.__path__ |
| and not module.__path__[0].startswith(install_dir)): |
| path = module.__path__[0] |
| else: |
| logging.warning('module %s has no __file__ or __path__', |
| self.module_name) |
| return True |
| logging.info( |
| 'Found %s installed in %s, installing our version in %s', |
| self.module_name, path, install_dir) |
| return True |
| self.installed_version = self._get_installed_version_from_module(module) |
| if not self.installed_version: |
| return True |
| |
| logging.info('imported %s version %s.', self.module_name, |
| self.installed_version) |
| if hasattr(self, 'minimum_version'): |
| return LooseVersion(self.minimum_version) > LooseVersion( |
| self.installed_version) |
| else: |
| return LooseVersion(self.version) > LooseVersion( |
| self.installed_version) |
| |
| |
| def _get_installed_version_from_module(self, module): |
| """Ask our module its version string and return it or '' if unknown.""" |
| try: |
| return module.__version__ |
| except AttributeError: |
| logging.error('could not get version from %s', module) |
| return '' |
| |
| |
| def _build_and_install(self, install_dir): |
| """Subclasses MUST provide their own implementation.""" |
| raise NotImplementedError |
| |
| |
| def _build_and_install_current_dir(self, install_dir): |
| """ |
| Subclasses that use _build_and_install_from_package() MUST provide |
| their own implementation of this method. |
| """ |
| raise NotImplementedError |
| |
| |
| def build_and_install(self, install_dir): |
| """ |
| Builds and installs the package. It must have been fetched already. |
| |
| @param install_dir - The package installation directory. If it does |
| not exist it will be created. |
| """ |
| if not self.verified_package: |
| raise Error('Must call fetch() first. - %s' % self.name) |
| self._check_os_requirements() |
| return self._build_and_install(install_dir) |
| |
| |
| def _check_os_requirements(self): |
| if not self.os_requirements: |
| return |
| failed = False |
| for file_names, package_name in self.os_requirements.iteritems(): |
| if not any(os.path.exists(file_name) for file_name in file_names): |
| failed = True |
| logging.error('Can\'t find %s, %s probably needs it.', |
| ' or '.join(file_names), self.name) |
| logging.error('Perhaps you need to install something similar ' |
| 'to the %s package for OS first.', package_name) |
| if failed: |
| raise Error('Missing OS requirements for %s. (see above)' % |
| self.name) |
| |
| |
| def _build_and_install_current_dir_setup_py(self, install_dir): |
| """For use as a _build_and_install_current_dir implementation.""" |
| egg_path = self._build_egg_using_setup_py(setup_py='setup.py') |
| if not egg_path: |
| return False |
| return self._install_from_egg(install_dir, egg_path) |
| |
| |
| def _build_and_install_current_dir_setupegg_py(self, install_dir): |
| """For use as a _build_and_install_current_dir implementation.""" |
| egg_path = self._build_egg_using_setup_py(setup_py='setupegg.py') |
| if not egg_path: |
| return False |
| return self._install_from_egg(install_dir, egg_path) |
| |
| |
| def _build_and_install_current_dir_noegg(self, install_dir): |
| if not self._build_using_setup_py(): |
| return False |
| return self._install_using_setup_py_and_rsync(install_dir) |
| |
| |
| def _get_extension(self, package): |
| """Get extension of package.""" |
| valid_package_extensions = ['.tar.gz', '.tar.bz2', '.zip'] |
| extension = None |
| |
| for ext in valid_package_extensions: |
| if package.endswith(ext): |
| extension = ext |
| break |
| |
| if not extension: |
| raise Error('Unexpected package file extension on %s' % package) |
| |
| return extension |
| |
| |
| def _build_and_install_from_package(self, install_dir): |
| """ |
| This method may be used as a _build_and_install() implementation |
| for subclasses if they implement _build_and_install_current_dir(). |
| |
| Extracts the .tar.gz file, chdirs into the extracted directory |
| (which is assumed to match the tar filename) and calls |
| _build_and_isntall_current_dir from there. |
| |
| Afterwards the build (regardless of failure) extracted .tar.gz |
| directory is cleaned up. |
| |
| @returns True on success, False otherwise. |
| |
| @raises OSError If the expected extraction directory does not exist. |
| """ |
| self._extract_compressed_package() |
| extension = self._get_extension(self.verified_package) |
| os.chdir(os.path.dirname(self.verified_package)) |
| os.chdir(self.extracted_package_path) |
| extracted_dir = os.getcwd() |
| try: |
| return self._build_and_install_current_dir(install_dir) |
| finally: |
| os.chdir(os.path.join(extracted_dir, '..')) |
| shutil.rmtree(extracted_dir) |
| |
| |
| def _extract_compressed_package(self): |
| """Extract the fetched compressed .tar or .zip within its directory.""" |
| if not self.verified_package: |
| raise Error('Package must have been fetched first.') |
| os.chdir(os.path.dirname(self.verified_package)) |
| if self.verified_package.endswith('gz'): |
| status = system("tar -xzf '%s'" % self.verified_package) |
| elif self.verified_package.endswith('bz2'): |
| status = system("tar -xjf '%s'" % self.verified_package) |
| elif self.verified_package.endswith('zip'): |
| status = system("unzip '%s'" % self.verified_package) |
| else: |
| raise Error('Unknown compression suffix on %s.' % |
| self.verified_package) |
| if status: |
| raise Error('tar failed with %s' % (status,)) |
| |
| |
| def _build_using_setup_py(self, setup_py='setup.py'): |
| """ |
| Assuming the cwd is the extracted python package, execute a simple |
| python setup.py build. |
| |
| @param setup_py - The name of the setup.py file to execute. |
| |
| @returns True on success, False otherwise. |
| """ |
| if not os.path.exists(setup_py): |
| raise Error('%s does not exist in %s' % (setup_py, os.getcwd())) |
| status = system("'%s' %s build" % (sys.executable, setup_py)) |
| if status: |
| logging.error('%s build failed.', self.name) |
| return False |
| return True |
| |
| |
| def _build_egg_using_setup_py(self, setup_py='setup.py'): |
| """ |
| Assuming the cwd is the extracted python package, execute a simple |
| python setup.py bdist_egg. |
| |
| @param setup_py - The name of the setup.py file to execute. |
| |
| @returns The relative path to the resulting egg file or '' on failure. |
| """ |
| if not os.path.exists(setup_py): |
| raise Error('%s does not exist in %s' % (setup_py, os.getcwd())) |
| egg_subdir = 'dist' |
| if os.path.isdir(egg_subdir): |
| shutil.rmtree(egg_subdir) |
| status = system("'%s' %s bdist_egg" % (sys.executable, setup_py)) |
| if status: |
| logging.error('bdist_egg of setuptools failed.') |
| return '' |
| # I've never seen a bdist_egg lay multiple .egg files. |
| for filename in os.listdir(egg_subdir): |
| if filename.endswith('.egg'): |
| return os.path.join(egg_subdir, filename) |
| |
| |
| def _install_from_egg(self, install_dir, egg_path): |
| """ |
| Install a module from an egg file by unzipping the necessary parts |
| into install_dir. |
| |
| @param install_dir - The installation directory. |
| @param egg_path - The pathname of the egg file. |
| """ |
| status = system("unzip -q -o -d '%s' '%s'" % (install_dir, egg_path)) |
| if status: |
| logging.error('unzip of %s failed', egg_path) |
| return False |
| egg_info_dir = os.path.join(install_dir, 'EGG-INFO') |
| if os.path.isdir(egg_info_dir): |
| egg_info_new_path = self._get_egg_info_path(install_dir) |
| if egg_info_new_path: |
| if os.path.exists(egg_info_new_path): |
| shutil.rmtree(egg_info_new_path) |
| os.rename(egg_info_dir, egg_info_new_path) |
| else: |
| shutil.rmtree(egg_info_dir) |
| return True |
| |
| |
| def _get_egg_info_path(self, install_dir): |
| """Get egg-info path for this package. |
| |
| Example path: install_dir/MySQL_python-1.2.3.egg-info |
| |
| """ |
| if self.dist_name: |
| egg_info_name_part = self.dist_name.replace('-', '_') |
| if self.version: |
| egg_info_filename = '%s-%s.egg-info' % (egg_info_name_part, |
| self.version) |
| else: |
| egg_info_filename = '%s.egg-info' % (egg_info_name_part,) |
| return os.path.join(install_dir, egg_info_filename) |
| else: |
| return None |
| |
| |
| def _get_temp_dir(self): |
| return tempfile.mkdtemp(dir='/var/tmp') |
| |
| |
| def _site_packages_path(self, temp_dir): |
| # This makes assumptions about what python setup.py install |
| # does when given a prefix. Is this always correct? |
| python_xy = 'python%s' % sys.version[:3] |
| return os.path.join(temp_dir, 'lib', python_xy, 'site-packages') |
| |
| |
| def _rsync (self, temp_site_dir, install_dir): |
| """Rsync contents. """ |
| status = system("rsync -r '%s/' '%s/'" % |
| (os.path.normpath(temp_site_dir), |
| os.path.normpath(install_dir))) |
| if status: |
| logging.error('%s rsync to install_dir failed.', self.name) |
| return False |
| return True |
| |
| |
| def _install_using_setup_py_and_rsync(self, install_dir, |
| setup_py='setup.py', |
| temp_dir=None): |
| """ |
| Assuming the cwd is the extracted python package, execute a simple: |
| |
| python setup.py install --prefix=BLA |
| |
| BLA will be a temporary directory that everything installed will |
| be picked out of and rsynced to the appropriate place under |
| install_dir afterwards. |
| |
| Afterwards, it deconstructs the extra lib/pythonX.Y/site-packages/ |
| directory tree that setuptools created and moves all installed |
| site-packages directly up into install_dir itself. |
| |
| @param install_dir the directory for the install to happen under. |
| @param setup_py - The name of the setup.py file to execute. |
| |
| @returns True on success, False otherwise. |
| """ |
| if not os.path.exists(setup_py): |
| raise Error('%s does not exist in %s' % (setup_py, os.getcwd())) |
| |
| if temp_dir is None: |
| temp_dir = self._get_temp_dir() |
| |
| try: |
| status = system("'%s' %s install --no-compile --prefix='%s'" |
| % (sys.executable, setup_py, temp_dir)) |
| if status: |
| logging.error('%s install failed.', self.name) |
| return False |
| |
| if os.path.isdir(os.path.join(temp_dir, 'lib')): |
| # NOTE: This ignores anything outside of the lib/ dir that |
| # was installed. |
| temp_site_dir = self._site_packages_path(temp_dir) |
| else: |
| temp_site_dir = temp_dir |
| |
| return self._rsync(temp_site_dir, install_dir) |
| finally: |
| shutil.rmtree(temp_dir) |
| |
| |
| |
| def _build_using_make(self, install_dir): |
| """Build the current package using configure/make. |
| |
| @returns True on success, False otherwise. |
| """ |
| install_prefix = os.path.join(install_dir, 'usr', 'local') |
| status = system('./configure --prefix=%s' % install_prefix) |
| if status: |
| logging.error('./configure failed for %s', self.name) |
| return False |
| status = system('make') |
| if status: |
| logging.error('make failed for %s', self.name) |
| return False |
| status = system('make check') |
| if status: |
| logging.error('make check failed for %s', self.name) |
| return False |
| return True |
| |
| |
| def _install_using_make(self): |
| """Install the current package using make install. |
| |
| Assumes the install path was set up while running ./configure (in |
| _build_using_make()). |
| |
| @returns True on success, False otherwise. |
| """ |
| status = system('make install') |
| return status == 0 |
| |
| |
| def fetch(self, dest_dir): |
| """ |
| Fetch the package from one its URLs and save it in dest_dir. |
| |
| If the the package already exists in dest_dir and the checksum |
| matches this code will not fetch it again. |
| |
| Sets the 'verified_package' attribute with the destination pathname. |
| |
| @param dest_dir - The destination directory to save the local file. |
| If it does not exist it will be created. |
| |
| @returns A boolean indicating if we the package is now in dest_dir. |
| @raises FetchError - When something unexpected happens. |
| """ |
| if not os.path.exists(dest_dir): |
| os.makedirs(dest_dir) |
| local_path = os.path.join(dest_dir, self.local_filename) |
| |
| # If the package exists, verify its checksum and be happy if it is good. |
| if os.path.exists(local_path): |
| actual_hex_sum = _checksum_file(local_path) |
| if self.hex_sum == actual_hex_sum: |
| logging.info('Good checksum for existing %s package.', |
| self.name) |
| self.verified_package = local_path |
| return True |
| logging.warning('Bad checksum for existing %s package. ' |
| 'Re-downloading', self.name) |
| os.rename(local_path, local_path + '.wrong-checksum') |
| |
| # Download the package from one of its urls, rejecting any if the |
| # checksum does not match. |
| for url in self.urls: |
| logging.info('Fetching %s', url) |
| try: |
| url_file = urllib2.urlopen(url) |
| except (urllib2.URLError, EnvironmentError): |
| logging.warning('Could not fetch %s package from %s.', |
| self.name, url) |
| continue |
| |
| data_length = int(url_file.info().get('Content-Length', |
| _MAX_PACKAGE_SIZE)) |
| if data_length <= 0 or data_length > _MAX_PACKAGE_SIZE: |
| raise FetchError('%s from %s fails Content-Length %d ' |
| 'sanity check.' % (self.name, url, |
| data_length)) |
| checksum = utils.hash('sha1') |
| total_read = 0 |
| output = open(local_path, 'wb') |
| try: |
| while total_read < data_length: |
| data = url_file.read(_READ_SIZE) |
| if not data: |
| break |
| output.write(data) |
| checksum.update(data) |
| total_read += len(data) |
| finally: |
| output.close() |
| if self.hex_sum != checksum.hexdigest(): |
| logging.warning('Bad checksum for %s fetched from %s.', |
| self.name, url) |
| logging.warning('Got %s', checksum.hexdigest()) |
| logging.warning('Expected %s', self.hex_sum) |
| os.unlink(local_path) |
| continue |
| logging.info('Good checksum.') |
| self.verified_package = local_path |
| return True |
| else: |
| return False |
| |
| |
| # NOTE: This class definition must come -before- all other ExternalPackage |
| # classes that need to use this version of setuptools so that is is inserted |
| # into the ExternalPackage.subclasses list before them. |
| class SetuptoolsPackage(ExternalPackage): |
| """setuptools package""" |
| # For all known setuptools releases a string compare works for the |
| # version string. Hopefully they never release a 0.10. (Their own |
| # version comparison code would break if they did.) |
| # Any system with setuptools > 18.0.1 is fine. If none installed, then |
| # try to install the latest found on the upstream. |
| minimum_version = '18.0.1' |
| version = '18.0.1' |
| urls = (_CHROMEOS_MIRROR + 'setuptools-%s.tar.gz' % (version,),) |
| local_filename = 'setuptools-%s.tar.gz' % version |
| hex_sum = 'ebc4fe81b7f6d61d923d9519f589903824044f52' |
| |
| SUDO_SLEEP_DELAY = 15 |
| |
| |
| def _build_and_install(self, install_dir): |
| """Install setuptools on the system.""" |
| logging.info('NOTE: setuptools install does not use install_dir.') |
| return self._build_and_install_from_package(install_dir) |
| |
| |
| def _build_and_install_current_dir(self, install_dir): |
| egg_path = self._build_egg_using_setup_py() |
| if not egg_path: |
| return False |
| |
| print '!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!\n' |
| print 'About to run sudo to install setuptools', self.version |
| print 'on your system for use by', sys.executable, '\n' |
| print '!! ^C within', self.SUDO_SLEEP_DELAY, 'seconds to abort.\n' |
| time.sleep(self.SUDO_SLEEP_DELAY) |
| |
| # Copy the egg to the local filesystem /var/tmp so that root can |
| # access it properly (avoid NFS squashroot issues). |
| temp_dir = self._get_temp_dir() |
| try: |
| shutil.copy(egg_path, temp_dir) |
| egg_name = os.path.split(egg_path)[1] |
| temp_egg = os.path.join(temp_dir, egg_name) |
| p = subprocess.Popen(['sudo', '/bin/sh', temp_egg], |
| stdout=subprocess.PIPE) |
| regex = re.compile('Copying (.*?) to (.*?)\n') |
| match = regex.search(p.communicate()[0]) |
| status = p.wait() |
| |
| if match: |
| compiled = os.path.join(match.group(2), match.group(1)) |
| os.system("sudo chmod a+r '%s'" % compiled) |
| finally: |
| shutil.rmtree(temp_dir) |
| |
| if status: |
| logging.error('install of setuptools from egg failed.') |
| return False |
| return True |
| |
| |
| class MySQLdbPackage(ExternalPackage): |
| """mysql package, used in scheduler.""" |
| module_name = 'MySQLdb' |
| version = '1.2.3' |
| local_filename = 'MySQL-python-%s.tar.gz' % version |
| urls = ('http://commondatastorage.googleapis.com/chromeos-mirror/gentoo/' |
| 'distfiles/%s' % local_filename,) |
| hex_sum = '3511bb8c57c6016eeafa531d5c3ea4b548915e3c' |
| |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| |
| def _build_and_install(self, install_dir): |
| if not os.path.exists('/usr/bin/mysql_config'): |
| error_msg = '''\ |
| You need to install /usr/bin/mysql_config. |
| On recent Debian based distros, run: \ |
| sudo apt-get install libmariadbclient-dev-compat |
| On older Debian based distros, run: sudo apt-get install libmysqlclient15-dev |
| ''' |
| logging.error(error_msg) |
| return False, error_msg |
| return self._build_and_install_from_package(install_dir) |
| |
| |
| class DjangoPackage(ExternalPackage): |
| """django package.""" |
| version = '1.5.1' |
| local_filename = 'Django-%s.tar.gz' % version |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = '0ab97b90c4c79636e56337f426f1e875faccbba1' |
| |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_noegg) |
| |
| |
| def _get_installed_version_from_module(self, module): |
| try: |
| return module.get_version().split()[0] |
| except AttributeError: |
| return '0.9.6' |
| |
| |
| |
| class NumpyPackage(ExternalPackage): |
| """numpy package, required by matploglib.""" |
| version = '1.7.0' |
| local_filename = 'numpy-%s.tar.gz' % version |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = 'ba328985f20390b0f969a5be2a6e1141d5752cf9' |
| |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setupegg_py) |
| |
| |
| |
| class JsonRPCLib(ExternalPackage): |
| """jsonrpclib package""" |
| version = '0.1.3' |
| module_name = 'jsonrpclib' |
| local_filename = '%s-%s.tar.gz' % (module_name, version) |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = '431714ed19ab677f641ce5d678a6a95016f5c452' |
| |
| def _get_installed_version_from_module(self, module): |
| # jsonrpclib doesn't contain a proper version |
| return self.version |
| |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_noegg) |
| |
| |
| class GwtPackage(ExternalPackage): |
| """Fetch and extract a local copy of GWT used to build the frontend.""" |
| |
| version = '2.3.0' |
| local_filename = 'gwt-%s.zip' % version |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = 'd51fce9166e6b31349659ffca89baf93e39bc84b' |
| name = 'gwt' |
| about_filename = 'about.txt' |
| module_name = None # Not a Python module. |
| |
| |
| def is_needed(self, install_dir): |
| gwt_dir = os.path.join(install_dir, self.name) |
| about_file = os.path.join(install_dir, self.name, self.about_filename) |
| |
| if not os.path.exists(gwt_dir) or not os.path.exists(about_file): |
| logging.info('gwt not installed for autotest') |
| return True |
| |
| f = open(about_file, 'r') |
| version_line = f.readline() |
| f.close() |
| |
| match = re.match(r'Google Web Toolkit (.*)', version_line) |
| if not match: |
| logging.info('did not find gwt version') |
| return True |
| |
| logging.info('found gwt version %s', match.group(1)) |
| return match.group(1) != self.version |
| |
| |
| def _build_and_install(self, install_dir): |
| os.chdir(install_dir) |
| self._extract_compressed_package() |
| extracted_dir = self.local_filename[:-len('.zip')] |
| target_dir = os.path.join(install_dir, self.name) |
| if os.path.exists(target_dir): |
| shutil.rmtree(target_dir) |
| os.rename(extracted_dir, target_dir) |
| return True |
| |
| |
| class PyudevPackage(ExternalPackage): |
| """ |
| pyudev module |
| |
| Used in unittests. |
| """ |
| version = '0.16.1' |
| url_filename = 'pyudev-%s.tar.gz' % version |
| local_filename = url_filename |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = 'b36bc5c553ce9b56d32a5e45063a2c88156771c0' |
| |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| |
| class PyMoxPackage(ExternalPackage): |
| """ |
| mox module |
| |
| Used in unittests. |
| """ |
| module_name = 'mox' |
| version = '0.5.3' |
| # Note: url_filename does not match local_filename, because of |
| # an uncontrolled fork at some point in time of mox versions. |
| url_filename = 'mox-%s-autotest.tar.gz' % version |
| local_filename = 'mox-%s.tar.gz' % version |
| urls = (_CHROMEOS_MIRROR + url_filename,) |
| hex_sum = '1c502d2c0a8aefbba2c7f385a83d33e7d822452a' |
| |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_noegg) |
| |
| def _get_installed_version_from_module(self, module): |
| # mox doesn't contain a proper version |
| return self.version |
| |
| |
| class PySeleniumPackage(ExternalPackage): |
| """ |
| selenium module |
| |
| Used in wifi_interop suite. |
| """ |
| module_name = 'selenium' |
| version = '2.37.2' |
| url_filename = 'selenium-%s.tar.gz' % version |
| local_filename = url_filename |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = '66946d5349e36d946daaad625c83c30c11609e36' |
| |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| |
| class FaultHandlerPackage(ExternalPackage): |
| """ |
| faulthandler module |
| """ |
| module_name = 'faulthandler' |
| version = '2.3' |
| url_filename = '%s-%s.tar.gz' % (module_name, version) |
| local_filename = url_filename |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = 'efb30c068414fba9df892e48fcf86170cbf53589' |
| |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_noegg) |
| |
| |
| class PsutilPackage(ExternalPackage): |
| """ |
| psutil module |
| """ |
| module_name = 'psutil' |
| version = '2.1.1' |
| url_filename = '%s-%s.tar.gz' % (module_name, version) |
| local_filename = url_filename |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = '0c20a20ed316e69f2b0881530439213988229916' |
| |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| |
| class ElasticSearchPackage(ExternalPackage): |
| """elasticsearch-py package.""" |
| version = '1.6.0' |
| url_filename = 'elasticsearch-%s.tar.gz' % version |
| local_filename = url_filename |
| urls = ('https://pypi.python.org/packages/source/e/elasticsearch/%s' % |
| (url_filename),) |
| hex_sum = '3e676c96f47935b1f52df82df3969564bd356b1c' |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| def _get_installed_version_from_module(self, module): |
| # Elastic's version format is like tuple (1, 6, 0), which needs to be |
| # transferred to 1.6.0. |
| try: |
| return '.'.join(str(i) for i in module.__version__) |
| except: |
| return self.version |
| |
| |
| class Urllib3Package(ExternalPackage): |
| """elasticsearch-py package.""" |
| version = '1.9' |
| url_filename = 'urllib3-%s.tar.gz' % version |
| local_filename = url_filename |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = '9522197efb2a2b49ce804de3a515f06d97b6602f' |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| class ImagingLibraryPackage(ExternalPackage): |
| """Python Imaging Library (PIL).""" |
| version = '1.1.7' |
| url_filename = 'Imaging-%s.tar.gz' % version |
| local_filename = url_filename |
| urls = ('http://commondatastorage.googleapis.com/chromeos-mirror/gentoo/' |
| 'distfiles/%s' % url_filename,) |
| hex_sum = '76c37504251171fda8da8e63ecb8bc42a69a5c81' |
| |
| def _build_and_install(self, install_dir): |
| #The path of zlib library might be different from what PIL setup.py is |
| #expected. Following change does the best attempt to link the library |
| #to a path PIL setup.py will try. |
| libz_possible_path = '/usr/lib/x86_64-linux-gnu/libz.so' |
| libz_expected_path = '/usr/lib/libz.so' |
| # TODO(crbug.com/957186): this sudo command fails if build_externals |
| # is running in non-interactive mode, and requires a workaround when |
| # running within a docker build process. Remove this operation, or |
| # remove this entire package. |
| if (os.path.exists(libz_possible_path) and |
| not os.path.exists(libz_expected_path)): |
| utils.run('sudo ln -s %s %s' % |
| (libz_possible_path, libz_expected_path)) |
| return self._build_and_install_from_package(install_dir) |
| |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_noegg) |
| |
| |
| class AstroidPackage(ExternalPackage): |
| """astroid package.""" |
| version = '1.5.3' |
| url_filename = 'astroid-%s.tar.gz' % version |
| local_filename = url_filename |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = 'e654225ab5bd2788e5e246b156910990bf33cde6' |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| |
| class LazyObjectProxyPackage(ExternalPackage): |
| """lazy-object-proxy package (dependency for astroid).""" |
| version = '1.3.1' |
| url_filename = 'lazy-object-proxy-%s.tar.gz' % version |
| local_filename = url_filename |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = '984828d8f672986ca926373986214d7057b772fb' |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| |
| class SingleDispatchPackage(ExternalPackage): |
| """singledispatch package (dependency for astroid).""" |
| version = '3.4.0.3' |
| url_filename = 'singledispatch-%s.tar.gz' % version |
| local_filename = url_filename |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = 'f93241b06754a612af8bb7aa208c4d1805637022' |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| |
| class Enum34Package(ExternalPackage): |
| """enum34 package (dependency for astroid).""" |
| version = '1.1.6' |
| url_filename = 'enum34-%s.tar.gz' % version |
| local_filename = url_filename |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = '014ef5878333ff91099893d615192c8cd0b1525a' |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| |
| class WraptPackage(ExternalPackage): |
| """wrapt package (dependency for astroid).""" |
| version = '1.10.10' |
| url_filename = 'wrapt-%s.tar.gz' % version |
| local_filename = url_filename |
| #md5=97365e906afa8b431f266866ec4e2e18 |
| urls = ('https://pypi.python.org/packages/a3/bb/' |
| '525e9de0a220060394f4aa34fdf6200853581803d92714ae41fc3556e7d7/%s' % |
| (url_filename),) |
| hex_sum = '6be4f1bb50db879863f4247692360eb830a3eb33' |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_noegg) |
| |
| |
| class SixPackage(ExternalPackage): |
| """six package (dependency for astroid).""" |
| version = '1.10.0' |
| url_filename = 'six-%s.tar.gz' % version |
| local_filename = url_filename |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = '30d480d2e352e8e4c2aae042cf1bf33368ff0920' |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| |
| class LruCachePackage(ExternalPackage): |
| """backports.functools_lru_cache package (dependency for astroid).""" |
| version = '1.4' |
| url_filename = 'backports.functools_lru_cache-%s.tar.gz' % version |
| local_filename = url_filename |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = '8a546e7887e961c2873c9b053f4e2cd2a96bd71d' |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| |
| class LogilabCommonPackage(ExternalPackage): |
| """logilab-common package.""" |
| version = '1.2.2' |
| module_name = 'logilab' |
| url_filename = 'logilab-common-%s.tar.gz' % version |
| local_filename = url_filename |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = 'ecad2d10c31dcf183c8bed87b6ec35e7ed397d27' |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| |
| class PyLintPackage(ExternalPackage): |
| """pylint package.""" |
| version = '1.7.2' |
| url_filename = 'pylint-%s.tar.gz' % version |
| local_filename = url_filename |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = '42d8b9394e5a485377ae128b01350f25d8b131e0' |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| |
| class ConfigParserPackage(ExternalPackage): |
| """configparser package (dependency for pylint).""" |
| version = '3.5.0' |
| url_filename = 'configparser-%s.tar.gz' % version |
| local_filename = url_filename |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = '8ee6b29c6a11977c0e094da1d4f5f71e7e7ac78b' |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| |
| class IsortPackage(ExternalPackage): |
| """isort package (dependency for pylint).""" |
| version = '4.2.15' |
| url_filename = 'isort-%s.tar.gz' % version |
| local_filename = url_filename |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = 'acacc36e476b70e13e6fda812c193f4c3c187781' |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| |
| class DateutilPackage(ExternalPackage): |
| """python-dateutil package.""" |
| version = '2.6.1' |
| local_filename = 'python-dateutil-%s.tar.gz' % version |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = 'db2ace298dee7e47fd720ed03eb790885347bf4e' |
| |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| |
| class Pytz(ExternalPackage): |
| """Pytz package.""" |
| version = '2016.10' |
| url_filename = 'pytz-%s.tar.gz' % version |
| local_filename = url_filename |
| #md5=cc9f16ba436efabdcef3c4d32ae4919c |
| urls = ('https://pypi.python.org/packages/42/00/' |
| '5c89fc6c9b305df84def61863528e899e9dccb196f8438f6cbe960758fc5/%s' % |
| (url_filename),) |
| hex_sum = '8d63f1e9b1ee862841b990a7d8ad1d4508d9f0be' |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| |
| class Tzlocal(ExternalPackage): |
| """Tzlocal package.""" |
| version = '1.3' |
| url_filename = 'tzlocal-%s.tar.gz' % version |
| local_filename = url_filename |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = '730e9d7112335865a1dcfabec69c8c3086be424f' |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| |
| class PyYAMLPackage(ExternalPackage): |
| """pyyaml package.""" |
| version = '3.12' |
| local_filename = 'PyYAML-%s.tar.gz' % version |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = 'cb7fd3e58c129494ee86e41baedfec69eb7dafbe' |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_noegg) |
| |
| |
| class GoogleAuthPackage(ExternalPackage): |
| """Google Auth Client.""" |
| version = '1.6.3' |
| local_filename = 'google-auth-%s.tar.gz' % version |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = 'a76f97686ebe42097d91e0996a72b26b54118f3b' |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| |
| class CachetoolsPackage(ExternalPackage): |
| """Cachetools package.""" |
| version = '3.1.1' |
| local_filename = 'cachetools-%s.tar.gz' % version |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| hex_sum = 'd030bfdfa91b0b1188993f5e8d7da077308c1eaf' |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| |
| class GrpcioPackage(ExternalPackage): |
| """GrpcioPackage package.""" |
| version = '1.26.0' |
| hex_sum = "b9a61f855bf3656d9b8ac305bd1e52442e120c48" |
| local_filename = 'grpcio-%s.tar.gz' % version |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| |
| class GrpcioToolsPackage(ExternalPackage): |
| """GrpcioPackage package.""" |
| version = '1.26.0' |
| hex_sum = "298724d8704523c6ff443303e0c26fc1d54f9acb" |
| local_filename = 'grpcio-tools-%s.tar.gz' % version |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| _build_and_install = ExternalPackage._build_and_install_from_package |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| |
| class Protobuf(ExternalPackage): |
| """GrpcioPackage package.""" |
| version = '3.11.2' |
| hex_sum = "e1f3ffa028ece5a529149dd56a3d64aea4ae1b1a" |
| local_filename = 'protobuf-%s.tar.gz' % version |
| urls = (_CHROMEOS_MIRROR + local_filename,) |
| _build_and_install_current_dir = ( |
| ExternalPackage._build_and_install_current_dir_setup_py) |
| |
| def _build_and_install(self, install_dir): |
| """ |
| This method may be used as a _build_and_install() implementation |
| for subclasses if they implement _build_and_install_current_dir(). |
| |
| Extracts the .tar.gz file, chdirs into the extracted directory |
| (which is assumed to match the tar filename) and calls |
| _build_and_isntall_current_dir from there. |
| |
| Afterwards the build (regardless of failure) extracted .tar.gz |
| directory is cleaned up. |
| |
| @returns True on success, False otherwise. |
| |
| @raises OSError If the expected extraction directory does not exist. |
| """ |
| self._extract_compressed_package() |
| extension = self._get_extension(self.verified_package) |
| os.chdir(os.path.dirname(self.verified_package)) |
| os.chdir(os.path.join(self.extracted_package_path, "python")) |
| extracted_dir = os.getcwd() |
| try: |
| return self._build_and_install_current_dir(install_dir) |
| finally: |
| os.chdir(os.path.join(extracted_dir, '..')) |
| shutil.rmtree(extracted_dir) |
| |
| |
| class _ExternalGitRepo(ExternalPackage): |
| """ |
| Parent class for any package which needs to pull a git repo. |
| |
| This class inherits from ExternalPackage only so we can sync git |
| repos through the build_externals script. We do not reuse any of |
| ExternalPackage's other methods. Any package that needs a git repo |
| should subclass this and override build_and_install or fetch as |
| they see appropriate. |
| """ |
| |
| os_requirements = {('/usr/bin/git') : 'git-core'} |
| |
| # All the chromiumos projects used on the lab servers should have a 'prod' |
| # branch used to track the software version deployed in prod. |
| PROD_BRANCH = 'prod' |
| MASTER_BRANCH = 'master' |
| |
| def is_needed(self, unused_install_dir): |
| """Tell build_externals that we need to fetch.""" |
| # TODO(beeps): check if we're already upto date. |
| return True |
| |
| |
| def build_and_install(self, unused_install_dir): |
| """ |
| Fall through method to install a package. |
| |
| Overwritten in base classes to pull a git repo. |
| """ |
| raise NotImplementedError |
| |
| |
| def fetch(self, unused_dest_dir): |
| """Fallthrough method to fetch a package.""" |
| return True |
| |
| |
| class HdctoolsRepo(_ExternalGitRepo): |
| """Clones or updates the hdctools repo.""" |
| |
| module_name = 'servo' |
| temp_hdctools_dir = tempfile.mktemp(suffix='hdctools') |
| _GIT_URL = ('https://chromium.googlesource.com/' |
| 'chromiumos/third_party/hdctools') |
| |
| def fetch(self, unused_dest_dir): |
| """ |
| Fetch repo to a temporary location. |
| |
| We use an intermediate temp directory to stage our |
| installation because we only care about the servo package. |
| If we can't get at the top commit hash after fetching |
| something is wrong. This can happen when we've cloned/pulled |
| an empty repo. Not something we expect to do. |
| |
| @parma unused_dest_dir: passed in because we inherit from |
| ExternalPackage. |
| |
| @return: True if repo sync was successful. |
| """ |
| git_repo = revision_control.GitRepo( |
| self.temp_hdctools_dir, |
| self._GIT_URL, |
| None, |
| abs_work_tree=self.temp_hdctools_dir) |
| git_repo.reinit_repo_at(self.PROD_BRANCH) |
| |
| if git_repo.get_latest_commit_hash(): |
| return True |
| return False |
| |
| |
| def build_and_install(self, install_dir): |
| """Reach into the hdctools repo and rsync only the servo directory.""" |
| |
| servo_dir = os.path.join(self.temp_hdctools_dir, 'servo') |
| if not os.path.exists(servo_dir): |
| return False |
| |
| rv = self._rsync(servo_dir, os.path.join(install_dir, 'servo')) |
| shutil.rmtree(self.temp_hdctools_dir) |
| return rv |
| |
| |
| class ChromiteRepo(_ExternalGitRepo): |
| """Clones or updates the chromite repo.""" |
| |
| _GIT_URL = ('https://chromium.googlesource.com/chromiumos/chromite') |
| |
| def build_and_install(self, install_dir, master_branch=False): |
| """ |
| Clone if the repo isn't initialized, pull clean bits if it is. |
| |
| Unlike it's hdctools counterpart the chromite repo clones master |
| directly into site-packages. It doesn't use an intermediate temp |
| directory because it doesn't need installation. |
| |
| @param install_dir: destination directory for chromite installation. |
| @param master_branch: if True, install master branch. Otherwise, |
| install prod branch. |
| """ |
| init_branch = (self.MASTER_BRANCH if master_branch |
| else self.PROD_BRANCH) |
| local_chromite_dir = os.path.join(install_dir, 'chromite') |
| git_repo = revision_control.GitRepo( |
| local_chromite_dir, |
| self._GIT_URL, |
| abs_work_tree=local_chromite_dir) |
| git_repo.reinit_repo_at(init_branch) |
| |
| |
| if git_repo.get_latest_commit_hash(): |
| return True |
| return False |
| |
| |
| class SuiteSchedulerRepo(_ExternalGitRepo): |
| """Clones or updates the suite_scheduler repo.""" |
| |
| _GIT_URL = ('https://chromium.googlesource.com/chromiumos/' |
| 'infra/suite_scheduler') |
| |
| def build_and_install(self, install_dir): |
| """ |
| Clone if the repo isn't initialized, pull clean bits if it is. |
| |
| @param install_dir: destination directory for suite_scheduler |
| installation. |
| @param master_branch: if True, install master branch. Otherwise, |
| install prod branch. |
| """ |
| local_dir = os.path.join(install_dir, 'suite_scheduler') |
| git_repo = revision_control.GitRepo( |
| local_dir, |
| self._GIT_URL, |
| abs_work_tree=local_dir) |
| git_repo.reinit_repo_at(self.MASTER_BRANCH) |
| |
| if git_repo.get_latest_commit_hash(): |
| return True |
| return False |
| |
| |
| class BtsocketRepo(_ExternalGitRepo): |
| """Clones or updates the btsocket repo.""" |
| |
| _GIT_URL = ('https://chromium.googlesource.com/' |
| 'chromiumos/platform/btsocket') |
| |
| def fetch(self, unused_dest_dir): |
| """ |
| Fetch repo to a temporary location. |
| |
| We use an intermediate temp directory because we have to build an |
| egg for installation. If we can't get at the top commit hash after |
| fetching something is wrong. This can happen when we've cloned/pulled |
| an empty repo. Not something we expect to do. |
| |
| @parma unused_dest_dir: passed in because we inherit from |
| ExternalPackage. |
| |
| @return: True if repo sync was successful. |
| """ |
| self.temp_btsocket_dir = autotemp.tempdir(unique_id='btsocket') |
| try: |
| git_repo = revision_control.GitRepo( |
| self.temp_btsocket_dir.name, |
| self._GIT_URL, |
| None, |
| abs_work_tree=self.temp_btsocket_dir.name) |
| git_repo.reinit_repo_at(self.PROD_BRANCH) |
| |
| if git_repo.get_latest_commit_hash(): |
| return True |
| except: |
| self.temp_btsocket_dir.clean() |
| raise |
| |
| self.temp_btsocket_dir.clean() |
| return False |
| |
| |
| def build_and_install(self, install_dir): |
| """ |
| Install the btsocket module using setup.py |
| |
| @param install_dir: Target installation directory. |
| |
| @return: A boolean indicating success of failure. |
| """ |
| work_dir = os.getcwd() |
| try: |
| os.chdir(self.temp_btsocket_dir.name) |
| rv = self._build_and_install_current_dir_setup_py(install_dir) |
| finally: |
| os.chdir(work_dir) |
| self.temp_btsocket_dir.clean() |
| return rv |
| |
| |
| class SkylabInventoryRepo(_ExternalGitRepo): |
| """Clones or updates the skylab_inventory repo.""" |
| |
| _GIT_URL = ('https://chromium.googlesource.com/chromiumos/infra/' |
| 'skylab_inventory') |
| |
| # TODO(nxia): create a prod branch for skylab_inventory. |
| def build_and_install(self, install_dir): |
| """ |
| @param install_dir: destination directory for skylab_inventory |
| installation. |
| """ |
| local_skylab_dir = os.path.join(install_dir, 'infra_skylab_inventory') |
| git_repo = revision_control.GitRepo( |
| local_skylab_dir, |
| self._GIT_URL, |
| abs_work_tree=local_skylab_dir) |
| git_repo.reinit_repo_at(self.MASTER_BRANCH) |
| |
| # The top-level __init__.py for skylab is at venv/skylab_inventory. |
| source = os.path.join(local_skylab_dir, 'venv', 'skylab_inventory') |
| link_name = os.path.join(install_dir, 'skylab_inventory') |
| |
| if (os.path.exists(link_name) and |
| os.path.realpath(link_name) != os.path.realpath(source)): |
| os.remove(link_name) |
| |
| if not os.path.exists(link_name): |
| os.symlink(source, link_name) |
| |
| if git_repo.get_latest_commit_hash(): |
| return True |
| return False |