# This file is merged from Debian's tests and Ubuntu's autopktest implementation # For Ubuntu's part Canonical is the original copyright holder. # # (C) 2015 Canonical Ltd. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 2 of the License, or # (at your option) any later version. ## Debian's part from britney2 import BinaryPackageId from britney2.installability.builder import InstallabilityTesterBuilder TEST_HINTER = 'test-hinter' HINTS_ALL = ('ALL') DEFAULT_URGENCY = 'medium' ## autopkgtest part import os import shutil import subprocess import tempfile import unittest PROJECT_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) architectures = ['amd64', 'arm64', 'armhf', 'i386', 'powerpc', 'ppc64el'] ## def new_pkg_universe_builder(): return UniverseBuilder() class MockObject(object): def __init__(self, **kwargs): for key, value in kwargs.items(): setattr(self, key, value) class PkgUniversePackageBuilder(object): def __init__(self, uni_builder, pkg_id): self._uni_builder = uni_builder self._pkg_id = pkg_id self._dependencies = set() self._conflicts = set() self._in_testing = True self._is_essential = False def is_essential(self): self._is_essential = True return self def in_testing(self): self._in_testing = True return self def not_in_testing(self): self._in_testing = False return self def depends_on(self, pkg): return self.depends_on_any_of(pkg) def depends_on_any_of(self, *pkgs): self._dependencies.add(frozenset(self._uni_builder._fetch_pkg_id(x) for x in pkgs)) return self def conflicts_with(self, *pkgs): self._conflicts.update(self._uni_builder._fetch_pkg_id(x) for x in pkgs) return self def new_package(self, *args, **kwargs): return self._uni_builder.new_package(*args, **kwargs) @property def pkg_id(self): return self._pkg_id def universe_builder(self): return self._uni_builder def build(self, *args, **kwargs): return self._uni_builder.build(*args, **kwargs) class UniverseBuilder(object): def __init__(self): self._cache = {} self._packages = {} self._default_version = '1.0-1' self._default_architecture = 'amd64' def _fetch_pkg_id(self, pkgish, version=None, architecture=None): if pkgish in self._cache: return self._cache[pkgish] if version is None: version = self._default_version if architecture is None: architecture = self._default_architecture if type(pkgish) == str: pkg_id = BinaryPackageId(pkgish, version, architecture) elif isinstance(pkgish, BinaryPackageId): pkg_id = pkgish elif type(pkgish) == tuple: if len(pkgish) == 2: pkg_id = BinaryPackageId(pkgish[0], pkgish[1], architecture) else: pkg_id = BinaryPackageId(*pkgish) elif isinstance(pkgish, PkgUniversePackageBuilder): pkg_id = pkgish._pkg_id else: raise ValueError("No clue on how to convert %s into a package id" % pkgish) self._cache[pkg_id] = pkg_id return pkg_id def new_package(self, raw_pkg_id_or_pkg, *, version=None, architecture=None): pkg_id = self._fetch_pkg_id(raw_pkg_id_or_pkg, version=version, architecture=architecture) pkg_builder = PkgUniversePackageBuilder(self, pkg_id) if pkg_id in self._packages: raise ValueError("Package %s already added previously" % pkg_id) self._packages[pkg_id] = pkg_builder return pkg_builder def build(self): builder = InstallabilityTesterBuilder() for pkg_id, pkg_builder in self._packages.items(): builder.add_binary(pkg_id, essential=pkg_builder._is_essential, in_testing=pkg_builder._in_testing, ) builder.set_relations(pkg_id, pkg_builder._dependencies, pkg_builder._conflicts) return builder.build() def pkg_id(self, pkgish): return self._fetch_pkg_id(pkgish) def update_package(self, pkgish): pkg_id = self._fetch_pkg_id(pkgish) if pkg_id not in self._packages: raise ValueError("Package %s has not been added yet" % pkg_id) return self._packages[pkg_id] # autopkgtest classes class TestData: def __init__(self): '''Construct local test package indexes. The archive is initially empty. You can create new packages with create_deb(). self.path contains the path of the archive, and self.apt_source provides an apt source "deb" line. It is kept in a temporary directory which gets removed when the Archive object gets deleted. ''' self.path = tempfile.mkdtemp(prefix='testarchive.') self.apt_source = 'deb file://%s /' % self.path self.suite_testing = 'testing' self.suite_unstable = 'unstable' self.compute_migrations = '' self.dirs = {False: os.path.join(self.path, 'data', self.suite_testing), True: os.path.join(self.path, 'data', self.suite_unstable)} os.makedirs(self.dirs[False]) os.mkdir(self.dirs[True]) self.added_sources = {False: set(), True: set()} self.added_binaries = {False: set(), True: set()} # pre-create all files for all architectures for arch in architectures: for dir in self.dirs.values(): with open(os.path.join(dir, 'Packages_' + arch), 'w'): pass for dir in self.dirs.values(): for fname in ['Dates', 'Blocks', 'Urgency', 'BugsV']: with open(os.path.join(dir, fname), 'w'): pass os.mkdir(os.path.join(self.path, 'data', 'hints')) shutil.copytree(os.path.join(PROJECT_DIR, 'tests', 'policy-test-data', 'piuparts', 'basic'), os.path.join(self.dirs[False], 'state')) os.mkdir(os.path.join(self.path, 'output')) # create temporary home dir for proposed-migration autopktest status self.home = os.path.join(self.path, 'home') os.environ['HOME'] = self.home os.makedirs(os.path.join(self.home, 'proposed-migration', 'autopkgtest', 'work')) def __del__(self): shutil.rmtree(self.path) def add(self, name, unstable, fields={}, add_src=True, testsuite=None, srcfields=None): '''Add a binary package to the index file. You need to specify at least the package name and in which list to put it (unstable==True for unstable/proposed, or False for testing/release). fields specifies all additional entries, e. g. {'Depends': 'foo, bar', 'Conflicts: baz'}. There are defaults for most fields. Unless add_src is set to False, this will also automatically create a source record, based on fields['Source'] and name. In that case, the "Testsuite:" field is set to the testsuite argument. ''' assert (name not in self.added_binaries[unstable]) self.added_binaries[unstable].add(name) fields.setdefault('Architecture', 'any') fields.setdefault('Version', '1') fields.setdefault('Priority', 'optional') fields.setdefault('Section', 'devel') fields.setdefault('Description', 'test pkg') if fields['Architecture'] == 'any': fields_local_copy = fields.copy() for a in architectures: fields_local_copy['Architecture'] = a self._append(name, unstable, 'Packages_' + a, fields_local_copy) elif fields['Architecture'] == 'all': for a in architectures: self._append(name, unstable, 'Packages_' + a, fields) else: self._append(name, unstable, 'Packages_' + fields['Architecture'], fields) if add_src: src = fields.get('Source', name) if src not in self.added_sources[unstable]: if srcfields is None: srcfields = {} srcfields['Version'] = fields['Version'] srcfields['Section'] = fields['Section'] if testsuite: srcfields['Testsuite'] = testsuite self.add_src(src, unstable, srcfields) def add_src(self, name, unstable, fields={}): '''Add a source package to the index file. You need to specify at least the package name and in which list to put it (unstable==True for unstable/proposed, or False for testing/release). fields specifies all additional entries, which can be Version (default: 1), Section (default: devel), Testsuite (default: none), and Extra-Source-Only. ''' assert (name not in self.added_sources[unstable]) self.added_sources[unstable].add(name) fields.setdefault('Version', '1') fields.setdefault('Section', 'devel') self._append(name, unstable, 'Sources', fields) def _append(self, name, unstable, file_name, fields): with open(os.path.join(self.dirs[unstable], file_name), 'a') as f: f.write('''Package: %s Maintainer: Joe ''' % name) for k, v in fields.items(): f.write('%s: %s\n' % (k, v)) f.write('\n') def remove_all(self, unstable): '''Remove all added packages''' self.added_binaries[unstable] = set() self.added_sources[unstable] = set() for a in architectures: open(os.path.join(self.dirs[unstable], 'Packages_' + a), 'w').close() open(os.path.join(self.dirs[unstable], 'Sources'), 'w').close() def add_default_packages(self, libc6=True, green=True, lightgreen=True, darkgreen=True, blue=True, black=True, grey=True): '''To avoid duplication, add packages we need all the time''' # libc6 (always) self.add('libc6', False) if (libc6 is True): self.add('libc6', True) # src:green self.add('libgreen1', False, {'Source': 'green', 'Depends': 'libc6 (>= 0.9)'}, testsuite='autopkgtest') if (green is True): self.add('libgreen1', True, {'Source': 'green', 'Depends': 'libc6 (>= 0.9)'}, testsuite='autopkgtest') self.add('green', False, {'Depends': 'libc6 (>= 0.9), libgreen1', 'Conflicts': 'blue'}, testsuite='autopkgtest') if (green is True): self.add('green', True, {'Depends': 'libc6 (>= 0.9), libgreen1', 'Conflicts': 'blue'}, testsuite='autopkgtest') # lightgreen self.add('lightgreen', False, {'Depends': 'libgreen1'}, testsuite='autopkgtest') if (lightgreen is True): self.add('lightgreen', True, {'Depends': 'libgreen1'}, testsuite='autopkgtest') ## autodep8 or similar test # darkgreen self.add('darkgreen', False, {'Depends': 'libgreen1'}, testsuite='autopkgtest-pkg-foo') if (darkgreen is True): self.add('darkgreen', True, {'Depends': 'libgreen1'}, testsuite='autopkgtest-pkg-foo') # blue self.add('blue', False, {'Depends': 'libc6 (>= 0.9)', 'Conflicts': 'green'}, testsuite='specialtest') if blue is True: self.add('blue', True, {'Depends': 'libc6 (>= 0.9)', 'Conflicts': 'green'}, testsuite='specialtest') # black self.add('black', False, {}, testsuite='autopkgtest') if black is True: self.add('black', True, {}, testsuite='autopkgtest') # grey self.add('grey', False, {}, testsuite='autopkgtest') if grey is True: self.add('grey', True, {}, testsuite='autopkgtest') class TestBase(unittest.TestCase): def setUp(self): super(TestBase, self).setUp() self.maxDiff = None self.data = TestData() self.britney = os.path.join(PROJECT_DIR, 'britney.py') # create temporary config so that tests can hack it self.britney_conf = os.path.join(self.data.path, 'britney.conf') with open(self.britney_conf, 'w') as f: f.write(''' TESTING = data/testing UNSTABLE = data/unstable NONINST_STATUS = data/testing/non-installable-status EXCUSES_OUTPUT = output/excuses.html EXCUSES_YAML_OUTPUT = output/excuses.yaml UPGRADE_OUTPUT = output/output.txt HEIDI_OUTPUT = output/HeidiResult STATIC_INPUT_DIR = data/testing/input STATE_DIR = data/testing/state ARCHITECTURES = amd64 arm64 armhf i386 powerpc ppc64el NOBREAKALL_ARCHES = amd64 arm64 armhf i386 powerpc ppc64el OUTOFSYNC_ARCHES = BREAK_ARCHES = NEW_ARCHES = MINDAYS_LOW = 0 MINDAYS_MEDIUM = 0 MINDAYS_HIGH = 0 MINDAYS_CRITICAL = 0 MINDAYS_EMERGENCY = 0 DEFAULT_URGENCY = medium NO_PENALTIES = high critical emergency BOUNTY_MIN_AGE = 8 HINTSDIR = data/hints HINTS_AUTOPKGTEST = ALL HINTS_FREEZE = block block-all block-udeb HINTS_FREEZE-EXCEPTION = unblock unblock-udeb HINTS_SATBRITNEY = easy HINTS_AUTO-REMOVALS = remove SMOOTH_UPDATES = badgers IGNORE_CRUFT = 0 REMOVE_OBSOLETE = no ADT_ENABLE = yes ADT_ARCHES = amd64 i386 ADT_AMQP = file://output/debci.input ADT_PPAS = ADT_SHARED_RESULTS_CACHE = ADT_SWIFT_URL = http://localhost:18085 ADT_CI_URL = https://autopkgtest.ubuntu.com/ ADT_HUGE = 20 ADT_DB_URL = ADT_SUCCESS_BOUNTY = ADT_REGRESSION_PENALTY = ADT_BASELINE = ADT_RETRY_OLDER_THAN = EMAIL_ENABLE = yes ''') assert os.path.exists(self.britney) def tearDown(self): del self.data def run_britney(self, args=[]): '''Run britney. Assert that it succeeds and does not produce anything on stderr. Return (excuses.yaml, excuses.html, britney_out). ''' britney = subprocess.Popen([self.britney, '-v', '-c', self.britney_conf, '%s' % self.data.compute_migrations], stdout=subprocess.PIPE, stderr=subprocess.PIPE, cwd=self.data.path, universal_newlines=True) (out, err) = britney.communicate() self.assertEqual(britney.returncode, 0, out + err) self.assertEqual(err, '') with open(os.path.join(self.data.path, 'output', 'excuses.yaml'), encoding='utf-8') as f: yaml = f.read() with open(os.path.join(self.data.path, 'output', 'excuses.html'), encoding='utf-8') as f: html = f.read() return (yaml, html, out) def create_hint(self, username, content): '''Create a hint file for the given username and content''' hints_path = os.path.join( self.data.path, 'data', 'hints', username) with open(hints_path, 'a') as fd: fd.write(content) fd.write('\n')