Use pkg_id instead pkg/arch in BINARIES

Signed-off-by: Niels Thykier <niels@thykier.net>
debian
Niels Thykier 9 years ago
parent aa34a47f8a
commit 12d9ae8fa3

@ -674,7 +674,6 @@ class Britney(object):
if "(" in source:
dpkg[SOURCEVER] = intern(source[source.find("(")+1:source.find(")")])
pkgarch = "%s/%s" % (pkg,arch)
# if the source package is available in the distribution, then register this binary package
if dpkg[SOURCE] in sources[distribution]:
# There may be multiple versions of any arch:all packages
@ -683,11 +682,11 @@ class Britney(object):
# source -> binary mapping once. It doesn't matter which
# of the versions we include as only the package name and
# architecture are recorded.
if pkgarch not in sources[distribution][dpkg[SOURCE]][BINARIES]:
sources[distribution][dpkg[SOURCE]][BINARIES].append(pkgarch)
if pkg_id not in sources[distribution][dpkg[SOURCE]][BINARIES]:
sources[distribution][dpkg[SOURCE]][BINARIES].append(pkg_id)
# if the source package doesn't exist, create a fake one
else:
sources[distribution][dpkg[SOURCE]] = [dpkg[SOURCEVER], 'faux', [pkgarch], None, True]
sources[distribution][dpkg[SOURCE]] = [dpkg[SOURCEVER], 'faux', [pkg_id], None, True]
# register virtual packages and real packages that provide them
if dpkg[PROVIDES]:
@ -1151,11 +1150,11 @@ class Britney(object):
anyworthdoing = False
# for every binary package produced by this source in unstable for this architecture
for pkg in sorted(filter(lambda x: x.endswith("/" + arch), source_u[BINARIES]), key=lambda x: x.split("/")[0]):
pkg_name = pkg.split("/")[0]
for pkg_id in sorted(x for x in source_u[BINARIES] if x[2] == arch):
pkg_name = pkg_id[0]
# retrieve the testing (if present) and unstable corresponding binary packages
binary_t = pkg in source_t[BINARIES] and self.binaries['testing'][arch][0][pkg_name] or None
binary_t = pkg_name in self.binaries['testing'][arch][0] and self.binaries['testing'][arch][0][pkg_name] or None
binary_u = self.binaries[suite][arch][0][pkg_name]
# this is the source version for the new binary package
@ -1228,7 +1227,8 @@ class Britney(object):
arch,
False)
for pkg in sorted(x.split("/")[0] for x in source_data[BINARIES] if x.endswith("/"+arch)):
for pkg_id in sorted(x for x in source_data[BINARIES] if x[2] == arch):
pkg = pkg_id[0]
# if the package is architecture-independent, then ignore it
tpkg_data = self.binaries['testing'][arch][0][pkg]
if tpkg_data[ARCHITECTURE] == 'all':
@ -1245,8 +1245,7 @@ class Britney(object):
# it "interesting" on its own. This case happens quite often with smooth updatable
# packages, where the old binary "survives" a full run because it still has
# reverse dependencies.
name = (pkg, tpkg_data[VERSION], tpkg_data[ARCHITECTURE])
if name not in smoothbins:
if pkg_id not in smoothbins:
anyworthdoing = True
# if there is nothing wrong and there is something worth doing, this is a valid candidate
@ -1407,7 +1406,7 @@ class Britney(object):
# if the package in testing has no binaries on this
# architecture, it can't be out-of-date
if not any(x for x in self.sources["testing"][src][BINARIES]
if x.endswith("/"+arch) and self.binaries["testing"][arch][0][x.split("/")[0]][ARCHITECTURE] != 'all'):
if x[2] == arch and self.binaries["testing"][arch][0][x[0]][ARCHITECTURE] != 'all'):
continue
# if the (t-)p-u package has produced any binaries on
@ -1443,7 +1442,8 @@ class Britney(object):
oodbins = {}
uptodatebins = False
# for every binary package produced by this source in the suite for this architecture
for pkg in sorted(x.split("/")[0] for x in self.sources[suite][src][BINARIES] if x.endswith("/"+arch)):
for pkg_id in sorted(x for x in self.sources[suite][src][BINARIES] if x[2] == arch):
pkg = pkg_id[0]
if pkg not in pkgs: pkgs[pkg] = []
pkgs[pkg].append(arch)
@ -1920,8 +1920,8 @@ class Britney(object):
# remove all the binaries
# first, build a list of eligible binaries
for p in source_data[BINARIES]:
binary, parch = p.split("/")
for pkg_id in source_data[BINARIES]:
binary, _, parch = pkg_id
if (migration_architecture != 'source'
and parch != migration_architecture):
continue
@ -1933,8 +1933,7 @@ class Britney(object):
if (not include_hijacked
and binaries_t[parch][0][binary][SOURCE] != source_name):
continue
version = binaries_t[parch][0][binary][VERSION]
bins.append((binary, version, parch))
bins.append(pkg_id)
for pkg_id in bins:
binary, _, parch = pkg_id
@ -2008,11 +2007,10 @@ class Britney(object):
# add the new binary packages (if we are not removing)
if not is_removal:
source_data = sources[suite][source_name]
for p in source_data[BINARIES]:
binary, parch = p.split("/")
for pkg_id in source_data[BINARIES]:
binary, _, parch = pkg_id
if migration_architecture not in ['source', parch]:
continue
version = self.binaries[suite][parch][0][binary][VERSION]
if (not include_hijacked
and self.binaries[suite][parch][0][binary][SOURCE] != source_name):
@ -2027,7 +2025,7 @@ class Britney(object):
rms.remove((rm_b, rm_v, rm_p))
continue
adds.add((binary, version, parch))
adds.add(pkg_id)
return (adds, rms, smoothbins)

@ -123,12 +123,11 @@ def undo_changes(lundo, inst_tester, sources, binaries, all_binary_packages,
# undo all new binaries (consequence of the above)
for (undo, item) in lundo:
if not item.is_removal and item.package in sources[item.suite]:
for p in sources[item.suite][item.package][BINARIES]:
binary, arch = p.split("/")
for pkg_id in sources[item.suite][item.package][BINARIES]:
binary, _, arch = pkg_id
if item.architecture in ['source', arch]:
version = binaries["testing"][arch][0][binary][VERSION]
del binaries["testing"][arch][0][binary]
inst_tester.remove_testing_binary((binary, version, arch))
inst_tester.remove_testing_binary(pkg_id)
# STEP 3

Loading…
Cancel
Save