diff options
| author | Fredrik Gustafsson <fredrik.gustafsson@axis.com> | 2020-07-24 16:42:29 +0200 |
|---|---|---|
| committer | Richard Purdie <richard.purdie@linuxfoundation.org> | 2020-07-27 19:58:09 +0100 |
| commit | 34dc8571d2d7c2bb9b5f92e157ab91e9d4f9e992 (patch) | |
| tree | de10e02d583cd12b2b319b7fdaf2951244fc3695 /meta/lib/oe/package_manager | |
| parent | 7b78e1b4de0421de9f876aeb4ff11d6d017a4b7b (diff) | |
| download | poky-34dc8571d2d7c2bb9b5f92e157ab91e9d4f9e992.tar.gz | |
package_manager: Move to package_manager/__init__.py
This is part of a refactor that will split the package manager
code so that it's possible to use other package managers in other
layers.
(From OE-Core rev: 66ccc7a228bf73df0a4dd846bf2c8e99eaa79580)
Signed-off-by: Fredrik Gustafsson <fredrigu@axis.com>
Signed-off-by: Richard Purdie <richard.purdie@linuxfoundation.org>
Diffstat (limited to 'meta/lib/oe/package_manager')
| -rw-r--r-- | meta/lib/oe/package_manager/__init__.py | 1863 |
1 files changed, 1863 insertions, 0 deletions
diff --git a/meta/lib/oe/package_manager/__init__.py b/meta/lib/oe/package_manager/__init__.py new file mode 100644 index 0000000000..35e5cff073 --- /dev/null +++ b/meta/lib/oe/package_manager/__init__.py | |||
| @@ -0,0 +1,1863 @@ | |||
| 1 | # | ||
| 2 | # SPDX-License-Identifier: GPL-2.0-only | ||
| 3 | # | ||
| 4 | |||
| 5 | from abc import ABCMeta, abstractmethod | ||
| 6 | import os | ||
| 7 | import glob | ||
| 8 | import subprocess | ||
| 9 | import shutil | ||
| 10 | import re | ||
| 11 | import collections | ||
| 12 | import bb | ||
| 13 | import tempfile | ||
| 14 | import oe.utils | ||
| 15 | import oe.path | ||
| 16 | import string | ||
| 17 | from oe.gpg_sign import get_signer | ||
| 18 | import hashlib | ||
| 19 | import fnmatch | ||
| 20 | |||
| 21 | # this can be used by all PM backends to create the index files in parallel | ||
| 22 | def create_index(arg): | ||
| 23 | index_cmd = arg | ||
| 24 | |||
| 25 | bb.note("Executing '%s' ..." % index_cmd) | ||
| 26 | result = subprocess.check_output(index_cmd, stderr=subprocess.STDOUT, shell=True).decode("utf-8") | ||
| 27 | if result: | ||
| 28 | bb.note(result) | ||
| 29 | |||
| 30 | def opkg_query(cmd_output): | ||
| 31 | """ | ||
| 32 | This method parse the output from the package managerand return | ||
| 33 | a dictionary with the information of the packages. This is used | ||
| 34 | when the packages are in deb or ipk format. | ||
| 35 | """ | ||
| 36 | verregex = re.compile(r' \([=<>]* [^ )]*\)') | ||
| 37 | output = dict() | ||
| 38 | pkg = "" | ||
| 39 | arch = "" | ||
| 40 | ver = "" | ||
| 41 | filename = "" | ||
| 42 | dep = [] | ||
| 43 | prov = [] | ||
| 44 | pkgarch = "" | ||
| 45 | for line in cmd_output.splitlines()+['']: | ||
| 46 | line = line.rstrip() | ||
| 47 | if ':' in line: | ||
| 48 | if line.startswith("Package: "): | ||
| 49 | pkg = line.split(": ")[1] | ||
| 50 | elif line.startswith("Architecture: "): | ||
| 51 | arch = line.split(": ")[1] | ||
| 52 | elif line.startswith("Version: "): | ||
| 53 | ver = line.split(": ")[1] | ||
| 54 | elif line.startswith("File: ") or line.startswith("Filename:"): | ||
| 55 | filename = line.split(": ")[1] | ||
| 56 | if "/" in filename: | ||
| 57 | filename = os.path.basename(filename) | ||
| 58 | elif line.startswith("Depends: "): | ||
| 59 | depends = verregex.sub('', line.split(": ")[1]) | ||
| 60 | for depend in depends.split(", "): | ||
| 61 | dep.append(depend) | ||
| 62 | elif line.startswith("Recommends: "): | ||
| 63 | recommends = verregex.sub('', line.split(": ")[1]) | ||
| 64 | for recommend in recommends.split(", "): | ||
| 65 | dep.append("%s [REC]" % recommend) | ||
| 66 | elif line.startswith("PackageArch: "): | ||
| 67 | pkgarch = line.split(": ")[1] | ||
| 68 | elif line.startswith("Provides: "): | ||
| 69 | provides = verregex.sub('', line.split(": ")[1]) | ||
| 70 | for provide in provides.split(", "): | ||
| 71 | prov.append(provide) | ||
| 72 | |||
| 73 | # When there is a blank line save the package information | ||
| 74 | elif not line: | ||
| 75 | # IPK doesn't include the filename | ||
| 76 | if not filename: | ||
| 77 | filename = "%s_%s_%s.ipk" % (pkg, ver, arch) | ||
| 78 | if pkg: | ||
| 79 | output[pkg] = {"arch":arch, "ver":ver, | ||
| 80 | "filename":filename, "deps": dep, "pkgarch":pkgarch, "provs": prov} | ||
| 81 | pkg = "" | ||
| 82 | arch = "" | ||
| 83 | ver = "" | ||
| 84 | filename = "" | ||
| 85 | dep = [] | ||
| 86 | prov = [] | ||
| 87 | pkgarch = "" | ||
| 88 | |||
| 89 | return output | ||
| 90 | |||
| 91 | def failed_postinsts_abort(pkgs, log_path): | ||
| 92 | bb.fatal("""Postinstall scriptlets of %s have failed. If the intention is to defer them to first boot, | ||
| 93 | then please place them into pkg_postinst_ontarget_${PN} (). | ||
| 94 | Deferring to first boot via 'exit 1' is no longer supported. | ||
| 95 | Details of the failure are in %s.""" %(pkgs, log_path)) | ||
| 96 | |||
| 97 | def generate_locale_archive(d, rootfs, target_arch, localedir): | ||
| 98 | # Pretty sure we don't need this for locale archive generation but | ||
| 99 | # keeping it to be safe... | ||
| 100 | locale_arch_options = { \ | ||
| 101 | "arc": ["--uint32-align=4", "--little-endian"], | ||
| 102 | "arceb": ["--uint32-align=4", "--big-endian"], | ||
| 103 | "arm": ["--uint32-align=4", "--little-endian"], | ||
| 104 | "armeb": ["--uint32-align=4", "--big-endian"], | ||
| 105 | "aarch64": ["--uint32-align=4", "--little-endian"], | ||
| 106 | "aarch64_be": ["--uint32-align=4", "--big-endian"], | ||
| 107 | "sh4": ["--uint32-align=4", "--big-endian"], | ||
| 108 | "powerpc": ["--uint32-align=4", "--big-endian"], | ||
| 109 | "powerpc64": ["--uint32-align=4", "--big-endian"], | ||
| 110 | "powerpc64le": ["--uint32-align=4", "--little-endian"], | ||
| 111 | "mips": ["--uint32-align=4", "--big-endian"], | ||
| 112 | "mipsisa32r6": ["--uint32-align=4", "--big-endian"], | ||
| 113 | "mips64": ["--uint32-align=4", "--big-endian"], | ||
| 114 | "mipsisa64r6": ["--uint32-align=4", "--big-endian"], | ||
| 115 | "mipsel": ["--uint32-align=4", "--little-endian"], | ||
| 116 | "mipsisa32r6el": ["--uint32-align=4", "--little-endian"], | ||
| 117 | "mips64el": ["--uint32-align=4", "--little-endian"], | ||
| 118 | "mipsisa64r6el": ["--uint32-align=4", "--little-endian"], | ||
| 119 | "riscv64": ["--uint32-align=4", "--little-endian"], | ||
| 120 | "riscv32": ["--uint32-align=4", "--little-endian"], | ||
| 121 | "i586": ["--uint32-align=4", "--little-endian"], | ||
| 122 | "i686": ["--uint32-align=4", "--little-endian"], | ||
| 123 | "x86_64": ["--uint32-align=4", "--little-endian"] | ||
| 124 | } | ||
| 125 | if target_arch in locale_arch_options: | ||
| 126 | arch_options = locale_arch_options[target_arch] | ||
| 127 | else: | ||
| 128 | bb.error("locale_arch_options not found for target_arch=" + target_arch) | ||
| 129 | bb.fatal("unknown arch:" + target_arch + " for locale_arch_options") | ||
| 130 | |||
| 131 | # Need to set this so cross-localedef knows where the archive is | ||
| 132 | env = dict(os.environ) | ||
| 133 | env["LOCALEARCHIVE"] = oe.path.join(localedir, "locale-archive") | ||
| 134 | |||
| 135 | for name in sorted(os.listdir(localedir)): | ||
| 136 | path = os.path.join(localedir, name) | ||
| 137 | if os.path.isdir(path): | ||
| 138 | cmd = ["cross-localedef", "--verbose"] | ||
| 139 | cmd += arch_options | ||
| 140 | cmd += ["--add-to-archive", path] | ||
| 141 | subprocess.check_output(cmd, env=env, stderr=subprocess.STDOUT) | ||
| 142 | |||
| 143 | class Indexer(object, metaclass=ABCMeta): | ||
| 144 | def __init__(self, d, deploy_dir): | ||
| 145 | self.d = d | ||
| 146 | self.deploy_dir = deploy_dir | ||
| 147 | |||
| 148 | @abstractmethod | ||
| 149 | def write_index(self): | ||
| 150 | pass | ||
| 151 | |||
| 152 | |||
| 153 | class RpmIndexer(Indexer): | ||
| 154 | def write_index(self): | ||
| 155 | self.do_write_index(self.deploy_dir) | ||
| 156 | |||
| 157 | def do_write_index(self, deploy_dir): | ||
| 158 | if self.d.getVar('PACKAGE_FEED_SIGN') == '1': | ||
| 159 | signer = get_signer(self.d, self.d.getVar('PACKAGE_FEED_GPG_BACKEND')) | ||
| 160 | else: | ||
| 161 | signer = None | ||
| 162 | |||
| 163 | createrepo_c = bb.utils.which(os.environ['PATH'], "createrepo_c") | ||
| 164 | result = create_index("%s --update -q %s" % (createrepo_c, deploy_dir)) | ||
| 165 | if result: | ||
| 166 | bb.fatal(result) | ||
| 167 | |||
| 168 | # Sign repomd | ||
| 169 | if signer: | ||
| 170 | sig_type = self.d.getVar('PACKAGE_FEED_GPG_SIGNATURE_TYPE') | ||
| 171 | is_ascii_sig = (sig_type.upper() != "BIN") | ||
| 172 | signer.detach_sign(os.path.join(deploy_dir, 'repodata', 'repomd.xml'), | ||
| 173 | self.d.getVar('PACKAGE_FEED_GPG_NAME'), | ||
| 174 | self.d.getVar('PACKAGE_FEED_GPG_PASSPHRASE_FILE'), | ||
| 175 | armor=is_ascii_sig) | ||
| 176 | |||
| 177 | class RpmSubdirIndexer(RpmIndexer): | ||
| 178 | def write_index(self): | ||
| 179 | bb.note("Generating package index for %s" %(self.deploy_dir)) | ||
| 180 | self.do_write_index(self.deploy_dir) | ||
| 181 | for entry in os.walk(self.deploy_dir): | ||
| 182 | if os.path.samefile(self.deploy_dir, entry[0]): | ||
| 183 | for dir in entry[1]: | ||
| 184 | if dir != 'repodata': | ||
| 185 | dir_path = oe.path.join(self.deploy_dir, dir) | ||
| 186 | bb.note("Generating package index for %s" %(dir_path)) | ||
| 187 | self.do_write_index(dir_path) | ||
| 188 | |||
| 189 | class OpkgIndexer(Indexer): | ||
| 190 | def write_index(self): | ||
| 191 | arch_vars = ["ALL_MULTILIB_PACKAGE_ARCHS", | ||
| 192 | "SDK_PACKAGE_ARCHS", | ||
| 193 | ] | ||
| 194 | |||
| 195 | opkg_index_cmd = bb.utils.which(os.getenv('PATH'), "opkg-make-index") | ||
| 196 | if self.d.getVar('PACKAGE_FEED_SIGN') == '1': | ||
| 197 | signer = get_signer(self.d, self.d.getVar('PACKAGE_FEED_GPG_BACKEND')) | ||
| 198 | else: | ||
| 199 | signer = None | ||
| 200 | |||
| 201 | if not os.path.exists(os.path.join(self.deploy_dir, "Packages")): | ||
| 202 | open(os.path.join(self.deploy_dir, "Packages"), "w").close() | ||
| 203 | |||
| 204 | index_cmds = set() | ||
| 205 | index_sign_files = set() | ||
| 206 | for arch_var in arch_vars: | ||
| 207 | archs = self.d.getVar(arch_var) | ||
| 208 | if archs is None: | ||
| 209 | continue | ||
| 210 | |||
| 211 | for arch in archs.split(): | ||
| 212 | pkgs_dir = os.path.join(self.deploy_dir, arch) | ||
| 213 | pkgs_file = os.path.join(pkgs_dir, "Packages") | ||
| 214 | |||
| 215 | if not os.path.isdir(pkgs_dir): | ||
| 216 | continue | ||
| 217 | |||
| 218 | if not os.path.exists(pkgs_file): | ||
| 219 | open(pkgs_file, "w").close() | ||
| 220 | |||
| 221 | index_cmds.add('%s --checksum md5 --checksum sha256 -r %s -p %s -m %s' % | ||
| 222 | (opkg_index_cmd, pkgs_file, pkgs_file, pkgs_dir)) | ||
| 223 | |||
| 224 | index_sign_files.add(pkgs_file) | ||
| 225 | |||
| 226 | if len(index_cmds) == 0: | ||
| 227 | bb.note("There are no packages in %s!" % self.deploy_dir) | ||
| 228 | return | ||
| 229 | |||
| 230 | oe.utils.multiprocess_launch(create_index, index_cmds, self.d) | ||
| 231 | |||
| 232 | if signer: | ||
| 233 | feed_sig_type = self.d.getVar('PACKAGE_FEED_GPG_SIGNATURE_TYPE') | ||
| 234 | is_ascii_sig = (feed_sig_type.upper() != "BIN") | ||
| 235 | for f in index_sign_files: | ||
| 236 | signer.detach_sign(f, | ||
| 237 | self.d.getVar('PACKAGE_FEED_GPG_NAME'), | ||
| 238 | self.d.getVar('PACKAGE_FEED_GPG_PASSPHRASE_FILE'), | ||
| 239 | armor=is_ascii_sig) | ||
| 240 | |||
| 241 | |||
| 242 | class DpkgIndexer(Indexer): | ||
| 243 | def _create_configs(self): | ||
| 244 | bb.utils.mkdirhier(self.apt_conf_dir) | ||
| 245 | bb.utils.mkdirhier(os.path.join(self.apt_conf_dir, "lists", "partial")) | ||
| 246 | bb.utils.mkdirhier(os.path.join(self.apt_conf_dir, "apt.conf.d")) | ||
| 247 | bb.utils.mkdirhier(os.path.join(self.apt_conf_dir, "preferences.d")) | ||
| 248 | |||
| 249 | with open(os.path.join(self.apt_conf_dir, "preferences"), | ||
| 250 | "w") as prefs_file: | ||
| 251 | pass | ||
| 252 | with open(os.path.join(self.apt_conf_dir, "sources.list"), | ||
| 253 | "w+") as sources_file: | ||
| 254 | pass | ||
| 255 | |||
| 256 | with open(self.apt_conf_file, "w") as apt_conf: | ||
| 257 | with open(os.path.join(self.d.expand("${STAGING_ETCDIR_NATIVE}"), | ||
| 258 | "apt", "apt.conf.sample")) as apt_conf_sample: | ||
| 259 | for line in apt_conf_sample.read().split("\n"): | ||
| 260 | line = re.sub(r"#ROOTFS#", "/dev/null", line) | ||
| 261 | line = re.sub(r"#APTCONF#", self.apt_conf_dir, line) | ||
| 262 | apt_conf.write(line + "\n") | ||
| 263 | |||
| 264 | def write_index(self): | ||
| 265 | self.apt_conf_dir = os.path.join(self.d.expand("${APTCONF_TARGET}"), | ||
| 266 | "apt-ftparchive") | ||
| 267 | self.apt_conf_file = os.path.join(self.apt_conf_dir, "apt.conf") | ||
| 268 | self._create_configs() | ||
| 269 | |||
| 270 | os.environ['APT_CONFIG'] = self.apt_conf_file | ||
| 271 | |||
| 272 | pkg_archs = self.d.getVar('PACKAGE_ARCHS') | ||
| 273 | if pkg_archs is not None: | ||
| 274 | arch_list = pkg_archs.split() | ||
| 275 | sdk_pkg_archs = self.d.getVar('SDK_PACKAGE_ARCHS') | ||
| 276 | if sdk_pkg_archs is not None: | ||
| 277 | for a in sdk_pkg_archs.split(): | ||
| 278 | if a not in pkg_archs: | ||
| 279 | arch_list.append(a) | ||
| 280 | |||
| 281 | all_mlb_pkg_arch_list = (self.d.getVar('ALL_MULTILIB_PACKAGE_ARCHS') or "").split() | ||
| 282 | arch_list.extend(arch for arch in all_mlb_pkg_arch_list if arch not in arch_list) | ||
| 283 | |||
| 284 | apt_ftparchive = bb.utils.which(os.getenv('PATH'), "apt-ftparchive") | ||
| 285 | gzip = bb.utils.which(os.getenv('PATH'), "gzip") | ||
| 286 | |||
| 287 | index_cmds = [] | ||
| 288 | deb_dirs_found = False | ||
| 289 | for arch in arch_list: | ||
| 290 | arch_dir = os.path.join(self.deploy_dir, arch) | ||
| 291 | if not os.path.isdir(arch_dir): | ||
| 292 | continue | ||
| 293 | |||
| 294 | cmd = "cd %s; PSEUDO_UNLOAD=1 %s packages . > Packages;" % (arch_dir, apt_ftparchive) | ||
| 295 | |||
| 296 | cmd += "%s -fcn Packages > Packages.gz;" % gzip | ||
| 297 | |||
| 298 | with open(os.path.join(arch_dir, "Release"), "w+") as release: | ||
| 299 | release.write("Label: %s\n" % arch) | ||
| 300 | |||
| 301 | cmd += "PSEUDO_UNLOAD=1 %s release . >> Release" % apt_ftparchive | ||
| 302 | |||
| 303 | index_cmds.append(cmd) | ||
| 304 | |||
| 305 | deb_dirs_found = True | ||
| 306 | |||
| 307 | if not deb_dirs_found: | ||
| 308 | bb.note("There are no packages in %s" % self.deploy_dir) | ||
| 309 | return | ||
| 310 | |||
| 311 | oe.utils.multiprocess_launch(create_index, index_cmds, self.d) | ||
| 312 | if self.d.getVar('PACKAGE_FEED_SIGN') == '1': | ||
| 313 | raise NotImplementedError('Package feed signing not implementd for dpkg') | ||
| 314 | |||
| 315 | |||
| 316 | |||
| 317 | class PkgsList(object, metaclass=ABCMeta): | ||
| 318 | def __init__(self, d, rootfs_dir): | ||
| 319 | self.d = d | ||
| 320 | self.rootfs_dir = rootfs_dir | ||
| 321 | |||
| 322 | @abstractmethod | ||
| 323 | def list_pkgs(self): | ||
| 324 | pass | ||
| 325 | |||
| 326 | class RpmPkgsList(PkgsList): | ||
| 327 | def list_pkgs(self): | ||
| 328 | return RpmPM(self.d, self.rootfs_dir, self.d.getVar('TARGET_VENDOR'), needfeed=False).list_installed() | ||
| 329 | |||
| 330 | class OpkgPkgsList(PkgsList): | ||
| 331 | def __init__(self, d, rootfs_dir, config_file): | ||
| 332 | super(OpkgPkgsList, self).__init__(d, rootfs_dir) | ||
| 333 | |||
| 334 | self.opkg_cmd = bb.utils.which(os.getenv('PATH'), "opkg") | ||
| 335 | self.opkg_args = "-f %s -o %s " % (config_file, rootfs_dir) | ||
| 336 | self.opkg_args += self.d.getVar("OPKG_ARGS") | ||
| 337 | |||
| 338 | def list_pkgs(self, format=None): | ||
| 339 | cmd = "%s %s status" % (self.opkg_cmd, self.opkg_args) | ||
| 340 | |||
| 341 | # opkg returns success even when it printed some | ||
| 342 | # "Collected errors:" report to stderr. Mixing stderr into | ||
| 343 | # stdout then leads to random failures later on when | ||
| 344 | # parsing the output. To avoid this we need to collect both | ||
| 345 | # output streams separately and check for empty stderr. | ||
| 346 | p = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE, shell=True) | ||
| 347 | cmd_output, cmd_stderr = p.communicate() | ||
| 348 | cmd_output = cmd_output.decode("utf-8") | ||
| 349 | cmd_stderr = cmd_stderr.decode("utf-8") | ||
| 350 | if p.returncode or cmd_stderr: | ||
| 351 | bb.fatal("Cannot get the installed packages list. Command '%s' " | ||
| 352 | "returned %d and stderr:\n%s" % (cmd, p.returncode, cmd_stderr)) | ||
| 353 | |||
| 354 | return opkg_query(cmd_output) | ||
| 355 | |||
| 356 | |||
| 357 | class DpkgPkgsList(PkgsList): | ||
| 358 | |||
| 359 | def list_pkgs(self): | ||
| 360 | cmd = [bb.utils.which(os.getenv('PATH'), "dpkg-query"), | ||
| 361 | "--admindir=%s/var/lib/dpkg" % self.rootfs_dir, | ||
| 362 | "-W"] | ||
| 363 | |||
| 364 | cmd.append("-f=Package: ${Package}\nArchitecture: ${PackageArch}\nVersion: ${Version}\nFile: ${Package}_${Version}_${Architecture}.deb\nDepends: ${Depends}\nRecommends: ${Recommends}\nProvides: ${Provides}\n\n") | ||
| 365 | |||
| 366 | try: | ||
| 367 | cmd_output = subprocess.check_output(cmd, stderr=subprocess.STDOUT).strip().decode("utf-8") | ||
| 368 | except subprocess.CalledProcessError as e: | ||
| 369 | bb.fatal("Cannot get the installed packages list. Command '%s' " | ||
| 370 | "returned %d:\n%s" % (' '.join(cmd), e.returncode, e.output.decode("utf-8"))) | ||
| 371 | |||
| 372 | return opkg_query(cmd_output) | ||
| 373 | |||
| 374 | |||
| 375 | class PackageManager(object, metaclass=ABCMeta): | ||
| 376 | """ | ||
| 377 | This is an abstract class. Do not instantiate this directly. | ||
| 378 | """ | ||
| 379 | |||
| 380 | def __init__(self, d, target_rootfs): | ||
| 381 | self.d = d | ||
| 382 | self.target_rootfs = target_rootfs | ||
| 383 | self.deploy_dir = None | ||
| 384 | self.deploy_lock = None | ||
| 385 | self._initialize_intercepts() | ||
| 386 | |||
| 387 | def _initialize_intercepts(self): | ||
| 388 | bb.note("Initializing intercept dir for %s" % self.target_rootfs) | ||
| 389 | # As there might be more than one instance of PackageManager operating at the same time | ||
| 390 | # we need to isolate the intercept_scripts directories from each other, | ||
| 391 | # hence the ugly hash digest in dir name. | ||
| 392 | self.intercepts_dir = os.path.join(self.d.getVar('WORKDIR'), "intercept_scripts-%s" % | ||
| 393 | (hashlib.sha256(self.target_rootfs.encode()).hexdigest())) | ||
| 394 | |||
| 395 | postinst_intercepts = (self.d.getVar("POSTINST_INTERCEPTS") or "").split() | ||
| 396 | if not postinst_intercepts: | ||
| 397 | postinst_intercepts_path = self.d.getVar("POSTINST_INTERCEPTS_PATH") | ||
| 398 | if not postinst_intercepts_path: | ||
| 399 | postinst_intercepts_path = self.d.getVar("POSTINST_INTERCEPTS_DIR") or self.d.expand("${COREBASE}/scripts/postinst-intercepts") | ||
| 400 | postinst_intercepts = oe.path.which_wild('*', postinst_intercepts_path) | ||
| 401 | |||
| 402 | bb.debug(1, 'Collected intercepts:\n%s' % ''.join(' %s\n' % i for i in postinst_intercepts)) | ||
| 403 | bb.utils.remove(self.intercepts_dir, True) | ||
| 404 | bb.utils.mkdirhier(self.intercepts_dir) | ||
| 405 | for intercept in postinst_intercepts: | ||
| 406 | bb.utils.copyfile(intercept, os.path.join(self.intercepts_dir, os.path.basename(intercept))) | ||
| 407 | |||
| 408 | @abstractmethod | ||
| 409 | def _handle_intercept_failure(self, failed_script): | ||
| 410 | pass | ||
| 411 | |||
| 412 | def _postpone_to_first_boot(self, postinst_intercept_hook): | ||
| 413 | with open(postinst_intercept_hook) as intercept: | ||
| 414 | registered_pkgs = None | ||
| 415 | for line in intercept.read().split("\n"): | ||
| 416 | m = re.match(r"^##PKGS:(.*)", line) | ||
| 417 | if m is not None: | ||
| 418 | registered_pkgs = m.group(1).strip() | ||
| 419 | break | ||
| 420 | |||
| 421 | if registered_pkgs is not None: | ||
| 422 | bb.note("If an image is being built, the postinstalls for the following packages " | ||
| 423 | "will be postponed for first boot: %s" % | ||
| 424 | registered_pkgs) | ||
| 425 | |||
| 426 | # call the backend dependent handler | ||
| 427 | self._handle_intercept_failure(registered_pkgs) | ||
| 428 | |||
| 429 | |||
| 430 | def run_intercepts(self, populate_sdk=None): | ||
| 431 | intercepts_dir = self.intercepts_dir | ||
| 432 | |||
| 433 | bb.note("Running intercept scripts:") | ||
| 434 | os.environ['D'] = self.target_rootfs | ||
| 435 | os.environ['STAGING_DIR_NATIVE'] = self.d.getVar('STAGING_DIR_NATIVE') | ||
| 436 | for script in os.listdir(intercepts_dir): | ||
| 437 | script_full = os.path.join(intercepts_dir, script) | ||
| 438 | |||
| 439 | if script == "postinst_intercept" or not os.access(script_full, os.X_OK): | ||
| 440 | continue | ||
| 441 | |||
| 442 | # we do not want to run any multilib variant of this | ||
| 443 | if script.startswith("delay_to_first_boot"): | ||
| 444 | self._postpone_to_first_boot(script_full) | ||
| 445 | continue | ||
| 446 | |||
| 447 | if populate_sdk == 'host' and self.d.getVar('SDK_OS') == 'mingw32': | ||
| 448 | bb.note("The postinstall intercept hook '%s' could not be executed due to missing wine support, details in %s/log.do_%s" | ||
| 449 | % (script, self.d.getVar('T'), self.d.getVar('BB_CURRENTTASK'))) | ||
| 450 | continue | ||
| 451 | |||
| 452 | bb.note("> Executing %s intercept ..." % script) | ||
| 453 | |||
| 454 | try: | ||
| 455 | output = subprocess.check_output(script_full, stderr=subprocess.STDOUT) | ||
| 456 | if output: bb.note(output.decode("utf-8")) | ||
| 457 | except subprocess.CalledProcessError as e: | ||
| 458 | bb.note("Exit code %d. Output:\n%s" % (e.returncode, e.output.decode("utf-8"))) | ||
| 459 | if populate_sdk == 'host': | ||
| 460 | bb.fatal("The postinstall intercept hook '%s' failed, details in %s/log.do_%s" % (script, self.d.getVar('T'), self.d.getVar('BB_CURRENTTASK'))) | ||
| 461 | elif populate_sdk == 'target': | ||
| 462 | if "qemuwrapper: qemu usermode is not supported" in e.output.decode("utf-8"): | ||
| 463 | bb.note("The postinstall intercept hook '%s' could not be executed due to missing qemu usermode support, details in %s/log.do_%s" | ||
| 464 | % (script, self.d.getVar('T'), self.d.getVar('BB_CURRENTTASK'))) | ||
| 465 | else: | ||
| 466 | bb.fatal("The postinstall intercept hook '%s' failed, details in %s/log.do_%s" % (script, self.d.getVar('T'), self.d.getVar('BB_CURRENTTASK'))) | ||
| 467 | else: | ||
| 468 | if "qemuwrapper: qemu usermode is not supported" in e.output.decode("utf-8"): | ||
| 469 | bb.note("The postinstall intercept hook '%s' could not be executed due to missing qemu usermode support, details in %s/log.do_%s" | ||
| 470 | % (script, self.d.getVar('T'), self.d.getVar('BB_CURRENTTASK'))) | ||
| 471 | self._postpone_to_first_boot(script_full) | ||
| 472 | else: | ||
| 473 | bb.fatal("The postinstall intercept hook '%s' failed, details in %s/log.do_%s" % (script, self.d.getVar('T'), self.d.getVar('BB_CURRENTTASK'))) | ||
| 474 | |||
| 475 | @abstractmethod | ||
| 476 | def update(self): | ||
| 477 | """ | ||
| 478 | Update the package manager package database. | ||
| 479 | """ | ||
| 480 | pass | ||
| 481 | |||
| 482 | @abstractmethod | ||
| 483 | def install(self, pkgs, attempt_only=False): | ||
| 484 | """ | ||
| 485 | Install a list of packages. 'pkgs' is a list object. If 'attempt_only' is | ||
| 486 | True, installation failures are ignored. | ||
| 487 | """ | ||
| 488 | pass | ||
| 489 | |||
| 490 | @abstractmethod | ||
| 491 | def remove(self, pkgs, with_dependencies=True): | ||
| 492 | """ | ||
| 493 | Remove a list of packages. 'pkgs' is a list object. If 'with_dependencies' | ||
| 494 | is False, then any dependencies are left in place. | ||
| 495 | """ | ||
| 496 | pass | ||
| 497 | |||
| 498 | @abstractmethod | ||
| 499 | def write_index(self): | ||
| 500 | """ | ||
| 501 | This function creates the index files | ||
| 502 | """ | ||
| 503 | pass | ||
| 504 | |||
| 505 | @abstractmethod | ||
| 506 | def remove_packaging_data(self): | ||
| 507 | pass | ||
| 508 | |||
| 509 | @abstractmethod | ||
| 510 | def list_installed(self): | ||
| 511 | pass | ||
| 512 | |||
| 513 | @abstractmethod | ||
| 514 | def extract(self, pkg): | ||
| 515 | """ | ||
| 516 | Returns the path to a tmpdir where resides the contents of a package. | ||
| 517 | Deleting the tmpdir is responsability of the caller. | ||
| 518 | """ | ||
| 519 | pass | ||
| 520 | |||
| 521 | @abstractmethod | ||
| 522 | def insert_feeds_uris(self, feed_uris, feed_base_paths, feed_archs): | ||
| 523 | """ | ||
| 524 | Add remote package feeds into repository manager configuration. The parameters | ||
| 525 | for the feeds are set by feed_uris, feed_base_paths and feed_archs. | ||
| 526 | See http://www.yoctoproject.org/docs/current/ref-manual/ref-manual.html#var-PACKAGE_FEED_URIS | ||
| 527 | for their description. | ||
| 528 | """ | ||
| 529 | pass | ||
| 530 | |||
| 531 | def install_glob(self, globs, sdk=False): | ||
| 532 | """ | ||
| 533 | Install all packages that match a glob. | ||
| 534 | """ | ||
| 535 | # TODO don't have sdk here but have a property on the superclass | ||
| 536 | # (and respect in install_complementary) | ||
| 537 | if sdk: | ||
| 538 | pkgdatadir = self.d.expand("${TMPDIR}/pkgdata/${SDK_SYS}") | ||
| 539 | else: | ||
| 540 | pkgdatadir = self.d.getVar("PKGDATA_DIR") | ||
| 541 | |||
| 542 | try: | ||
| 543 | bb.note("Installing globbed packages...") | ||
| 544 | cmd = ["oe-pkgdata-util", "-p", pkgdatadir, "list-pkgs", globs] | ||
| 545 | pkgs = subprocess.check_output(cmd, stderr=subprocess.STDOUT).decode("utf-8") | ||
| 546 | self.install(pkgs.split(), attempt_only=True) | ||
| 547 | except subprocess.CalledProcessError as e: | ||
| 548 | # Return code 1 means no packages matched | ||
| 549 | if e.returncode != 1: | ||
| 550 | bb.fatal("Could not compute globbed packages list. Command " | ||
| 551 | "'%s' returned %d:\n%s" % | ||
| 552 | (' '.join(cmd), e.returncode, e.output.decode("utf-8"))) | ||
| 553 | |||
| 554 | def install_complementary(self, globs=None): | ||
| 555 | """ | ||
| 556 | Install complementary packages based upon the list of currently installed | ||
| 557 | packages e.g. locales, *-dev, *-dbg, etc. This will only attempt to install | ||
| 558 | these packages, if they don't exist then no error will occur. Note: every | ||
| 559 | backend needs to call this function explicitly after the normal package | ||
| 560 | installation | ||
| 561 | """ | ||
| 562 | if globs is None: | ||
| 563 | globs = self.d.getVar('IMAGE_INSTALL_COMPLEMENTARY') | ||
| 564 | split_linguas = set() | ||
| 565 | |||
| 566 | for translation in self.d.getVar('IMAGE_LINGUAS').split(): | ||
| 567 | split_linguas.add(translation) | ||
| 568 | split_linguas.add(translation.split('-')[0]) | ||
| 569 | |||
| 570 | split_linguas = sorted(split_linguas) | ||
| 571 | |||
| 572 | for lang in split_linguas: | ||
| 573 | globs += " *-locale-%s" % lang | ||
| 574 | for complementary_linguas in (self.d.getVar('IMAGE_LINGUAS_COMPLEMENTARY') or "").split(): | ||
| 575 | globs += (" " + complementary_linguas) % lang | ||
| 576 | |||
| 577 | if globs is None: | ||
| 578 | return | ||
| 579 | |||
| 580 | # we need to write the list of installed packages to a file because the | ||
| 581 | # oe-pkgdata-util reads it from a file | ||
| 582 | with tempfile.NamedTemporaryFile(mode="w+", prefix="installed-pkgs") as installed_pkgs: | ||
| 583 | pkgs = self.list_installed() | ||
| 584 | |||
| 585 | provided_pkgs = set() | ||
| 586 | for pkg in pkgs.values(): | ||
| 587 | provided_pkgs |= set(pkg.get('provs', [])) | ||
| 588 | |||
| 589 | output = oe.utils.format_pkg_list(pkgs, "arch") | ||
| 590 | installed_pkgs.write(output) | ||
| 591 | installed_pkgs.flush() | ||
| 592 | |||
| 593 | cmd = ["oe-pkgdata-util", | ||
| 594 | "-p", self.d.getVar('PKGDATA_DIR'), "glob", installed_pkgs.name, | ||
| 595 | globs] | ||
| 596 | exclude = self.d.getVar('PACKAGE_EXCLUDE_COMPLEMENTARY') | ||
| 597 | if exclude: | ||
| 598 | cmd.extend(['--exclude=' + '|'.join(exclude.split())]) | ||
| 599 | try: | ||
| 600 | bb.note('Running %s' % cmd) | ||
| 601 | complementary_pkgs = subprocess.check_output(cmd, stderr=subprocess.STDOUT).decode("utf-8") | ||
| 602 | complementary_pkgs = set(complementary_pkgs.split()) | ||
| 603 | skip_pkgs = sorted(complementary_pkgs & provided_pkgs) | ||
| 604 | install_pkgs = sorted(complementary_pkgs - provided_pkgs) | ||
| 605 | bb.note("Installing complementary packages ... %s (skipped already provided packages %s)" % ( | ||
| 606 | ' '.join(install_pkgs), | ||
| 607 | ' '.join(skip_pkgs))) | ||
| 608 | self.install(install_pkgs, attempt_only=True) | ||
| 609 | except subprocess.CalledProcessError as e: | ||
| 610 | bb.fatal("Could not compute complementary packages list. Command " | ||
| 611 | "'%s' returned %d:\n%s" % | ||
| 612 | (' '.join(cmd), e.returncode, e.output.decode("utf-8"))) | ||
| 613 | |||
| 614 | target_arch = self.d.getVar('TARGET_ARCH') | ||
| 615 | localedir = oe.path.join(self.target_rootfs, self.d.getVar("libdir"), "locale") | ||
| 616 | if os.path.exists(localedir) and os.listdir(localedir): | ||
| 617 | generate_locale_archive(self.d, self.target_rootfs, target_arch, localedir) | ||
| 618 | # And now delete the binary locales | ||
| 619 | self.remove(fnmatch.filter(self.list_installed(), "glibc-binary-localedata-*"), False) | ||
| 620 | |||
| 621 | def deploy_dir_lock(self): | ||
| 622 | if self.deploy_dir is None: | ||
| 623 | raise RuntimeError("deploy_dir is not set!") | ||
| 624 | |||
| 625 | lock_file_name = os.path.join(self.deploy_dir, "deploy.lock") | ||
| 626 | |||
| 627 | self.deploy_lock = bb.utils.lockfile(lock_file_name) | ||
| 628 | |||
| 629 | def deploy_dir_unlock(self): | ||
| 630 | if self.deploy_lock is None: | ||
| 631 | return | ||
| 632 | |||
| 633 | bb.utils.unlockfile(self.deploy_lock) | ||
| 634 | |||
| 635 | self.deploy_lock = None | ||
| 636 | |||
| 637 | def construct_uris(self, uris, base_paths): | ||
| 638 | """ | ||
| 639 | Construct URIs based on the following pattern: uri/base_path where 'uri' | ||
| 640 | and 'base_path' correspond to each element of the corresponding array | ||
| 641 | argument leading to len(uris) x len(base_paths) elements on the returned | ||
| 642 | array | ||
| 643 | """ | ||
| 644 | def _append(arr1, arr2, sep='/'): | ||
| 645 | res = [] | ||
| 646 | narr1 = [a.rstrip(sep) for a in arr1] | ||
| 647 | narr2 = [a.rstrip(sep).lstrip(sep) for a in arr2] | ||
| 648 | for a1 in narr1: | ||
| 649 | if arr2: | ||
| 650 | for a2 in narr2: | ||
| 651 | res.append("%s%s%s" % (a1, sep, a2)) | ||
| 652 | else: | ||
| 653 | res.append(a1) | ||
| 654 | return res | ||
| 655 | return _append(uris, base_paths) | ||
| 656 | |||
| 657 | def create_packages_dir(d, subrepo_dir, deploydir, taskname, filterbydependencies): | ||
| 658 | """ | ||
| 659 | Go through our do_package_write_X dependencies and hardlink the packages we depend | ||
| 660 | upon into the repo directory. This prevents us seeing other packages that may | ||
| 661 | have been built that we don't depend upon and also packages for architectures we don't | ||
| 662 | support. | ||
| 663 | """ | ||
| 664 | import errno | ||
| 665 | |||
| 666 | taskdepdata = d.getVar("BB_TASKDEPDATA", False) | ||
| 667 | mytaskname = d.getVar("BB_RUNTASK") | ||
| 668 | pn = d.getVar("PN") | ||
| 669 | seendirs = set() | ||
| 670 | multilibs = {} | ||
| 671 | |||
| 672 | bb.utils.remove(subrepo_dir, recurse=True) | ||
| 673 | bb.utils.mkdirhier(subrepo_dir) | ||
| 674 | |||
| 675 | # Detect bitbake -b usage | ||
| 676 | nodeps = d.getVar("BB_LIMITEDDEPS") or False | ||
| 677 | if nodeps or not filterbydependencies: | ||
| 678 | oe.path.symlink(deploydir, subrepo_dir, True) | ||
| 679 | return | ||
| 680 | |||
| 681 | start = None | ||
| 682 | for dep in taskdepdata: | ||
| 683 | data = taskdepdata[dep] | ||
| 684 | if data[1] == mytaskname and data[0] == pn: | ||
| 685 | start = dep | ||
| 686 | break | ||
| 687 | if start is None: | ||
| 688 | bb.fatal("Couldn't find ourself in BB_TASKDEPDATA?") | ||
| 689 | pkgdeps = set() | ||
| 690 | start = [start] | ||
| 691 | seen = set(start) | ||
| 692 | # Support direct dependencies (do_rootfs -> do_package_write_X) | ||
| 693 | # or indirect dependencies within PN (do_populate_sdk_ext -> do_rootfs -> do_package_write_X) | ||
| 694 | while start: | ||
| 695 | next = [] | ||
| 696 | for dep2 in start: | ||
| 697 | for dep in taskdepdata[dep2][3]: | ||
| 698 | if taskdepdata[dep][0] != pn: | ||
| 699 | if "do_" + taskname in dep: | ||
| 700 | pkgdeps.add(dep) | ||
| 701 | elif dep not in seen: | ||
| 702 | next.append(dep) | ||
| 703 | seen.add(dep) | ||
| 704 | start = next | ||
| 705 | |||
| 706 | for dep in pkgdeps: | ||
| 707 | c = taskdepdata[dep][0] | ||
| 708 | manifest, d2 = oe.sstatesig.find_sstate_manifest(c, taskdepdata[dep][2], taskname, d, multilibs) | ||
| 709 | if not manifest: | ||
| 710 | bb.fatal("No manifest generated from: %s in %s" % (c, taskdepdata[dep][2])) | ||
| 711 | if not os.path.exists(manifest): | ||
| 712 | continue | ||
| 713 | with open(manifest, "r") as f: | ||
| 714 | for l in f: | ||
| 715 | l = l.strip() | ||
| 716 | deploydir = os.path.normpath(deploydir) | ||
| 717 | if bb.data.inherits_class('packagefeed-stability', d): | ||
| 718 | dest = l.replace(deploydir + "-prediff", "") | ||
| 719 | else: | ||
| 720 | dest = l.replace(deploydir, "") | ||
| 721 | dest = subrepo_dir + dest | ||
| 722 | if l.endswith("/"): | ||
| 723 | if dest not in seendirs: | ||
| 724 | bb.utils.mkdirhier(dest) | ||
| 725 | seendirs.add(dest) | ||
| 726 | continue | ||
| 727 | # Try to hardlink the file, copy if that fails | ||
| 728 | destdir = os.path.dirname(dest) | ||
| 729 | if destdir not in seendirs: | ||
| 730 | bb.utils.mkdirhier(destdir) | ||
| 731 | seendirs.add(destdir) | ||
| 732 | try: | ||
| 733 | os.link(l, dest) | ||
| 734 | except OSError as err: | ||
| 735 | if err.errno == errno.EXDEV: | ||
| 736 | bb.utils.copyfile(l, dest) | ||
| 737 | else: | ||
| 738 | raise | ||
| 739 | |||
| 740 | class RpmPM(PackageManager): | ||
| 741 | def __init__(self, | ||
| 742 | d, | ||
| 743 | target_rootfs, | ||
| 744 | target_vendor, | ||
| 745 | task_name='target', | ||
| 746 | arch_var=None, | ||
| 747 | os_var=None, | ||
| 748 | rpm_repo_workdir="oe-rootfs-repo", | ||
| 749 | filterbydependencies=True, | ||
| 750 | needfeed=True): | ||
| 751 | super(RpmPM, self).__init__(d, target_rootfs) | ||
| 752 | self.target_vendor = target_vendor | ||
| 753 | self.task_name = task_name | ||
| 754 | if arch_var == None: | ||
| 755 | self.archs = self.d.getVar('ALL_MULTILIB_PACKAGE_ARCHS').replace("-","_") | ||
| 756 | else: | ||
| 757 | self.archs = self.d.getVar(arch_var).replace("-","_") | ||
| 758 | if task_name == "host": | ||
| 759 | self.primary_arch = self.d.getVar('SDK_ARCH') | ||
| 760 | else: | ||
| 761 | self.primary_arch = self.d.getVar('MACHINE_ARCH') | ||
| 762 | |||
| 763 | if needfeed: | ||
| 764 | self.rpm_repo_dir = oe.path.join(self.d.getVar('WORKDIR'), rpm_repo_workdir) | ||
| 765 | create_packages_dir(self.d, oe.path.join(self.rpm_repo_dir, "rpm"), d.getVar("DEPLOY_DIR_RPM"), "package_write_rpm", filterbydependencies) | ||
| 766 | |||
| 767 | self.saved_packaging_data = self.d.expand('${T}/saved_packaging_data/%s' % self.task_name) | ||
| 768 | if not os.path.exists(self.d.expand('${T}/saved_packaging_data')): | ||
| 769 | bb.utils.mkdirhier(self.d.expand('${T}/saved_packaging_data')) | ||
| 770 | self.packaging_data_dirs = ['etc/rpm', 'etc/rpmrc', 'etc/dnf', 'var/lib/rpm', 'var/lib/dnf', 'var/cache/dnf'] | ||
| 771 | self.solution_manifest = self.d.expand('${T}/saved/%s_solution' % | ||
| 772 | self.task_name) | ||
| 773 | if not os.path.exists(self.d.expand('${T}/saved')): | ||
| 774 | bb.utils.mkdirhier(self.d.expand('${T}/saved')) | ||
| 775 | |||
| 776 | def _configure_dnf(self): | ||
| 777 | # libsolv handles 'noarch' internally, we don't need to specify it explicitly | ||
| 778 | archs = [i for i in reversed(self.archs.split()) if i not in ["any", "all", "noarch"]] | ||
| 779 | # This prevents accidental matching against libsolv's built-in policies | ||
| 780 | if len(archs) <= 1: | ||
| 781 | archs = archs + ["bogusarch"] | ||
| 782 | # This architecture needs to be upfront so that packages using it are properly prioritized | ||
| 783 | archs = ["sdk_provides_dummy_target"] + archs | ||
| 784 | confdir = "%s/%s" %(self.target_rootfs, "etc/dnf/vars/") | ||
| 785 | bb.utils.mkdirhier(confdir) | ||
| 786 | open(confdir + "arch", 'w').write(":".join(archs)) | ||
| 787 | distro_codename = self.d.getVar('DISTRO_CODENAME') | ||
| 788 | open(confdir + "releasever", 'w').write(distro_codename if distro_codename is not None else '') | ||
| 789 | |||
| 790 | open(oe.path.join(self.target_rootfs, "etc/dnf/dnf.conf"), 'w').write("") | ||
| 791 | |||
| 792 | |||
| 793 | def _configure_rpm(self): | ||
| 794 | # We need to configure rpm to use our primary package architecture as the installation architecture, | ||
| 795 | # and to make it compatible with other package architectures that we use. | ||
| 796 | # Otherwise it will refuse to proceed with packages installation. | ||
| 797 | platformconfdir = "%s/%s" %(self.target_rootfs, "etc/rpm/") | ||
| 798 | rpmrcconfdir = "%s/%s" %(self.target_rootfs, "etc/") | ||
| 799 | bb.utils.mkdirhier(platformconfdir) | ||
| 800 | open(platformconfdir + "platform", 'w').write("%s-pc-linux" % self.primary_arch) | ||
| 801 | with open(rpmrcconfdir + "rpmrc", 'w') as f: | ||
| 802 | f.write("arch_compat: %s: %s\n" % (self.primary_arch, self.archs if len(self.archs) > 0 else self.primary_arch)) | ||
| 803 | f.write("buildarch_compat: %s: noarch\n" % self.primary_arch) | ||
| 804 | |||
| 805 | open(platformconfdir + "macros", 'w').write("%_transaction_color 7\n") | ||
| 806 | if self.d.getVar('RPM_PREFER_ELF_ARCH'): | ||
| 807 | open(platformconfdir + "macros", 'a').write("%%_prefer_color %s" % (self.d.getVar('RPM_PREFER_ELF_ARCH'))) | ||
| 808 | |||
| 809 | if self.d.getVar('RPM_SIGN_PACKAGES') == '1': | ||
| 810 | signer = get_signer(self.d, self.d.getVar('RPM_GPG_BACKEND')) | ||
| 811 | pubkey_path = oe.path.join(self.d.getVar('B'), 'rpm-key') | ||
| 812 | signer.export_pubkey(pubkey_path, self.d.getVar('RPM_GPG_NAME')) | ||
| 813 | rpm_bin = bb.utils.which(os.getenv('PATH'), "rpmkeys") | ||
| 814 | cmd = [rpm_bin, '--root=%s' % self.target_rootfs, '--import', pubkey_path] | ||
| 815 | try: | ||
| 816 | subprocess.check_output(cmd, stderr=subprocess.STDOUT) | ||
| 817 | except subprocess.CalledProcessError as e: | ||
| 818 | bb.fatal("Importing GPG key failed. Command '%s' " | ||
| 819 | "returned %d:\n%s" % (' '.join(cmd), e.returncode, e.output.decode("utf-8"))) | ||
| 820 | |||
| 821 | def create_configs(self): | ||
| 822 | self._configure_dnf() | ||
| 823 | self._configure_rpm() | ||
| 824 | |||
| 825 | def write_index(self): | ||
| 826 | lockfilename = self.d.getVar('DEPLOY_DIR_RPM') + "/rpm.lock" | ||
| 827 | lf = bb.utils.lockfile(lockfilename, False) | ||
| 828 | RpmIndexer(self.d, self.rpm_repo_dir).write_index() | ||
| 829 | bb.utils.unlockfile(lf) | ||
| 830 | |||
| 831 | def insert_feeds_uris(self, feed_uris, feed_base_paths, feed_archs): | ||
| 832 | from urllib.parse import urlparse | ||
| 833 | |||
| 834 | if feed_uris == "": | ||
| 835 | return | ||
| 836 | |||
| 837 | gpg_opts = '' | ||
| 838 | if self.d.getVar('PACKAGE_FEED_SIGN') == '1': | ||
| 839 | gpg_opts += 'repo_gpgcheck=1\n' | ||
| 840 | gpg_opts += 'gpgkey=file://%s/pki/packagefeed-gpg/PACKAGEFEED-GPG-KEY-%s-%s\n' % (self.d.getVar('sysconfdir'), self.d.getVar('DISTRO'), self.d.getVar('DISTRO_CODENAME')) | ||
| 841 | |||
| 842 | if self.d.getVar('RPM_SIGN_PACKAGES') != '1': | ||
| 843 | gpg_opts += 'gpgcheck=0\n' | ||
| 844 | |||
| 845 | bb.utils.mkdirhier(oe.path.join(self.target_rootfs, "etc", "yum.repos.d")) | ||
| 846 | remote_uris = self.construct_uris(feed_uris.split(), feed_base_paths.split()) | ||
| 847 | for uri in remote_uris: | ||
| 848 | repo_base = "oe-remote-repo" + "-".join(urlparse(uri).path.split("/")) | ||
| 849 | if feed_archs is not None: | ||
| 850 | for arch in feed_archs.split(): | ||
| 851 | repo_uri = uri + "/" + arch | ||
| 852 | repo_id = "oe-remote-repo" + "-".join(urlparse(repo_uri).path.split("/")) | ||
| 853 | repo_name = "OE Remote Repo:" + " ".join(urlparse(repo_uri).path.split("/")) | ||
| 854 | open(oe.path.join(self.target_rootfs, "etc", "yum.repos.d", repo_base + ".repo"), 'a').write( | ||
| 855 | "[%s]\nname=%s\nbaseurl=%s\n%s\n" % (repo_id, repo_name, repo_uri, gpg_opts)) | ||
| 856 | else: | ||
| 857 | repo_name = "OE Remote Repo:" + " ".join(urlparse(uri).path.split("/")) | ||
| 858 | repo_uri = uri | ||
| 859 | open(oe.path.join(self.target_rootfs, "etc", "yum.repos.d", repo_base + ".repo"), 'w').write( | ||
| 860 | "[%s]\nname=%s\nbaseurl=%s\n%s" % (repo_base, repo_name, repo_uri, gpg_opts)) | ||
| 861 | |||
| 862 | def _prepare_pkg_transaction(self): | ||
| 863 | os.environ['D'] = self.target_rootfs | ||
| 864 | os.environ['OFFLINE_ROOT'] = self.target_rootfs | ||
| 865 | os.environ['IPKG_OFFLINE_ROOT'] = self.target_rootfs | ||
| 866 | os.environ['OPKG_OFFLINE_ROOT'] = self.target_rootfs | ||
| 867 | os.environ['INTERCEPT_DIR'] = self.intercepts_dir | ||
| 868 | os.environ['NATIVE_ROOT'] = self.d.getVar('STAGING_DIR_NATIVE') | ||
| 869 | |||
| 870 | |||
| 871 | def install(self, pkgs, attempt_only = False): | ||
| 872 | if len(pkgs) == 0: | ||
| 873 | return | ||
| 874 | self._prepare_pkg_transaction() | ||
| 875 | |||
| 876 | bad_recommendations = self.d.getVar('BAD_RECOMMENDATIONS') | ||
| 877 | package_exclude = self.d.getVar('PACKAGE_EXCLUDE') | ||
| 878 | exclude_pkgs = (bad_recommendations.split() if bad_recommendations else []) + (package_exclude.split() if package_exclude else []) | ||
| 879 | |||
| 880 | output = self._invoke_dnf((["--skip-broken"] if attempt_only else []) + | ||
| 881 | (["-x", ",".join(exclude_pkgs)] if len(exclude_pkgs) > 0 else []) + | ||
| 882 | (["--setopt=install_weak_deps=False"] if self.d.getVar('NO_RECOMMENDATIONS') == "1" else []) + | ||
| 883 | (["--nogpgcheck"] if self.d.getVar('RPM_SIGN_PACKAGES') != '1' else ["--setopt=gpgcheck=True"]) + | ||
| 884 | ["install"] + | ||
| 885 | pkgs) | ||
| 886 | |||
| 887 | failed_scriptlets_pkgnames = collections.OrderedDict() | ||
| 888 | for line in output.splitlines(): | ||
| 889 | if line.startswith("Error in POSTIN scriptlet in rpm package"): | ||
| 890 | failed_scriptlets_pkgnames[line.split()[-1]] = True | ||
| 891 | |||
| 892 | if len(failed_scriptlets_pkgnames) > 0: | ||
| 893 | failed_postinsts_abort(list(failed_scriptlets_pkgnames.keys()), self.d.expand("${T}/log.do_${BB_CURRENTTASK}")) | ||
| 894 | |||
| 895 | def remove(self, pkgs, with_dependencies = True): | ||
| 896 | if not pkgs: | ||
| 897 | return | ||
| 898 | |||
| 899 | self._prepare_pkg_transaction() | ||
| 900 | |||
| 901 | if with_dependencies: | ||
| 902 | self._invoke_dnf(["remove"] + pkgs) | ||
| 903 | else: | ||
| 904 | cmd = bb.utils.which(os.getenv('PATH'), "rpm") | ||
| 905 | args = ["-e", "-v", "--nodeps", "--root=%s" %self.target_rootfs] | ||
| 906 | |||
| 907 | try: | ||
| 908 | bb.note("Running %s" % ' '.join([cmd] + args + pkgs)) | ||
| 909 | output = subprocess.check_output([cmd] + args + pkgs, stderr=subprocess.STDOUT).decode("utf-8") | ||
| 910 | bb.note(output) | ||
| 911 | except subprocess.CalledProcessError as e: | ||
| 912 | bb.fatal("Could not invoke rpm. Command " | ||
| 913 | "'%s' returned %d:\n%s" % (' '.join([cmd] + args + pkgs), e.returncode, e.output.decode("utf-8"))) | ||
| 914 | |||
| 915 | def upgrade(self): | ||
| 916 | self._prepare_pkg_transaction() | ||
| 917 | self._invoke_dnf(["upgrade"]) | ||
| 918 | |||
| 919 | def autoremove(self): | ||
| 920 | self._prepare_pkg_transaction() | ||
| 921 | self._invoke_dnf(["autoremove"]) | ||
| 922 | |||
| 923 | def remove_packaging_data(self): | ||
| 924 | self._invoke_dnf(["clean", "all"]) | ||
| 925 | for dir in self.packaging_data_dirs: | ||
| 926 | bb.utils.remove(oe.path.join(self.target_rootfs, dir), True) | ||
| 927 | |||
| 928 | def backup_packaging_data(self): | ||
| 929 | # Save the packaging dirs for increment rpm image generation | ||
| 930 | if os.path.exists(self.saved_packaging_data): | ||
| 931 | bb.utils.remove(self.saved_packaging_data, True) | ||
| 932 | for i in self.packaging_data_dirs: | ||
| 933 | source_dir = oe.path.join(self.target_rootfs, i) | ||
| 934 | target_dir = oe.path.join(self.saved_packaging_data, i) | ||
| 935 | if os.path.isdir(source_dir): | ||
| 936 | shutil.copytree(source_dir, target_dir, symlinks=True) | ||
| 937 | elif os.path.isfile(source_dir): | ||
| 938 | shutil.copy2(source_dir, target_dir) | ||
| 939 | |||
| 940 | def recovery_packaging_data(self): | ||
| 941 | # Move the rpmlib back | ||
| 942 | if os.path.exists(self.saved_packaging_data): | ||
| 943 | for i in self.packaging_data_dirs: | ||
| 944 | target_dir = oe.path.join(self.target_rootfs, i) | ||
| 945 | if os.path.exists(target_dir): | ||
| 946 | bb.utils.remove(target_dir, True) | ||
| 947 | source_dir = oe.path.join(self.saved_packaging_data, i) | ||
| 948 | if os.path.isdir(source_dir): | ||
| 949 | shutil.copytree(source_dir, target_dir, symlinks=True) | ||
| 950 | elif os.path.isfile(source_dir): | ||
| 951 | shutil.copy2(source_dir, target_dir) | ||
| 952 | |||
| 953 | def list_installed(self): | ||
| 954 | output = self._invoke_dnf(["repoquery", "--installed", "--queryformat", "Package: %{name} %{arch} %{version} %{name}-%{version}-%{release}.%{arch}.rpm\nDependencies:\n%{requires}\nRecommendations:\n%{recommends}\nDependenciesEndHere:\n"], | ||
| 955 | print_output = False) | ||
| 956 | packages = {} | ||
| 957 | current_package = None | ||
| 958 | current_deps = None | ||
| 959 | current_state = "initial" | ||
| 960 | for line in output.splitlines(): | ||
| 961 | if line.startswith("Package:"): | ||
| 962 | package_info = line.split(" ")[1:] | ||
| 963 | current_package = package_info[0] | ||
| 964 | package_arch = package_info[1] | ||
| 965 | package_version = package_info[2] | ||
| 966 | package_rpm = package_info[3] | ||
| 967 | packages[current_package] = {"arch":package_arch, "ver":package_version, "filename":package_rpm} | ||
| 968 | current_deps = [] | ||
| 969 | elif line.startswith("Dependencies:"): | ||
| 970 | current_state = "dependencies" | ||
| 971 | elif line.startswith("Recommendations"): | ||
| 972 | current_state = "recommendations" | ||
| 973 | elif line.startswith("DependenciesEndHere:"): | ||
| 974 | current_state = "initial" | ||
| 975 | packages[current_package]["deps"] = current_deps | ||
| 976 | elif len(line) > 0: | ||
| 977 | if current_state == "dependencies": | ||
| 978 | current_deps.append(line) | ||
| 979 | elif current_state == "recommendations": | ||
| 980 | current_deps.append("%s [REC]" % line) | ||
| 981 | |||
| 982 | return packages | ||
| 983 | |||
| 984 | def update(self): | ||
| 985 | self._invoke_dnf(["makecache", "--refresh"]) | ||
| 986 | |||
| 987 | def _invoke_dnf(self, dnf_args, fatal = True, print_output = True ): | ||
| 988 | os.environ['RPM_ETCCONFIGDIR'] = self.target_rootfs | ||
| 989 | |||
| 990 | dnf_cmd = bb.utils.which(os.getenv('PATH'), "dnf") | ||
| 991 | standard_dnf_args = ["-v", "--rpmverbosity=info", "-y", | ||
| 992 | "-c", oe.path.join(self.target_rootfs, "etc/dnf/dnf.conf"), | ||
| 993 | "--setopt=reposdir=%s" %(oe.path.join(self.target_rootfs, "etc/yum.repos.d")), | ||
| 994 | "--installroot=%s" % (self.target_rootfs), | ||
| 995 | "--setopt=logdir=%s" % (self.d.getVar('T')) | ||
| 996 | ] | ||
| 997 | if hasattr(self, "rpm_repo_dir"): | ||
| 998 | standard_dnf_args.append("--repofrompath=oe-repo,%s" % (self.rpm_repo_dir)) | ||
| 999 | cmd = [dnf_cmd] + standard_dnf_args + dnf_args | ||
| 1000 | bb.note('Running %s' % ' '.join(cmd)) | ||
| 1001 | try: | ||
| 1002 | output = subprocess.check_output(cmd,stderr=subprocess.STDOUT).decode("utf-8") | ||
| 1003 | if print_output: | ||
| 1004 | bb.debug(1, output) | ||
| 1005 | return output | ||
| 1006 | except subprocess.CalledProcessError as e: | ||
| 1007 | if print_output: | ||
| 1008 | (bb.note, bb.fatal)[fatal]("Could not invoke dnf. Command " | ||
| 1009 | "'%s' returned %d:\n%s" % (' '.join(cmd), e.returncode, e.output.decode("utf-8"))) | ||
| 1010 | else: | ||
| 1011 | (bb.note, bb.fatal)[fatal]("Could not invoke dnf. Command " | ||
| 1012 | "'%s' returned %d:" % (' '.join(cmd), e.returncode)) | ||
| 1013 | return e.output.decode("utf-8") | ||
| 1014 | |||
| 1015 | def dump_install_solution(self, pkgs): | ||
| 1016 | open(self.solution_manifest, 'w').write(" ".join(pkgs)) | ||
| 1017 | return pkgs | ||
| 1018 | |||
| 1019 | def load_old_install_solution(self): | ||
| 1020 | if not os.path.exists(self.solution_manifest): | ||
| 1021 | return [] | ||
| 1022 | with open(self.solution_manifest, 'r') as fd: | ||
| 1023 | return fd.read().split() | ||
| 1024 | |||
| 1025 | def _script_num_prefix(self, path): | ||
| 1026 | files = os.listdir(path) | ||
| 1027 | numbers = set() | ||
| 1028 | numbers.add(99) | ||
| 1029 | for f in files: | ||
| 1030 | numbers.add(int(f.split("-")[0])) | ||
| 1031 | return max(numbers) + 1 | ||
| 1032 | |||
| 1033 | def save_rpmpostinst(self, pkg): | ||
| 1034 | bb.note("Saving postinstall script of %s" % (pkg)) | ||
| 1035 | cmd = bb.utils.which(os.getenv('PATH'), "rpm") | ||
| 1036 | args = ["-q", "--root=%s" % self.target_rootfs, "--queryformat", "%{postin}", pkg] | ||
| 1037 | |||
| 1038 | try: | ||
| 1039 | output = subprocess.check_output([cmd] + args,stderr=subprocess.STDOUT).decode("utf-8") | ||
| 1040 | except subprocess.CalledProcessError as e: | ||
| 1041 | bb.fatal("Could not invoke rpm. Command " | ||
| 1042 | "'%s' returned %d:\n%s" % (' '.join([cmd] + args), e.returncode, e.output.decode("utf-8"))) | ||
| 1043 | |||
| 1044 | # may need to prepend #!/bin/sh to output | ||
| 1045 | |||
| 1046 | target_path = oe.path.join(self.target_rootfs, self.d.expand('${sysconfdir}/rpm-postinsts/')) | ||
| 1047 | bb.utils.mkdirhier(target_path) | ||
| 1048 | num = self._script_num_prefix(target_path) | ||
| 1049 | saved_script_name = oe.path.join(target_path, "%d-%s" % (num, pkg)) | ||
| 1050 | open(saved_script_name, 'w').write(output) | ||
| 1051 | os.chmod(saved_script_name, 0o755) | ||
| 1052 | |||
| 1053 | def _handle_intercept_failure(self, registered_pkgs): | ||
| 1054 | rpm_postinsts_dir = self.target_rootfs + self.d.expand('${sysconfdir}/rpm-postinsts/') | ||
| 1055 | bb.utils.mkdirhier(rpm_postinsts_dir) | ||
| 1056 | |||
| 1057 | # Save the package postinstalls in /etc/rpm-postinsts | ||
| 1058 | for pkg in registered_pkgs.split(): | ||
| 1059 | self.save_rpmpostinst(pkg) | ||
| 1060 | |||
| 1061 | def extract(self, pkg): | ||
| 1062 | output = self._invoke_dnf(["repoquery", "--queryformat", "%{location}", pkg]) | ||
| 1063 | pkg_name = output.splitlines()[-1] | ||
| 1064 | if not pkg_name.endswith(".rpm"): | ||
| 1065 | bb.fatal("dnf could not find package %s in repository: %s" %(pkg, output)) | ||
| 1066 | pkg_path = oe.path.join(self.rpm_repo_dir, pkg_name) | ||
| 1067 | |||
| 1068 | cpio_cmd = bb.utils.which(os.getenv("PATH"), "cpio") | ||
| 1069 | rpm2cpio_cmd = bb.utils.which(os.getenv("PATH"), "rpm2cpio") | ||
| 1070 | |||
| 1071 | if not os.path.isfile(pkg_path): | ||
| 1072 | bb.fatal("Unable to extract package for '%s'." | ||
| 1073 | "File %s doesn't exists" % (pkg, pkg_path)) | ||
| 1074 | |||
| 1075 | tmp_dir = tempfile.mkdtemp() | ||
| 1076 | current_dir = os.getcwd() | ||
| 1077 | os.chdir(tmp_dir) | ||
| 1078 | |||
| 1079 | try: | ||
| 1080 | cmd = "%s %s | %s -idmv" % (rpm2cpio_cmd, pkg_path, cpio_cmd) | ||
| 1081 | output = subprocess.check_output(cmd, stderr=subprocess.STDOUT, shell=True) | ||
| 1082 | except subprocess.CalledProcessError as e: | ||
| 1083 | bb.utils.remove(tmp_dir, recurse=True) | ||
| 1084 | bb.fatal("Unable to extract %s package. Command '%s' " | ||
| 1085 | "returned %d:\n%s" % (pkg_path, cmd, e.returncode, e.output.decode("utf-8"))) | ||
| 1086 | except OSError as e: | ||
| 1087 | bb.utils.remove(tmp_dir, recurse=True) | ||
| 1088 | bb.fatal("Unable to extract %s package. Command '%s' " | ||
| 1089 | "returned %d:\n%s at %s" % (pkg_path, cmd, e.errno, e.strerror, e.filename)) | ||
| 1090 | |||
| 1091 | bb.note("Extracted %s to %s" % (pkg_path, tmp_dir)) | ||
| 1092 | os.chdir(current_dir) | ||
| 1093 | |||
| 1094 | return tmp_dir | ||
| 1095 | |||
| 1096 | |||
| 1097 | class OpkgDpkgPM(PackageManager): | ||
| 1098 | def __init__(self, d, target_rootfs): | ||
| 1099 | """ | ||
| 1100 | This is an abstract class. Do not instantiate this directly. | ||
| 1101 | """ | ||
| 1102 | super(OpkgDpkgPM, self).__init__(d, target_rootfs) | ||
| 1103 | |||
| 1104 | def package_info(self, pkg, cmd): | ||
| 1105 | """ | ||
| 1106 | Returns a dictionary with the package info. | ||
| 1107 | |||
| 1108 | This method extracts the common parts for Opkg and Dpkg | ||
| 1109 | """ | ||
| 1110 | |||
| 1111 | try: | ||
| 1112 | output = subprocess.check_output(cmd, stderr=subprocess.STDOUT, shell=True).decode("utf-8") | ||
| 1113 | except subprocess.CalledProcessError as e: | ||
| 1114 | bb.fatal("Unable to list available packages. Command '%s' " | ||
| 1115 | "returned %d:\n%s" % (cmd, e.returncode, e.output.decode("utf-8"))) | ||
| 1116 | return opkg_query(output) | ||
| 1117 | |||
| 1118 | def extract(self, pkg, pkg_info): | ||
| 1119 | """ | ||
| 1120 | Returns the path to a tmpdir where resides the contents of a package. | ||
| 1121 | |||
| 1122 | Deleting the tmpdir is responsability of the caller. | ||
| 1123 | |||
| 1124 | This method extracts the common parts for Opkg and Dpkg | ||
| 1125 | """ | ||
| 1126 | |||
| 1127 | ar_cmd = bb.utils.which(os.getenv("PATH"), "ar") | ||
| 1128 | tar_cmd = bb.utils.which(os.getenv("PATH"), "tar") | ||
| 1129 | pkg_path = pkg_info[pkg]["filepath"] | ||
| 1130 | |||
| 1131 | if not os.path.isfile(pkg_path): | ||
| 1132 | bb.fatal("Unable to extract package for '%s'." | ||
| 1133 | "File %s doesn't exists" % (pkg, pkg_path)) | ||
| 1134 | |||
| 1135 | tmp_dir = tempfile.mkdtemp() | ||
| 1136 | current_dir = os.getcwd() | ||
| 1137 | os.chdir(tmp_dir) | ||
| 1138 | data_tar = 'data.tar.xz' | ||
| 1139 | |||
| 1140 | try: | ||
| 1141 | cmd = [ar_cmd, 'x', pkg_path] | ||
| 1142 | output = subprocess.check_output(cmd, stderr=subprocess.STDOUT) | ||
| 1143 | cmd = [tar_cmd, 'xf', data_tar] | ||
| 1144 | output = subprocess.check_output(cmd, stderr=subprocess.STDOUT) | ||
| 1145 | except subprocess.CalledProcessError as e: | ||
| 1146 | bb.utils.remove(tmp_dir, recurse=True) | ||
| 1147 | bb.fatal("Unable to extract %s package. Command '%s' " | ||
| 1148 | "returned %d:\n%s" % (pkg_path, ' '.join(cmd), e.returncode, e.output.decode("utf-8"))) | ||
| 1149 | except OSError as e: | ||
| 1150 | bb.utils.remove(tmp_dir, recurse=True) | ||
| 1151 | bb.fatal("Unable to extract %s package. Command '%s' " | ||
| 1152 | "returned %d:\n%s at %s" % (pkg_path, ' '.join(cmd), e.errno, e.strerror, e.filename)) | ||
| 1153 | |||
| 1154 | bb.note("Extracted %s to %s" % (pkg_path, tmp_dir)) | ||
| 1155 | bb.utils.remove(os.path.join(tmp_dir, "debian-binary")) | ||
| 1156 | bb.utils.remove(os.path.join(tmp_dir, "control.tar.gz")) | ||
| 1157 | os.chdir(current_dir) | ||
| 1158 | |||
| 1159 | return tmp_dir | ||
| 1160 | |||
| 1161 | def _handle_intercept_failure(self, registered_pkgs): | ||
| 1162 | self.mark_packages("unpacked", registered_pkgs.split()) | ||
| 1163 | |||
| 1164 | class OpkgPM(OpkgDpkgPM): | ||
| 1165 | def __init__(self, d, target_rootfs, config_file, archs, task_name='target', ipk_repo_workdir="oe-rootfs-repo", filterbydependencies=True, prepare_index=True): | ||
| 1166 | super(OpkgPM, self).__init__(d, target_rootfs) | ||
| 1167 | |||
| 1168 | self.config_file = config_file | ||
| 1169 | self.pkg_archs = archs | ||
| 1170 | self.task_name = task_name | ||
| 1171 | |||
| 1172 | self.deploy_dir = oe.path.join(self.d.getVar('WORKDIR'), ipk_repo_workdir) | ||
| 1173 | self.deploy_lock_file = os.path.join(self.deploy_dir, "deploy.lock") | ||
| 1174 | self.opkg_cmd = bb.utils.which(os.getenv('PATH'), "opkg") | ||
| 1175 | self.opkg_args = "--volatile-cache -f %s -t %s -o %s " % (self.config_file, self.d.expand('${T}/ipktemp/') ,target_rootfs) | ||
| 1176 | self.opkg_args += self.d.getVar("OPKG_ARGS") | ||
| 1177 | |||
| 1178 | if prepare_index: | ||
| 1179 | create_packages_dir(self.d, self.deploy_dir, d.getVar("DEPLOY_DIR_IPK"), "package_write_ipk", filterbydependencies) | ||
| 1180 | |||
| 1181 | opkg_lib_dir = self.d.getVar('OPKGLIBDIR') | ||
| 1182 | if opkg_lib_dir[0] == "/": | ||
| 1183 | opkg_lib_dir = opkg_lib_dir[1:] | ||
| 1184 | |||
| 1185 | self.opkg_dir = os.path.join(target_rootfs, opkg_lib_dir, "opkg") | ||
| 1186 | |||
| 1187 | bb.utils.mkdirhier(self.opkg_dir) | ||
| 1188 | |||
| 1189 | self.saved_opkg_dir = self.d.expand('${T}/saved/%s' % self.task_name) | ||
| 1190 | if not os.path.exists(self.d.expand('${T}/saved')): | ||
| 1191 | bb.utils.mkdirhier(self.d.expand('${T}/saved')) | ||
| 1192 | |||
| 1193 | self.from_feeds = (self.d.getVar('BUILD_IMAGES_FROM_FEEDS') or "") == "1" | ||
| 1194 | if self.from_feeds: | ||
| 1195 | self._create_custom_config() | ||
| 1196 | else: | ||
| 1197 | self._create_config() | ||
| 1198 | |||
| 1199 | self.indexer = OpkgIndexer(self.d, self.deploy_dir) | ||
| 1200 | |||
| 1201 | def mark_packages(self, status_tag, packages=None): | ||
| 1202 | """ | ||
| 1203 | This function will change a package's status in /var/lib/opkg/status file. | ||
| 1204 | If 'packages' is None then the new_status will be applied to all | ||
| 1205 | packages | ||
| 1206 | """ | ||
| 1207 | status_file = os.path.join(self.opkg_dir, "status") | ||
| 1208 | |||
| 1209 | with open(status_file, "r") as sf: | ||
| 1210 | with open(status_file + ".tmp", "w+") as tmp_sf: | ||
| 1211 | if packages is None: | ||
| 1212 | tmp_sf.write(re.sub(r"Package: (.*?)\n((?:[^\n]+\n)*?)Status: (.*)(?:unpacked|installed)", | ||
| 1213 | r"Package: \1\n\2Status: \3%s" % status_tag, | ||
| 1214 | sf.read())) | ||
| 1215 | else: | ||
| 1216 | if type(packages).__name__ != "list": | ||
| 1217 | raise TypeError("'packages' should be a list object") | ||
| 1218 | |||
| 1219 | status = sf.read() | ||
| 1220 | for pkg in packages: | ||
| 1221 | status = re.sub(r"Package: %s\n((?:[^\n]+\n)*?)Status: (.*)(?:unpacked|installed)" % pkg, | ||
| 1222 | r"Package: %s\n\1Status: \2%s" % (pkg, status_tag), | ||
| 1223 | status) | ||
| 1224 | |||
| 1225 | tmp_sf.write(status) | ||
| 1226 | |||
| 1227 | os.rename(status_file + ".tmp", status_file) | ||
| 1228 | |||
| 1229 | def _create_custom_config(self): | ||
| 1230 | bb.note("Building from feeds activated!") | ||
| 1231 | |||
| 1232 | with open(self.config_file, "w+") as config_file: | ||
| 1233 | priority = 1 | ||
| 1234 | for arch in self.pkg_archs.split(): | ||
| 1235 | config_file.write("arch %s %d\n" % (arch, priority)) | ||
| 1236 | priority += 5 | ||
| 1237 | |||
| 1238 | for line in (self.d.getVar('IPK_FEED_URIS') or "").split(): | ||
| 1239 | feed_match = re.match(r"^[ \t]*(.*)##([^ \t]*)[ \t]*$", line) | ||
| 1240 | |||
| 1241 | if feed_match is not None: | ||
| 1242 | feed_name = feed_match.group(1) | ||
| 1243 | feed_uri = feed_match.group(2) | ||
| 1244 | |||
| 1245 | bb.note("Add %s feed with URL %s" % (feed_name, feed_uri)) | ||
| 1246 | |||
| 1247 | config_file.write("src/gz %s %s\n" % (feed_name, feed_uri)) | ||
| 1248 | |||
| 1249 | """ | ||
| 1250 | Allow to use package deploy directory contents as quick devel-testing | ||
| 1251 | feed. This creates individual feed configs for each arch subdir of those | ||
| 1252 | specified as compatible for the current machine. | ||
| 1253 | NOTE: Development-helper feature, NOT a full-fledged feed. | ||
| 1254 | """ | ||
| 1255 | if (self.d.getVar('FEED_DEPLOYDIR_BASE_URI') or "") != "": | ||
| 1256 | for arch in self.pkg_archs.split(): | ||
| 1257 | cfg_file_name = os.path.join(self.target_rootfs, | ||
| 1258 | self.d.getVar("sysconfdir"), | ||
| 1259 | "opkg", | ||
| 1260 | "local-%s-feed.conf" % arch) | ||
| 1261 | |||
| 1262 | with open(cfg_file_name, "w+") as cfg_file: | ||
| 1263 | cfg_file.write("src/gz local-%s %s/%s" % | ||
| 1264 | (arch, | ||
| 1265 | self.d.getVar('FEED_DEPLOYDIR_BASE_URI'), | ||
| 1266 | arch)) | ||
| 1267 | |||
| 1268 | if self.d.getVar('OPKGLIBDIR') != '/var/lib': | ||
| 1269 | # There is no command line option for this anymore, we need to add | ||
| 1270 | # info_dir and status_file to config file, if OPKGLIBDIR doesn't have | ||
| 1271 | # the default value of "/var/lib" as defined in opkg: | ||
| 1272 | # libopkg/opkg_conf.h:#define OPKG_CONF_DEFAULT_LISTS_DIR VARDIR "/lib/opkg/lists" | ||
| 1273 | # libopkg/opkg_conf.h:#define OPKG_CONF_DEFAULT_INFO_DIR VARDIR "/lib/opkg/info" | ||
| 1274 | # libopkg/opkg_conf.h:#define OPKG_CONF_DEFAULT_STATUS_FILE VARDIR "/lib/opkg/status" | ||
| 1275 | cfg_file.write("option info_dir %s\n" % os.path.join(self.d.getVar('OPKGLIBDIR'), 'opkg', 'info')) | ||
| 1276 | cfg_file.write("option lists_dir %s\n" % os.path.join(self.d.getVar('OPKGLIBDIR'), 'opkg', 'lists')) | ||
| 1277 | cfg_file.write("option status_file %s\n" % os.path.join(self.d.getVar('OPKGLIBDIR'), 'opkg', 'status')) | ||
| 1278 | |||
| 1279 | |||
| 1280 | def _create_config(self): | ||
| 1281 | with open(self.config_file, "w+") as config_file: | ||
| 1282 | priority = 1 | ||
| 1283 | for arch in self.pkg_archs.split(): | ||
| 1284 | config_file.write("arch %s %d\n" % (arch, priority)) | ||
| 1285 | priority += 5 | ||
| 1286 | |||
| 1287 | config_file.write("src oe file:%s\n" % self.deploy_dir) | ||
| 1288 | |||
| 1289 | for arch in self.pkg_archs.split(): | ||
| 1290 | pkgs_dir = os.path.join(self.deploy_dir, arch) | ||
| 1291 | if os.path.isdir(pkgs_dir): | ||
| 1292 | config_file.write("src oe-%s file:%s\n" % | ||
| 1293 | (arch, pkgs_dir)) | ||
| 1294 | |||
| 1295 | if self.d.getVar('OPKGLIBDIR') != '/var/lib': | ||
| 1296 | # There is no command line option for this anymore, we need to add | ||
| 1297 | # info_dir and status_file to config file, if OPKGLIBDIR doesn't have | ||
| 1298 | # the default value of "/var/lib" as defined in opkg: | ||
| 1299 | # libopkg/opkg_conf.h:#define OPKG_CONF_DEFAULT_LISTS_DIR VARDIR "/lib/opkg/lists" | ||
| 1300 | # libopkg/opkg_conf.h:#define OPKG_CONF_DEFAULT_INFO_DIR VARDIR "/lib/opkg/info" | ||
| 1301 | # libopkg/opkg_conf.h:#define OPKG_CONF_DEFAULT_STATUS_FILE VARDIR "/lib/opkg/status" | ||
| 1302 | config_file.write("option info_dir %s\n" % os.path.join(self.d.getVar('OPKGLIBDIR'), 'opkg', 'info')) | ||
| 1303 | config_file.write("option lists_dir %s\n" % os.path.join(self.d.getVar('OPKGLIBDIR'), 'opkg', 'lists')) | ||
| 1304 | config_file.write("option status_file %s\n" % os.path.join(self.d.getVar('OPKGLIBDIR'), 'opkg', 'status')) | ||
| 1305 | |||
| 1306 | def insert_feeds_uris(self, feed_uris, feed_base_paths, feed_archs): | ||
| 1307 | if feed_uris == "": | ||
| 1308 | return | ||
| 1309 | |||
| 1310 | rootfs_config = os.path.join('%s/etc/opkg/base-feeds.conf' | ||
| 1311 | % self.target_rootfs) | ||
| 1312 | |||
| 1313 | os.makedirs('%s/etc/opkg' % self.target_rootfs, exist_ok=True) | ||
| 1314 | |||
| 1315 | feed_uris = self.construct_uris(feed_uris.split(), feed_base_paths.split()) | ||
| 1316 | archs = self.pkg_archs.split() if feed_archs is None else feed_archs.split() | ||
| 1317 | |||
| 1318 | with open(rootfs_config, "w+") as config_file: | ||
| 1319 | uri_iterator = 0 | ||
| 1320 | for uri in feed_uris: | ||
| 1321 | if archs: | ||
| 1322 | for arch in archs: | ||
| 1323 | if (feed_archs is None) and (not os.path.exists(oe.path.join(self.deploy_dir, arch))): | ||
| 1324 | continue | ||
| 1325 | bb.note('Adding opkg feed url-%s-%d (%s)' % | ||
| 1326 | (arch, uri_iterator, uri)) | ||
| 1327 | config_file.write("src/gz uri-%s-%d %s/%s\n" % | ||
| 1328 | (arch, uri_iterator, uri, arch)) | ||
| 1329 | else: | ||
| 1330 | bb.note('Adding opkg feed url-%d (%s)' % | ||
| 1331 | (uri_iterator, uri)) | ||
| 1332 | config_file.write("src/gz uri-%d %s\n" % | ||
| 1333 | (uri_iterator, uri)) | ||
| 1334 | |||
| 1335 | uri_iterator += 1 | ||
| 1336 | |||
| 1337 | def update(self): | ||
| 1338 | self.deploy_dir_lock() | ||
| 1339 | |||
| 1340 | cmd = "%s %s update" % (self.opkg_cmd, self.opkg_args) | ||
| 1341 | |||
| 1342 | try: | ||
| 1343 | subprocess.check_output(cmd.split(), stderr=subprocess.STDOUT) | ||
| 1344 | except subprocess.CalledProcessError as e: | ||
| 1345 | self.deploy_dir_unlock() | ||
| 1346 | bb.fatal("Unable to update the package index files. Command '%s' " | ||
| 1347 | "returned %d:\n%s" % (cmd, e.returncode, e.output.decode("utf-8"))) | ||
| 1348 | |||
| 1349 | self.deploy_dir_unlock() | ||
| 1350 | |||
| 1351 | def install(self, pkgs, attempt_only=False): | ||
| 1352 | if not pkgs: | ||
| 1353 | return | ||
| 1354 | |||
| 1355 | cmd = "%s %s" % (self.opkg_cmd, self.opkg_args) | ||
| 1356 | for exclude in (self.d.getVar("PACKAGE_EXCLUDE") or "").split(): | ||
| 1357 | cmd += " --add-exclude %s" % exclude | ||
| 1358 | for bad_recommendation in (self.d.getVar("BAD_RECOMMENDATIONS") or "").split(): | ||
| 1359 | cmd += " --add-ignore-recommends %s" % bad_recommendation | ||
| 1360 | cmd += " install " | ||
| 1361 | cmd += " ".join(pkgs) | ||
| 1362 | |||
| 1363 | os.environ['D'] = self.target_rootfs | ||
| 1364 | os.environ['OFFLINE_ROOT'] = self.target_rootfs | ||
| 1365 | os.environ['IPKG_OFFLINE_ROOT'] = self.target_rootfs | ||
| 1366 | os.environ['OPKG_OFFLINE_ROOT'] = self.target_rootfs | ||
| 1367 | os.environ['INTERCEPT_DIR'] = self.intercepts_dir | ||
| 1368 | os.environ['NATIVE_ROOT'] = self.d.getVar('STAGING_DIR_NATIVE') | ||
| 1369 | |||
| 1370 | try: | ||
| 1371 | bb.note("Installing the following packages: %s" % ' '.join(pkgs)) | ||
| 1372 | bb.note(cmd) | ||
| 1373 | output = subprocess.check_output(cmd.split(), stderr=subprocess.STDOUT).decode("utf-8") | ||
| 1374 | bb.note(output) | ||
| 1375 | failed_pkgs = [] | ||
| 1376 | for line in output.split('\n'): | ||
| 1377 | if line.endswith("configuration required on target."): | ||
| 1378 | bb.warn(line) | ||
| 1379 | failed_pkgs.append(line.split(".")[0]) | ||
| 1380 | if failed_pkgs: | ||
| 1381 | failed_postinsts_abort(failed_pkgs, self.d.expand("${T}/log.do_${BB_CURRENTTASK}")) | ||
| 1382 | except subprocess.CalledProcessError as e: | ||
| 1383 | (bb.fatal, bb.warn)[attempt_only]("Unable to install packages. " | ||
| 1384 | "Command '%s' returned %d:\n%s" % | ||
| 1385 | (cmd, e.returncode, e.output.decode("utf-8"))) | ||
| 1386 | |||
| 1387 | def remove(self, pkgs, with_dependencies=True): | ||
| 1388 | if not pkgs: | ||
| 1389 | return | ||
| 1390 | |||
| 1391 | if with_dependencies: | ||
| 1392 | cmd = "%s %s --force-remove --force-removal-of-dependent-packages remove %s" % \ | ||
| 1393 | (self.opkg_cmd, self.opkg_args, ' '.join(pkgs)) | ||
| 1394 | else: | ||
| 1395 | cmd = "%s %s --force-depends remove %s" % \ | ||
| 1396 | (self.opkg_cmd, self.opkg_args, ' '.join(pkgs)) | ||
| 1397 | |||
| 1398 | try: | ||
| 1399 | bb.note(cmd) | ||
| 1400 | output = subprocess.check_output(cmd.split(), stderr=subprocess.STDOUT).decode("utf-8") | ||
| 1401 | bb.note(output) | ||
| 1402 | except subprocess.CalledProcessError as e: | ||
| 1403 | bb.fatal("Unable to remove packages. Command '%s' " | ||
| 1404 | "returned %d:\n%s" % (e.cmd, e.returncode, e.output.decode("utf-8"))) | ||
| 1405 | |||
| 1406 | def write_index(self): | ||
| 1407 | self.deploy_dir_lock() | ||
| 1408 | |||
| 1409 | result = self.indexer.write_index() | ||
| 1410 | |||
| 1411 | self.deploy_dir_unlock() | ||
| 1412 | |||
| 1413 | if result is not None: | ||
| 1414 | bb.fatal(result) | ||
| 1415 | |||
| 1416 | def remove_packaging_data(self): | ||
| 1417 | bb.utils.remove(self.opkg_dir, True) | ||
| 1418 | # create the directory back, it's needed by PM lock | ||
| 1419 | bb.utils.mkdirhier(self.opkg_dir) | ||
| 1420 | |||
| 1421 | def remove_lists(self): | ||
| 1422 | if not self.from_feeds: | ||
| 1423 | bb.utils.remove(os.path.join(self.opkg_dir, "lists"), True) | ||
| 1424 | |||
| 1425 | def list_installed(self): | ||
| 1426 | return OpkgPkgsList(self.d, self.target_rootfs, self.config_file).list_pkgs() | ||
| 1427 | |||
| 1428 | def dummy_install(self, pkgs): | ||
| 1429 | """ | ||
| 1430 | The following function dummy installs pkgs and returns the log of output. | ||
| 1431 | """ | ||
| 1432 | if len(pkgs) == 0: | ||
| 1433 | return | ||
| 1434 | |||
| 1435 | # Create an temp dir as opkg root for dummy installation | ||
| 1436 | temp_rootfs = self.d.expand('${T}/opkg') | ||
| 1437 | opkg_lib_dir = self.d.getVar('OPKGLIBDIR') | ||
| 1438 | if opkg_lib_dir[0] == "/": | ||
| 1439 | opkg_lib_dir = opkg_lib_dir[1:] | ||
| 1440 | temp_opkg_dir = os.path.join(temp_rootfs, opkg_lib_dir, 'opkg') | ||
| 1441 | bb.utils.mkdirhier(temp_opkg_dir) | ||
| 1442 | |||
| 1443 | opkg_args = "-f %s -o %s " % (self.config_file, temp_rootfs) | ||
| 1444 | opkg_args += self.d.getVar("OPKG_ARGS") | ||
| 1445 | |||
| 1446 | cmd = "%s %s update" % (self.opkg_cmd, opkg_args) | ||
| 1447 | try: | ||
| 1448 | subprocess.check_output(cmd, stderr=subprocess.STDOUT, shell=True) | ||
| 1449 | except subprocess.CalledProcessError as e: | ||
| 1450 | bb.fatal("Unable to update. Command '%s' " | ||
| 1451 | "returned %d:\n%s" % (cmd, e.returncode, e.output.decode("utf-8"))) | ||
| 1452 | |||
| 1453 | # Dummy installation | ||
| 1454 | cmd = "%s %s --noaction install %s " % (self.opkg_cmd, | ||
| 1455 | opkg_args, | ||
| 1456 | ' '.join(pkgs)) | ||
| 1457 | try: | ||
| 1458 | output = subprocess.check_output(cmd, stderr=subprocess.STDOUT, shell=True) | ||
| 1459 | except subprocess.CalledProcessError as e: | ||
| 1460 | bb.fatal("Unable to dummy install packages. Command '%s' " | ||
| 1461 | "returned %d:\n%s" % (cmd, e.returncode, e.output.decode("utf-8"))) | ||
| 1462 | |||
| 1463 | bb.utils.remove(temp_rootfs, True) | ||
| 1464 | |||
| 1465 | return output | ||
| 1466 | |||
| 1467 | def backup_packaging_data(self): | ||
| 1468 | # Save the opkglib for increment ipk image generation | ||
| 1469 | if os.path.exists(self.saved_opkg_dir): | ||
| 1470 | bb.utils.remove(self.saved_opkg_dir, True) | ||
| 1471 | shutil.copytree(self.opkg_dir, | ||
| 1472 | self.saved_opkg_dir, | ||
| 1473 | symlinks=True) | ||
| 1474 | |||
| 1475 | def recover_packaging_data(self): | ||
| 1476 | # Move the opkglib back | ||
| 1477 | if os.path.exists(self.saved_opkg_dir): | ||
| 1478 | if os.path.exists(self.opkg_dir): | ||
| 1479 | bb.utils.remove(self.opkg_dir, True) | ||
| 1480 | |||
| 1481 | bb.note('Recover packaging data') | ||
| 1482 | shutil.copytree(self.saved_opkg_dir, | ||
| 1483 | self.opkg_dir, | ||
| 1484 | symlinks=True) | ||
| 1485 | |||
| 1486 | def package_info(self, pkg): | ||
| 1487 | """ | ||
| 1488 | Returns a dictionary with the package info. | ||
| 1489 | """ | ||
| 1490 | cmd = "%s %s info %s" % (self.opkg_cmd, self.opkg_args, pkg) | ||
| 1491 | pkg_info = super(OpkgPM, self).package_info(pkg, cmd) | ||
| 1492 | |||
| 1493 | pkg_arch = pkg_info[pkg]["arch"] | ||
| 1494 | pkg_filename = pkg_info[pkg]["filename"] | ||
| 1495 | pkg_info[pkg]["filepath"] = \ | ||
| 1496 | os.path.join(self.deploy_dir, pkg_arch, pkg_filename) | ||
| 1497 | |||
| 1498 | return pkg_info | ||
| 1499 | |||
| 1500 | def extract(self, pkg): | ||
| 1501 | """ | ||
| 1502 | Returns the path to a tmpdir where resides the contents of a package. | ||
| 1503 | |||
| 1504 | Deleting the tmpdir is responsability of the caller. | ||
| 1505 | """ | ||
| 1506 | pkg_info = self.package_info(pkg) | ||
| 1507 | if not pkg_info: | ||
| 1508 | bb.fatal("Unable to get information for package '%s' while " | ||
| 1509 | "trying to extract the package." % pkg) | ||
| 1510 | |||
| 1511 | tmp_dir = super(OpkgPM, self).extract(pkg, pkg_info) | ||
| 1512 | bb.utils.remove(os.path.join(tmp_dir, "data.tar.xz")) | ||
| 1513 | |||
| 1514 | return tmp_dir | ||
| 1515 | |||
| 1516 | class DpkgPM(OpkgDpkgPM): | ||
| 1517 | def __init__(self, d, target_rootfs, archs, base_archs, apt_conf_dir=None, deb_repo_workdir="oe-rootfs-repo", filterbydependencies=True): | ||
| 1518 | super(DpkgPM, self).__init__(d, target_rootfs) | ||
| 1519 | self.deploy_dir = oe.path.join(self.d.getVar('WORKDIR'), deb_repo_workdir) | ||
| 1520 | |||
| 1521 | create_packages_dir(self.d, self.deploy_dir, d.getVar("DEPLOY_DIR_DEB"), "package_write_deb", filterbydependencies) | ||
| 1522 | |||
| 1523 | if apt_conf_dir is None: | ||
| 1524 | self.apt_conf_dir = self.d.expand("${APTCONF_TARGET}/apt") | ||
| 1525 | else: | ||
| 1526 | self.apt_conf_dir = apt_conf_dir | ||
| 1527 | self.apt_conf_file = os.path.join(self.apt_conf_dir, "apt.conf") | ||
| 1528 | self.apt_get_cmd = bb.utils.which(os.getenv('PATH'), "apt-get") | ||
| 1529 | self.apt_cache_cmd = bb.utils.which(os.getenv('PATH'), "apt-cache") | ||
| 1530 | |||
| 1531 | self.apt_args = d.getVar("APT_ARGS") | ||
| 1532 | |||
| 1533 | self.all_arch_list = archs.split() | ||
| 1534 | all_mlb_pkg_arch_list = (self.d.getVar('ALL_MULTILIB_PACKAGE_ARCHS') or "").split() | ||
| 1535 | self.all_arch_list.extend(arch for arch in all_mlb_pkg_arch_list if arch not in self.all_arch_list) | ||
| 1536 | |||
| 1537 | self._create_configs(archs, base_archs) | ||
| 1538 | |||
| 1539 | self.indexer = DpkgIndexer(self.d, self.deploy_dir) | ||
| 1540 | |||
| 1541 | def mark_packages(self, status_tag, packages=None): | ||
| 1542 | """ | ||
| 1543 | This function will change a package's status in /var/lib/dpkg/status file. | ||
| 1544 | If 'packages' is None then the new_status will be applied to all | ||
| 1545 | packages | ||
| 1546 | """ | ||
| 1547 | status_file = self.target_rootfs + "/var/lib/dpkg/status" | ||
| 1548 | |||
| 1549 | with open(status_file, "r") as sf: | ||
| 1550 | with open(status_file + ".tmp", "w+") as tmp_sf: | ||
| 1551 | if packages is None: | ||
| 1552 | tmp_sf.write(re.sub(r"Package: (.*?)\n((?:[^\n]+\n)*?)Status: (.*)(?:unpacked|installed)", | ||
| 1553 | r"Package: \1\n\2Status: \3%s" % status_tag, | ||
| 1554 | sf.read())) | ||
| 1555 | else: | ||
| 1556 | if type(packages).__name__ != "list": | ||
| 1557 | raise TypeError("'packages' should be a list object") | ||
| 1558 | |||
| 1559 | status = sf.read() | ||
| 1560 | for pkg in packages: | ||
| 1561 | status = re.sub(r"Package: %s\n((?:[^\n]+\n)*?)Status: (.*)(?:unpacked|installed)" % pkg, | ||
| 1562 | r"Package: %s\n\1Status: \2%s" % (pkg, status_tag), | ||
| 1563 | status) | ||
| 1564 | |||
| 1565 | tmp_sf.write(status) | ||
| 1566 | |||
| 1567 | os.rename(status_file + ".tmp", status_file) | ||
| 1568 | |||
| 1569 | def run_pre_post_installs(self, package_name=None): | ||
| 1570 | """ | ||
| 1571 | Run the pre/post installs for package "package_name". If package_name is | ||
| 1572 | None, then run all pre/post install scriptlets. | ||
| 1573 | """ | ||
| 1574 | info_dir = self.target_rootfs + "/var/lib/dpkg/info" | ||
| 1575 | ControlScript = collections.namedtuple("ControlScript", ["suffix", "name", "argument"]) | ||
| 1576 | control_scripts = [ | ||
| 1577 | ControlScript(".preinst", "Preinstall", "install"), | ||
| 1578 | ControlScript(".postinst", "Postinstall", "configure")] | ||
| 1579 | status_file = self.target_rootfs + "/var/lib/dpkg/status" | ||
| 1580 | installed_pkgs = [] | ||
| 1581 | |||
| 1582 | with open(status_file, "r") as status: | ||
| 1583 | for line in status.read().split('\n'): | ||
| 1584 | m = re.match(r"^Package: (.*)", line) | ||
| 1585 | if m is not None: | ||
| 1586 | installed_pkgs.append(m.group(1)) | ||
| 1587 | |||
| 1588 | if package_name is not None and not package_name in installed_pkgs: | ||
| 1589 | return | ||
| 1590 | |||
| 1591 | os.environ['D'] = self.target_rootfs | ||
| 1592 | os.environ['OFFLINE_ROOT'] = self.target_rootfs | ||
| 1593 | os.environ['IPKG_OFFLINE_ROOT'] = self.target_rootfs | ||
| 1594 | os.environ['OPKG_OFFLINE_ROOT'] = self.target_rootfs | ||
| 1595 | os.environ['INTERCEPT_DIR'] = self.intercepts_dir | ||
| 1596 | os.environ['NATIVE_ROOT'] = self.d.getVar('STAGING_DIR_NATIVE') | ||
| 1597 | |||
| 1598 | for pkg_name in installed_pkgs: | ||
| 1599 | for control_script in control_scripts: | ||
| 1600 | p_full = os.path.join(info_dir, pkg_name + control_script.suffix) | ||
| 1601 | if os.path.exists(p_full): | ||
| 1602 | try: | ||
| 1603 | bb.note("Executing %s for package: %s ..." % | ||
| 1604 | (control_script.name.lower(), pkg_name)) | ||
| 1605 | output = subprocess.check_output([p_full, control_script.argument], | ||
| 1606 | stderr=subprocess.STDOUT).decode("utf-8") | ||
| 1607 | bb.note(output) | ||
| 1608 | except subprocess.CalledProcessError as e: | ||
| 1609 | bb.warn("%s for package %s failed with %d:\n%s" % | ||
| 1610 | (control_script.name, pkg_name, e.returncode, | ||
| 1611 | e.output.decode("utf-8"))) | ||
| 1612 | failed_postinsts_abort([pkg_name], self.d.expand("${T}/log.do_${BB_CURRENTTASK}")) | ||
| 1613 | |||
| 1614 | def update(self): | ||
| 1615 | os.environ['APT_CONFIG'] = self.apt_conf_file | ||
| 1616 | |||
| 1617 | self.deploy_dir_lock() | ||
| 1618 | |||
| 1619 | cmd = "%s update" % self.apt_get_cmd | ||
| 1620 | |||
| 1621 | try: | ||
| 1622 | subprocess.check_output(cmd.split(), stderr=subprocess.STDOUT) | ||
| 1623 | except subprocess.CalledProcessError as e: | ||
| 1624 | bb.fatal("Unable to update the package index files. Command '%s' " | ||
| 1625 | "returned %d:\n%s" % (e.cmd, e.returncode, e.output.decode("utf-8"))) | ||
| 1626 | |||
| 1627 | self.deploy_dir_unlock() | ||
| 1628 | |||
| 1629 | def install(self, pkgs, attempt_only=False): | ||
| 1630 | if attempt_only and len(pkgs) == 0: | ||
| 1631 | return | ||
| 1632 | |||
| 1633 | os.environ['APT_CONFIG'] = self.apt_conf_file | ||
| 1634 | |||
| 1635 | cmd = "%s %s install --force-yes --allow-unauthenticated --no-remove %s" % \ | ||
| 1636 | (self.apt_get_cmd, self.apt_args, ' '.join(pkgs)) | ||
| 1637 | |||
| 1638 | try: | ||
| 1639 | bb.note("Installing the following packages: %s" % ' '.join(pkgs)) | ||
| 1640 | subprocess.check_output(cmd.split(), stderr=subprocess.STDOUT) | ||
| 1641 | except subprocess.CalledProcessError as e: | ||
| 1642 | (bb.fatal, bb.warn)[attempt_only]("Unable to install packages. " | ||
| 1643 | "Command '%s' returned %d:\n%s" % | ||
| 1644 | (cmd, e.returncode, e.output.decode("utf-8"))) | ||
| 1645 | |||
| 1646 | # rename *.dpkg-new files/dirs | ||
| 1647 | for root, dirs, files in os.walk(self.target_rootfs): | ||
| 1648 | for dir in dirs: | ||
| 1649 | new_dir = re.sub(r"\.dpkg-new", "", dir) | ||
| 1650 | if dir != new_dir: | ||
| 1651 | os.rename(os.path.join(root, dir), | ||
| 1652 | os.path.join(root, new_dir)) | ||
| 1653 | |||
| 1654 | for file in files: | ||
| 1655 | new_file = re.sub(r"\.dpkg-new", "", file) | ||
| 1656 | if file != new_file: | ||
| 1657 | os.rename(os.path.join(root, file), | ||
| 1658 | os.path.join(root, new_file)) | ||
| 1659 | |||
| 1660 | |||
| 1661 | def remove(self, pkgs, with_dependencies=True): | ||
| 1662 | if not pkgs: | ||
| 1663 | return | ||
| 1664 | |||
| 1665 | if with_dependencies: | ||
| 1666 | os.environ['APT_CONFIG'] = self.apt_conf_file | ||
| 1667 | cmd = "%s purge %s" % (self.apt_get_cmd, ' '.join(pkgs)) | ||
| 1668 | else: | ||
| 1669 | cmd = "%s --admindir=%s/var/lib/dpkg --instdir=%s" \ | ||
| 1670 | " -P --force-depends %s" % \ | ||
| 1671 | (bb.utils.which(os.getenv('PATH'), "dpkg"), | ||
| 1672 | self.target_rootfs, self.target_rootfs, ' '.join(pkgs)) | ||
| 1673 | |||
| 1674 | try: | ||
| 1675 | subprocess.check_output(cmd.split(), stderr=subprocess.STDOUT) | ||
| 1676 | except subprocess.CalledProcessError as e: | ||
| 1677 | bb.fatal("Unable to remove packages. Command '%s' " | ||
| 1678 | "returned %d:\n%s" % (e.cmd, e.returncode, e.output.decode("utf-8"))) | ||
| 1679 | |||
| 1680 | def write_index(self): | ||
| 1681 | self.deploy_dir_lock() | ||
| 1682 | |||
| 1683 | result = self.indexer.write_index() | ||
| 1684 | |||
| 1685 | self.deploy_dir_unlock() | ||
| 1686 | |||
| 1687 | if result is not None: | ||
| 1688 | bb.fatal(result) | ||
| 1689 | |||
| 1690 | def insert_feeds_uris(self, feed_uris, feed_base_paths, feed_archs): | ||
| 1691 | if feed_uris == "": | ||
| 1692 | return | ||
| 1693 | |||
| 1694 | sources_conf = os.path.join("%s/etc/apt/sources.list" | ||
| 1695 | % self.target_rootfs) | ||
| 1696 | arch_list = [] | ||
| 1697 | |||
| 1698 | if feed_archs is None: | ||
| 1699 | for arch in self.all_arch_list: | ||
| 1700 | if not os.path.exists(os.path.join(self.deploy_dir, arch)): | ||
| 1701 | continue | ||
| 1702 | arch_list.append(arch) | ||
| 1703 | else: | ||
| 1704 | arch_list = feed_archs.split() | ||
| 1705 | |||
| 1706 | feed_uris = self.construct_uris(feed_uris.split(), feed_base_paths.split()) | ||
| 1707 | |||
| 1708 | with open(sources_conf, "w+") as sources_file: | ||
| 1709 | for uri in feed_uris: | ||
| 1710 | if arch_list: | ||
| 1711 | for arch in arch_list: | ||
| 1712 | bb.note('Adding dpkg channel at (%s)' % uri) | ||
| 1713 | sources_file.write("deb %s/%s ./\n" % | ||
| 1714 | (uri, arch)) | ||
| 1715 | else: | ||
| 1716 | bb.note('Adding dpkg channel at (%s)' % uri) | ||
| 1717 | sources_file.write("deb %s ./\n" % uri) | ||
| 1718 | |||
| 1719 | def _create_configs(self, archs, base_archs): | ||
| 1720 | base_archs = re.sub(r"_", r"-", base_archs) | ||
| 1721 | |||
| 1722 | if os.path.exists(self.apt_conf_dir): | ||
| 1723 | bb.utils.remove(self.apt_conf_dir, True) | ||
| 1724 | |||
| 1725 | bb.utils.mkdirhier(self.apt_conf_dir) | ||
| 1726 | bb.utils.mkdirhier(self.apt_conf_dir + "/lists/partial/") | ||
| 1727 | bb.utils.mkdirhier(self.apt_conf_dir + "/apt.conf.d/") | ||
| 1728 | bb.utils.mkdirhier(self.apt_conf_dir + "/preferences.d/") | ||
| 1729 | |||
| 1730 | arch_list = [] | ||
| 1731 | for arch in self.all_arch_list: | ||
| 1732 | if not os.path.exists(os.path.join(self.deploy_dir, arch)): | ||
| 1733 | continue | ||
| 1734 | arch_list.append(arch) | ||
| 1735 | |||
| 1736 | with open(os.path.join(self.apt_conf_dir, "preferences"), "w+") as prefs_file: | ||
| 1737 | priority = 801 | ||
| 1738 | for arch in arch_list: | ||
| 1739 | prefs_file.write( | ||
| 1740 | "Package: *\n" | ||
| 1741 | "Pin: release l=%s\n" | ||
| 1742 | "Pin-Priority: %d\n\n" % (arch, priority)) | ||
| 1743 | |||
| 1744 | priority += 5 | ||
| 1745 | |||
| 1746 | pkg_exclude = self.d.getVar('PACKAGE_EXCLUDE') or "" | ||
| 1747 | for pkg in pkg_exclude.split(): | ||
| 1748 | prefs_file.write( | ||
| 1749 | "Package: %s\n" | ||
| 1750 | "Pin: release *\n" | ||
| 1751 | "Pin-Priority: -1\n\n" % pkg) | ||
| 1752 | |||
| 1753 | arch_list.reverse() | ||
| 1754 | |||
| 1755 | with open(os.path.join(self.apt_conf_dir, "sources.list"), "w+") as sources_file: | ||
| 1756 | for arch in arch_list: | ||
| 1757 | sources_file.write("deb file:%s/ ./\n" % | ||
| 1758 | os.path.join(self.deploy_dir, arch)) | ||
| 1759 | |||
| 1760 | base_arch_list = base_archs.split() | ||
| 1761 | multilib_variants = self.d.getVar("MULTILIB_VARIANTS"); | ||
| 1762 | for variant in multilib_variants.split(): | ||
| 1763 | localdata = bb.data.createCopy(self.d) | ||
| 1764 | variant_tune = localdata.getVar("DEFAULTTUNE_virtclass-multilib-" + variant, False) | ||
| 1765 | orig_arch = localdata.getVar("DPKG_ARCH") | ||
| 1766 | localdata.setVar("DEFAULTTUNE", variant_tune) | ||
| 1767 | variant_arch = localdata.getVar("DPKG_ARCH") | ||
| 1768 | if variant_arch not in base_arch_list: | ||
| 1769 | base_arch_list.append(variant_arch) | ||
| 1770 | |||
| 1771 | with open(self.apt_conf_file, "w+") as apt_conf: | ||
| 1772 | with open(self.d.expand("${STAGING_ETCDIR_NATIVE}/apt/apt.conf.sample")) as apt_conf_sample: | ||
| 1773 | for line in apt_conf_sample.read().split("\n"): | ||
| 1774 | match_arch = re.match(r" Architecture \".*\";$", line) | ||
| 1775 | architectures = "" | ||
| 1776 | if match_arch: | ||
| 1777 | for base_arch in base_arch_list: | ||
| 1778 | architectures += "\"%s\";" % base_arch | ||
| 1779 | apt_conf.write(" Architectures {%s};\n" % architectures); | ||
| 1780 | apt_conf.write(" Architecture \"%s\";\n" % base_archs) | ||
| 1781 | else: | ||
| 1782 | line = re.sub(r"#ROOTFS#", self.target_rootfs, line) | ||
| 1783 | line = re.sub(r"#APTCONF#", self.apt_conf_dir, line) | ||
| 1784 | apt_conf.write(line + "\n") | ||
| 1785 | |||
| 1786 | target_dpkg_dir = "%s/var/lib/dpkg" % self.target_rootfs | ||
| 1787 | bb.utils.mkdirhier(os.path.join(target_dpkg_dir, "info")) | ||
| 1788 | |||
| 1789 | bb.utils.mkdirhier(os.path.join(target_dpkg_dir, "updates")) | ||
| 1790 | |||
| 1791 | if not os.path.exists(os.path.join(target_dpkg_dir, "status")): | ||
| 1792 | open(os.path.join(target_dpkg_dir, "status"), "w+").close() | ||
| 1793 | if not os.path.exists(os.path.join(target_dpkg_dir, "available")): | ||
| 1794 | open(os.path.join(target_dpkg_dir, "available"), "w+").close() | ||
| 1795 | |||
| 1796 | def remove_packaging_data(self): | ||
| 1797 | bb.utils.remove(self.target_rootfs + self.d.getVar('opkglibdir'), True) | ||
| 1798 | bb.utils.remove(self.target_rootfs + "/var/lib/dpkg/", True) | ||
| 1799 | |||
| 1800 | def fix_broken_dependencies(self): | ||
| 1801 | os.environ['APT_CONFIG'] = self.apt_conf_file | ||
| 1802 | |||
| 1803 | cmd = "%s %s --allow-unauthenticated -f install" % (self.apt_get_cmd, self.apt_args) | ||
| 1804 | |||
| 1805 | try: | ||
| 1806 | subprocess.check_output(cmd.split(), stderr=subprocess.STDOUT) | ||
| 1807 | except subprocess.CalledProcessError as e: | ||
| 1808 | bb.fatal("Cannot fix broken dependencies. Command '%s' " | ||
| 1809 | "returned %d:\n%s" % (cmd, e.returncode, e.output.decode("utf-8"))) | ||
| 1810 | |||
| 1811 | def list_installed(self): | ||
| 1812 | return DpkgPkgsList(self.d, self.target_rootfs).list_pkgs() | ||
| 1813 | |||
| 1814 | def package_info(self, pkg): | ||
| 1815 | """ | ||
| 1816 | Returns a dictionary with the package info. | ||
| 1817 | """ | ||
| 1818 | cmd = "%s show %s" % (self.apt_cache_cmd, pkg) | ||
| 1819 | pkg_info = super(DpkgPM, self).package_info(pkg, cmd) | ||
| 1820 | |||
| 1821 | pkg_arch = pkg_info[pkg]["pkgarch"] | ||
| 1822 | pkg_filename = pkg_info[pkg]["filename"] | ||
| 1823 | pkg_info[pkg]["filepath"] = \ | ||
| 1824 | os.path.join(self.deploy_dir, pkg_arch, pkg_filename) | ||
| 1825 | |||
| 1826 | return pkg_info | ||
| 1827 | |||
| 1828 | def extract(self, pkg): | ||
| 1829 | """ | ||
| 1830 | Returns the path to a tmpdir where resides the contents of a package. | ||
| 1831 | |||
| 1832 | Deleting the tmpdir is responsability of the caller. | ||
| 1833 | """ | ||
| 1834 | pkg_info = self.package_info(pkg) | ||
| 1835 | if not pkg_info: | ||
| 1836 | bb.fatal("Unable to get information for package '%s' while " | ||
| 1837 | "trying to extract the package." % pkg) | ||
| 1838 | |||
| 1839 | tmp_dir = super(DpkgPM, self).extract(pkg, pkg_info) | ||
| 1840 | bb.utils.remove(os.path.join(tmp_dir, "data.tar.xz")) | ||
| 1841 | |||
| 1842 | return tmp_dir | ||
| 1843 | |||
| 1844 | def generate_index_files(d): | ||
| 1845 | classes = d.getVar('PACKAGE_CLASSES').replace("package_", "").split() | ||
| 1846 | |||
| 1847 | indexer_map = { | ||
| 1848 | "rpm": (RpmSubdirIndexer, d.getVar('DEPLOY_DIR_RPM')), | ||
| 1849 | "ipk": (OpkgIndexer, d.getVar('DEPLOY_DIR_IPK')), | ||
| 1850 | "deb": (DpkgIndexer, d.getVar('DEPLOY_DIR_DEB')) | ||
| 1851 | } | ||
| 1852 | |||
| 1853 | result = None | ||
| 1854 | |||
| 1855 | for pkg_class in classes: | ||
| 1856 | if not pkg_class in indexer_map: | ||
| 1857 | continue | ||
| 1858 | |||
| 1859 | if os.path.exists(indexer_map[pkg_class][1]): | ||
| 1860 | result = indexer_map[pkg_class][0](d, indexer_map[pkg_class][1]).write_index() | ||
| 1861 | |||
| 1862 | if result is not None: | ||
| 1863 | bb.fatal(result) | ||
