diff options
| author | Richard Purdie <richard.purdie@linuxfoundation.org> | 2012-07-11 17:33:43 +0000 |
|---|---|---|
| committer | Richard Purdie <richard.purdie@linuxfoundation.org> | 2012-07-19 10:24:50 +0100 |
| commit | bfd279de3275abbfaf3e630383ec244131e0375f (patch) | |
| tree | 0d1c90461a890d21444f5d2afb13c52b302427f1 /meta/classes/package.bbclass | |
| parent | 99203edda6f0b09d817454d656c100b7a8806b18 (diff) | |
| download | poky-bfd279de3275abbfaf3e630383ec244131e0375f.tar.gz | |
Convert tab indentation in python functions into four-space
(From OE-Core rev: 604d46c686d06d62d5a07b9c7f4fa170f99307d8)
Signed-off-by: Richard Purdie <richard.purdie@linuxfoundation.org>
Diffstat (limited to 'meta/classes/package.bbclass')
| -rw-r--r-- | meta/classes/package.bbclass | 2930 |
1 files changed, 1465 insertions, 1465 deletions
diff --git a/meta/classes/package.bbclass b/meta/classes/package.bbclass index dfd42117c5..a51e955325 100644 --- a/meta/classes/package.bbclass +++ b/meta/classes/package.bbclass | |||
| @@ -25,8 +25,8 @@ | |||
| 25 | # The data is stores in FILER{PROVIDES,DEPENDS}_file_pkg variables with | 25 | # The data is stores in FILER{PROVIDES,DEPENDS}_file_pkg variables with |
| 26 | # a list of affected files in FILER{PROVIDES,DEPENDS}FLIST_pkg | 26 | # a list of affected files in FILER{PROVIDES,DEPENDS}FLIST_pkg |
| 27 | # | 27 | # |
| 28 | # h) package_do_shlibs - Look at the shared libraries generated and autotmatically add any | 28 | # h) package_do_shlibs - Look at the shared libraries generated and autotmatically add any |
| 29 | # depenedencies found. Also stores the package name so anyone else using this library | 29 | # depenedencies found. Also stores the package name so anyone else using this library |
| 30 | # knows which package to depend on. | 30 | # knows which package to depend on. |
| 31 | # | 31 | # |
| 32 | # i) package_do_pkgconfig - Keep track of which packages need and provide which .pc files | 32 | # i) package_do_pkgconfig - Keep track of which packages need and provide which .pc files |
| @@ -35,7 +35,7 @@ | |||
| 35 | # | 35 | # |
| 36 | # k) package_depchains - Adds automatic dependencies to -dbg and -dev packages | 36 | # k) package_depchains - Adds automatic dependencies to -dbg and -dev packages |
| 37 | # | 37 | # |
| 38 | # l) emit_pkgdata - saves the packaging data into PKGDATA_DIR for use in later | 38 | # l) emit_pkgdata - saves the packaging data into PKGDATA_DIR for use in later |
| 39 | # packaging steps | 39 | # packaging steps |
| 40 | 40 | ||
| 41 | inherit packagedata | 41 | inherit packagedata |
| @@ -52,112 +52,112 @@ ALL_MULTILIB_PACKAGE_ARCHS = "${@all_multilib_tune_values(d, 'PACKAGE_ARCHS')}" | |||
| 52 | PACKAGE_DEPENDS += "rpm-native" | 52 | PACKAGE_DEPENDS += "rpm-native" |
| 53 | 53 | ||
| 54 | def legitimize_package_name(s): | 54 | def legitimize_package_name(s): |
| 55 | """ | 55 | """ |
| 56 | Make sure package names are legitimate strings | 56 | Make sure package names are legitimate strings |
| 57 | """ | 57 | """ |
| 58 | import re | 58 | import re |
| 59 | 59 | ||
| 60 | def fixutf(m): | 60 | def fixutf(m): |
| 61 | cp = m.group(1) | 61 | cp = m.group(1) |
| 62 | if cp: | 62 | if cp: |
| 63 | return ('\u%s' % cp).decode('unicode_escape').encode('utf-8') | 63 | return ('\u%s' % cp).decode('unicode_escape').encode('utf-8') |
| 64 | 64 | ||
| 65 | # Handle unicode codepoints encoded as <U0123>, as in glibc locale files. | 65 | # Handle unicode codepoints encoded as <U0123>, as in glibc locale files. |
| 66 | s = re.sub('<U([0-9A-Fa-f]{1,4})>', fixutf, s) | 66 | s = re.sub('<U([0-9A-Fa-f]{1,4})>', fixutf, s) |
| 67 | 67 | ||
| 68 | # Remaining package name validity fixes | 68 | # Remaining package name validity fixes |
| 69 | return s.lower().replace('_', '-').replace('@', '+').replace(',', '+').replace('/', '-') | 69 | return s.lower().replace('_', '-').replace('@', '+').replace(',', '+').replace('/', '-') |
| 70 | 70 | ||
| 71 | def do_split_packages(d, root, file_regex, output_pattern, description, postinst=None, recursive=False, hook=None, extra_depends=None, aux_files_pattern=None, postrm=None, allow_dirs=False, prepend=False, match_path=False, aux_files_pattern_verbatim=None, allow_links=False): | 71 | def do_split_packages(d, root, file_regex, output_pattern, description, postinst=None, recursive=False, hook=None, extra_depends=None, aux_files_pattern=None, postrm=None, allow_dirs=False, prepend=False, match_path=False, aux_files_pattern_verbatim=None, allow_links=False): |
| 72 | """ | 72 | """ |
| 73 | Used in .bb files to split up dynamically generated subpackages of a | 73 | Used in .bb files to split up dynamically generated subpackages of a |
| 74 | given package, usually plugins or modules. | 74 | given package, usually plugins or modules. |
| 75 | """ | 75 | """ |
| 76 | 76 | ||
| 77 | ml = d.getVar("MLPREFIX", True) | 77 | ml = d.getVar("MLPREFIX", True) |
| 78 | if ml: | 78 | if ml: |
| 79 | if not output_pattern.startswith(ml): | 79 | if not output_pattern.startswith(ml): |
| 80 | output_pattern = ml + output_pattern | 80 | output_pattern = ml + output_pattern |
| 81 | 81 | ||
| 82 | newdeps = [] | 82 | newdeps = [] |
| 83 | for dep in (extra_depends or "").split(): | 83 | for dep in (extra_depends or "").split(): |
| 84 | if dep.startswith(ml): | 84 | if dep.startswith(ml): |
| 85 | newdeps.append(dep) | 85 | newdeps.append(dep) |
| 86 | else: | 86 | else: |
| 87 | newdeps.append(ml + dep) | 87 | newdeps.append(ml + dep) |
| 88 | if newdeps: | 88 | if newdeps: |
| 89 | extra_depends = " ".join(newdeps) | 89 | extra_depends = " ".join(newdeps) |
| 90 | 90 | ||
| 91 | dvar = d.getVar('PKGD', True) | 91 | dvar = d.getVar('PKGD', True) |
| 92 | 92 | ||
| 93 | packages = d.getVar('PACKAGES', True).split() | 93 | packages = d.getVar('PACKAGES', True).split() |
| 94 | 94 | ||
| 95 | if postinst: | 95 | if postinst: |
| 96 | postinst = '#!/bin/sh\n' + postinst + '\n' | 96 | postinst = '#!/bin/sh\n' + postinst + '\n' |
| 97 | if postrm: | 97 | if postrm: |
| 98 | postrm = '#!/bin/sh\n' + postrm + '\n' | 98 | postrm = '#!/bin/sh\n' + postrm + '\n' |
| 99 | if not recursive: | 99 | if not recursive: |
| 100 | objs = os.listdir(dvar + root) | 100 | objs = os.listdir(dvar + root) |
| 101 | else: | 101 | else: |
| 102 | objs = [] | 102 | objs = [] |
| 103 | for walkroot, dirs, files in os.walk(dvar + root): | 103 | for walkroot, dirs, files in os.walk(dvar + root): |
| 104 | for file in files: | 104 | for file in files: |
| 105 | relpath = os.path.join(walkroot, file).replace(dvar + root + '/', '', 1) | 105 | relpath = os.path.join(walkroot, file).replace(dvar + root + '/', '', 1) |
| 106 | if relpath: | 106 | if relpath: |
| 107 | objs.append(relpath) | 107 | objs.append(relpath) |
| 108 | 108 | ||
| 109 | if extra_depends == None: | 109 | if extra_depends == None: |
| 110 | extra_depends = d.getVar("PN", True) | 110 | extra_depends = d.getVar("PN", True) |
| 111 | 111 | ||
| 112 | for o in sorted(objs): | 112 | for o in sorted(objs): |
| 113 | import re, stat | 113 | import re, stat |
| 114 | if match_path: | 114 | if match_path: |
| 115 | m = re.match(file_regex, o) | 115 | m = re.match(file_regex, o) |
| 116 | else: | 116 | else: |
| 117 | m = re.match(file_regex, os.path.basename(o)) | 117 | m = re.match(file_regex, os.path.basename(o)) |
| 118 | 118 | ||
| 119 | if not m: | 119 | if not m: |
| 120 | continue | 120 | continue |
| 121 | f = os.path.join(dvar + root, o) | 121 | f = os.path.join(dvar + root, o) |
| 122 | mode = os.lstat(f).st_mode | 122 | mode = os.lstat(f).st_mode |
| 123 | if not (stat.S_ISREG(mode) or (allow_links and stat.S_ISLNK(mode)) or (allow_dirs and stat.S_ISDIR(mode))): | 123 | if not (stat.S_ISREG(mode) or (allow_links and stat.S_ISLNK(mode)) or (allow_dirs and stat.S_ISDIR(mode))): |
| 124 | continue | 124 | continue |
| 125 | on = legitimize_package_name(m.group(1)) | 125 | on = legitimize_package_name(m.group(1)) |
| 126 | pkg = output_pattern % on | 126 | pkg = output_pattern % on |
| 127 | if not pkg in packages: | 127 | if not pkg in packages: |
| 128 | if prepend: | 128 | if prepend: |
| 129 | packages = [pkg] + packages | 129 | packages = [pkg] + packages |
| 130 | else: | 130 | else: |
| 131 | packages.append(pkg) | 131 | packages.append(pkg) |
| 132 | oldfiles = d.getVar('FILES_' + pkg, True) | 132 | oldfiles = d.getVar('FILES_' + pkg, True) |
| 133 | if not oldfiles: | 133 | if not oldfiles: |
| 134 | the_files = [os.path.join(root, o)] | 134 | the_files = [os.path.join(root, o)] |
| 135 | if aux_files_pattern: | 135 | if aux_files_pattern: |
| 136 | if type(aux_files_pattern) is list: | 136 | if type(aux_files_pattern) is list: |
| 137 | for fp in aux_files_pattern: | 137 | for fp in aux_files_pattern: |
| 138 | the_files.append(fp % on) | 138 | the_files.append(fp % on) |
| 139 | else: | 139 | else: |
| 140 | the_files.append(aux_files_pattern % on) | 140 | the_files.append(aux_files_pattern % on) |
| 141 | if aux_files_pattern_verbatim: | 141 | if aux_files_pattern_verbatim: |
| 142 | if type(aux_files_pattern_verbatim) is list: | 142 | if type(aux_files_pattern_verbatim) is list: |
| 143 | for fp in aux_files_pattern_verbatim: | 143 | for fp in aux_files_pattern_verbatim: |
| 144 | the_files.append(fp % m.group(1)) | 144 | the_files.append(fp % m.group(1)) |
| 145 | else: | 145 | else: |
| 146 | the_files.append(aux_files_pattern_verbatim % m.group(1)) | 146 | the_files.append(aux_files_pattern_verbatim % m.group(1)) |
| 147 | d.setVar('FILES_' + pkg, " ".join(the_files)) | 147 | d.setVar('FILES_' + pkg, " ".join(the_files)) |
| 148 | if extra_depends != '': | 148 | if extra_depends != '': |
| 149 | d.appendVar('RDEPENDS_' + pkg, ' ' + extra_depends) | 149 | d.appendVar('RDEPENDS_' + pkg, ' ' + extra_depends) |
| 150 | d.setVar('DESCRIPTION_' + pkg, description % on) | 150 | d.setVar('DESCRIPTION_' + pkg, description % on) |
| 151 | if postinst: | 151 | if postinst: |
| 152 | d.setVar('pkg_postinst_' + pkg, postinst) | 152 | d.setVar('pkg_postinst_' + pkg, postinst) |
| 153 | if postrm: | 153 | if postrm: |
| 154 | d.setVar('pkg_postrm_' + pkg, postrm) | 154 | d.setVar('pkg_postrm_' + pkg, postrm) |
| 155 | else: | 155 | else: |
| 156 | d.setVar('FILES_' + pkg, oldfiles + " " + os.path.join(root, o)) | 156 | d.setVar('FILES_' + pkg, oldfiles + " " + os.path.join(root, o)) |
| 157 | if callable(hook): | 157 | if callable(hook): |
| 158 | hook(f, pkg, file_regex, output_pattern, m.group(1)) | 158 | hook(f, pkg, file_regex, output_pattern, m.group(1)) |
| 159 | 159 | ||
| 160 | d.setVar('PACKAGES', ' '.join(packages)) | 160 | d.setVar('PACKAGES', ' '.join(packages)) |
| 161 | 161 | ||
| 162 | PACKAGE_DEPENDS += "file-native" | 162 | PACKAGE_DEPENDS += "file-native" |
| 163 | 163 | ||
| @@ -195,7 +195,7 @@ def splitfile(file, debugfile, debugsrcdir, d): | |||
| 195 | 195 | ||
| 196 | # We ignore kernel modules, we don't generate debug info files. | 196 | # We ignore kernel modules, we don't generate debug info files. |
| 197 | if file.find("/lib/modules/") != -1 and file.endswith(".ko"): | 197 | if file.find("/lib/modules/") != -1 and file.endswith(".ko"): |
| 198 | return 1 | 198 | return 1 |
| 199 | 199 | ||
| 200 | newmode = None | 200 | newmode = None |
| 201 | if not os.access(file, os.W_OK) or os.access(file, os.R_OK): | 201 | if not os.access(file, os.W_OK) or os.access(file, os.R_OK): |
| @@ -205,7 +205,7 @@ def splitfile(file, debugfile, debugsrcdir, d): | |||
| 205 | 205 | ||
| 206 | # We need to extract the debug src information here... | 206 | # We need to extract the debug src information here... |
| 207 | if debugsrcdir: | 207 | if debugsrcdir: |
| 208 | subprocess.call("%s'%s' -b '%s' -d '%s' -i -l '%s' '%s'" % (pathprefix, debugedit, workparentdir, debugsrcdir, sourcefile, file), shell=True) | 208 | subprocess.call("%s'%s' -b '%s' -d '%s' -i -l '%s' '%s'" % (pathprefix, debugedit, workparentdir, debugsrcdir, sourcefile, file), shell=True) |
| 209 | 209 | ||
| 210 | bb.mkdirhier(os.path.dirname(debugfile)) | 210 | bb.mkdirhier(os.path.dirname(debugfile)) |
| 211 | 211 | ||
| @@ -316,826 +316,826 @@ def runstrip(file, elftype, d): | |||
| 316 | # | 316 | # |
| 317 | 317 | ||
| 318 | def get_package_mapping (pkg, d): | 318 | def get_package_mapping (pkg, d): |
| 319 | import oe.packagedata | 319 | import oe.packagedata |
| 320 | 320 | ||
| 321 | data = oe.packagedata.read_subpkgdata(pkg, d) | 321 | data = oe.packagedata.read_subpkgdata(pkg, d) |
| 322 | key = "PKG_%s" % pkg | 322 | key = "PKG_%s" % pkg |
| 323 | 323 | ||
| 324 | if key in data: | 324 | if key in data: |
| 325 | return data[key] | 325 | return data[key] |
| 326 | 326 | ||
| 327 | return pkg | 327 | return pkg |
| 328 | 328 | ||
| 329 | def runtime_mapping_rename (varname, d): | 329 | def runtime_mapping_rename (varname, d): |
| 330 | #bb.note("%s before: %s" % (varname, d.getVar(varname, True))) | 330 | #bb.note("%s before: %s" % (varname, d.getVar(varname, True))) |
| 331 | 331 | ||
| 332 | new_depends = [] | 332 | new_depends = [] |
| 333 | deps = bb.utils.explode_dep_versions(d.getVar(varname, True) or "") | 333 | deps = bb.utils.explode_dep_versions(d.getVar(varname, True) or "") |
| 334 | for depend in deps: | 334 | for depend in deps: |
| 335 | # Have to be careful with any version component of the depend | 335 | # Have to be careful with any version component of the depend |
| 336 | new_depend = get_package_mapping(depend, d) | 336 | new_depend = get_package_mapping(depend, d) |
| 337 | if deps[depend]: | 337 | if deps[depend]: |
| 338 | new_depends.append("%s (%s)" % (new_depend, deps[depend])) | 338 | new_depends.append("%s (%s)" % (new_depend, deps[depend])) |
| 339 | else: | 339 | else: |
| 340 | new_depends.append(new_depend) | 340 | new_depends.append(new_depend) |
| 341 | 341 | ||
| 342 | d.setVar(varname, " ".join(new_depends) or None) | 342 | d.setVar(varname, " ".join(new_depends) or None) |
| 343 | 343 | ||
| 344 | #bb.note("%s after: %s" % (varname, d.getVar(varname, True))) | 344 | #bb.note("%s after: %s" % (varname, d.getVar(varname, True))) |
| 345 | 345 | ||
| 346 | # | 346 | # |
| 347 | # Package functions suitable for inclusion in PACKAGEFUNCS | 347 | # Package functions suitable for inclusion in PACKAGEFUNCS |
| 348 | # | 348 | # |
| 349 | 349 | ||
| 350 | python package_get_auto_pr() { | 350 | python package_get_auto_pr() { |
| 351 | # per recipe PRSERV_HOST PRSERV_PORT | 351 | # per recipe PRSERV_HOST PRSERV_PORT |
| 352 | pn = d.getVar('PN', True) | 352 | pn = d.getVar('PN', True) |
| 353 | host = d.getVar("PRSERV_HOST_" + pn, True) | 353 | host = d.getVar("PRSERV_HOST_" + pn, True) |
| 354 | port = d.getVar("PRSERV_PORT_" + pn, True) | 354 | port = d.getVar("PRSERV_PORT_" + pn, True) |
| 355 | if not (host is None): | 355 | if not (host is None): |
| 356 | d.setVar("PRSERV_HOST", host) | 356 | d.setVar("PRSERV_HOST", host) |
| 357 | if not (port is None): | 357 | if not (port is None): |
| 358 | d.setVar("PRSERV_PORT", port) | 358 | d.setVar("PRSERV_PORT", port) |
| 359 | if d.getVar('USE_PR_SERV', True) != "0": | 359 | if d.getVar('USE_PR_SERV', True) != "0": |
| 360 | try: | 360 | try: |
| 361 | auto_pr=prserv_get_pr_auto(d) | 361 | auto_pr=prserv_get_pr_auto(d) |
| 362 | except Exception as e: | 362 | except Exception as e: |
| 363 | bb.fatal("Can NOT get PRAUTO, exception %s" % str(e)) | 363 | bb.fatal("Can NOT get PRAUTO, exception %s" % str(e)) |
| 364 | return | 364 | return |
| 365 | if auto_pr is None: | 365 | if auto_pr is None: |
| 366 | if d.getVar('PRSERV_LOCKDOWN', True): | 366 | if d.getVar('PRSERV_LOCKDOWN', True): |
| 367 | bb.fatal("Can NOT get PRAUTO from lockdown exported file") | 367 | bb.fatal("Can NOT get PRAUTO from lockdown exported file") |
| 368 | else: | 368 | else: |
| 369 | bb.fatal("Can NOT get PRAUTO from remote PR service") | 369 | bb.fatal("Can NOT get PRAUTO from remote PR service") |
| 370 | return | 370 | return |
| 371 | d.setVar('PRAUTO',str(auto_pr)) | 371 | d.setVar('PRAUTO',str(auto_pr)) |
| 372 | } | 372 | } |
| 373 | 373 | ||
| 374 | python package_do_split_locales() { | 374 | python package_do_split_locales() { |
| 375 | if (d.getVar('PACKAGE_NO_LOCALE', True) == '1'): | 375 | if (d.getVar('PACKAGE_NO_LOCALE', True) == '1'): |
| 376 | bb.debug(1, "package requested not splitting locales") | 376 | bb.debug(1, "package requested not splitting locales") |
| 377 | return | 377 | return |
| 378 | 378 | ||
| 379 | packages = (d.getVar('PACKAGES', True) or "").split() | 379 | packages = (d.getVar('PACKAGES', True) or "").split() |
| 380 | 380 | ||
| 381 | datadir = d.getVar('datadir', True) | 381 | datadir = d.getVar('datadir', True) |
| 382 | if not datadir: | 382 | if not datadir: |
| 383 | bb.note("datadir not defined") | 383 | bb.note("datadir not defined") |
| 384 | return | 384 | return |
| 385 | 385 | ||
| 386 | dvar = d.getVar('PKGD', True) | 386 | dvar = d.getVar('PKGD', True) |
| 387 | pn = d.getVar('PN', True) | 387 | pn = d.getVar('PN', True) |
| 388 | 388 | ||
| 389 | if pn + '-locale' in packages: | 389 | if pn + '-locale' in packages: |
| 390 | packages.remove(pn + '-locale') | 390 | packages.remove(pn + '-locale') |
| 391 | 391 | ||
| 392 | localedir = os.path.join(dvar + datadir, 'locale') | 392 | localedir = os.path.join(dvar + datadir, 'locale') |
| 393 | 393 | ||
| 394 | if not os.path.isdir(localedir): | 394 | if not os.path.isdir(localedir): |
| 395 | bb.debug(1, "No locale files in this package") | 395 | bb.debug(1, "No locale files in this package") |
| 396 | return | 396 | return |
| 397 | 397 | ||
| 398 | locales = os.listdir(localedir) | 398 | locales = os.listdir(localedir) |
| 399 | 399 | ||
| 400 | summary = d.getVar('SUMMARY', True) or pn | 400 | summary = d.getVar('SUMMARY', True) or pn |
| 401 | description = d.getVar('DESCRIPTION', True) or "" | 401 | description = d.getVar('DESCRIPTION', True) or "" |
| 402 | locale_section = d.getVar('LOCALE_SECTION', True) | 402 | locale_section = d.getVar('LOCALE_SECTION', True) |
| 403 | mlprefix = d.getVar('MLPREFIX', True) or "" | 403 | mlprefix = d.getVar('MLPREFIX', True) or "" |
| 404 | for l in sorted(locales): | 404 | for l in sorted(locales): |
| 405 | ln = legitimize_package_name(l) | 405 | ln = legitimize_package_name(l) |
| 406 | pkg = pn + '-locale-' + ln | 406 | pkg = pn + '-locale-' + ln |
| 407 | packages.append(pkg) | 407 | packages.append(pkg) |
| 408 | d.setVar('FILES_' + pkg, os.path.join(datadir, 'locale', l)) | 408 | d.setVar('FILES_' + pkg, os.path.join(datadir, 'locale', l)) |
| 409 | d.setVar('RDEPENDS_' + pkg, '%s %svirtual-locale-%s' % (pn, mlprefix, ln)) | 409 | d.setVar('RDEPENDS_' + pkg, '%s %svirtual-locale-%s' % (pn, mlprefix, ln)) |
| 410 | d.setVar('RPROVIDES_' + pkg, '%s-locale %s%s-translation' % (pn, mlprefix, ln)) | 410 | d.setVar('RPROVIDES_' + pkg, '%s-locale %s%s-translation' % (pn, mlprefix, ln)) |
| 411 | d.setVar('SUMMARY_' + pkg, '%s - %s translations' % (summary, l)) | 411 | d.setVar('SUMMARY_' + pkg, '%s - %s translations' % (summary, l)) |
| 412 | d.setVar('DESCRIPTION_' + pkg, '%s This package contains language translation files for the %s locale.' % (description, l)) | 412 | d.setVar('DESCRIPTION_' + pkg, '%s This package contains language translation files for the %s locale.' % (description, l)) |
| 413 | if locale_section: | 413 | if locale_section: |
| 414 | d.setVar('SECTION_' + pkg, locale_section) | 414 | d.setVar('SECTION_' + pkg, locale_section) |
| 415 | 415 | ||
| 416 | d.setVar('PACKAGES', ' '.join(packages)) | 416 | d.setVar('PACKAGES', ' '.join(packages)) |
| 417 | 417 | ||
| 418 | # Disabled by RP 18/06/07 | 418 | # Disabled by RP 18/06/07 |
| 419 | # Wildcards aren't supported in debian | 419 | # Wildcards aren't supported in debian |
| 420 | # They break with ipkg since glibc-locale* will mean that | 420 | # They break with ipkg since glibc-locale* will mean that |
| 421 | # glibc-localedata-translit* won't install as a dependency | 421 | # glibc-localedata-translit* won't install as a dependency |
| 422 | # for some other package which breaks meta-toolchain | 422 | # for some other package which breaks meta-toolchain |
| 423 | # Probably breaks since virtual-locale- isn't provided anywhere | 423 | # Probably breaks since virtual-locale- isn't provided anywhere |
| 424 | #rdep = (d.getVar('RDEPENDS_%s' % pn, True) or d.getVar('RDEPENDS', True) or "").split() | 424 | #rdep = (d.getVar('RDEPENDS_%s' % pn, True) or d.getVar('RDEPENDS', True) or "").split() |
| 425 | #rdep.append('%s-locale*' % pn) | 425 | #rdep.append('%s-locale*' % pn) |
| 426 | #d.setVar('RDEPENDS_%s' % pn, ' '.join(rdep)) | 426 | #d.setVar('RDEPENDS_%s' % pn, ' '.join(rdep)) |
| 427 | } | 427 | } |
| 428 | 428 | ||
| 429 | python perform_packagecopy () { | 429 | python perform_packagecopy () { |
| 430 | import subprocess | 430 | import subprocess |
| 431 | dest = d.getVar('D', True) | 431 | dest = d.getVar('D', True) |
| 432 | dvar = d.getVar('PKGD', True) | 432 | dvar = d.getVar('PKGD', True) |
| 433 | 433 | ||
| 434 | bb.mkdirhier(dvar) | 434 | bb.mkdirhier(dvar) |
| 435 | 435 | ||
| 436 | # Start by package population by taking a copy of the installed | 436 | # Start by package population by taking a copy of the installed |
| 437 | # files to operate on | 437 | # files to operate on |
| 438 | subprocess.call('rm -rf %s/*' % (dvar), shell=True) | 438 | subprocess.call('rm -rf %s/*' % (dvar), shell=True) |
| 439 | # Preserve sparse files and hard links | 439 | # Preserve sparse files and hard links |
| 440 | subprocess.call('tar -cf - -C %s -ps . | tar -xf - -C %s' % (dest, dvar), shell=True) | 440 | subprocess.call('tar -cf - -C %s -ps . | tar -xf - -C %s' % (dest, dvar), shell=True) |
| 441 | } | 441 | } |
| 442 | 442 | ||
| 443 | # We generate a master list of directories to process, we start by | 443 | # We generate a master list of directories to process, we start by |
| 444 | # seeding this list with reasonable defaults, then load from | 444 | # seeding this list with reasonable defaults, then load from |
| 445 | # the fs-perms.txt files | 445 | # the fs-perms.txt files |
| 446 | python fixup_perms () { | 446 | python fixup_perms () { |
| 447 | import os, pwd, grp | 447 | import os, pwd, grp |
| 448 | 448 | ||
| 449 | # init using a string with the same format as a line as documented in | 449 | # init using a string with the same format as a line as documented in |
| 450 | # the fs-perms.txt file | 450 | # the fs-perms.txt file |
| 451 | # <path> <mode> <uid> <gid> <walk> <fmode> <fuid> <fgid> | 451 | # <path> <mode> <uid> <gid> <walk> <fmode> <fuid> <fgid> |
| 452 | # <path> link <link target> | 452 | # <path> link <link target> |
| 453 | # | 453 | # |
| 454 | # __str__ can be used to print out an entry in the input format | 454 | # __str__ can be used to print out an entry in the input format |
| 455 | # | 455 | # |
| 456 | # if fs_perms_entry.path is None: | 456 | # if fs_perms_entry.path is None: |
| 457 | # an error occured | 457 | # an error occured |
| 458 | # if fs_perms_entry.link, you can retrieve: | 458 | # if fs_perms_entry.link, you can retrieve: |
| 459 | # fs_perms_entry.path = path | 459 | # fs_perms_entry.path = path |
| 460 | # fs_perms_entry.link = target of link | 460 | # fs_perms_entry.link = target of link |
| 461 | # if not fs_perms_entry.link, you can retrieve: | 461 | # if not fs_perms_entry.link, you can retrieve: |
| 462 | # fs_perms_entry.path = path | 462 | # fs_perms_entry.path = path |
| 463 | # fs_perms_entry.mode = expected dir mode or None | 463 | # fs_perms_entry.mode = expected dir mode or None |
| 464 | # fs_perms_entry.uid = expected uid or -1 | 464 | # fs_perms_entry.uid = expected uid or -1 |
| 465 | # fs_perms_entry.gid = expected gid or -1 | 465 | # fs_perms_entry.gid = expected gid or -1 |
| 466 | # fs_perms_entry.walk = 'true' or something else | 466 | # fs_perms_entry.walk = 'true' or something else |
| 467 | # fs_perms_entry.fmode = expected file mode or None | 467 | # fs_perms_entry.fmode = expected file mode or None |
| 468 | # fs_perms_entry.fuid = expected file uid or -1 | 468 | # fs_perms_entry.fuid = expected file uid or -1 |
| 469 | # fs_perms_entry_fgid = expected file gid or -1 | 469 | # fs_perms_entry_fgid = expected file gid or -1 |
| 470 | class fs_perms_entry(): | 470 | class fs_perms_entry(): |
| 471 | def __init__(self, line): | 471 | def __init__(self, line): |
| 472 | lsplit = line.split() | 472 | lsplit = line.split() |
| 473 | if len(lsplit) == 3 and lsplit[1].lower() == "link": | 473 | if len(lsplit) == 3 and lsplit[1].lower() == "link": |
| 474 | self._setlink(lsplit[0], lsplit[2]) | 474 | self._setlink(lsplit[0], lsplit[2]) |
| 475 | elif len(lsplit) == 8: | 475 | elif len(lsplit) == 8: |
| 476 | self._setdir(lsplit[0], lsplit[1], lsplit[2], lsplit[3], lsplit[4], lsplit[5], lsplit[6], lsplit[7]) | 476 | self._setdir(lsplit[0], lsplit[1], lsplit[2], lsplit[3], lsplit[4], lsplit[5], lsplit[6], lsplit[7]) |
| 477 | else: | 477 | else: |
| 478 | bb.error("Fixup Perms: invalid config line %s" % line) | 478 | bb.error("Fixup Perms: invalid config line %s" % line) |
| 479 | self.path = None | 479 | self.path = None |
| 480 | self.link = None | 480 | self.link = None |
| 481 | 481 | ||
| 482 | def _setdir(self, path, mode, uid, gid, walk, fmode, fuid, fgid): | 482 | def _setdir(self, path, mode, uid, gid, walk, fmode, fuid, fgid): |
| 483 | self.path = os.path.normpath(path) | 483 | self.path = os.path.normpath(path) |
| 484 | self.link = None | 484 | self.link = None |
| 485 | self.mode = self._procmode(mode) | 485 | self.mode = self._procmode(mode) |
| 486 | self.uid = self._procuid(uid) | 486 | self.uid = self._procuid(uid) |
| 487 | self.gid = self._procgid(gid) | 487 | self.gid = self._procgid(gid) |
| 488 | self.walk = walk.lower() | 488 | self.walk = walk.lower() |
| 489 | self.fmode = self._procmode(fmode) | 489 | self.fmode = self._procmode(fmode) |
| 490 | self.fuid = self._procuid(fuid) | 490 | self.fuid = self._procuid(fuid) |
| 491 | self.fgid = self._procgid(fgid) | 491 | self.fgid = self._procgid(fgid) |
| 492 | 492 | ||
| 493 | def _setlink(self, path, link): | 493 | def _setlink(self, path, link): |
| 494 | self.path = os.path.normpath(path) | 494 | self.path = os.path.normpath(path) |
| 495 | self.link = link | 495 | self.link = link |
| 496 | 496 | ||
| 497 | def _procmode(self, mode): | 497 | def _procmode(self, mode): |
| 498 | if not mode or (mode and mode == "-"): | 498 | if not mode or (mode and mode == "-"): |
| 499 | return None | 499 | return None |
| 500 | else: | 500 | else: |
| 501 | return int(mode,8) | 501 | return int(mode,8) |
| 502 | 502 | ||
| 503 | # Note uid/gid -1 has special significance in os.lchown | 503 | # Note uid/gid -1 has special significance in os.lchown |
| 504 | def _procuid(self, uid): | 504 | def _procuid(self, uid): |
| 505 | if uid is None or uid == "-": | 505 | if uid is None or uid == "-": |
| 506 | return -1 | 506 | return -1 |
| 507 | elif uid.isdigit(): | 507 | elif uid.isdigit(): |
| 508 | return int(uid) | 508 | return int(uid) |
| 509 | else: | 509 | else: |
| 510 | return pwd.getpwnam(uid).pw_uid | 510 | return pwd.getpwnam(uid).pw_uid |
| 511 | 511 | ||
| 512 | def _procgid(self, gid): | 512 | def _procgid(self, gid): |
| 513 | if gid is None or gid == "-": | 513 | if gid is None or gid == "-": |
| 514 | return -1 | 514 | return -1 |
| 515 | elif gid.isdigit(): | 515 | elif gid.isdigit(): |
| 516 | return int(gid) | 516 | return int(gid) |
| 517 | else: | 517 | else: |
| 518 | return grp.getgrnam(gid).gr_gid | 518 | return grp.getgrnam(gid).gr_gid |
| 519 | 519 | ||
| 520 | # Use for debugging the entries | 520 | # Use for debugging the entries |
| 521 | def __str__(self): | 521 | def __str__(self): |
| 522 | if self.link: | 522 | if self.link: |
| 523 | return "%s link %s" % (self.path, self.link) | 523 | return "%s link %s" % (self.path, self.link) |
| 524 | else: | 524 | else: |
| 525 | mode = "-" | 525 | mode = "-" |
| 526 | if self.mode: | 526 | if self.mode: |
| 527 | mode = "0%o" % self.mode | 527 | mode = "0%o" % self.mode |
| 528 | fmode = "-" | 528 | fmode = "-" |
| 529 | if self.fmode: | 529 | if self.fmode: |
| 530 | fmode = "0%o" % self.fmode | 530 | fmode = "0%o" % self.fmode |
| 531 | uid = self._mapugid(self.uid) | 531 | uid = self._mapugid(self.uid) |
| 532 | gid = self._mapugid(self.gid) | 532 | gid = self._mapugid(self.gid) |
| 533 | fuid = self._mapugid(self.fuid) | 533 | fuid = self._mapugid(self.fuid) |
| 534 | fgid = self._mapugid(self.fgid) | 534 | fgid = self._mapugid(self.fgid) |
| 535 | return "%s %s %s %s %s %s %s %s" % (self.path, mode, uid, gid, self.walk, fmode, fuid, fgid) | 535 | return "%s %s %s %s %s %s %s %s" % (self.path, mode, uid, gid, self.walk, fmode, fuid, fgid) |
| 536 | 536 | ||
| 537 | def _mapugid(self, id): | 537 | def _mapugid(self, id): |
| 538 | if id is None or id == -1: | 538 | if id is None or id == -1: |
| 539 | return "-" | 539 | return "-" |
| 540 | else: | 540 | else: |
| 541 | return "%d" % id | 541 | return "%d" % id |
| 542 | 542 | ||
| 543 | # Fix the permission, owner and group of path | 543 | # Fix the permission, owner and group of path |
| 544 | def fix_perms(path, mode, uid, gid, dir): | 544 | def fix_perms(path, mode, uid, gid, dir): |
| 545 | if mode and not os.path.islink(path): | 545 | if mode and not os.path.islink(path): |
| 546 | #bb.note("Fixup Perms: chmod 0%o %s" % (mode, dir)) | 546 | #bb.note("Fixup Perms: chmod 0%o %s" % (mode, dir)) |
| 547 | os.chmod(path, mode) | 547 | os.chmod(path, mode) |
| 548 | # -1 is a special value that means don't change the uid/gid | 548 | # -1 is a special value that means don't change the uid/gid |
| 549 | # if they are BOTH -1, don't bother to lchown | 549 | # if they are BOTH -1, don't bother to lchown |
| 550 | if not (uid == -1 and gid == -1): | 550 | if not (uid == -1 and gid == -1): |
| 551 | #bb.note("Fixup Perms: lchown %d:%d %s" % (uid, gid, dir)) | 551 | #bb.note("Fixup Perms: lchown %d:%d %s" % (uid, gid, dir)) |
| 552 | os.lchown(path, uid, gid) | 552 | os.lchown(path, uid, gid) |
| 553 | 553 | ||
| 554 | # Return a list of configuration files based on either the default | 554 | # Return a list of configuration files based on either the default |
| 555 | # files/fs-perms.txt or the contents of FILESYSTEM_PERMS_TABLES | 555 | # files/fs-perms.txt or the contents of FILESYSTEM_PERMS_TABLES |
| 556 | # paths are resolved via BBPATH | 556 | # paths are resolved via BBPATH |
| 557 | def get_fs_perms_list(d): | 557 | def get_fs_perms_list(d): |
| 558 | str = "" | 558 | str = "" |
| 559 | fs_perms_tables = d.getVar('FILESYSTEM_PERMS_TABLES', True) | 559 | fs_perms_tables = d.getVar('FILESYSTEM_PERMS_TABLES', True) |
| 560 | if not fs_perms_tables: | 560 | if not fs_perms_tables: |
| 561 | fs_perms_tables = 'files/fs-perms.txt' | 561 | fs_perms_tables = 'files/fs-perms.txt' |
| 562 | for conf_file in fs_perms_tables.split(): | 562 | for conf_file in fs_perms_tables.split(): |
| 563 | str += " %s" % bb.which(d.getVar('BBPATH', True), conf_file) | 563 | str += " %s" % bb.which(d.getVar('BBPATH', True), conf_file) |
| 564 | return str | 564 | return str |
| 565 | 565 | ||
| 566 | 566 | ||
| 567 | 567 | ||
| 568 | dvar = d.getVar('PKGD', True) | 568 | dvar = d.getVar('PKGD', True) |
| 569 | 569 | ||
| 570 | fs_perms_table = {} | 570 | fs_perms_table = {} |
| 571 | 571 | ||
| 572 | # By default all of the standard directories specified in | 572 | # By default all of the standard directories specified in |
| 573 | # bitbake.conf will get 0755 root:root. | 573 | # bitbake.conf will get 0755 root:root. |
| 574 | target_path_vars = [ 'base_prefix', | 574 | target_path_vars = [ 'base_prefix', |
| 575 | 'prefix', | 575 | 'prefix', |
| 576 | 'exec_prefix', | 576 | 'exec_prefix', |
| 577 | 'base_bindir', | 577 | 'base_bindir', |
| 578 | 'base_sbindir', | 578 | 'base_sbindir', |
| 579 | 'base_libdir', | 579 | 'base_libdir', |
| 580 | 'datadir', | 580 | 'datadir', |
| 581 | 'sysconfdir', | 581 | 'sysconfdir', |
| 582 | 'servicedir', | 582 | 'servicedir', |
| 583 | 'sharedstatedir', | 583 | 'sharedstatedir', |
| 584 | 'localstatedir', | 584 | 'localstatedir', |
| 585 | 'infodir', | 585 | 'infodir', |
| 586 | 'mandir', | 586 | 'mandir', |
| 587 | 'docdir', | 587 | 'docdir', |
| 588 | 'bindir', | 588 | 'bindir', |
| 589 | 'sbindir', | 589 | 'sbindir', |
| 590 | 'libexecdir', | 590 | 'libexecdir', |
| 591 | 'libdir', | 591 | 'libdir', |
| 592 | 'includedir', | 592 | 'includedir', |
| 593 | 'oldincludedir' ] | 593 | 'oldincludedir' ] |
| 594 | 594 | ||
| 595 | for path in target_path_vars: | 595 | for path in target_path_vars: |
| 596 | dir = d.getVar(path, True) or "" | 596 | dir = d.getVar(path, True) or "" |
| 597 | if dir == "": | 597 | if dir == "": |
| 598 | continue | 598 | continue |
| 599 | fs_perms_table[dir] = fs_perms_entry(bb.data.expand("%s 0755 root root false - - -" % (dir), d)) | 599 | fs_perms_table[dir] = fs_perms_entry(bb.data.expand("%s 0755 root root false - - -" % (dir), d)) |
| 600 | 600 | ||
| 601 | # Now we actually load from the configuration files | 601 | # Now we actually load from the configuration files |
| 602 | for conf in get_fs_perms_list(d).split(): | 602 | for conf in get_fs_perms_list(d).split(): |
| 603 | if os.path.exists(conf): | 603 | if os.path.exists(conf): |
| 604 | f = open(conf) | 604 | f = open(conf) |
| 605 | for line in f: | 605 | for line in f: |
| 606 | if line.startswith('#'): | 606 | if line.startswith('#'): |
| 607 | continue | 607 | continue |
| 608 | lsplit = line.split() | 608 | lsplit = line.split() |
| 609 | if len(lsplit) == 0: | 609 | if len(lsplit) == 0: |
| 610 | continue | 610 | continue |
| 611 | if len(lsplit) != 8 and not (len(lsplit) == 3 and lsplit[1].lower() == "link"): | 611 | if len(lsplit) != 8 and not (len(lsplit) == 3 and lsplit[1].lower() == "link"): |
| 612 | bb.error("Fixup perms: %s invalid line: %s" % (conf, line)) | 612 | bb.error("Fixup perms: %s invalid line: %s" % (conf, line)) |
| 613 | continue | 613 | continue |
| 614 | entry = fs_perms_entry(d.expand(line)) | 614 | entry = fs_perms_entry(d.expand(line)) |
| 615 | if entry and entry.path: | 615 | if entry and entry.path: |
| 616 | fs_perms_table[entry.path] = entry | 616 | fs_perms_table[entry.path] = entry |
| 617 | f.close() | 617 | f.close() |
| 618 | 618 | ||
| 619 | # Debug -- list out in-memory table | 619 | # Debug -- list out in-memory table |
| 620 | #for dir in fs_perms_table: | 620 | #for dir in fs_perms_table: |
| 621 | # bb.note("Fixup Perms: %s: %s" % (dir, str(fs_perms_table[dir]))) | 621 | # bb.note("Fixup Perms: %s: %s" % (dir, str(fs_perms_table[dir]))) |
| 622 | 622 | ||
| 623 | # We process links first, so we can go back and fixup directory ownership | 623 | # We process links first, so we can go back and fixup directory ownership |
| 624 | # for any newly created directories | 624 | # for any newly created directories |
| 625 | for dir in fs_perms_table: | 625 | for dir in fs_perms_table: |
| 626 | if not fs_perms_table[dir].link: | 626 | if not fs_perms_table[dir].link: |
| 627 | continue | 627 | continue |
| 628 | 628 | ||
| 629 | origin = dvar + dir | 629 | origin = dvar + dir |
| 630 | if not (os.path.exists(origin) and os.path.isdir(origin) and not os.path.islink(origin)): | 630 | if not (os.path.exists(origin) and os.path.isdir(origin) and not os.path.islink(origin)): |
| 631 | continue | 631 | continue |
| 632 | 632 | ||
| 633 | link = fs_perms_table[dir].link | 633 | link = fs_perms_table[dir].link |
| 634 | if link[0] == "/": | 634 | if link[0] == "/": |
| 635 | target = dvar + link | 635 | target = dvar + link |
| 636 | ptarget = link | 636 | ptarget = link |
| 637 | else: | 637 | else: |
| 638 | target = os.path.join(os.path.dirname(origin), link) | 638 | target = os.path.join(os.path.dirname(origin), link) |
| 639 | ptarget = os.path.join(os.path.dirname(dir), link) | 639 | ptarget = os.path.join(os.path.dirname(dir), link) |
| 640 | if os.path.exists(target): | 640 | if os.path.exists(target): |
| 641 | bb.error("Fixup Perms: Unable to correct directory link, target already exists: %s -> %s" % (dir, ptarget)) | 641 | bb.error("Fixup Perms: Unable to correct directory link, target already exists: %s -> %s" % (dir, ptarget)) |
| 642 | continue | 642 | continue |
| 643 | 643 | ||
| 644 | # Create path to move directory to, move it, and then setup the symlink | 644 | # Create path to move directory to, move it, and then setup the symlink |
| 645 | bb.mkdirhier(os.path.dirname(target)) | 645 | bb.mkdirhier(os.path.dirname(target)) |
| 646 | #bb.note("Fixup Perms: Rename %s -> %s" % (dir, ptarget)) | 646 | #bb.note("Fixup Perms: Rename %s -> %s" % (dir, ptarget)) |
| 647 | os.rename(origin, target) | 647 | os.rename(origin, target) |
| 648 | #bb.note("Fixup Perms: Link %s -> %s" % (dir, link)) | 648 | #bb.note("Fixup Perms: Link %s -> %s" % (dir, link)) |
| 649 | os.symlink(link, origin) | 649 | os.symlink(link, origin) |
| 650 | 650 | ||
| 651 | for dir in fs_perms_table: | 651 | for dir in fs_perms_table: |
| 652 | if fs_perms_table[dir].link: | 652 | if fs_perms_table[dir].link: |
| 653 | continue | 653 | continue |
| 654 | 654 | ||
| 655 | origin = dvar + dir | 655 | origin = dvar + dir |
| 656 | if not (os.path.exists(origin) and os.path.isdir(origin)): | 656 | if not (os.path.exists(origin) and os.path.isdir(origin)): |
| 657 | continue | 657 | continue |
| 658 | 658 | ||
| 659 | fix_perms(origin, fs_perms_table[dir].mode, fs_perms_table[dir].uid, fs_perms_table[dir].gid, dir) | 659 | fix_perms(origin, fs_perms_table[dir].mode, fs_perms_table[dir].uid, fs_perms_table[dir].gid, dir) |
| 660 | 660 | ||
| 661 | if fs_perms_table[dir].walk == 'true': | 661 | if fs_perms_table[dir].walk == 'true': |
| 662 | for root, dirs, files in os.walk(origin): | 662 | for root, dirs, files in os.walk(origin): |
| 663 | for dr in dirs: | 663 | for dr in dirs: |
| 664 | each_dir = os.path.join(root, dr) | 664 | each_dir = os.path.join(root, dr) |
| 665 | fix_perms(each_dir, fs_perms_table[dir].mode, fs_perms_table[dir].uid, fs_perms_table[dir].gid, dir) | 665 | fix_perms(each_dir, fs_perms_table[dir].mode, fs_perms_table[dir].uid, fs_perms_table[dir].gid, dir) |
| 666 | for f in files: | 666 | for f in files: |
| 667 | each_file = os.path.join(root, f) | 667 | each_file = os.path.join(root, f) |
| 668 | fix_perms(each_file, fs_perms_table[dir].fmode, fs_perms_table[dir].fuid, fs_perms_table[dir].fgid, dir) | 668 | fix_perms(each_file, fs_perms_table[dir].fmode, fs_perms_table[dir].fuid, fs_perms_table[dir].fgid, dir) |
| 669 | } | 669 | } |
| 670 | 670 | ||
| 671 | python split_and_strip_files () { | 671 | python split_and_strip_files () { |
| 672 | import commands, stat, errno, subprocess | 672 | import commands, stat, errno, subprocess |
| 673 | 673 | ||
| 674 | dvar = d.getVar('PKGD', True) | 674 | dvar = d.getVar('PKGD', True) |
| 675 | pn = d.getVar('PN', True) | 675 | pn = d.getVar('PN', True) |
| 676 | 676 | ||
| 677 | # We default to '.debug' style | 677 | # We default to '.debug' style |
| 678 | if d.getVar('PACKAGE_DEBUG_SPLIT_STYLE', True) == 'debug-file-directory': | 678 | if d.getVar('PACKAGE_DEBUG_SPLIT_STYLE', True) == 'debug-file-directory': |
| 679 | # Single debug-file-directory style debug info | 679 | # Single debug-file-directory style debug info |
| 680 | debugappend = ".debug" | 680 | debugappend = ".debug" |
| 681 | debugdir = "" | 681 | debugdir = "" |
| 682 | debuglibdir = "/usr/lib/debug" | 682 | debuglibdir = "/usr/lib/debug" |
| 683 | debugsrcdir = "/usr/src/debug" | 683 | debugsrcdir = "/usr/src/debug" |
| 684 | else: | 684 | else: |
| 685 | # Original OE-core, a.k.a. ".debug", style debug info | 685 | # Original OE-core, a.k.a. ".debug", style debug info |
| 686 | debugappend = "" | 686 | debugappend = "" |
| 687 | debugdir = "/.debug" | 687 | debugdir = "/.debug" |
| 688 | debuglibdir = "" | 688 | debuglibdir = "" |
| 689 | debugsrcdir = "/usr/src/debug" | 689 | debugsrcdir = "/usr/src/debug" |
| 690 | 690 | ||
| 691 | os.chdir(dvar) | 691 | os.chdir(dvar) |
| 692 | 692 | ||
| 693 | # Return type (bits): | 693 | # Return type (bits): |
| 694 | # 0 - not elf | 694 | # 0 - not elf |
| 695 | # 1 - ELF | 695 | # 1 - ELF |
| 696 | # 2 - stripped | 696 | # 2 - stripped |
| 697 | # 4 - executable | 697 | # 4 - executable |
| 698 | # 8 - shared library | 698 | # 8 - shared library |
| 699 | def isELF(path): | 699 | def isELF(path): |
| 700 | type = 0 | 700 | type = 0 |
| 701 | pathprefix = "export PATH=%s; " % d.getVar('PATH', True) | 701 | pathprefix = "export PATH=%s; " % d.getVar('PATH', True) |
| 702 | ret, result = commands.getstatusoutput("%sfile '%s'" % (pathprefix, path)) | 702 | ret, result = commands.getstatusoutput("%sfile '%s'" % (pathprefix, path)) |
| 703 | 703 | ||
| 704 | if ret: | 704 | if ret: |
| 705 | bb.error("split_and_strip_files: 'file %s' failed" % path) | 705 | bb.error("split_and_strip_files: 'file %s' failed" % path) |
| 706 | return type | 706 | return type |
| 707 | 707 | ||
| 708 | # Not stripped | 708 | # Not stripped |
| 709 | if "ELF" in result: | 709 | if "ELF" in result: |
| 710 | type |= 1 | 710 | type |= 1 |
| 711 | if "not stripped" not in result: | 711 | if "not stripped" not in result: |
| 712 | type |= 2 | 712 | type |= 2 |
| 713 | if "executable" in result: | 713 | if "executable" in result: |
| 714 | type |= 4 | 714 | type |= 4 |
| 715 | if "shared" in result: | 715 | if "shared" in result: |
| 716 | type |= 8 | 716 | type |= 8 |
| 717 | return type | 717 | return type |
| 718 | 718 | ||
| 719 | 719 | ||
| 720 | # | 720 | # |
| 721 | # First lets figure out all of the files we may have to process ... do this only once! | 721 | # First lets figure out all of the files we may have to process ... do this only once! |
| 722 | # | 722 | # |
| 723 | file_list = {} | 723 | file_list = {} |
| 724 | file_links = {} | 724 | file_links = {} |
| 725 | if (d.getVar('INHIBIT_PACKAGE_DEBUG_SPLIT', True) != '1') and \ | 725 | if (d.getVar('INHIBIT_PACKAGE_DEBUG_SPLIT', True) != '1') and \ |
| 726 | (d.getVar('INHIBIT_PACKAGE_STRIP', True) != '1'): | 726 | (d.getVar('INHIBIT_PACKAGE_STRIP', True) != '1'): |
| 727 | for root, dirs, files in os.walk(dvar): | 727 | for root, dirs, files in os.walk(dvar): |
| 728 | for f in files: | 728 | for f in files: |
| 729 | file = os.path.join(root, f) | 729 | file = os.path.join(root, f) |
| 730 | # Only process files (and symlinks)... Skip files that are obviously debug files | 730 | # Only process files (and symlinks)... Skip files that are obviously debug files |
| 731 | if not (debugappend != "" and file.endswith(debugappend)) and \ | 731 | if not (debugappend != "" and file.endswith(debugappend)) and \ |
| 732 | not (debugdir != "" and debugdir in os.path.dirname(file[len(dvar):])) and \ | 732 | not (debugdir != "" and debugdir in os.path.dirname(file[len(dvar):])) and \ |
| 733 | os.path.isfile(file): | 733 | os.path.isfile(file): |
| 734 | try: | 734 | try: |
| 735 | s = os.stat(file) | 735 | s = os.stat(file) |
| 736 | except OSError, (err, strerror): | 736 | except OSError, (err, strerror): |
| 737 | if err != errno.ENOENT: | 737 | if err != errno.ENOENT: |
| 738 | raise | 738 | raise |
| 739 | # Skip broken symlinks | 739 | # Skip broken symlinks |
| 740 | continue | 740 | continue |
| 741 | # Is the item excutable? Then we need to process it. | 741 | # Is the item excutable? Then we need to process it. |
| 742 | if (s[stat.ST_MODE] & stat.S_IXUSR) or \ | 742 | if (s[stat.ST_MODE] & stat.S_IXUSR) or \ |
| 743 | (s[stat.ST_MODE] & stat.S_IXGRP) or \ | 743 | (s[stat.ST_MODE] & stat.S_IXGRP) or \ |
| 744 | (s[stat.ST_MODE] & stat.S_IXOTH): | 744 | (s[stat.ST_MODE] & stat.S_IXOTH): |
| 745 | # If it's a symlink, and points to an ELF file, we capture the readlink target | 745 | # If it's a symlink, and points to an ELF file, we capture the readlink target |
| 746 | if os.path.islink(file): | 746 | if os.path.islink(file): |
| 747 | target = os.readlink(file) | 747 | target = os.readlink(file) |
| 748 | if not os.path.isabs(target): | 748 | if not os.path.isabs(target): |
| 749 | ltarget = os.path.join(os.path.dirname(file), target) | 749 | ltarget = os.path.join(os.path.dirname(file), target) |
| 750 | else: | 750 | else: |
| 751 | ltarget = target | 751 | ltarget = target |
| 752 | 752 | ||
| 753 | if isELF(ltarget): | 753 | if isELF(ltarget): |
| 754 | #bb.note("Sym: %s (%d)" % (ltarget, isELF(ltarget))) | 754 | #bb.note("Sym: %s (%d)" % (ltarget, isELF(ltarget))) |
| 755 | file_list[file] = "sym: " + target | 755 | file_list[file] = "sym: " + target |
| 756 | continue | 756 | continue |
| 757 | # It's a file (or hardlink), not a link | 757 | # It's a file (or hardlink), not a link |
| 758 | # ...but is it ELF, and is it already stripped? | 758 | # ...but is it ELF, and is it already stripped? |
| 759 | elf_file = isELF(file) | 759 | elf_file = isELF(file) |
| 760 | if elf_file & 1: | 760 | if elf_file & 1: |
| 761 | # Check if it's a hard link to something else | 761 | # Check if it's a hard link to something else |
| 762 | if s.st_nlink > 1: | 762 | if s.st_nlink > 1: |
| 763 | file_reference = "%d_%d" % (s.st_dev, s.st_ino) | 763 | file_reference = "%d_%d" % (s.st_dev, s.st_ino) |
| 764 | # Hard link to something else | 764 | # Hard link to something else |
| 765 | file_list[file] = "hard: " + file_reference | 765 | file_list[file] = "hard: " + file_reference |
| 766 | continue | 766 | continue |
| 767 | 767 | ||
| 768 | file_list[file] = "ELF: %d" % elf_file | 768 | file_list[file] = "ELF: %d" % elf_file |
| 769 | 769 | ||
| 770 | 770 | ||
| 771 | # | 771 | # |
| 772 | # First lets process debug splitting | 772 | # First lets process debug splitting |
| 773 | # | 773 | # |
| 774 | if (d.getVar('INHIBIT_PACKAGE_DEBUG_SPLIT', True) != '1'): | 774 | if (d.getVar('INHIBIT_PACKAGE_DEBUG_SPLIT', True) != '1'): |
| 775 | for file in file_list: | 775 | for file in file_list: |
| 776 | src = file[len(dvar):] | 776 | src = file[len(dvar):] |
| 777 | dest = debuglibdir + os.path.dirname(src) + debugdir + "/" + os.path.basename(src) + debugappend | 777 | dest = debuglibdir + os.path.dirname(src) + debugdir + "/" + os.path.basename(src) + debugappend |
| 778 | fpath = dvar + dest | 778 | fpath = dvar + dest |
| 779 | # Preserve symlinks in debug area... | 779 | # Preserve symlinks in debug area... |
| 780 | if file_list[file].startswith("sym: "): | 780 | if file_list[file].startswith("sym: "): |
| 781 | ltarget = file_list[file][5:] | 781 | ltarget = file_list[file][5:] |
| 782 | lpath = os.path.dirname(ltarget) | 782 | lpath = os.path.dirname(ltarget) |
| 783 | lbase = os.path.basename(ltarget) | 783 | lbase = os.path.basename(ltarget) |
| 784 | ftarget = "" | 784 | ftarget = "" |
| 785 | if lpath and lpath != ".": | 785 | if lpath and lpath != ".": |
| 786 | ftarget += lpath + debugdir + "/" | 786 | ftarget += lpath + debugdir + "/" |
| 787 | ftarget += lbase + debugappend | 787 | ftarget += lbase + debugappend |
| 788 | if lpath.startswith(".."): | 788 | if lpath.startswith(".."): |
| 789 | ftarget = os.path.join("..", ftarget) | 789 | ftarget = os.path.join("..", ftarget) |
| 790 | bb.mkdirhier(os.path.dirname(fpath)) | 790 | bb.mkdirhier(os.path.dirname(fpath)) |
| 791 | #bb.note("Symlink %s -> %s" % (fpath, ftarget)) | 791 | #bb.note("Symlink %s -> %s" % (fpath, ftarget)) |
| 792 | os.symlink(ftarget, fpath) | 792 | os.symlink(ftarget, fpath) |
| 793 | continue | 793 | continue |
| 794 | 794 | ||
| 795 | # Preserve hard links in debug area... | 795 | # Preserve hard links in debug area... |
| 796 | file_reference = "" | 796 | file_reference = "" |
| 797 | if file_list[file].startswith("hard: "): | 797 | if file_list[file].startswith("hard: "): |
| 798 | file_reference = file_list[file][6:] | 798 | file_reference = file_list[file][6:] |
| 799 | if file_reference not in file_links: | 799 | if file_reference not in file_links: |
| 800 | # If this is a new file, add it as a reference, and | 800 | # If this is a new file, add it as a reference, and |
| 801 | # update it's type, so we can fall through and split | 801 | # update it's type, so we can fall through and split |
| 802 | file_list[file] = "ELF: %d" % (isELF(file)) | 802 | file_list[file] = "ELF: %d" % (isELF(file)) |
| 803 | else: | 803 | else: |
| 804 | target = file_links[file_reference][len(dvar):] | 804 | target = file_links[file_reference][len(dvar):] |
| 805 | ftarget = dvar + debuglibdir + os.path.dirname(target) + debugdir + "/" + os.path.basename(target) + debugappend | 805 | ftarget = dvar + debuglibdir + os.path.dirname(target) + debugdir + "/" + os.path.basename(target) + debugappend |
| 806 | bb.mkdirhier(os.path.dirname(fpath)) | 806 | bb.mkdirhier(os.path.dirname(fpath)) |
| 807 | #bb.note("Link %s -> %s" % (fpath, ftarget)) | 807 | #bb.note("Link %s -> %s" % (fpath, ftarget)) |
| 808 | os.link(ftarget, fpath) | 808 | os.link(ftarget, fpath) |
| 809 | continue | 809 | continue |
| 810 | 810 | ||
| 811 | # It's ELF... | 811 | # It's ELF... |
| 812 | if file_list[file].startswith("ELF: "): | 812 | if file_list[file].startswith("ELF: "): |
| 813 | elf_file = int(file_list[file][5:]) | 813 | elf_file = int(file_list[file][5:]) |
| 814 | if elf_file & 2: | 814 | if elf_file & 2: |
| 815 | bb.warn("File '%s' from %s was already stripped, this will prevent future debugging!" % (src, pn)) | 815 | bb.warn("File '%s' from %s was already stripped, this will prevent future debugging!" % (src, pn)) |
| 816 | continue | 816 | continue |
| 817 | 817 | ||
| 818 | # Split the file... | 818 | # Split the file... |
| 819 | bb.mkdirhier(os.path.dirname(fpath)) | 819 | bb.mkdirhier(os.path.dirname(fpath)) |
| 820 | #bb.note("Split %s -> %s" % (file, fpath)) | 820 | #bb.note("Split %s -> %s" % (file, fpath)) |
| 821 | # Only store off the hard link reference if we successfully split! | 821 | # Only store off the hard link reference if we successfully split! |
| 822 | if splitfile(file, fpath, debugsrcdir, d) == 0 and file_reference != "": | 822 | if splitfile(file, fpath, debugsrcdir, d) == 0 and file_reference != "": |
| 823 | file_links[file_reference] = file | 823 | file_links[file_reference] = file |
| 824 | 824 | ||
| 825 | # The above may have generated dangling symlinks, remove them! | 825 | # The above may have generated dangling symlinks, remove them! |
| 826 | # Dangling symlinks are a result of something NOT being split, such as a stripped binary. | 826 | # Dangling symlinks are a result of something NOT being split, such as a stripped binary. |
| 827 | # This should be a rare occurance, but we want to clean up anyway. | 827 | # This should be a rare occurance, but we want to clean up anyway. |
| 828 | for file in file_list: | 828 | for file in file_list: |
| 829 | if file_list[file].startswith("sym: "): | 829 | if file_list[file].startswith("sym: "): |
| 830 | src = file[len(dvar):] | 830 | src = file[len(dvar):] |
| 831 | dest = debuglibdir + os.path.dirname(src) + debugdir + "/" + os.path.basename(src) + debugappend | 831 | dest = debuglibdir + os.path.dirname(src) + debugdir + "/" + os.path.basename(src) + debugappend |
| 832 | fpath = dvar + dest | 832 | fpath = dvar + dest |
| 833 | try: | 833 | try: |
| 834 | s = os.stat(fpath) | 834 | s = os.stat(fpath) |
| 835 | except OSError, (err, strerror): | 835 | except OSError, (err, strerror): |
| 836 | if err != errno.ENOENT: | 836 | if err != errno.ENOENT: |
| 837 | raise | 837 | raise |
| 838 | #bb.note("Remove dangling link %s -> %s" % (fpath, os.readlink(fpath))) | 838 | #bb.note("Remove dangling link %s -> %s" % (fpath, os.readlink(fpath))) |
| 839 | os.unlink(fpath) | 839 | os.unlink(fpath) |
| 840 | # This could leave an empty debug directory laying around | 840 | # This could leave an empty debug directory laying around |
| 841 | # take care of the obvious case... | 841 | # take care of the obvious case... |
| 842 | subprocess.call("rmdir %s 2>/dev/null" % os.path.dirname(fpath), shell=True) | 842 | subprocess.call("rmdir %s 2>/dev/null" % os.path.dirname(fpath), shell=True) |
| 843 | 843 | ||
| 844 | # Process the debugsrcdir if requested... | 844 | # Process the debugsrcdir if requested... |
| 845 | # This copies and places the referenced sources for later debugging... | 845 | # This copies and places the referenced sources for later debugging... |
| 846 | splitfile2(debugsrcdir, d) | 846 | splitfile2(debugsrcdir, d) |
| 847 | # | 847 | # |
| 848 | # End of debug splitting | 848 | # End of debug splitting |
| 849 | # | 849 | # |
| 850 | 850 | ||
| 851 | # | 851 | # |
| 852 | # Now lets go back over things and strip them | 852 | # Now lets go back over things and strip them |
| 853 | # | 853 | # |
| 854 | if (d.getVar('INHIBIT_PACKAGE_STRIP', True) != '1'): | 854 | if (d.getVar('INHIBIT_PACKAGE_STRIP', True) != '1'): |
| 855 | for file in file_list: | 855 | for file in file_list: |
| 856 | if file_list[file].startswith("ELF: "): | 856 | if file_list[file].startswith("ELF: "): |
| 857 | elf_file = int(file_list[file][5:]) | 857 | elf_file = int(file_list[file][5:]) |
| 858 | #bb.note("Strip %s" % file) | 858 | #bb.note("Strip %s" % file) |
| 859 | runstrip(file, elf_file, d) | 859 | runstrip(file, elf_file, d) |
| 860 | 860 | ||
| 861 | 861 | ||
| 862 | if (d.getVar('INHIBIT_PACKAGE_STRIP', True) != '1'): | 862 | if (d.getVar('INHIBIT_PACKAGE_STRIP', True) != '1'): |
| 863 | for root, dirs, files in os.walk(dvar): | 863 | for root, dirs, files in os.walk(dvar): |
| 864 | for f in files: | 864 | for f in files: |
| 865 | if not f.endswith(".ko"): | 865 | if not f.endswith(".ko"): |
| 866 | continue | 866 | continue |
| 867 | runstrip(os.path.join(root, f), None, d) | 867 | runstrip(os.path.join(root, f), None, d) |
| 868 | # | 868 | # |
| 869 | # End of strip | 869 | # End of strip |
| 870 | # | 870 | # |
| 871 | } | 871 | } |
| 872 | 872 | ||
| 873 | python populate_packages () { | 873 | python populate_packages () { |
| 874 | import glob, stat, errno, re, subprocess | 874 | import glob, stat, errno, re, subprocess |
| 875 | 875 | ||
| 876 | workdir = d.getVar('WORKDIR', True) | 876 | workdir = d.getVar('WORKDIR', True) |
| 877 | outdir = d.getVar('DEPLOY_DIR', True) | 877 | outdir = d.getVar('DEPLOY_DIR', True) |
| 878 | dvar = d.getVar('PKGD', True) | 878 | dvar = d.getVar('PKGD', True) |
| 879 | packages = d.getVar('PACKAGES', True) | 879 | packages = d.getVar('PACKAGES', True) |
| 880 | pn = d.getVar('PN', True) | 880 | pn = d.getVar('PN', True) |
| 881 | 881 | ||
| 882 | bb.mkdirhier(outdir) | 882 | bb.mkdirhier(outdir) |
| 883 | os.chdir(dvar) | 883 | os.chdir(dvar) |
| 884 | 884 | ||
| 885 | # Sanity check PACKAGES for duplicates and for LICENSE_EXCLUSION | 885 | # Sanity check PACKAGES for duplicates and for LICENSE_EXCLUSION |
| 886 | # Sanity should be moved to sanity.bbclass once we have the infrastucture | 886 | # Sanity should be moved to sanity.bbclass once we have the infrastucture |
| 887 | package_list = [] | 887 | package_list = [] |
| 888 | 888 | ||
| 889 | for pkg in packages.split(): | 889 | for pkg in packages.split(): |
| 890 | if d.getVar('LICENSE_EXCLUSION-' + pkg, True): | 890 | if d.getVar('LICENSE_EXCLUSION-' + pkg, True): |
| 891 | bb.warn("%s has an incompatible license. Excluding from packaging." % pkg) | 891 | bb.warn("%s has an incompatible license. Excluding from packaging." % pkg) |
| 892 | packages.remove(pkg) | 892 | packages.remove(pkg) |
| 893 | else: | 893 | else: |
| 894 | if pkg in package_list: | 894 | if pkg in package_list: |
| 895 | bb.error("%s is listed in PACKAGES multiple times, this leads to packaging errors." % pkg) | 895 | bb.error("%s is listed in PACKAGES multiple times, this leads to packaging errors." % pkg) |
| 896 | else: | 896 | else: |
| 897 | package_list.append(pkg) | 897 | package_list.append(pkg) |
| 898 | d.setVar('PACKAGES', ' '.join(package_list)) | 898 | d.setVar('PACKAGES', ' '.join(package_list)) |
| 899 | pkgdest = d.getVar('PKGDEST', True) | 899 | pkgdest = d.getVar('PKGDEST', True) |
| 900 | subprocess.call('rm -rf %s' % pkgdest, shell=True) | 900 | subprocess.call('rm -rf %s' % pkgdest, shell=True) |
| 901 | 901 | ||
| 902 | seen = [] | 902 | seen = [] |
| 903 | 903 | ||
| 904 | for pkg in package_list: | 904 | for pkg in package_list: |
| 905 | localdata = bb.data.createCopy(d) | 905 | localdata = bb.data.createCopy(d) |
| 906 | root = os.path.join(pkgdest, pkg) | 906 | root = os.path.join(pkgdest, pkg) |
| 907 | bb.mkdirhier(root) | 907 | bb.mkdirhier(root) |
| 908 | 908 | ||
| 909 | localdata.setVar('PKG', pkg) | 909 | localdata.setVar('PKG', pkg) |
| 910 | overrides = localdata.getVar('OVERRIDES', True) | 910 | overrides = localdata.getVar('OVERRIDES', True) |
| 911 | if not overrides: | 911 | if not overrides: |
| 912 | raise bb.build.FuncFailed('OVERRIDES not defined') | 912 | raise bb.build.FuncFailed('OVERRIDES not defined') |
| 913 | localdata.setVar('OVERRIDES', overrides + ':' + pkg) | 913 | localdata.setVar('OVERRIDES', overrides + ':' + pkg) |
| 914 | bb.data.update_data(localdata) | 914 | bb.data.update_data(localdata) |
| 915 | 915 | ||
| 916 | filesvar = localdata.getVar('FILES', True) or "" | 916 | filesvar = localdata.getVar('FILES', True) or "" |
| 917 | files = filesvar.split() | 917 | files = filesvar.split() |
| 918 | file_links = {} | 918 | file_links = {} |
| 919 | for file in files: | 919 | for file in files: |
| 920 | if file.find("//") != -1: | 920 | if file.find("//") != -1: |
| 921 | bb.warn("FILES variable for package %s contains '//' which is invalid. Attempting to fix this but you should correct the metadata.\n" % pkg) | 921 | bb.warn("FILES variable for package %s contains '//' which is invalid. Attempting to fix this but you should correct the metadata.\n" % pkg) |
| 922 | file.replace("//", "/") | 922 | file.replace("//", "/") |
| 923 | if os.path.isabs(file): | 923 | if os.path.isabs(file): |
| 924 | file = '.' + file | 924 | file = '.' + file |
| 925 | if not os.path.islink(file): | 925 | if not os.path.islink(file): |
| 926 | if os.path.isdir(file): | 926 | if os.path.isdir(file): |
| 927 | newfiles = [ os.path.join(file,x) for x in os.listdir(file) ] | 927 | newfiles = [ os.path.join(file,x) for x in os.listdir(file) ] |
| 928 | if newfiles: | 928 | if newfiles: |
| 929 | files += newfiles | 929 | files += newfiles |
| 930 | continue | 930 | continue |
| 931 | globbed = glob.glob(file) | 931 | globbed = glob.glob(file) |
| 932 | if globbed: | 932 | if globbed: |
| 933 | if [ file ] != globbed: | 933 | if [ file ] != globbed: |
| 934 | files += globbed | 934 | files += globbed |
| 935 | continue | 935 | continue |
| 936 | if (not os.path.islink(file)) and (not os.path.exists(file)): | 936 | if (not os.path.islink(file)) and (not os.path.exists(file)): |
| 937 | continue | 937 | continue |
| 938 | if file in seen: | 938 | if file in seen: |
| 939 | continue | 939 | continue |
| 940 | seen.append(file) | 940 | seen.append(file) |
| 941 | 941 | ||
| 942 | def mkdir(src, dest, p): | 942 | def mkdir(src, dest, p): |
| 943 | src = os.path.join(src, p) | 943 | src = os.path.join(src, p) |
| 944 | dest = os.path.join(dest, p) | 944 | dest = os.path.join(dest, p) |
| 945 | bb.mkdirhier(dest) | 945 | bb.mkdirhier(dest) |
| 946 | fstat = os.stat(src) | 946 | fstat = os.stat(src) |
| 947 | os.chmod(dest, fstat.st_mode) | 947 | os.chmod(dest, fstat.st_mode) |
| 948 | os.chown(dest, fstat.st_uid, fstat.st_gid) | 948 | os.chown(dest, fstat.st_uid, fstat.st_gid) |
| 949 | if p not in seen: | 949 | if p not in seen: |
| 950 | seen.append(p) | 950 | seen.append(p) |
| 951 | 951 | ||
| 952 | def mkdir_recurse(src, dest, paths): | 952 | def mkdir_recurse(src, dest, paths): |
| 953 | while paths.startswith("./"): | 953 | while paths.startswith("./"): |
| 954 | paths = paths[2:] | 954 | paths = paths[2:] |
| 955 | p = "." | 955 | p = "." |
| 956 | for c in paths.split("/"): | 956 | for c in paths.split("/"): |
| 957 | p = os.path.join(p, c) | 957 | p = os.path.join(p, c) |
| 958 | if not os.path.exists(os.path.join(dest, p)): | 958 | if not os.path.exists(os.path.join(dest, p)): |
| 959 | mkdir(src, dest, p) | 959 | mkdir(src, dest, p) |
| 960 | 960 | ||
| 961 | if os.path.isdir(file) and not os.path.islink(file): | 961 | if os.path.isdir(file) and not os.path.islink(file): |
| 962 | mkdir_recurse(dvar, root, file) | 962 | mkdir_recurse(dvar, root, file) |
| 963 | continue | 963 | continue |
| 964 | 964 | ||
| 965 | mkdir_recurse(dvar, root, os.path.dirname(file)) | 965 | mkdir_recurse(dvar, root, os.path.dirname(file)) |
| 966 | fpath = os.path.join(root,file) | 966 | fpath = os.path.join(root,file) |
| 967 | if not os.path.islink(file): | 967 | if not os.path.islink(file): |
| 968 | os.link(file, fpath) | 968 | os.link(file, fpath) |
| 969 | fstat = os.stat(file) | 969 | fstat = os.stat(file) |
| 970 | os.chmod(fpath, fstat.st_mode) | 970 | os.chmod(fpath, fstat.st_mode) |
| 971 | os.chown(fpath, fstat.st_uid, fstat.st_gid) | 971 | os.chown(fpath, fstat.st_uid, fstat.st_gid) |
| 972 | continue | 972 | continue |
| 973 | ret = bb.copyfile(file, fpath) | 973 | ret = bb.copyfile(file, fpath) |
| 974 | if ret is False or ret == 0: | 974 | if ret is False or ret == 0: |
| 975 | raise bb.build.FuncFailed("File population failed") | 975 | raise bb.build.FuncFailed("File population failed") |
| 976 | 976 | ||
| 977 | del localdata | 977 | del localdata |
| 978 | os.chdir(workdir) | 978 | os.chdir(workdir) |
| 979 | 979 | ||
| 980 | unshipped = [] | 980 | unshipped = [] |
| 981 | for root, dirs, files in os.walk(dvar): | 981 | for root, dirs, files in os.walk(dvar): |
| 982 | dir = root[len(dvar):] | 982 | dir = root[len(dvar):] |
| 983 | if not dir: | 983 | if not dir: |
| 984 | dir = os.sep | 984 | dir = os.sep |
| 985 | for f in (files + dirs): | 985 | for f in (files + dirs): |
| 986 | path = os.path.join(dir, f) | 986 | path = os.path.join(dir, f) |
| 987 | if ('.' + path) not in seen: | 987 | if ('.' + path) not in seen: |
| 988 | unshipped.append(path) | 988 | unshipped.append(path) |
| 989 | 989 | ||
| 990 | if unshipped != []: | 990 | if unshipped != []: |
| 991 | msg = pn + ": Files/directories were installed but not shipped" | 991 | msg = pn + ": Files/directories were installed but not shipped" |
| 992 | if "installed_vs_shipped" in (d.getVar('INSANE_SKIP_' + pn, True) or "").split(): | 992 | if "installed_vs_shipped" in (d.getVar('INSANE_SKIP_' + pn, True) or "").split(): |
| 993 | bb.note("Package %s skipping QA tests: installed_vs_shipped" % pn) | 993 | bb.note("Package %s skipping QA tests: installed_vs_shipped" % pn) |
| 994 | else: | 994 | else: |
| 995 | for f in unshipped: | 995 | for f in unshipped: |
| 996 | msg = msg + "\n " + f | 996 | msg = msg + "\n " + f |
| 997 | package_qa_handle_error("installed_vs_shipped", msg, d) | 997 | package_qa_handle_error("installed_vs_shipped", msg, d) |
| 998 | 998 | ||
| 999 | bb.build.exec_func("package_name_hook", d) | 999 | bb.build.exec_func("package_name_hook", d) |
| 1000 | 1000 | ||
| 1001 | for pkg in package_list: | 1001 | for pkg in package_list: |
| 1002 | pkgname = d.getVar('PKG_%s' % pkg, True) | 1002 | pkgname = d.getVar('PKG_%s' % pkg, True) |
| 1003 | if pkgname is None: | 1003 | if pkgname is None: |
| 1004 | d.setVar('PKG_%s' % pkg, pkg) | 1004 | d.setVar('PKG_%s' % pkg, pkg) |
| 1005 | 1005 | ||
| 1006 | dangling_links = {} | 1006 | dangling_links = {} |
| 1007 | pkg_files = {} | 1007 | pkg_files = {} |
| 1008 | for pkg in package_list: | 1008 | for pkg in package_list: |
| 1009 | dangling_links[pkg] = [] | 1009 | dangling_links[pkg] = [] |
| 1010 | pkg_files[pkg] = [] | 1010 | pkg_files[pkg] = [] |
| 1011 | inst_root = os.path.join(pkgdest, pkg) | 1011 | inst_root = os.path.join(pkgdest, pkg) |
| 1012 | for root, dirs, files in os.walk(inst_root): | 1012 | for root, dirs, files in os.walk(inst_root): |
| 1013 | for f in files: | 1013 | for f in files: |
| 1014 | path = os.path.join(root, f) | 1014 | path = os.path.join(root, f) |
| 1015 | rpath = path[len(inst_root):] | 1015 | rpath = path[len(inst_root):] |
| 1016 | pkg_files[pkg].append(rpath) | 1016 | pkg_files[pkg].append(rpath) |
| 1017 | try: | 1017 | try: |
| 1018 | s = os.stat(path) | 1018 | s = os.stat(path) |
| 1019 | except OSError, (err, strerror): | 1019 | except OSError, (err, strerror): |
| 1020 | if err != errno.ENOENT: | 1020 | if err != errno.ENOENT: |
| 1021 | raise | 1021 | raise |
| 1022 | target = os.readlink(path) | 1022 | target = os.readlink(path) |
| 1023 | if target[0] != '/': | 1023 | if target[0] != '/': |
| 1024 | target = os.path.join(root[len(inst_root):], target) | 1024 | target = os.path.join(root[len(inst_root):], target) |
| 1025 | dangling_links[pkg].append(os.path.normpath(target)) | 1025 | dangling_links[pkg].append(os.path.normpath(target)) |
| 1026 | 1026 | ||
| 1027 | for pkg in package_list: | 1027 | for pkg in package_list: |
| 1028 | rdepends = bb.utils.explode_dep_versions(d.getVar('RDEPENDS_' + pkg, True) or d.getVar('RDEPENDS', True) or "") | 1028 | rdepends = bb.utils.explode_dep_versions(d.getVar('RDEPENDS_' + pkg, True) or d.getVar('RDEPENDS', True) or "") |
| 1029 | 1029 | ||
| 1030 | for l in dangling_links[pkg]: | 1030 | for l in dangling_links[pkg]: |
| 1031 | found = False | 1031 | found = False |
| 1032 | bb.debug(1, "%s contains dangling link %s" % (pkg, l)) | 1032 | bb.debug(1, "%s contains dangling link %s" % (pkg, l)) |
| 1033 | for p in package_list: | 1033 | for p in package_list: |
| 1034 | for f in pkg_files[p]: | 1034 | for f in pkg_files[p]: |
| 1035 | if f == l: | 1035 | if f == l: |
| 1036 | found = True | 1036 | found = True |
| 1037 | bb.debug(1, "target found in %s" % p) | 1037 | bb.debug(1, "target found in %s" % p) |
| 1038 | if p == pkg: | 1038 | if p == pkg: |
| 1039 | break | 1039 | break |
| 1040 | if p not in rdepends: | 1040 | if p not in rdepends: |
| 1041 | rdepends[p] = "" | 1041 | rdepends[p] = "" |
| 1042 | break | 1042 | break |
| 1043 | if found == False: | 1043 | if found == False: |
| 1044 | bb.note("%s contains dangling symlink to %s" % (pkg, l)) | 1044 | bb.note("%s contains dangling symlink to %s" % (pkg, l)) |
| 1045 | d.setVar('RDEPENDS_' + pkg, bb.utils.join_deps(rdepends, commasep=False)) | 1045 | d.setVar('RDEPENDS_' + pkg, bb.utils.join_deps(rdepends, commasep=False)) |
| 1046 | } | 1046 | } |
| 1047 | populate_packages[dirs] = "${D}" | 1047 | populate_packages[dirs] = "${D}" |
| 1048 | 1048 | ||
| 1049 | PKGDESTWORK = "${WORKDIR}/pkgdata" | 1049 | PKGDESTWORK = "${WORKDIR}/pkgdata" |
| 1050 | 1050 | ||
| 1051 | python emit_pkgdata() { | 1051 | python emit_pkgdata() { |
| 1052 | from glob import glob | 1052 | from glob import glob |
| 1053 | 1053 | ||
| 1054 | def write_if_exists(f, pkg, var): | 1054 | def write_if_exists(f, pkg, var): |
| 1055 | def encode(str): | 1055 | def encode(str): |
| 1056 | import codecs | 1056 | import codecs |
| 1057 | c = codecs.getencoder("string_escape") | 1057 | c = codecs.getencoder("string_escape") |
| 1058 | return c(str)[0] | 1058 | return c(str)[0] |
| 1059 | 1059 | ||
| 1060 | val = d.getVar('%s_%s' % (var, pkg), True) | 1060 | val = d.getVar('%s_%s' % (var, pkg), True) |
| 1061 | if val: | 1061 | if val: |
| 1062 | f.write('%s_%s: %s\n' % (var, pkg, encode(val))) | 1062 | f.write('%s_%s: %s\n' % (var, pkg, encode(val))) |
| 1063 | return | 1063 | return |
| 1064 | val = d.getVar('%s' % (var), True) | 1064 | val = d.getVar('%s' % (var), True) |
| 1065 | if val: | 1065 | if val: |
| 1066 | f.write('%s: %s\n' % (var, encode(val))) | 1066 | f.write('%s: %s\n' % (var, encode(val))) |
| 1067 | return | 1067 | return |
| 1068 | 1068 | ||
| 1069 | def get_directory_size(dir): | 1069 | def get_directory_size(dir): |
| 1070 | if os.listdir(dir): | 1070 | if os.listdir(dir): |
| 1071 | size = int(os.popen('du -sk %s' % dir).readlines()[0].split('\t')[0]) | 1071 | size = int(os.popen('du -sk %s' % dir).readlines()[0].split('\t')[0]) |
| 1072 | else: | 1072 | else: |
| 1073 | size = 0 | 1073 | size = 0 |
| 1074 | return size | 1074 | return size |
| 1075 | 1075 | ||
| 1076 | packages = d.getVar('PACKAGES', True) | 1076 | packages = d.getVar('PACKAGES', True) |
| 1077 | pkgdest = d.getVar('PKGDEST', True) | 1077 | pkgdest = d.getVar('PKGDEST', True) |
| 1078 | pkgdatadir = d.getVar('PKGDESTWORK', True) | 1078 | pkgdatadir = d.getVar('PKGDESTWORK', True) |
| 1079 | 1079 | ||
| 1080 | # Take shared lock since we're only reading, not writing | 1080 | # Take shared lock since we're only reading, not writing |
| 1081 | lf = bb.utils.lockfile(d.expand("${PACKAGELOCK}"), True) | 1081 | lf = bb.utils.lockfile(d.expand("${PACKAGELOCK}"), True) |
| 1082 | 1082 | ||
| 1083 | data_file = pkgdatadir + d.expand("/${PN}" ) | 1083 | data_file = pkgdatadir + d.expand("/${PN}" ) |
| 1084 | f = open(data_file, 'w') | 1084 | f = open(data_file, 'w') |
| 1085 | f.write("PACKAGES: %s\n" % packages) | 1085 | f.write("PACKAGES: %s\n" % packages) |
| 1086 | f.close() | 1086 | f.close() |
| 1087 | 1087 | ||
| 1088 | workdir = d.getVar('WORKDIR', True) | 1088 | workdir = d.getVar('WORKDIR', True) |
| 1089 | 1089 | ||
| 1090 | for pkg in packages.split(): | 1090 | for pkg in packages.split(): |
| 1091 | subdata_file = pkgdatadir + "/runtime/%s" % pkg | 1091 | subdata_file = pkgdatadir + "/runtime/%s" % pkg |
| 1092 | 1092 | ||
| 1093 | sf = open(subdata_file, 'w') | 1093 | sf = open(subdata_file, 'w') |
| 1094 | write_if_exists(sf, pkg, 'PN') | 1094 | write_if_exists(sf, pkg, 'PN') |
| 1095 | write_if_exists(sf, pkg, 'PV') | 1095 | write_if_exists(sf, pkg, 'PV') |
| 1096 | write_if_exists(sf, pkg, 'PR') | 1096 | write_if_exists(sf, pkg, 'PR') |
| 1097 | write_if_exists(sf, pkg, 'PKGV') | 1097 | write_if_exists(sf, pkg, 'PKGV') |
| 1098 | write_if_exists(sf, pkg, 'PKGR') | 1098 | write_if_exists(sf, pkg, 'PKGR') |
| 1099 | write_if_exists(sf, pkg, 'LICENSE') | 1099 | write_if_exists(sf, pkg, 'LICENSE') |
| 1100 | write_if_exists(sf, pkg, 'DESCRIPTION') | 1100 | write_if_exists(sf, pkg, 'DESCRIPTION') |
| 1101 | write_if_exists(sf, pkg, 'SUMMARY') | 1101 | write_if_exists(sf, pkg, 'SUMMARY') |
| 1102 | write_if_exists(sf, pkg, 'RDEPENDS') | 1102 | write_if_exists(sf, pkg, 'RDEPENDS') |
| 1103 | write_if_exists(sf, pkg, 'RPROVIDES') | 1103 | write_if_exists(sf, pkg, 'RPROVIDES') |
| 1104 | write_if_exists(sf, pkg, 'RRECOMMENDS') | 1104 | write_if_exists(sf, pkg, 'RRECOMMENDS') |
| 1105 | write_if_exists(sf, pkg, 'RSUGGESTS') | 1105 | write_if_exists(sf, pkg, 'RSUGGESTS') |
| 1106 | write_if_exists(sf, pkg, 'RREPLACES') | 1106 | write_if_exists(sf, pkg, 'RREPLACES') |
| 1107 | write_if_exists(sf, pkg, 'RCONFLICTS') | 1107 | write_if_exists(sf, pkg, 'RCONFLICTS') |
| 1108 | write_if_exists(sf, pkg, 'SECTION') | 1108 | write_if_exists(sf, pkg, 'SECTION') |
| 1109 | write_if_exists(sf, pkg, 'PKG') | 1109 | write_if_exists(sf, pkg, 'PKG') |
| 1110 | write_if_exists(sf, pkg, 'ALLOW_EMPTY') | 1110 | write_if_exists(sf, pkg, 'ALLOW_EMPTY') |
| 1111 | write_if_exists(sf, pkg, 'FILES') | 1111 | write_if_exists(sf, pkg, 'FILES') |
| 1112 | write_if_exists(sf, pkg, 'pkg_postinst') | 1112 | write_if_exists(sf, pkg, 'pkg_postinst') |
| 1113 | write_if_exists(sf, pkg, 'pkg_postrm') | 1113 | write_if_exists(sf, pkg, 'pkg_postrm') |
| 1114 | write_if_exists(sf, pkg, 'pkg_preinst') | 1114 | write_if_exists(sf, pkg, 'pkg_preinst') |
| 1115 | write_if_exists(sf, pkg, 'pkg_prerm') | 1115 | write_if_exists(sf, pkg, 'pkg_prerm') |
| 1116 | write_if_exists(sf, pkg, 'FILERPROVIDESFLIST') | 1116 | write_if_exists(sf, pkg, 'FILERPROVIDESFLIST') |
| 1117 | for dfile in (d.getVar('FILERPROVIDESFLIST_' + pkg, True) or "").split(): | 1117 | for dfile in (d.getVar('FILERPROVIDESFLIST_' + pkg, True) or "").split(): |
| 1118 | write_if_exists(sf, pkg, 'FILERPROVIDES_' + dfile) | 1118 | write_if_exists(sf, pkg, 'FILERPROVIDES_' + dfile) |
| 1119 | 1119 | ||
| 1120 | write_if_exists(sf, pkg, 'FILERDEPENDSFLIST') | 1120 | write_if_exists(sf, pkg, 'FILERDEPENDSFLIST') |
| 1121 | for dfile in (d.getVar('FILERDEPENDSFLIST_' + pkg, True) or "").split(): | 1121 | for dfile in (d.getVar('FILERDEPENDSFLIST_' + pkg, True) or "").split(): |
| 1122 | write_if_exists(sf, pkg, 'FILERDEPENDS_' + dfile) | 1122 | write_if_exists(sf, pkg, 'FILERDEPENDS_' + dfile) |
| 1123 | 1123 | ||
| 1124 | sf.write('%s_%s: %s\n' % ('PKGSIZE', pkg, get_directory_size(pkgdest + "/%s" % pkg))) | 1124 | sf.write('%s_%s: %s\n' % ('PKGSIZE', pkg, get_directory_size(pkgdest + "/%s" % pkg))) |
| 1125 | sf.close() | 1125 | sf.close() |
| 1126 | 1126 | ||
| 1127 | 1127 | ||
| 1128 | allow_empty = d.getVar('ALLOW_EMPTY_%s' % pkg, True) | 1128 | allow_empty = d.getVar('ALLOW_EMPTY_%s' % pkg, True) |
| 1129 | if not allow_empty: | 1129 | if not allow_empty: |
| 1130 | allow_empty = d.getVar('ALLOW_EMPTY', True) | 1130 | allow_empty = d.getVar('ALLOW_EMPTY', True) |
| 1131 | root = "%s/%s" % (pkgdest, pkg) | 1131 | root = "%s/%s" % (pkgdest, pkg) |
| 1132 | os.chdir(root) | 1132 | os.chdir(root) |
| 1133 | g = glob('*') | 1133 | g = glob('*') |
| 1134 | if g or allow_empty == "1": | 1134 | if g or allow_empty == "1": |
| 1135 | packagedfile = pkgdatadir + '/runtime/%s.packaged' % pkg | 1135 | packagedfile = pkgdatadir + '/runtime/%s.packaged' % pkg |
| 1136 | file(packagedfile, 'w').close() | 1136 | file(packagedfile, 'w').close() |
| 1137 | 1137 | ||
| 1138 | bb.utils.unlockfile(lf) | 1138 | bb.utils.unlockfile(lf) |
| 1139 | } | 1139 | } |
| 1140 | emit_pkgdata[dirs] = "${PKGDESTWORK}/runtime" | 1140 | emit_pkgdata[dirs] = "${PKGDESTWORK}/runtime" |
| 1141 | 1141 | ||
| @@ -1156,557 +1156,557 @@ RPMDEPS = "${STAGING_LIBDIR_NATIVE}/rpm/bin/rpmdeps-oecore --macros ${STAGING_LI | |||
| 1156 | # FILERDEPENDS_filepath_pkg - per file dep | 1156 | # FILERDEPENDS_filepath_pkg - per file dep |
| 1157 | 1157 | ||
| 1158 | python package_do_filedeps() { | 1158 | python package_do_filedeps() { |
| 1159 | import re | 1159 | import re |
| 1160 | 1160 | ||
| 1161 | if d.getVar('SKIP_FILEDEPS', True) == '1': | 1161 | if d.getVar('SKIP_FILEDEPS', True) == '1': |
| 1162 | return | 1162 | return |
| 1163 | 1163 | ||
| 1164 | pkgdest = d.getVar('PKGDEST', True) | 1164 | pkgdest = d.getVar('PKGDEST', True) |
| 1165 | packages = d.getVar('PACKAGES', True) | 1165 | packages = d.getVar('PACKAGES', True) |
| 1166 | 1166 | ||
| 1167 | rpmdeps = d.expand("${RPMDEPS}") | 1167 | rpmdeps = d.expand("${RPMDEPS}") |
| 1168 | r = re.compile(r'[<>=]+ +[^ ]*') | 1168 | r = re.compile(r'[<>=]+ +[^ ]*') |
| 1169 | 1169 | ||
| 1170 | def file_translate(file): | 1170 | def file_translate(file): |
| 1171 | ft = file.replace("@", "@at@") | 1171 | ft = file.replace("@", "@at@") |
| 1172 | ft = ft.replace(" ", "@space@") | 1172 | ft = ft.replace(" ", "@space@") |
| 1173 | ft = ft.replace("\t", "@tab@") | 1173 | ft = ft.replace("\t", "@tab@") |
| 1174 | ft = ft.replace("[", "@openbrace@") | 1174 | ft = ft.replace("[", "@openbrace@") |
| 1175 | ft = ft.replace("]", "@closebrace@") | 1175 | ft = ft.replace("]", "@closebrace@") |
| 1176 | ft = ft.replace("_", "@underscore@") | 1176 | ft = ft.replace("_", "@underscore@") |
| 1177 | return ft | 1177 | return ft |
| 1178 | 1178 | ||
| 1179 | # Quick routine to process the results of the rpmdeps call... | 1179 | # Quick routine to process the results of the rpmdeps call... |
| 1180 | def process_deps(pipe, pkg, provides_files, requires_files): | 1180 | def process_deps(pipe, pkg, provides_files, requires_files): |
| 1181 | provides = {} | 1181 | provides = {} |
| 1182 | requires = {} | 1182 | requires = {} |
| 1183 | 1183 | ||
| 1184 | for line in pipe: | 1184 | for line in pipe: |
| 1185 | f = line.split(" ", 1)[0].strip() | 1185 | f = line.split(" ", 1)[0].strip() |
| 1186 | line = line.split(" ", 1)[1].strip() | 1186 | line = line.split(" ", 1)[1].strip() |
| 1187 | 1187 | ||
| 1188 | if line.startswith("Requires:"): | 1188 | if line.startswith("Requires:"): |
| 1189 | i = requires | 1189 | i = requires |
| 1190 | elif line.startswith("Provides:"): | 1190 | elif line.startswith("Provides:"): |
| 1191 | i = provides | 1191 | i = provides |
| 1192 | else: | 1192 | else: |
| 1193 | continue | 1193 | continue |
| 1194 | 1194 | ||
| 1195 | file = f.replace(pkgdest + "/" + pkg, "") | 1195 | file = f.replace(pkgdest + "/" + pkg, "") |
| 1196 | file = file_translate(file) | 1196 | file = file_translate(file) |
| 1197 | value = line.split(":", 1)[1].strip() | 1197 | value = line.split(":", 1)[1].strip() |
| 1198 | value = r.sub(r'(\g<0>)', value) | 1198 | value = r.sub(r'(\g<0>)', value) |
| 1199 | 1199 | ||
| 1200 | if value.startswith("rpmlib("): | 1200 | if value.startswith("rpmlib("): |
| 1201 | continue | 1201 | continue |
| 1202 | if value == "python": | 1202 | if value == "python": |
| 1203 | continue | 1203 | continue |
| 1204 | if file not in i: | 1204 | if file not in i: |
| 1205 | i[file] = [] | 1205 | i[file] = [] |
| 1206 | i[file].append(value) | 1206 | i[file].append(value) |
| 1207 | 1207 | ||
| 1208 | for file in provides: | 1208 | for file in provides: |
| 1209 | provides_files.append(file) | 1209 | provides_files.append(file) |
| 1210 | key = "FILERPROVIDES_" + file + "_" + pkg | 1210 | key = "FILERPROVIDES_" + file + "_" + pkg |
| 1211 | d.setVar(key, " ".join(provides[file])) | 1211 | d.setVar(key, " ".join(provides[file])) |
| 1212 | 1212 | ||
| 1213 | for file in requires: | 1213 | for file in requires: |
| 1214 | requires_files.append(file) | 1214 | requires_files.append(file) |
| 1215 | key = "FILERDEPENDS_" + file + "_" + pkg | 1215 | key = "FILERDEPENDS_" + file + "_" + pkg |
| 1216 | d.setVar(key, " ".join(requires[file])) | 1216 | d.setVar(key, " ".join(requires[file])) |
| 1217 | 1217 | ||
| 1218 | def chunks(files, n): | 1218 | def chunks(files, n): |
| 1219 | return [files[i:i+n] for i in range(0, len(files), n)] | 1219 | return [files[i:i+n] for i in range(0, len(files), n)] |
| 1220 | 1220 | ||
| 1221 | # Determine dependencies | 1221 | # Determine dependencies |
| 1222 | for pkg in packages.split(): | 1222 | for pkg in packages.split(): |
| 1223 | if pkg.endswith('-dbg') or pkg.endswith('-doc') or pkg.find('-locale-') != -1 or pkg.find('-localedata-') != -1 or pkg.find('-gconv-') != -1 or pkg.find('-charmap-') != -1 or pkg.startswith('kernel-module-'): | 1223 | if pkg.endswith('-dbg') or pkg.endswith('-doc') or pkg.find('-locale-') != -1 or pkg.find('-localedata-') != -1 or pkg.find('-gconv-') != -1 or pkg.find('-charmap-') != -1 or pkg.startswith('kernel-module-'): |
| 1224 | continue | 1224 | continue |
| 1225 | 1225 | ||
| 1226 | provides_files = [] | 1226 | provides_files = [] |
| 1227 | requires_files = [] | 1227 | requires_files = [] |
| 1228 | rpfiles = [] | 1228 | rpfiles = [] |
| 1229 | for root, dirs, files in os.walk(pkgdest + "/" + pkg): | 1229 | for root, dirs, files in os.walk(pkgdest + "/" + pkg): |
| 1230 | for file in files: | 1230 | for file in files: |
| 1231 | rpfiles.append(os.path.join(root, file)) | 1231 | rpfiles.append(os.path.join(root, file)) |
| 1232 | 1232 | ||
| 1233 | for files in chunks(rpfiles, 100): | 1233 | for files in chunks(rpfiles, 100): |
| 1234 | dep_pipe = os.popen(rpmdeps + " " + " ".join(files)) | 1234 | dep_pipe = os.popen(rpmdeps + " " + " ".join(files)) |
| 1235 | 1235 | ||
| 1236 | process_deps(dep_pipe, pkg, provides_files, requires_files) | 1236 | process_deps(dep_pipe, pkg, provides_files, requires_files) |
| 1237 | 1237 | ||
| 1238 | d.setVar("FILERDEPENDSFLIST_" + pkg, " ".join(requires_files)) | 1238 | d.setVar("FILERDEPENDSFLIST_" + pkg, " ".join(requires_files)) |
| 1239 | d.setVar("FILERPROVIDESFLIST_" + pkg, " ".join(provides_files)) | 1239 | d.setVar("FILERPROVIDESFLIST_" + pkg, " ".join(provides_files)) |
| 1240 | } | 1240 | } |
| 1241 | 1241 | ||
| 1242 | SHLIBSDIR = "${STAGING_DIR_HOST}/shlibs" | 1242 | SHLIBSDIR = "${STAGING_DIR_HOST}/shlibs" |
| 1243 | SHLIBSWORKDIR = "${WORKDIR}/shlibs" | 1243 | SHLIBSWORKDIR = "${WORKDIR}/shlibs" |
| 1244 | 1244 | ||
| 1245 | python package_do_shlibs() { | 1245 | python package_do_shlibs() { |
| 1246 | import re, pipes | 1246 | import re, pipes |
| 1247 | 1247 | ||
| 1248 | exclude_shlibs = d.getVar('EXCLUDE_FROM_SHLIBS', 0) | 1248 | exclude_shlibs = d.getVar('EXCLUDE_FROM_SHLIBS', 0) |
| 1249 | if exclude_shlibs: | 1249 | if exclude_shlibs: |
| 1250 | bb.note("not generating shlibs") | 1250 | bb.note("not generating shlibs") |
| 1251 | return | 1251 | return |
| 1252 | 1252 | ||
| 1253 | lib_re = re.compile("^.*\.so") | 1253 | lib_re = re.compile("^.*\.so") |
| 1254 | libdir_re = re.compile(".*/%s$" % d.getVar('baselib', True)) | 1254 | libdir_re = re.compile(".*/%s$" % d.getVar('baselib', True)) |
| 1255 | 1255 | ||
| 1256 | packages = d.getVar('PACKAGES', True) | 1256 | packages = d.getVar('PACKAGES', True) |
| 1257 | targetos = d.getVar('TARGET_OS', True) | 1257 | targetos = d.getVar('TARGET_OS', True) |
| 1258 | 1258 | ||
| 1259 | workdir = d.getVar('WORKDIR', True) | 1259 | workdir = d.getVar('WORKDIR', True) |
| 1260 | 1260 | ||
| 1261 | ver = d.getVar('PKGV', True) | 1261 | ver = d.getVar('PKGV', True) |
| 1262 | if not ver: | 1262 | if not ver: |
| 1263 | bb.error("PKGV not defined") | 1263 | bb.error("PKGV not defined") |
| 1264 | return | 1264 | return |
| 1265 | 1265 | ||
| 1266 | pkgdest = d.getVar('PKGDEST', True) | 1266 | pkgdest = d.getVar('PKGDEST', True) |
| 1267 | 1267 | ||
| 1268 | shlibs_dir = d.getVar('SHLIBSDIR', True) | 1268 | shlibs_dir = d.getVar('SHLIBSDIR', True) |
| 1269 | shlibswork_dir = d.getVar('SHLIBSWORKDIR', True) | 1269 | shlibswork_dir = d.getVar('SHLIBSWORKDIR', True) |
| 1270 | 1270 | ||
| 1271 | # Take shared lock since we're only reading, not writing | 1271 | # Take shared lock since we're only reading, not writing |
| 1272 | lf = bb.utils.lockfile(d.expand("${PACKAGELOCK}")) | 1272 | lf = bb.utils.lockfile(d.expand("${PACKAGELOCK}")) |
| 1273 | 1273 | ||
| 1274 | def linux_so(root, path, file): | 1274 | def linux_so(root, path, file): |
| 1275 | needs_ldconfig = False | 1275 | needs_ldconfig = False |
| 1276 | cmd = d.getVar('OBJDUMP', True) + " -p " + pipes.quote(os.path.join(root, file)) + " 2>/dev/null" | 1276 | cmd = d.getVar('OBJDUMP', True) + " -p " + pipes.quote(os.path.join(root, file)) + " 2>/dev/null" |
| 1277 | cmd = "PATH=\"%s\" %s" % (d.getVar('PATH', True), cmd) | 1277 | cmd = "PATH=\"%s\" %s" % (d.getVar('PATH', True), cmd) |
| 1278 | fd = os.popen(cmd) | 1278 | fd = os.popen(cmd) |
| 1279 | lines = fd.readlines() | 1279 | lines = fd.readlines() |
| 1280 | fd.close() | 1280 | fd.close() |
| 1281 | for l in lines: | 1281 | for l in lines: |
| 1282 | m = re.match("\s+NEEDED\s+([^\s]*)", l) | 1282 | m = re.match("\s+NEEDED\s+([^\s]*)", l) |
| 1283 | if m: | 1283 | if m: |
| 1284 | if m.group(1) not in needed[pkg]: | 1284 | if m.group(1) not in needed[pkg]: |
| 1285 | needed[pkg].append(m.group(1)) | 1285 | needed[pkg].append(m.group(1)) |
| 1286 | m = re.match("\s+SONAME\s+([^\s]*)", l) | 1286 | m = re.match("\s+SONAME\s+([^\s]*)", l) |
| 1287 | if m: | 1287 | if m: |
| 1288 | this_soname = m.group(1) | 1288 | this_soname = m.group(1) |
| 1289 | if not this_soname in sonames: | 1289 | if not this_soname in sonames: |
| 1290 | # if library is private (only used by package) then do not build shlib for it | 1290 | # if library is private (only used by package) then do not build shlib for it |
| 1291 | if not private_libs or -1 == private_libs.find(this_soname): | 1291 | if not private_libs or -1 == private_libs.find(this_soname): |
| 1292 | sonames.append(this_soname) | 1292 | sonames.append(this_soname) |
| 1293 | if libdir_re.match(root): | 1293 | if libdir_re.match(root): |
| 1294 | needs_ldconfig = True | 1294 | needs_ldconfig = True |
| 1295 | if snap_symlinks and (file != this_soname): | 1295 | if snap_symlinks and (file != this_soname): |
| 1296 | renames.append((os.path.join(root, file), os.path.join(root, this_soname))) | 1296 | renames.append((os.path.join(root, file), os.path.join(root, this_soname))) |
| 1297 | return needs_ldconfig | 1297 | return needs_ldconfig |
| 1298 | 1298 | ||
| 1299 | def darwin_so(root, path, file): | 1299 | def darwin_so(root, path, file): |
| 1300 | fullpath = os.path.join(root, file) | 1300 | fullpath = os.path.join(root, file) |
| 1301 | if not os.path.exists(fullpath): | 1301 | if not os.path.exists(fullpath): |
| 1302 | return | 1302 | return |
| 1303 | 1303 | ||
| 1304 | def get_combinations(base): | 1304 | def get_combinations(base): |
| 1305 | # | 1305 | # |
| 1306 | # Given a base library name, find all combinations of this split by "." and "-" | 1306 | # Given a base library name, find all combinations of this split by "." and "-" |
| 1307 | # | 1307 | # |
| 1308 | combos = [] | 1308 | combos = [] |
| 1309 | options = base.split(".") | 1309 | options = base.split(".") |
| 1310 | for i in range(1, len(options) + 1): | 1310 | for i in range(1, len(options) + 1): |
| 1311 | combos.append(".".join(options[0:i])) | 1311 | combos.append(".".join(options[0:i])) |
| 1312 | options = base.split("-") | 1312 | options = base.split("-") |
| 1313 | for i in range(1, len(options) + 1): | 1313 | for i in range(1, len(options) + 1): |
| 1314 | combos.append("-".join(options[0:i])) | 1314 | combos.append("-".join(options[0:i])) |
| 1315 | return combos | 1315 | return combos |
| 1316 | 1316 | ||
| 1317 | if (file.endswith('.dylib') or file.endswith('.so')) and not pkg.endswith('-dev') and not pkg.endswith('-dbg'): | 1317 | if (file.endswith('.dylib') or file.endswith('.so')) and not pkg.endswith('-dev') and not pkg.endswith('-dbg'): |
| 1318 | # Drop suffix | 1318 | # Drop suffix |
| 1319 | name = file.rsplit(".",1)[0] | 1319 | name = file.rsplit(".",1)[0] |
| 1320 | # Find all combinations | 1320 | # Find all combinations |
| 1321 | combos = get_combinations(name) | 1321 | combos = get_combinations(name) |
| 1322 | for combo in combos: | 1322 | for combo in combos: |
| 1323 | if not combo in sonames: | 1323 | if not combo in sonames: |
| 1324 | sonames.append(combo) | 1324 | sonames.append(combo) |
| 1325 | if file.endswith('.dylib') or file.endswith('.so'): | 1325 | if file.endswith('.dylib') or file.endswith('.so'): |
| 1326 | lafile = fullpath.replace(os.path.join(pkgdest, pkg), d.getVar('PKGD', True)) | 1326 | lafile = fullpath.replace(os.path.join(pkgdest, pkg), d.getVar('PKGD', True)) |
| 1327 | # Drop suffix | 1327 | # Drop suffix |
| 1328 | lafile = lafile.rsplit(".",1)[0] | 1328 | lafile = lafile.rsplit(".",1)[0] |
| 1329 | lapath = os.path.dirname(lafile) | 1329 | lapath = os.path.dirname(lafile) |
| 1330 | lafile = os.path.basename(lafile) | 1330 | lafile = os.path.basename(lafile) |
| 1331 | # Find all combinations | 1331 | # Find all combinations |
| 1332 | combos = get_combinations(lafile) | 1332 | combos = get_combinations(lafile) |
| 1333 | for combo in combos: | 1333 | for combo in combos: |
| 1334 | if os.path.exists(lapath + '/' + combo + '.la'): | 1334 | if os.path.exists(lapath + '/' + combo + '.la'): |
| 1335 | break | 1335 | break |
| 1336 | lafile = lapath + '/' + combo + '.la' | 1336 | lafile = lapath + '/' + combo + '.la' |
| 1337 | 1337 | ||
| 1338 | #bb.note("Foo2: %s" % lafile) | 1338 | #bb.note("Foo2: %s" % lafile) |
| 1339 | #bb.note("Foo %s %s" % (file, fullpath)) | 1339 | #bb.note("Foo %s %s" % (file, fullpath)) |
| 1340 | if os.path.exists(lafile): | 1340 | if os.path.exists(lafile): |
| 1341 | fd = open(lafile, 'r') | 1341 | fd = open(lafile, 'r') |
| 1342 | lines = fd.readlines() | 1342 | lines = fd.readlines() |
| 1343 | fd.close() | 1343 | fd.close() |
| 1344 | for l in lines: | 1344 | for l in lines: |
| 1345 | m = re.match("\s*dependency_libs=\s*'(.*)'", l) | 1345 | m = re.match("\s*dependency_libs=\s*'(.*)'", l) |
| 1346 | if m: | 1346 | if m: |
| 1347 | deps = m.group(1).split(" ") | 1347 | deps = m.group(1).split(" ") |
| 1348 | for dep in deps: | 1348 | for dep in deps: |
| 1349 | #bb.note("Trying %s for %s" % (dep, pkg)) | 1349 | #bb.note("Trying %s for %s" % (dep, pkg)) |
| 1350 | name = None | 1350 | name = None |
| 1351 | if dep.endswith(".la"): | 1351 | if dep.endswith(".la"): |
| 1352 | name = os.path.basename(dep).replace(".la", "") | 1352 | name = os.path.basename(dep).replace(".la", "") |
| 1353 | elif dep.startswith("-l"): | 1353 | elif dep.startswith("-l"): |
| 1354 | name = dep.replace("-l", "lib") | 1354 | name = dep.replace("-l", "lib") |
| 1355 | if pkg not in needed: | 1355 | if pkg not in needed: |
| 1356 | needed[pkg] = [] | 1356 | needed[pkg] = [] |
| 1357 | if name and name not in needed[pkg]: | 1357 | if name and name not in needed[pkg]: |
| 1358 | needed[pkg].append(name) | 1358 | needed[pkg].append(name) |
| 1359 | #bb.note("Adding %s for %s" % (name, pkg)) | 1359 | #bb.note("Adding %s for %s" % (name, pkg)) |
| 1360 | 1360 | ||
| 1361 | if d.getVar('PACKAGE_SNAP_LIB_SYMLINKS', True) == "1": | 1361 | if d.getVar('PACKAGE_SNAP_LIB_SYMLINKS', True) == "1": |
| 1362 | snap_symlinks = True | 1362 | snap_symlinks = True |
| 1363 | else: | 1363 | else: |
| 1364 | snap_symlinks = False | 1364 | snap_symlinks = False |
| 1365 | 1365 | ||
| 1366 | if (d.getVar('USE_LDCONFIG', True) or "1") == "1": | 1366 | if (d.getVar('USE_LDCONFIG', True) or "1") == "1": |
| 1367 | use_ldconfig = True | 1367 | use_ldconfig = True |
| 1368 | else: | 1368 | else: |
| 1369 | use_ldconfig = False | 1369 | use_ldconfig = False |
| 1370 | 1370 | ||
| 1371 | needed = {} | 1371 | needed = {} |
| 1372 | shlib_provider = {} | 1372 | shlib_provider = {} |
| 1373 | for pkg in packages.split(): | 1373 | for pkg in packages.split(): |
| 1374 | private_libs = d.getVar('PRIVATE_LIBS_' + pkg, True) or d.getVar('PRIVATE_LIBS', True) | 1374 | private_libs = d.getVar('PRIVATE_LIBS_' + pkg, True) or d.getVar('PRIVATE_LIBS', True) |
| 1375 | needs_ldconfig = False | 1375 | needs_ldconfig = False |
| 1376 | bb.debug(2, "calculating shlib provides for %s" % pkg) | 1376 | bb.debug(2, "calculating shlib provides for %s" % pkg) |
| 1377 | 1377 | ||
| 1378 | pkgver = d.getVar('PKGV_' + pkg, True) | 1378 | pkgver = d.getVar('PKGV_' + pkg, True) |
| 1379 | if not pkgver: | 1379 | if not pkgver: |
| 1380 | pkgver = d.getVar('PV_' + pkg, True) | 1380 | pkgver = d.getVar('PV_' + pkg, True) |
| 1381 | if not pkgver: | 1381 | if not pkgver: |
| 1382 | pkgver = ver | 1382 | pkgver = ver |
| 1383 | 1383 | ||
| 1384 | needed[pkg] = [] | 1384 | needed[pkg] = [] |
| 1385 | sonames = list() | 1385 | sonames = list() |
| 1386 | renames = list() | 1386 | renames = list() |
| 1387 | top = os.path.join(pkgdest, pkg) | 1387 | top = os.path.join(pkgdest, pkg) |
| 1388 | for root, dirs, files in os.walk(top): | 1388 | for root, dirs, files in os.walk(top): |
| 1389 | for file in files: | 1389 | for file in files: |
| 1390 | soname = None | 1390 | soname = None |
| 1391 | path = os.path.join(root, file) | 1391 | path = os.path.join(root, file) |
| 1392 | if os.path.islink(path): | 1392 | if os.path.islink(path): |
| 1393 | continue | 1393 | continue |
| 1394 | if targetos == "darwin" or targetos == "darwin8": | 1394 | if targetos == "darwin" or targetos == "darwin8": |
| 1395 | darwin_so(root, dirs, file) | 1395 | darwin_so(root, dirs, file) |
| 1396 | elif os.access(path, os.X_OK) or lib_re.match(file): | 1396 | elif os.access(path, os.X_OK) or lib_re.match(file): |
| 1397 | ldconfig = linux_so(root, dirs, file) | 1397 | ldconfig = linux_so(root, dirs, file) |
| 1398 | needs_ldconfig = needs_ldconfig or ldconfig | 1398 | needs_ldconfig = needs_ldconfig or ldconfig |
| 1399 | for (old, new) in renames: | 1399 | for (old, new) in renames: |
| 1400 | bb.note("Renaming %s to %s" % (old, new)) | 1400 | bb.note("Renaming %s to %s" % (old, new)) |
| 1401 | os.rename(old, new) | 1401 | os.rename(old, new) |
| 1402 | shlibs_file = os.path.join(shlibswork_dir, pkg + ".list") | 1402 | shlibs_file = os.path.join(shlibswork_dir, pkg + ".list") |
| 1403 | shver_file = os.path.join(shlibswork_dir, pkg + ".ver") | 1403 | shver_file = os.path.join(shlibswork_dir, pkg + ".ver") |
| 1404 | if len(sonames): | 1404 | if len(sonames): |
| 1405 | fd = open(shlibs_file, 'w') | 1405 | fd = open(shlibs_file, 'w') |
| 1406 | for s in sonames: | 1406 | for s in sonames: |
| 1407 | fd.write(s + '\n') | 1407 | fd.write(s + '\n') |
| 1408 | shlib_provider[s] = (pkg, pkgver) | 1408 | shlib_provider[s] = (pkg, pkgver) |
| 1409 | fd.close() | 1409 | fd.close() |
| 1410 | fd = open(shver_file, 'w') | 1410 | fd = open(shver_file, 'w') |
| 1411 | fd.write(pkgver + '\n') | 1411 | fd.write(pkgver + '\n') |
| 1412 | fd.close() | 1412 | fd.close() |
| 1413 | if needs_ldconfig and use_ldconfig: | 1413 | if needs_ldconfig and use_ldconfig: |
| 1414 | bb.debug(1, 'adding ldconfig call to postinst for %s' % pkg) | 1414 | bb.debug(1, 'adding ldconfig call to postinst for %s' % pkg) |
| 1415 | postinst = d.getVar('pkg_postinst_%s' % pkg, True) or d.getVar('pkg_postinst', True) | 1415 | postinst = d.getVar('pkg_postinst_%s' % pkg, True) or d.getVar('pkg_postinst', True) |
| 1416 | if not postinst: | 1416 | if not postinst: |
| 1417 | postinst = '#!/bin/sh\n' | 1417 | postinst = '#!/bin/sh\n' |
| 1418 | postinst += d.getVar('ldconfig_postinst_fragment', True) | 1418 | postinst += d.getVar('ldconfig_postinst_fragment', True) |
| 1419 | d.setVar('pkg_postinst_%s' % pkg, postinst) | 1419 | d.setVar('pkg_postinst_%s' % pkg, postinst) |
| 1420 | 1420 | ||
| 1421 | list_re = re.compile('^(.*)\.list$') | 1421 | list_re = re.compile('^(.*)\.list$') |
| 1422 | for dir in [shlibs_dir]: | 1422 | for dir in [shlibs_dir]: |
| 1423 | if not os.path.exists(dir): | 1423 | if not os.path.exists(dir): |
| 1424 | continue | 1424 | continue |
| 1425 | for file in os.listdir(dir): | 1425 | for file in os.listdir(dir): |
| 1426 | m = list_re.match(file) | 1426 | m = list_re.match(file) |
| 1427 | if m: | 1427 | if m: |
| 1428 | dep_pkg = m.group(1) | 1428 | dep_pkg = m.group(1) |
| 1429 | fd = open(os.path.join(dir, file)) | 1429 | fd = open(os.path.join(dir, file)) |
| 1430 | lines = fd.readlines() | 1430 | lines = fd.readlines() |
| 1431 | fd.close() | 1431 | fd.close() |
| 1432 | ver_file = os.path.join(dir, dep_pkg + '.ver') | 1432 | ver_file = os.path.join(dir, dep_pkg + '.ver') |
| 1433 | lib_ver = None | 1433 | lib_ver = None |
| 1434 | if os.path.exists(ver_file): | 1434 | if os.path.exists(ver_file): |
| 1435 | fd = open(ver_file) | 1435 | fd = open(ver_file) |
| 1436 | lib_ver = fd.readline().rstrip() | 1436 | lib_ver = fd.readline().rstrip() |
| 1437 | fd.close() | 1437 | fd.close() |
| 1438 | for l in lines: | 1438 | for l in lines: |
| 1439 | shlib_provider[l.rstrip()] = (dep_pkg, lib_ver) | 1439 | shlib_provider[l.rstrip()] = (dep_pkg, lib_ver) |
| 1440 | 1440 | ||
| 1441 | bb.utils.unlockfile(lf) | 1441 | bb.utils.unlockfile(lf) |
| 1442 | 1442 | ||
| 1443 | assumed_libs = d.getVar('ASSUME_SHLIBS', True) | 1443 | assumed_libs = d.getVar('ASSUME_SHLIBS', True) |
| 1444 | if assumed_libs: | 1444 | if assumed_libs: |
| 1445 | for e in assumed_libs.split(): | 1445 | for e in assumed_libs.split(): |
| 1446 | l, dep_pkg = e.split(":") | 1446 | l, dep_pkg = e.split(":") |
| 1447 | lib_ver = None | 1447 | lib_ver = None |
| 1448 | dep_pkg = dep_pkg.rsplit("_", 1) | 1448 | dep_pkg = dep_pkg.rsplit("_", 1) |
| 1449 | if len(dep_pkg) == 2: | 1449 | if len(dep_pkg) == 2: |
| 1450 | lib_ver = dep_pkg[1] | 1450 | lib_ver = dep_pkg[1] |
| 1451 | dep_pkg = dep_pkg[0] | 1451 | dep_pkg = dep_pkg[0] |
| 1452 | shlib_provider[l] = (dep_pkg, lib_ver) | 1452 | shlib_provider[l] = (dep_pkg, lib_ver) |
| 1453 | 1453 | ||
| 1454 | for pkg in packages.split(): | 1454 | for pkg in packages.split(): |
| 1455 | bb.debug(2, "calculating shlib requirements for %s" % pkg) | 1455 | bb.debug(2, "calculating shlib requirements for %s" % pkg) |
| 1456 | 1456 | ||
| 1457 | deps = list() | 1457 | deps = list() |
| 1458 | for n in needed[pkg]: | 1458 | for n in needed[pkg]: |
| 1459 | if n in shlib_provider.keys(): | 1459 | if n in shlib_provider.keys(): |
| 1460 | (dep_pkg, ver_needed) = shlib_provider[n] | 1460 | (dep_pkg, ver_needed) = shlib_provider[n] |
| 1461 | 1461 | ||
| 1462 | bb.debug(2, '%s: Dependency %s requires package %s' % (pkg, n, dep_pkg)) | 1462 | bb.debug(2, '%s: Dependency %s requires package %s' % (pkg, n, dep_pkg)) |
| 1463 | 1463 | ||
| 1464 | if dep_pkg == pkg: | 1464 | if dep_pkg == pkg: |
| 1465 | continue | 1465 | continue |
| 1466 | 1466 | ||
| 1467 | if ver_needed: | 1467 | if ver_needed: |
| 1468 | dep = "%s (>= %s)" % (dep_pkg, ver_needed) | 1468 | dep = "%s (>= %s)" % (dep_pkg, ver_needed) |
| 1469 | else: | 1469 | else: |
| 1470 | dep = dep_pkg | 1470 | dep = dep_pkg |
| 1471 | if not dep in deps: | 1471 | if not dep in deps: |
| 1472 | deps.append(dep) | 1472 | deps.append(dep) |
| 1473 | else: | 1473 | else: |
| 1474 | bb.note("Couldn't find shared library provider for %s" % n) | 1474 | bb.note("Couldn't find shared library provider for %s" % n) |
| 1475 | 1475 | ||
| 1476 | deps_file = os.path.join(pkgdest, pkg + ".shlibdeps") | 1476 | deps_file = os.path.join(pkgdest, pkg + ".shlibdeps") |
| 1477 | if os.path.exists(deps_file): | 1477 | if os.path.exists(deps_file): |
| 1478 | os.remove(deps_file) | 1478 | os.remove(deps_file) |
| 1479 | if len(deps): | 1479 | if len(deps): |
| 1480 | fd = open(deps_file, 'w') | 1480 | fd = open(deps_file, 'w') |
| 1481 | for dep in deps: | 1481 | for dep in deps: |
| 1482 | fd.write(dep + '\n') | 1482 | fd.write(dep + '\n') |
| 1483 | fd.close() | 1483 | fd.close() |
| 1484 | } | 1484 | } |
| 1485 | 1485 | ||
| 1486 | python package_do_pkgconfig () { | 1486 | python package_do_pkgconfig () { |
| 1487 | import re | 1487 | import re |
| 1488 | 1488 | ||
| 1489 | packages = d.getVar('PACKAGES', True) | 1489 | packages = d.getVar('PACKAGES', True) |
| 1490 | workdir = d.getVar('WORKDIR', True) | 1490 | workdir = d.getVar('WORKDIR', True) |
| 1491 | pkgdest = d.getVar('PKGDEST', True) | 1491 | pkgdest = d.getVar('PKGDEST', True) |
| 1492 | 1492 | ||
| 1493 | shlibs_dir = d.getVar('SHLIBSDIR', True) | 1493 | shlibs_dir = d.getVar('SHLIBSDIR', True) |
| 1494 | shlibswork_dir = d.getVar('SHLIBSWORKDIR', True) | 1494 | shlibswork_dir = d.getVar('SHLIBSWORKDIR', True) |
| 1495 | 1495 | ||
| 1496 | pc_re = re.compile('(.*)\.pc$') | 1496 | pc_re = re.compile('(.*)\.pc$') |
| 1497 | var_re = re.compile('(.*)=(.*)') | 1497 | var_re = re.compile('(.*)=(.*)') |
| 1498 | field_re = re.compile('(.*): (.*)') | 1498 | field_re = re.compile('(.*): (.*)') |
| 1499 | 1499 | ||
| 1500 | pkgconfig_provided = {} | 1500 | pkgconfig_provided = {} |
| 1501 | pkgconfig_needed = {} | 1501 | pkgconfig_needed = {} |
| 1502 | for pkg in packages.split(): | 1502 | for pkg in packages.split(): |
| 1503 | pkgconfig_provided[pkg] = [] | 1503 | pkgconfig_provided[pkg] = [] |
| 1504 | pkgconfig_needed[pkg] = [] | 1504 | pkgconfig_needed[pkg] = [] |
| 1505 | top = os.path.join(pkgdest, pkg) | 1505 | top = os.path.join(pkgdest, pkg) |
| 1506 | for root, dirs, files in os.walk(top): | 1506 | for root, dirs, files in os.walk(top): |
| 1507 | for file in files: | 1507 | for file in files: |
| 1508 | m = pc_re.match(file) | 1508 | m = pc_re.match(file) |
| 1509 | if m: | 1509 | if m: |
| 1510 | pd = bb.data.init() | 1510 | pd = bb.data.init() |
| 1511 | name = m.group(1) | 1511 | name = m.group(1) |
| 1512 | pkgconfig_provided[pkg].append(name) | 1512 | pkgconfig_provided[pkg].append(name) |
| 1513 | path = os.path.join(root, file) | 1513 | path = os.path.join(root, file) |
| 1514 | if not os.access(path, os.R_OK): | 1514 | if not os.access(path, os.R_OK): |
| 1515 | continue | 1515 | continue |
| 1516 | f = open(path, 'r') | 1516 | f = open(path, 'r') |
| 1517 | lines = f.readlines() | 1517 | lines = f.readlines() |
| 1518 | f.close() | 1518 | f.close() |
| 1519 | for l in lines: | 1519 | for l in lines: |
| 1520 | m = var_re.match(l) | 1520 | m = var_re.match(l) |
| 1521 | if m: | 1521 | if m: |
| 1522 | name = m.group(1) | 1522 | name = m.group(1) |
| 1523 | val = m.group(2) | 1523 | val = m.group(2) |
| 1524 | pd.setVar(name, pd.expand(val)) | 1524 | pd.setVar(name, pd.expand(val)) |
| 1525 | continue | 1525 | continue |
| 1526 | m = field_re.match(l) | 1526 | m = field_re.match(l) |
| 1527 | if m: | 1527 | if m: |
| 1528 | hdr = m.group(1) | 1528 | hdr = m.group(1) |
| 1529 | exp = bb.data.expand(m.group(2), pd) | 1529 | exp = bb.data.expand(m.group(2), pd) |
| 1530 | if hdr == 'Requires': | 1530 | if hdr == 'Requires': |
| 1531 | pkgconfig_needed[pkg] += exp.replace(',', ' ').split() | 1531 | pkgconfig_needed[pkg] += exp.replace(',', ' ').split() |
| 1532 | 1532 | ||
| 1533 | # Take shared lock since we're only reading, not writing | 1533 | # Take shared lock since we're only reading, not writing |
| 1534 | lf = bb.utils.lockfile(d.expand("${PACKAGELOCK}")) | 1534 | lf = bb.utils.lockfile(d.expand("${PACKAGELOCK}")) |
| 1535 | 1535 | ||
| 1536 | for pkg in packages.split(): | 1536 | for pkg in packages.split(): |
| 1537 | pkgs_file = os.path.join(shlibswork_dir, pkg + ".pclist") | 1537 | pkgs_file = os.path.join(shlibswork_dir, pkg + ".pclist") |
| 1538 | if pkgconfig_provided[pkg] != []: | 1538 | if pkgconfig_provided[pkg] != []: |
| 1539 | f = open(pkgs_file, 'w') | 1539 | f = open(pkgs_file, 'w') |
| 1540 | for p in pkgconfig_provided[pkg]: | 1540 | for p in pkgconfig_provided[pkg]: |
| 1541 | f.write('%s\n' % p) | 1541 | f.write('%s\n' % p) |
| 1542 | f.close() | 1542 | f.close() |
| 1543 | 1543 | ||
| 1544 | for dir in [shlibs_dir]: | 1544 | for dir in [shlibs_dir]: |
| 1545 | if not os.path.exists(dir): | 1545 | if not os.path.exists(dir): |
| 1546 | continue | 1546 | continue |
| 1547 | for file in os.listdir(dir): | 1547 | for file in os.listdir(dir): |
| 1548 | m = re.match('^(.*)\.pclist$', file) | 1548 | m = re.match('^(.*)\.pclist$', file) |
| 1549 | if m: | 1549 | if m: |
| 1550 | pkg = m.group(1) | 1550 | pkg = m.group(1) |
| 1551 | fd = open(os.path.join(dir, file)) | 1551 | fd = open(os.path.join(dir, file)) |
| 1552 | lines = fd.readlines() | 1552 | lines = fd.readlines() |
| 1553 | fd.close() | 1553 | fd.close() |
| 1554 | pkgconfig_provided[pkg] = [] | 1554 | pkgconfig_provided[pkg] = [] |
| 1555 | for l in lines: | 1555 | for l in lines: |
| 1556 | pkgconfig_provided[pkg].append(l.rstrip()) | 1556 | pkgconfig_provided[pkg].append(l.rstrip()) |
| 1557 | 1557 | ||
| 1558 | for pkg in packages.split(): | 1558 | for pkg in packages.split(): |
| 1559 | deps = [] | 1559 | deps = [] |
| 1560 | for n in pkgconfig_needed[pkg]: | 1560 | for n in pkgconfig_needed[pkg]: |
| 1561 | found = False | 1561 | found = False |
| 1562 | for k in pkgconfig_provided.keys(): | 1562 | for k in pkgconfig_provided.keys(): |
| 1563 | if n in pkgconfig_provided[k]: | 1563 | if n in pkgconfig_provided[k]: |
| 1564 | if k != pkg and not (k in deps): | 1564 | if k != pkg and not (k in deps): |
| 1565 | deps.append(k) | 1565 | deps.append(k) |
| 1566 | found = True | 1566 | found = True |
| 1567 | if found == False: | 1567 | if found == False: |
| 1568 | bb.note("couldn't find pkgconfig module '%s' in any package" % n) | 1568 | bb.note("couldn't find pkgconfig module '%s' in any package" % n) |
| 1569 | deps_file = os.path.join(pkgdest, pkg + ".pcdeps") | 1569 | deps_file = os.path.join(pkgdest, pkg + ".pcdeps") |
| 1570 | if len(deps): | 1570 | if len(deps): |
| 1571 | fd = open(deps_file, 'w') | 1571 | fd = open(deps_file, 'w') |
| 1572 | for dep in deps: | 1572 | for dep in deps: |
| 1573 | fd.write(dep + '\n') | 1573 | fd.write(dep + '\n') |
| 1574 | fd.close() | 1574 | fd.close() |
| 1575 | 1575 | ||
| 1576 | bb.utils.unlockfile(lf) | 1576 | bb.utils.unlockfile(lf) |
| 1577 | } | 1577 | } |
| 1578 | 1578 | ||
| 1579 | python read_shlibdeps () { | 1579 | python read_shlibdeps () { |
| 1580 | packages = d.getVar('PACKAGES', True).split() | 1580 | packages = d.getVar('PACKAGES', True).split() |
| 1581 | for pkg in packages: | 1581 | for pkg in packages: |
| 1582 | rdepends = bb.utils.explode_dep_versions(d.getVar('RDEPENDS_' + pkg, False) or d.getVar('RDEPENDS', False) or "") | 1582 | rdepends = bb.utils.explode_dep_versions(d.getVar('RDEPENDS_' + pkg, False) or d.getVar('RDEPENDS', False) or "") |
| 1583 | 1583 | ||
| 1584 | for extension in ".shlibdeps", ".pcdeps", ".clilibdeps": | 1584 | for extension in ".shlibdeps", ".pcdeps", ".clilibdeps": |
| 1585 | depsfile = d.expand("${PKGDEST}/" + pkg + extension) | 1585 | depsfile = d.expand("${PKGDEST}/" + pkg + extension) |
| 1586 | if os.access(depsfile, os.R_OK): | 1586 | if os.access(depsfile, os.R_OK): |
| 1587 | fd = file(depsfile) | 1587 | fd = file(depsfile) |
| 1588 | lines = fd.readlines() | 1588 | lines = fd.readlines() |
| 1589 | fd.close() | 1589 | fd.close() |
| 1590 | for l in lines: | 1590 | for l in lines: |
| 1591 | rdepends[l.rstrip()] = "" | 1591 | rdepends[l.rstrip()] = "" |
| 1592 | d.setVar('RDEPENDS_' + pkg, bb.utils.join_deps(rdepends, commasep=False)) | 1592 | d.setVar('RDEPENDS_' + pkg, bb.utils.join_deps(rdepends, commasep=False)) |
| 1593 | } | 1593 | } |
| 1594 | 1594 | ||
| 1595 | python package_depchains() { | 1595 | python package_depchains() { |
| 1596 | """ | 1596 | """ |
| 1597 | For a given set of prefix and postfix modifiers, make those packages | 1597 | For a given set of prefix and postfix modifiers, make those packages |
| 1598 | RRECOMMENDS on the corresponding packages for its RDEPENDS. | 1598 | RRECOMMENDS on the corresponding packages for its RDEPENDS. |
| 1599 | 1599 | ||
| 1600 | Example: If package A depends upon package B, and A's .bb emits an | 1600 | Example: If package A depends upon package B, and A's .bb emits an |
| 1601 | A-dev package, this would make A-dev Recommends: B-dev. | 1601 | A-dev package, this would make A-dev Recommends: B-dev. |
| 1602 | 1602 | ||
| 1603 | If only one of a given suffix is specified, it will take the RRECOMMENDS | 1603 | If only one of a given suffix is specified, it will take the RRECOMMENDS |
| 1604 | based on the RDEPENDS of *all* other packages. If more than one of a given | 1604 | based on the RDEPENDS of *all* other packages. If more than one of a given |
| 1605 | suffix is specified, its will only use the RDEPENDS of the single parent | 1605 | suffix is specified, its will only use the RDEPENDS of the single parent |
| 1606 | package. | 1606 | package. |
| 1607 | """ | 1607 | """ |
| 1608 | 1608 | ||
| 1609 | packages = d.getVar('PACKAGES', True) | 1609 | packages = d.getVar('PACKAGES', True) |
| 1610 | postfixes = (d.getVar('DEPCHAIN_POST', True) or '').split() | 1610 | postfixes = (d.getVar('DEPCHAIN_POST', True) or '').split() |
| 1611 | prefixes = (d.getVar('DEPCHAIN_PRE', True) or '').split() | 1611 | prefixes = (d.getVar('DEPCHAIN_PRE', True) or '').split() |
| 1612 | 1612 | ||
| 1613 | def pkg_adddeprrecs(pkg, base, suffix, getname, depends, d): | 1613 | def pkg_adddeprrecs(pkg, base, suffix, getname, depends, d): |
| 1614 | 1614 | ||
| 1615 | #bb.note('depends for %s is %s' % (base, depends)) | 1615 | #bb.note('depends for %s is %s' % (base, depends)) |
| 1616 | rreclist = bb.utils.explode_dep_versions(d.getVar('RRECOMMENDS_' + pkg, True) or d.getVar('RRECOMMENDS', True) or "") | 1616 | rreclist = bb.utils.explode_dep_versions(d.getVar('RRECOMMENDS_' + pkg, True) or d.getVar('RRECOMMENDS', True) or "") |
| 1617 | 1617 | ||
| 1618 | for depend in depends: | 1618 | for depend in depends: |
| 1619 | if depend.find('-native') != -1 or depend.find('-cross') != -1 or depend.startswith('virtual/'): | 1619 | if depend.find('-native') != -1 or depend.find('-cross') != -1 or depend.startswith('virtual/'): |
| 1620 | #bb.note("Skipping %s" % depend) | 1620 | #bb.note("Skipping %s" % depend) |
| 1621 | continue | 1621 | continue |
| 1622 | if depend.endswith('-dev'): | 1622 | if depend.endswith('-dev'): |
| 1623 | depend = depend.replace('-dev', '') | 1623 | depend = depend.replace('-dev', '') |
| 1624 | if depend.endswith('-dbg'): | 1624 | if depend.endswith('-dbg'): |
| 1625 | depend = depend.replace('-dbg', '') | 1625 | depend = depend.replace('-dbg', '') |
| 1626 | pkgname = getname(depend, suffix) | 1626 | pkgname = getname(depend, suffix) |
| 1627 | #bb.note("Adding %s for %s" % (pkgname, depend)) | 1627 | #bb.note("Adding %s for %s" % (pkgname, depend)) |
| 1628 | if pkgname not in rreclist: | 1628 | if pkgname not in rreclist: |
| 1629 | rreclist[pkgname] = "" | 1629 | rreclist[pkgname] = "" |
| 1630 | 1630 | ||
| 1631 | #bb.note('setting: RRECOMMENDS_%s=%s' % (pkg, ' '.join(rreclist))) | 1631 | #bb.note('setting: RRECOMMENDS_%s=%s' % (pkg, ' '.join(rreclist))) |
| 1632 | d.setVar('RRECOMMENDS_%s' % pkg, bb.utils.join_deps(rreclist, commasep=False)) | 1632 | d.setVar('RRECOMMENDS_%s' % pkg, bb.utils.join_deps(rreclist, commasep=False)) |
| 1633 | 1633 | ||
| 1634 | def pkg_addrrecs(pkg, base, suffix, getname, rdepends, d): | 1634 | def pkg_addrrecs(pkg, base, suffix, getname, rdepends, d): |
| 1635 | 1635 | ||
| 1636 | #bb.note('rdepends for %s is %s' % (base, rdepends)) | 1636 | #bb.note('rdepends for %s is %s' % (base, rdepends)) |
| 1637 | rreclist = bb.utils.explode_dep_versions(d.getVar('RRECOMMENDS_' + pkg, True) or d.getVar('RRECOMMENDS', True) or "") | 1637 | rreclist = bb.utils.explode_dep_versions(d.getVar('RRECOMMENDS_' + pkg, True) or d.getVar('RRECOMMENDS', True) or "") |
| 1638 | 1638 | ||
| 1639 | for depend in rdepends: | 1639 | for depend in rdepends: |
| 1640 | if depend.find('virtual-locale-') != -1: | 1640 | if depend.find('virtual-locale-') != -1: |
| 1641 | #bb.note("Skipping %s" % depend) | 1641 | #bb.note("Skipping %s" % depend) |
| 1642 | continue | 1642 | continue |
| 1643 | if depend.endswith('-dev'): | 1643 | if depend.endswith('-dev'): |
| 1644 | depend = depend.replace('-dev', '') | 1644 | depend = depend.replace('-dev', '') |
| 1645 | if depend.endswith('-dbg'): | 1645 | if depend.endswith('-dbg'): |
| 1646 | depend = depend.replace('-dbg', '') | 1646 | depend = depend.replace('-dbg', '') |
| 1647 | pkgname = getname(depend, suffix) | 1647 | pkgname = getname(depend, suffix) |
| 1648 | #bb.note("Adding %s for %s" % (pkgname, depend)) | 1648 | #bb.note("Adding %s for %s" % (pkgname, depend)) |
| 1649 | if pkgname not in rreclist: | 1649 | if pkgname not in rreclist: |
| 1650 | rreclist[pkgname] = "" | 1650 | rreclist[pkgname] = "" |
| 1651 | 1651 | ||
| 1652 | #bb.note('setting: RRECOMMENDS_%s=%s' % (pkg, ' '.join(rreclist))) | 1652 | #bb.note('setting: RRECOMMENDS_%s=%s' % (pkg, ' '.join(rreclist))) |
| 1653 | d.setVar('RRECOMMENDS_%s' % pkg, bb.utils.join_deps(rreclist, commasep=False)) | 1653 | d.setVar('RRECOMMENDS_%s' % pkg, bb.utils.join_deps(rreclist, commasep=False)) |
| 1654 | 1654 | ||
| 1655 | def add_dep(list, dep): | 1655 | def add_dep(list, dep): |
| 1656 | dep = dep.split(' (')[0].strip() | 1656 | dep = dep.split(' (')[0].strip() |
| 1657 | if dep not in list: | 1657 | if dep not in list: |
| 1658 | list.append(dep) | 1658 | list.append(dep) |
| 1659 | 1659 | ||
| 1660 | depends = [] | 1660 | depends = [] |
| 1661 | for dep in bb.utils.explode_deps(d.getVar('DEPENDS', True) or ""): | 1661 | for dep in bb.utils.explode_deps(d.getVar('DEPENDS', True) or ""): |
| 1662 | add_dep(depends, dep) | 1662 | add_dep(depends, dep) |
| 1663 | 1663 | ||
| 1664 | rdepends = [] | 1664 | rdepends = [] |
| 1665 | for dep in bb.utils.explode_deps(d.getVar('RDEPENDS', True) or ""): | 1665 | for dep in bb.utils.explode_deps(d.getVar('RDEPENDS', True) or ""): |
| 1666 | add_dep(rdepends, dep) | 1666 | add_dep(rdepends, dep) |
| 1667 | 1667 | ||
| 1668 | for pkg in packages.split(): | 1668 | for pkg in packages.split(): |
| 1669 | for dep in bb.utils.explode_deps(d.getVar('RDEPENDS_' + pkg, True) or ""): | 1669 | for dep in bb.utils.explode_deps(d.getVar('RDEPENDS_' + pkg, True) or ""): |
| 1670 | add_dep(rdepends, dep) | 1670 | add_dep(rdepends, dep) |
| 1671 | 1671 | ||
| 1672 | #bb.note('rdepends is %s' % rdepends) | 1672 | #bb.note('rdepends is %s' % rdepends) |
| 1673 | 1673 | ||
| 1674 | def post_getname(name, suffix): | 1674 | def post_getname(name, suffix): |
| 1675 | return '%s%s' % (name, suffix) | 1675 | return '%s%s' % (name, suffix) |
| 1676 | def pre_getname(name, suffix): | 1676 | def pre_getname(name, suffix): |
| 1677 | return '%s%s' % (suffix, name) | 1677 | return '%s%s' % (suffix, name) |
| 1678 | 1678 | ||
| 1679 | pkgs = {} | 1679 | pkgs = {} |
| 1680 | for pkg in packages.split(): | 1680 | for pkg in packages.split(): |
| 1681 | for postfix in postfixes: | 1681 | for postfix in postfixes: |
| 1682 | if pkg.endswith(postfix): | 1682 | if pkg.endswith(postfix): |
| 1683 | if not postfix in pkgs: | 1683 | if not postfix in pkgs: |
| 1684 | pkgs[postfix] = {} | 1684 | pkgs[postfix] = {} |
| 1685 | pkgs[postfix][pkg] = (pkg[:-len(postfix)], post_getname) | 1685 | pkgs[postfix][pkg] = (pkg[:-len(postfix)], post_getname) |
| 1686 | 1686 | ||
| 1687 | for prefix in prefixes: | 1687 | for prefix in prefixes: |
| 1688 | if pkg.startswith(prefix): | 1688 | if pkg.startswith(prefix): |
| 1689 | if not prefix in pkgs: | 1689 | if not prefix in pkgs: |
| 1690 | pkgs[prefix] = {} | 1690 | pkgs[prefix] = {} |
| 1691 | pkgs[prefix][pkg] = (pkg[:-len(prefix)], pre_getname) | 1691 | pkgs[prefix][pkg] = (pkg[:-len(prefix)], pre_getname) |
| 1692 | 1692 | ||
| 1693 | for suffix in pkgs: | 1693 | for suffix in pkgs: |
| 1694 | for pkg in pkgs[suffix]: | 1694 | for pkg in pkgs[suffix]: |
| 1695 | if d.getVarFlag('RRECOMMENDS_' + pkg, 'nodeprrecs'): | 1695 | if d.getVarFlag('RRECOMMENDS_' + pkg, 'nodeprrecs'): |
| 1696 | continue | 1696 | continue |
| 1697 | (base, func) = pkgs[suffix][pkg] | 1697 | (base, func) = pkgs[suffix][pkg] |
| 1698 | if suffix == "-dev": | 1698 | if suffix == "-dev": |
| 1699 | pkg_adddeprrecs(pkg, base, suffix, func, depends, d) | 1699 | pkg_adddeprrecs(pkg, base, suffix, func, depends, d) |
| 1700 | if len(pkgs[suffix]) == 1: | 1700 | if len(pkgs[suffix]) == 1: |
| 1701 | pkg_addrrecs(pkg, base, suffix, func, rdepends, d) | 1701 | pkg_addrrecs(pkg, base, suffix, func, rdepends, d) |
| 1702 | else: | 1702 | else: |
| 1703 | rdeps = [] | 1703 | rdeps = [] |
| 1704 | for dep in bb.utils.explode_deps(d.getVar('RDEPENDS_' + base, True) or d.getVar('RDEPENDS', True) or ""): | 1704 | for dep in bb.utils.explode_deps(d.getVar('RDEPENDS_' + base, True) or d.getVar('RDEPENDS', True) or ""): |
| 1705 | add_dep(rdeps, dep) | 1705 | add_dep(rdeps, dep) |
| 1706 | pkg_addrrecs(pkg, base, suffix, func, rdeps, d) | 1706 | pkg_addrrecs(pkg, base, suffix, func, rdeps, d) |
| 1707 | } | 1707 | } |
| 1708 | 1708 | ||
| 1709 | # Since bitbake can't determine which variables are accessed during package | 1709 | # Since bitbake can't determine which variables are accessed during package |
| 1710 | # iteration, we need to list them here: | 1710 | # iteration, we need to list them here: |
| 1711 | PACKAGEVARS = "FILES RDEPENDS RRECOMMENDS SUMMARY DESCRIPTION RSUGGESTS RPROVIDES RCONFLICTS PKG ALLOW_EMPTY pkg_postinst pkg_postrm INITSCRIPT_NAME INITSCRIPT_PARAMS DEBIAN_NOAUTONAME ALTERNATIVE PKGE PKGV PKGR" | 1711 | PACKAGEVARS = "FILES RDEPENDS RRECOMMENDS SUMMARY DESCRIPTION RSUGGESTS RPROVIDES RCONFLICTS PKG ALLOW_EMPTY pkg_postinst pkg_postrm INITSCRIPT_NAME INITSCRIPT_PARAMS DEBIAN_NOAUTONAME ALTERNATIVE PKGE PKGV PKGR" |
| 1712 | 1712 | ||
| @@ -1720,44 +1720,44 @@ def gen_packagevar(d): | |||
| 1720 | return " ".join(ret) | 1720 | return " ".join(ret) |
| 1721 | 1721 | ||
| 1722 | PACKAGE_PREPROCESS_FUNCS ?= "" | 1722 | PACKAGE_PREPROCESS_FUNCS ?= "" |
| 1723 | PACKAGEFUNCS ?= "package_get_auto_pr \ | 1723 | PACKAGEFUNCS ?= "package_get_auto_pr \ |
| 1724 | perform_packagecopy \ | 1724 | perform_packagecopy \ |
| 1725 | ${PACKAGE_PREPROCESS_FUNCS} \ | 1725 | ${PACKAGE_PREPROCESS_FUNCS} \ |
| 1726 | package_do_split_locales \ | 1726 | package_do_split_locales \ |
| 1727 | split_and_strip_files \ | 1727 | split_and_strip_files \ |
| 1728 | fixup_perms \ | 1728 | fixup_perms \ |
| 1729 | populate_packages \ | 1729 | populate_packages \ |
| 1730 | package_do_filedeps \ | 1730 | package_do_filedeps \ |
| 1731 | package_do_shlibs \ | 1731 | package_do_shlibs \ |
| 1732 | package_do_pkgconfig \ | 1732 | package_do_pkgconfig \ |
| 1733 | read_shlibdeps \ | 1733 | read_shlibdeps \ |
| 1734 | package_depchains \ | 1734 | package_depchains \ |
| 1735 | emit_pkgdata" | 1735 | emit_pkgdata" |
| 1736 | 1736 | ||
| 1737 | python do_package () { | 1737 | python do_package () { |
| 1738 | # Change the following version to cause sstate to invalidate the package | 1738 | # Change the following version to cause sstate to invalidate the package |
| 1739 | # cache. This is useful if an item this class depends on changes in a | 1739 | # cache. This is useful if an item this class depends on changes in a |
| 1740 | # way that the output of this class changes. rpmdeps is a good example | 1740 | # way that the output of this class changes. rpmdeps is a good example |
| 1741 | # as any change to rpmdeps requires this to be rerun. | 1741 | # as any change to rpmdeps requires this to be rerun. |
| 1742 | # PACKAGE_BBCLASS_VERSION = "1" | 1742 | # PACKAGE_BBCLASS_VERSION = "1" |
| 1743 | 1743 | ||
| 1744 | packages = (d.getVar('PACKAGES', True) or "").split() | 1744 | packages = (d.getVar('PACKAGES', True) or "").split() |
| 1745 | if len(packages) < 1: | 1745 | if len(packages) < 1: |
| 1746 | bb.debug(1, "No packages to build, skipping do_package") | 1746 | bb.debug(1, "No packages to build, skipping do_package") |
| 1747 | return | 1747 | return |
| 1748 | 1748 | ||
| 1749 | workdir = d.getVar('WORKDIR', True) | 1749 | workdir = d.getVar('WORKDIR', True) |
| 1750 | outdir = d.getVar('DEPLOY_DIR', True) | 1750 | outdir = d.getVar('DEPLOY_DIR', True) |
| 1751 | dest = d.getVar('D', True) | 1751 | dest = d.getVar('D', True) |
| 1752 | dvar = d.getVar('PKGD', True) | 1752 | dvar = d.getVar('PKGD', True) |
| 1753 | pn = d.getVar('PN', True) | 1753 | pn = d.getVar('PN', True) |
| 1754 | 1754 | ||
| 1755 | if not workdir or not outdir or not dest or not dvar or not pn or not packages: | 1755 | if not workdir or not outdir or not dest or not dvar or not pn or not packages: |
| 1756 | bb.error("WORKDIR, DEPLOY_DIR, D, PN and PKGD all must be defined, unable to package") | 1756 | bb.error("WORKDIR, DEPLOY_DIR, D, PN and PKGD all must be defined, unable to package") |
| 1757 | return | 1757 | return |
| 1758 | 1758 | ||
| 1759 | for f in (d.getVar('PACKAGEFUNCS', True) or '').split(): | 1759 | for f in (d.getVar('PACKAGEFUNCS', True) or '').split(): |
| 1760 | bb.build.exec_func(f, d) | 1760 | bb.build.exec_func(f, d) |
| 1761 | } | 1761 | } |
| 1762 | 1762 | ||
| 1763 | do_package[dirs] = "${SHLIBSWORKDIR} ${PKGDESTWORK} ${D}" | 1763 | do_package[dirs] = "${SHLIBSWORKDIR} ${PKGDESTWORK} ${D}" |
| @@ -1775,7 +1775,7 @@ do_package[stamp-extra-info] = "${MACHINE}" | |||
| 1775 | do_package_setscene[dirs] = "${STAGING_DIR}" | 1775 | do_package_setscene[dirs] = "${STAGING_DIR}" |
| 1776 | 1776 | ||
| 1777 | python do_package_setscene () { | 1777 | python do_package_setscene () { |
| 1778 | sstate_setscene(d) | 1778 | sstate_setscene(d) |
| 1779 | } | 1779 | } |
| 1780 | addtask do_package_setscene | 1780 | addtask do_package_setscene |
| 1781 | 1781 | ||
| @@ -1793,14 +1793,14 @@ addtask package_write before do_build after do_package | |||
| 1793 | # | 1793 | # |
| 1794 | 1794 | ||
| 1795 | def mapping_rename_hook(d): | 1795 | def mapping_rename_hook(d): |
| 1796 | """ | 1796 | """ |
| 1797 | Rewrite variables to account for package renaming in things | 1797 | Rewrite variables to account for package renaming in things |
| 1798 | like debian.bbclass or manual PKG variable name changes | 1798 | like debian.bbclass or manual PKG variable name changes |
| 1799 | """ | 1799 | """ |
| 1800 | runtime_mapping_rename("RDEPENDS", d) | 1800 | runtime_mapping_rename("RDEPENDS", d) |
| 1801 | runtime_mapping_rename("RRECOMMENDS", d) | 1801 | runtime_mapping_rename("RRECOMMENDS", d) |
| 1802 | runtime_mapping_rename("RSUGGESTS", d) | 1802 | runtime_mapping_rename("RSUGGESTS", d) |
| 1803 | runtime_mapping_rename("RPROVIDES", d) | 1803 | runtime_mapping_rename("RPROVIDES", d) |
| 1804 | runtime_mapping_rename("RREPLACES", d) | 1804 | runtime_mapping_rename("RREPLACES", d) |
| 1805 | runtime_mapping_rename("RCONFLICTS", d) | 1805 | runtime_mapping_rename("RCONFLICTS", d) |
| 1806 | 1806 | ||
