diff options
Diffstat (limited to 'meta/classes/base.bbclass')
-rw-r--r-- | meta/classes/base.bbclass | 114 |
1 files changed, 57 insertions, 57 deletions
diff --git a/meta/classes/base.bbclass b/meta/classes/base.bbclass index f5397446dd..f3f798f9bf 100644 --- a/meta/classes/base.bbclass +++ b/meta/classes/base.bbclass | |||
@@ -60,9 +60,9 @@ def base_dep_prepend(d): | |||
60 | # INHIBIT_DEFAULT_DEPS doesn't apply to the patch command. Whether or not | 60 | # INHIBIT_DEFAULT_DEPS doesn't apply to the patch command. Whether or not |
61 | # we need that built is the responsibility of the patch function / class, not | 61 | # we need that built is the responsibility of the patch function / class, not |
62 | # the application. | 62 | # the application. |
63 | if not bb.data.getVar('INHIBIT_DEFAULT_DEPS', d): | 63 | if not d.getVar('INHIBIT_DEFAULT_DEPS'): |
64 | if (bb.data.getVar('HOST_SYS', d, 1) != | 64 | if (d.getVar('HOST_SYS', 1) != |
65 | bb.data.getVar('BUILD_SYS', d, 1)): | 65 | d.getVar('BUILD_SYS', 1)): |
66 | deps += " virtual/${TARGET_PREFIX}gcc virtual/${TARGET_PREFIX}compilerlibs virtual/libc " | 66 | deps += " virtual/${TARGET_PREFIX}gcc virtual/${TARGET_PREFIX}compilerlibs virtual/libc " |
67 | return deps | 67 | return deps |
68 | 68 | ||
@@ -73,13 +73,13 @@ DEPENDS_prepend="${BASEDEPENDS} " | |||
73 | FILESPATH = "${@base_set_filespath([ "${FILE_DIRNAME}/${PF}", "${FILE_DIRNAME}/${P}", "${FILE_DIRNAME}/${PN}", "${FILE_DIRNAME}/${BP}", "${FILE_DIRNAME}/${BPN}", "${FILE_DIRNAME}/files", "${FILE_DIRNAME}" ], d)}" | 73 | FILESPATH = "${@base_set_filespath([ "${FILE_DIRNAME}/${PF}", "${FILE_DIRNAME}/${P}", "${FILE_DIRNAME}/${PN}", "${FILE_DIRNAME}/${BP}", "${FILE_DIRNAME}/${BPN}", "${FILE_DIRNAME}/files", "${FILE_DIRNAME}" ], d)}" |
74 | # THISDIR only works properly with imediate expansion as it has to run | 74 | # THISDIR only works properly with imediate expansion as it has to run |
75 | # in the context of the location its used (:=) | 75 | # in the context of the location its used (:=) |
76 | THISDIR = "${@os.path.dirname(bb.data.getVar('FILE', d, True))}" | 76 | THISDIR = "${@os.path.dirname(d.getVar('FILE', True))}" |
77 | 77 | ||
78 | addtask fetch | 78 | addtask fetch |
79 | do_fetch[dirs] = "${DL_DIR}" | 79 | do_fetch[dirs] = "${DL_DIR}" |
80 | python base_do_fetch() { | 80 | python base_do_fetch() { |
81 | 81 | ||
82 | src_uri = (bb.data.getVar('SRC_URI', d, True) or "").split() | 82 | src_uri = (d.getVar('SRC_URI', True) or "").split() |
83 | if len(src_uri) == 0: | 83 | if len(src_uri) == 0: |
84 | return | 84 | return |
85 | 85 | ||
@@ -96,14 +96,14 @@ python base_do_fetch() { | |||
96 | addtask unpack after do_fetch | 96 | addtask unpack after do_fetch |
97 | do_unpack[dirs] = "${WORKDIR}" | 97 | do_unpack[dirs] = "${WORKDIR}" |
98 | python base_do_unpack() { | 98 | python base_do_unpack() { |
99 | src_uri = (bb.data.getVar('SRC_URI', d, True) or "").split() | 99 | src_uri = (d.getVar('SRC_URI', True) or "").split() |
100 | if len(src_uri) == 0: | 100 | if len(src_uri) == 0: |
101 | return | 101 | return |
102 | 102 | ||
103 | localdata = bb.data.createCopy(d) | 103 | localdata = bb.data.createCopy(d) |
104 | bb.data.update_data(localdata) | 104 | bb.data.update_data(localdata) |
105 | 105 | ||
106 | rootdir = bb.data.getVar('WORKDIR', localdata, True) | 106 | rootdir = localdata.getVar('WORKDIR', True) |
107 | 107 | ||
108 | try: | 108 | try: |
109 | fetcher = bb.fetch2.Fetch(src_uri, localdata) | 109 | fetcher = bb.fetch2.Fetch(src_uri, localdata) |
@@ -118,7 +118,7 @@ def generate_git_config(e): | |||
118 | from bb import data | 118 | from bb import data |
119 | 119 | ||
120 | if data.getVar('GIT_CORE_CONFIG', e.data, True): | 120 | if data.getVar('GIT_CORE_CONFIG', e.data, True): |
121 | gitconfig_path = bb.data.getVar('GIT_CONFIG', e.data, True) | 121 | gitconfig_path = e.data.getVar('GIT_CONFIG', True) |
122 | proxy_command = " gitproxy = %s\n" % data.getVar('GIT_PROXY_COMMAND', e.data, True) | 122 | proxy_command = " gitproxy = %s\n" % data.getVar('GIT_PROXY_COMMAND', e.data, True) |
123 | 123 | ||
124 | bb.mkdirhier(bb.data.expand("${STAGING_DIR_NATIVE}/usr/etc/", e.data)) | 124 | bb.mkdirhier(bb.data.expand("${STAGING_DIR_NATIVE}/usr/etc/", e.data)) |
@@ -207,11 +207,11 @@ python base_eventhandler() { | |||
207 | name = getName(e) | 207 | name = getName(e) |
208 | 208 | ||
209 | if name.startswith("BuildStarted"): | 209 | if name.startswith("BuildStarted"): |
210 | bb.data.setVar( 'BB_VERSION', bb.__version__, e.data ) | 210 | e.data.setVar( 'BB_VERSION', bb.__version__) |
211 | statusvars = ['BB_VERSION', 'TARGET_ARCH', 'TARGET_OS', 'MACHINE', 'DISTRO', 'DISTRO_VERSION','TUNE_FEATURES', 'TARGET_FPU'] | 211 | statusvars = ['BB_VERSION', 'TARGET_ARCH', 'TARGET_OS', 'MACHINE', 'DISTRO', 'DISTRO_VERSION','TUNE_FEATURES', 'TARGET_FPU'] |
212 | statuslines = ["%-17s = \"%s\"" % (i, bb.data.getVar(i, e.data, 1) or '') for i in statusvars] | 212 | statuslines = ["%-17s = \"%s\"" % (i, e.data.getVar(i, 1) or '') for i in statusvars] |
213 | 213 | ||
214 | layers = (bb.data.getVar("BBLAYERS", e.data, 1) or "").split() | 214 | layers = (e.data.getVar("BBLAYERS", 1) or "").split() |
215 | layers_branch_rev = ["%-17s = \"%s:%s\"" % (os.path.basename(i), \ | 215 | layers_branch_rev = ["%-17s = \"%s:%s\"" % (os.path.basename(i), \ |
216 | base_get_metadata_git_branch(i, None).strip(), \ | 216 | base_get_metadata_git_branch(i, None).strip(), \ |
217 | base_get_metadata_git_revision(i, None)) \ | 217 | base_get_metadata_git_revision(i, None)) \ |
@@ -237,7 +237,7 @@ python base_eventhandler() { | |||
237 | needed_vars = [ "TARGET_ARCH", "TARGET_OS" ] | 237 | needed_vars = [ "TARGET_ARCH", "TARGET_OS" ] |
238 | pesteruser = [] | 238 | pesteruser = [] |
239 | for v in needed_vars: | 239 | for v in needed_vars: |
240 | val = bb.data.getVar(v, e.data, 1) | 240 | val = e.data.getVar(v, 1) |
241 | if not val or val == 'INVALID': | 241 | if not val or val == 'INVALID': |
242 | pesteruser.append(v) | 242 | pesteruser.append(v) |
243 | if pesteruser: | 243 | if pesteruser: |
@@ -330,23 +330,23 @@ python () { | |||
330 | appendVar('EXTRA_OECONF', extraconf) | 330 | appendVar('EXTRA_OECONF', extraconf) |
331 | 331 | ||
332 | # If PRINC is set, try and increase the PR value by the amount specified | 332 | # If PRINC is set, try and increase the PR value by the amount specified |
333 | princ = bb.data.getVar('PRINC', d, True) | 333 | princ = d.getVar('PRINC', True) |
334 | if princ: | 334 | if princ: |
335 | pr = bb.data.getVar('PR', d, True) | 335 | pr = d.getVar('PR', True) |
336 | pr_prefix = re.search("\D+",pr) | 336 | pr_prefix = re.search("\D+",pr) |
337 | prval = re.search("\d+",pr) | 337 | prval = re.search("\d+",pr) |
338 | if pr_prefix is None or prval is None: | 338 | if pr_prefix is None or prval is None: |
339 | bb.error("Unable to analyse format of PR variable: %s" % pr) | 339 | bb.error("Unable to analyse format of PR variable: %s" % pr) |
340 | nval = int(prval.group(0)) + int(princ) | 340 | nval = int(prval.group(0)) + int(princ) |
341 | pr = pr_prefix.group(0) + str(nval) + pr[prval.end():] | 341 | pr = pr_prefix.group(0) + str(nval) + pr[prval.end():] |
342 | bb.data.setVar('PR', pr, d) | 342 | d.setVar('PR', pr) |
343 | 343 | ||
344 | pn = bb.data.getVar('PN', d, 1) | 344 | pn = d.getVar('PN', 1) |
345 | license = bb.data.getVar('LICENSE', d, True) | 345 | license = d.getVar('LICENSE', True) |
346 | if license == "INVALID": | 346 | if license == "INVALID": |
347 | bb.fatal('This recipe does not have the LICENSE field set (%s)' % pn) | 347 | bb.fatal('This recipe does not have the LICENSE field set (%s)' % pn) |
348 | 348 | ||
349 | commercial_license = " %s " % bb.data.getVar('COMMERCIAL_LICENSE', d, 1) | 349 | commercial_license = " %s " % d.getVar('COMMERCIAL_LICENSE', 1) |
350 | import re | 350 | import re |
351 | pnr = "[ \t]%s[ \t]" % pn.replace('+', "\+") | 351 | pnr = "[ \t]%s[ \t]" % pn.replace('+', "\+") |
352 | if commercial_license and re.search(pnr, commercial_license): | 352 | if commercial_license and re.search(pnr, commercial_license): |
@@ -356,86 +356,86 @@ python () { | |||
356 | # If we're building a target package we need to use fakeroot (pseudo) | 356 | # If we're building a target package we need to use fakeroot (pseudo) |
357 | # in order to capture permissions, owners, groups and special files | 357 | # in order to capture permissions, owners, groups and special files |
358 | if not bb.data.inherits_class('native', d) and not bb.data.inherits_class('cross', d): | 358 | if not bb.data.inherits_class('native', d) and not bb.data.inherits_class('cross', d): |
359 | bb.data.setVarFlag('do_configure', 'umask', 022, d) | 359 | d.setVarFlag('do_configure', 'umask', 022) |
360 | bb.data.setVarFlag('do_compile', 'umask', 022, d) | 360 | d.setVarFlag('do_compile', 'umask', 022) |
361 | deps = (bb.data.getVarFlag('do_install', 'depends', d) or "").split() | 361 | deps = (d.getVarFlag('do_install', 'depends') or "").split() |
362 | deps.append('virtual/fakeroot-native:do_populate_sysroot') | 362 | deps.append('virtual/fakeroot-native:do_populate_sysroot') |
363 | bb.data.setVarFlag('do_install', 'depends', " ".join(deps),d) | 363 | bb.data.setVarFlag('do_install', 'depends', " ".join(deps),d) |
364 | bb.data.setVarFlag('do_install', 'fakeroot', 1, d) | 364 | d.setVarFlag('do_install', 'fakeroot', 1) |
365 | bb.data.setVarFlag('do_install', 'umask', 022, d) | 365 | d.setVarFlag('do_install', 'umask', 022) |
366 | deps = (bb.data.getVarFlag('do_package', 'depends', d) or "").split() | 366 | deps = (d.getVarFlag('do_package', 'depends') or "").split() |
367 | deps.append('virtual/fakeroot-native:do_populate_sysroot') | 367 | deps.append('virtual/fakeroot-native:do_populate_sysroot') |
368 | bb.data.setVarFlag('do_package', 'depends', " ".join(deps),d) | 368 | bb.data.setVarFlag('do_package', 'depends', " ".join(deps),d) |
369 | bb.data.setVarFlag('do_package', 'fakeroot', 1, d) | 369 | d.setVarFlag('do_package', 'fakeroot', 1) |
370 | bb.data.setVarFlag('do_package', 'umask', 022, d) | 370 | d.setVarFlag('do_package', 'umask', 022) |
371 | bb.data.setVarFlag('do_package_setscene', 'fakeroot', 1, d) | 371 | d.setVarFlag('do_package_setscene', 'fakeroot', 1) |
372 | source_mirror_fetch = bb.data.getVar('SOURCE_MIRROR_FETCH', d, 0) | 372 | source_mirror_fetch = d.getVar('SOURCE_MIRROR_FETCH', 0) |
373 | if not source_mirror_fetch: | 373 | if not source_mirror_fetch: |
374 | need_host = bb.data.getVar('COMPATIBLE_HOST', d, 1) | 374 | need_host = d.getVar('COMPATIBLE_HOST', 1) |
375 | if need_host: | 375 | if need_host: |
376 | import re | 376 | import re |
377 | this_host = bb.data.getVar('HOST_SYS', d, 1) | 377 | this_host = d.getVar('HOST_SYS', 1) |
378 | if not re.match(need_host, this_host): | 378 | if not re.match(need_host, this_host): |
379 | raise bb.parse.SkipPackage("incompatible with host %s" % this_host) | 379 | raise bb.parse.SkipPackage("incompatible with host %s" % this_host) |
380 | 380 | ||
381 | need_machine = bb.data.getVar('COMPATIBLE_MACHINE', d, 1) | 381 | need_machine = d.getVar('COMPATIBLE_MACHINE', 1) |
382 | if need_machine: | 382 | if need_machine: |
383 | import re | 383 | import re |
384 | this_machine = bb.data.getVar('MACHINE', d, 1) | 384 | this_machine = d.getVar('MACHINE', 1) |
385 | if this_machine and not re.match(need_machine, this_machine): | 385 | if this_machine and not re.match(need_machine, this_machine): |
386 | this_soc_family = bb.data.getVar('SOC_FAMILY', d, 1) | 386 | this_soc_family = d.getVar('SOC_FAMILY', 1) |
387 | if (this_soc_family and not re.match(need_machine, this_soc_family)) or not this_soc_family: | 387 | if (this_soc_family and not re.match(need_machine, this_soc_family)) or not this_soc_family: |
388 | raise bb.parse.SkipPackage("incompatible with machine %s" % this_machine) | 388 | raise bb.parse.SkipPackage("incompatible with machine %s" % this_machine) |
389 | 389 | ||
390 | 390 | ||
391 | dont_want_license = bb.data.getVar('INCOMPATIBLE_LICENSE', d, 1) | 391 | dont_want_license = d.getVar('INCOMPATIBLE_LICENSE', 1) |
392 | if dont_want_license and not pn.endswith("-native") and not pn.endswith("-cross") and not pn.endswith("-cross-initial") and not pn.endswith("-cross-intermediate") and not pn.endswith("-crosssdk-intermediate") and not pn.endswith("-crosssdk") and not pn.endswith("-crosssdk-initial"): | 392 | if dont_want_license and not pn.endswith("-native") and not pn.endswith("-cross") and not pn.endswith("-cross-initial") and not pn.endswith("-cross-intermediate") and not pn.endswith("-crosssdk-intermediate") and not pn.endswith("-crosssdk") and not pn.endswith("-crosssdk-initial"): |
393 | hosttools_whitelist = (bb.data.getVar('HOSTTOOLS_WHITELIST_%s' % dont_want_license, d, 1) or "").split() | 393 | hosttools_whitelist = (d.getVar('HOSTTOOLS_WHITELIST_%s' % dont_want_license, 1) or "").split() |
394 | lgplv2_whitelist = (bb.data.getVar('LGPLv2_WHITELIST_%s' % dont_want_license, d, 1) or "").split() | 394 | lgplv2_whitelist = (d.getVar('LGPLv2_WHITELIST_%s' % dont_want_license, 1) or "").split() |
395 | dont_want_whitelist = (bb.data.getVar('WHITELIST_%s' % dont_want_license, d, 1) or "").split() | 395 | dont_want_whitelist = (d.getVar('WHITELIST_%s' % dont_want_license, 1) or "").split() |
396 | if pn not in hosttools_whitelist and pn not in lgplv2_whitelist and pn not in dont_want_whitelist: | 396 | if pn not in hosttools_whitelist and pn not in lgplv2_whitelist and pn not in dont_want_whitelist: |
397 | 397 | ||
398 | import re | 398 | import re |
399 | this_license = bb.data.getVar('LICENSE', d, 1) | 399 | this_license = d.getVar('LICENSE', 1) |
400 | if this_license and re.search(dont_want_license, this_license): | 400 | if this_license and re.search(dont_want_license, this_license): |
401 | bb.note("SKIPPING %s because it's %s" % (pn, this_license)) | 401 | bb.note("SKIPPING %s because it's %s" % (pn, this_license)) |
402 | raise bb.parse.SkipPackage("incompatible with license %s" % this_license) | 402 | raise bb.parse.SkipPackage("incompatible with license %s" % this_license) |
403 | 403 | ||
404 | # Git packages should DEPEND on git-native | 404 | # Git packages should DEPEND on git-native |
405 | srcuri = bb.data.getVar('SRC_URI', d, 1) | 405 | srcuri = d.getVar('SRC_URI', 1) |
406 | if "git://" in srcuri: | 406 | if "git://" in srcuri: |
407 | depends = bb.data.getVarFlag('do_fetch', 'depends', d) or "" | 407 | depends = d.getVarFlag('do_fetch', 'depends') or "" |
408 | depends = depends + " git-native:do_populate_sysroot" | 408 | depends = depends + " git-native:do_populate_sysroot" |
409 | bb.data.setVarFlag('do_fetch', 'depends', depends, d) | 409 | d.setVarFlag('do_fetch', 'depends', depends) |
410 | 410 | ||
411 | # Mercurial packages should DEPEND on mercurial-native | 411 | # Mercurial packages should DEPEND on mercurial-native |
412 | elif "hg://" in srcuri: | 412 | elif "hg://" in srcuri: |
413 | depends = bb.data.getVarFlag('do_fetch', 'depends', d) or "" | 413 | depends = d.getVarFlag('do_fetch', 'depends') or "" |
414 | depends = depends + " mercurial-native:do_populate_sysroot" | 414 | depends = depends + " mercurial-native:do_populate_sysroot" |
415 | bb.data.setVarFlag('do_fetch', 'depends', depends, d) | 415 | d.setVarFlag('do_fetch', 'depends', depends) |
416 | 416 | ||
417 | # OSC packages should DEPEND on osc-native | 417 | # OSC packages should DEPEND on osc-native |
418 | elif "osc://" in srcuri: | 418 | elif "osc://" in srcuri: |
419 | depends = bb.data.getVarFlag('do_fetch', 'depends', d) or "" | 419 | depends = d.getVarFlag('do_fetch', 'depends') or "" |
420 | depends = depends + " osc-native:do_populate_sysroot" | 420 | depends = depends + " osc-native:do_populate_sysroot" |
421 | bb.data.setVarFlag('do_fetch', 'depends', depends, d) | 421 | d.setVarFlag('do_fetch', 'depends', depends) |
422 | 422 | ||
423 | # *.xz should depends on xz-native for unpacking | 423 | # *.xz should depends on xz-native for unpacking |
424 | # Not endswith because of "*.patch.xz;patch=1". Need bb.decodeurl in future | 424 | # Not endswith because of "*.patch.xz;patch=1". Need bb.decodeurl in future |
425 | if '.xz' in srcuri: | 425 | if '.xz' in srcuri: |
426 | depends = bb.data.getVarFlag('do_unpack', 'depends', d) or "" | 426 | depends = d.getVarFlag('do_unpack', 'depends') or "" |
427 | depends = depends + " xz-native:do_populate_sysroot" | 427 | depends = depends + " xz-native:do_populate_sysroot" |
428 | bb.data.setVarFlag('do_unpack', 'depends', depends, d) | 428 | d.setVarFlag('do_unpack', 'depends', depends) |
429 | 429 | ||
430 | # unzip-native should already be staged before unpacking ZIP recipes | 430 | # unzip-native should already be staged before unpacking ZIP recipes |
431 | if ".zip" in srcuri: | 431 | if ".zip" in srcuri: |
432 | depends = bb.data.getVarFlag('do_unpack', 'depends', d) or "" | 432 | depends = d.getVarFlag('do_unpack', 'depends') or "" |
433 | depends = depends + " unzip-native:do_populate_sysroot" | 433 | depends = depends + " unzip-native:do_populate_sysroot" |
434 | bb.data.setVarFlag('do_unpack', 'depends', depends, d) | 434 | d.setVarFlag('do_unpack', 'depends', depends) |
435 | 435 | ||
436 | # 'multimachine' handling | 436 | # 'multimachine' handling |
437 | mach_arch = bb.data.getVar('MACHINE_ARCH', d, 1) | 437 | mach_arch = d.getVar('MACHINE_ARCH', 1) |
438 | pkg_arch = bb.data.getVar('PACKAGE_ARCH', d, 1) | 438 | pkg_arch = d.getVar('PACKAGE_ARCH', 1) |
439 | 439 | ||
440 | if (pkg_arch == mach_arch): | 440 | if (pkg_arch == mach_arch): |
441 | # Already machine specific - nothing further to do | 441 | # Already machine specific - nothing further to do |
@@ -445,7 +445,7 @@ python () { | |||
445 | # We always try to scan SRC_URI for urls with machine overrides | 445 | # We always try to scan SRC_URI for urls with machine overrides |
446 | # unless the package sets SRC_URI_OVERRIDES_PACKAGE_ARCH=0 | 446 | # unless the package sets SRC_URI_OVERRIDES_PACKAGE_ARCH=0 |
447 | # | 447 | # |
448 | override = bb.data.getVar('SRC_URI_OVERRIDES_PACKAGE_ARCH', d, 1) | 448 | override = d.getVar('SRC_URI_OVERRIDES_PACKAGE_ARCH', 1) |
449 | if override != '0': | 449 | if override != '0': |
450 | paths = [] | 450 | paths = [] |
451 | for p in [ "${PF}", "${P}", "${PN}", "files", "" ]: | 451 | for p in [ "${PF}", "${P}", "${PN}", "files", "" ]: |
@@ -461,18 +461,18 @@ python () { | |||
461 | for mp in paths: | 461 | for mp in paths: |
462 | if local.startswith(mp): | 462 | if local.startswith(mp): |
463 | #bb.note("overriding PACKAGE_ARCH from %s to %s" % (pkg_arch, mach_arch)) | 463 | #bb.note("overriding PACKAGE_ARCH from %s to %s" % (pkg_arch, mach_arch)) |
464 | bb.data.setVar('PACKAGE_ARCH', "${MACHINE_ARCH}", d) | 464 | d.setVar('PACKAGE_ARCH', "${MACHINE_ARCH}") |
465 | return | 465 | return |
466 | 466 | ||
467 | packages = bb.data.getVar('PACKAGES', d, 1).split() | 467 | packages = d.getVar('PACKAGES', 1).split() |
468 | for pkg in packages: | 468 | for pkg in packages: |
469 | pkgarch = bb.data.getVar("PACKAGE_ARCH_%s" % pkg, d, 1) | 469 | pkgarch = d.getVar("PACKAGE_ARCH_%s" % pkg, 1) |
470 | 470 | ||
471 | # We could look for != PACKAGE_ARCH here but how to choose | 471 | # We could look for != PACKAGE_ARCH here but how to choose |
472 | # if multiple differences are present? | 472 | # if multiple differences are present? |
473 | # Look through PACKAGE_ARCHS for the priority order? | 473 | # Look through PACKAGE_ARCHS for the priority order? |
474 | if pkgarch and pkgarch == mach_arch: | 474 | if pkgarch and pkgarch == mach_arch: |
475 | bb.data.setVar('PACKAGE_ARCH', "${MACHINE_ARCH}", d) | 475 | d.setVar('PACKAGE_ARCH', "${MACHINE_ARCH}") |
476 | bb.warn("Recipe %s is marked as only being architecture specific but seems to have machine specific packages?! The recipe may as well mark itself as machine specific directly." % d.getVar("PN", True)) | 476 | bb.warn("Recipe %s is marked as only being architecture specific but seems to have machine specific packages?! The recipe may as well mark itself as machine specific directly." % d.getVar("PN", True)) |
477 | } | 477 | } |
478 | 478 | ||
@@ -483,7 +483,7 @@ python do_cleansstate() { | |||
483 | 483 | ||
484 | addtask cleanall after do_cleansstate | 484 | addtask cleanall after do_cleansstate |
485 | python do_cleanall() { | 485 | python do_cleanall() { |
486 | src_uri = (bb.data.getVar('SRC_URI', d, True) or "").split() | 486 | src_uri = (d.getVar('SRC_URI', True) or "").split() |
487 | if len(src_uri) == 0: | 487 | if len(src_uri) == 0: |
488 | return | 488 | return |
489 | 489 | ||