diff options
Diffstat (limited to 'meta/classes')
39 files changed, 208 insertions, 208 deletions
diff --git a/meta/classes/autotools.bbclass b/meta/classes/autotools.bbclass index a8d2b5f32e..66eba9fad0 100644 --- a/meta/classes/autotools.bbclass +++ b/meta/classes/autotools.bbclass | |||
@@ -1,8 +1,8 @@ | |||
1 | def autotools_dep_prepend(d): | 1 | def autotools_dep_prepend(d): |
2 | if d.getVar('INHIBIT_AUTOTOOLS_DEPS', 1): | 2 | if d.getVar('INHIBIT_AUTOTOOLS_DEPS', True): |
3 | return '' | 3 | return '' |
4 | 4 | ||
5 | pn = d.getVar('PN', 1) | 5 | pn = d.getVar('PN', True) |
6 | deps = '' | 6 | deps = '' |
7 | 7 | ||
8 | if pn in ['autoconf-native', 'automake-native', 'help2man-native']: | 8 | if pn in ['autoconf-native', 'automake-native', 'help2man-native']: |
@@ -13,7 +13,7 @@ def autotools_dep_prepend(d): | |||
13 | deps += 'libtool-native ' | 13 | deps += 'libtool-native ' |
14 | if not bb.data.inherits_class('native', d) \ | 14 | if not bb.data.inherits_class('native', d) \ |
15 | and not bb.data.inherits_class('cross', d) \ | 15 | and not bb.data.inherits_class('cross', d) \ |
16 | and not d.getVar('INHIBIT_DEFAULT_DEPS', 1): | 16 | and not d.getVar('INHIBIT_DEFAULT_DEPS', True): |
17 | deps += 'libtool-cross ' | 17 | deps += 'libtool-cross ' |
18 | 18 | ||
19 | return deps + 'gnu-config-native ' | 19 | return deps + 'gnu-config-native ' |
diff --git a/meta/classes/base.bbclass b/meta/classes/base.bbclass index cab56deb39..48e4a28d83 100644 --- a/meta/classes/base.bbclass +++ b/meta/classes/base.bbclass | |||
@@ -60,8 +60,8 @@ def base_dep_prepend(d): | |||
60 | # we need that built is the responsibility of the patch function / class, not | 60 | # we need that built is the responsibility of the patch function / class, not |
61 | # the application. | 61 | # the application. |
62 | if not d.getVar('INHIBIT_DEFAULT_DEPS'): | 62 | if not d.getVar('INHIBIT_DEFAULT_DEPS'): |
63 | if (d.getVar('HOST_SYS', 1) != | 63 | if (d.getVar('HOST_SYS', True) != |
64 | d.getVar('BUILD_SYS', 1)): | 64 | d.getVar('BUILD_SYS', True)): |
65 | deps += " virtual/${TARGET_PREFIX}gcc virtual/${TARGET_PREFIX}compilerlibs virtual/libc " | 65 | deps += " virtual/${TARGET_PREFIX}gcc virtual/${TARGET_PREFIX}compilerlibs virtual/libc " |
66 | return deps | 66 | return deps |
67 | 67 | ||
@@ -203,7 +203,7 @@ def preferred_ml_updates(d): | |||
203 | 203 | ||
204 | 204 | ||
205 | def get_layers_branch_rev(d): | 205 | def get_layers_branch_rev(d): |
206 | layers = (d.getVar("BBLAYERS", 1) or "").split() | 206 | layers = (d.getVar("BBLAYERS", True) or "").split() |
207 | layers_branch_rev = ["%-17s = \"%s:%s\"" % (os.path.basename(i), \ | 207 | layers_branch_rev = ["%-17s = \"%s:%s\"" % (os.path.basename(i), \ |
208 | base_get_metadata_git_branch(i, None).strip(), \ | 208 | base_get_metadata_git_branch(i, None).strip(), \ |
209 | base_get_metadata_git_revision(i, None)) \ | 209 | base_get_metadata_git_revision(i, None)) \ |
@@ -233,7 +233,7 @@ python base_eventhandler() { | |||
233 | if name.startswith("BuildStarted"): | 233 | if name.startswith("BuildStarted"): |
234 | e.data.setVar( 'BB_VERSION', bb.__version__) | 234 | e.data.setVar( 'BB_VERSION', bb.__version__) |
235 | statusvars = ['BB_VERSION', 'TARGET_ARCH', 'TARGET_OS', 'MACHINE', 'DISTRO', 'DISTRO_VERSION','TUNE_FEATURES', 'TARGET_FPU'] | 235 | statusvars = ['BB_VERSION', 'TARGET_ARCH', 'TARGET_OS', 'MACHINE', 'DISTRO', 'DISTRO_VERSION','TUNE_FEATURES', 'TARGET_FPU'] |
236 | statuslines = ["%-17s = \"%s\"" % (i, e.data.getVar(i, 1) or '') for i in statusvars] | 236 | statuslines = ["%-17s = \"%s\"" % (i, e.data.getVar(i, True) or '') for i in statusvars] |
237 | 237 | ||
238 | statuslines += get_layers_branch_rev(e.data) | 238 | statuslines += get_layers_branch_rev(e.data) |
239 | statusmsg = "\nOE Build Configuration:\n%s\n" % '\n'.join(statuslines) | 239 | statusmsg = "\nOE Build Configuration:\n%s\n" % '\n'.join(statuslines) |
@@ -242,7 +242,7 @@ python base_eventhandler() { | |||
242 | needed_vars = [ "TARGET_ARCH", "TARGET_OS" ] | 242 | needed_vars = [ "TARGET_ARCH", "TARGET_OS" ] |
243 | pesteruser = [] | 243 | pesteruser = [] |
244 | for v in needed_vars: | 244 | for v in needed_vars: |
245 | val = e.data.getVar(v, 1) | 245 | val = e.data.getVar(v, True) |
246 | if not val or val == 'INVALID': | 246 | if not val or val == 'INVALID': |
247 | pesteruser.append(v) | 247 | pesteruser.append(v) |
248 | if pesteruser: | 248 | if pesteruser: |
@@ -344,7 +344,7 @@ python () { | |||
344 | pr = pr_prefix.group(0) + str(nval) + pr[prval.end():] | 344 | pr = pr_prefix.group(0) + str(nval) + pr[prval.end():] |
345 | d.setVar('PR', pr) | 345 | d.setVar('PR', pr) |
346 | 346 | ||
347 | pn = d.getVar('PN', 1) | 347 | pn = d.getVar('PN', True) |
348 | license = d.getVar('LICENSE', True) | 348 | license = d.getVar('LICENSE', True) |
349 | if license == "INVALID": | 349 | if license == "INVALID": |
350 | bb.fatal('This recipe does not have the LICENSE field set (%s)' % pn) | 350 | bb.fatal('This recipe does not have the LICENSE field set (%s)' % pn) |
@@ -370,36 +370,36 @@ python () { | |||
370 | d.setVarFlag('do_package_setscene', 'fakeroot', 1) | 370 | d.setVarFlag('do_package_setscene', 'fakeroot', 1) |
371 | source_mirror_fetch = d.getVar('SOURCE_MIRROR_FETCH', 0) | 371 | source_mirror_fetch = d.getVar('SOURCE_MIRROR_FETCH', 0) |
372 | if not source_mirror_fetch: | 372 | if not source_mirror_fetch: |
373 | need_host = d.getVar('COMPATIBLE_HOST', 1) | 373 | need_host = d.getVar('COMPATIBLE_HOST', True) |
374 | if need_host: | 374 | if need_host: |
375 | import re | 375 | import re |
376 | this_host = d.getVar('HOST_SYS', 1) | 376 | this_host = d.getVar('HOST_SYS', True) |
377 | if not re.match(need_host, this_host): | 377 | if not re.match(need_host, this_host): |
378 | raise bb.parse.SkipPackage("incompatible with host %s (not in COMPATIBLE_HOST)" % this_host) | 378 | raise bb.parse.SkipPackage("incompatible with host %s (not in COMPATIBLE_HOST)" % this_host) |
379 | 379 | ||
380 | need_machine = d.getVar('COMPATIBLE_MACHINE', 1) | 380 | need_machine = d.getVar('COMPATIBLE_MACHINE', True) |
381 | if need_machine: | 381 | if need_machine: |
382 | import re | 382 | import re |
383 | this_machine = d.getVar('MACHINE', 1) | 383 | this_machine = d.getVar('MACHINE', True) |
384 | if this_machine and not re.match(need_machine, this_machine): | 384 | if this_machine and not re.match(need_machine, this_machine): |
385 | this_soc_family = d.getVar('SOC_FAMILY', 1) | 385 | this_soc_family = d.getVar('SOC_FAMILY', True) |
386 | if (this_soc_family and not re.match(need_machine, this_soc_family)) or not this_soc_family: | 386 | if (this_soc_family and not re.match(need_machine, this_soc_family)) or not this_soc_family: |
387 | raise bb.parse.SkipPackage("incompatible with machine %s (not in COMPATIBLE_MACHINE)" % this_machine) | 387 | raise bb.parse.SkipPackage("incompatible with machine %s (not in COMPATIBLE_MACHINE)" % this_machine) |
388 | 388 | ||
389 | 389 | ||
390 | dont_want_license = d.getVar('INCOMPATIBLE_LICENSE', 1) | 390 | dont_want_license = d.getVar('INCOMPATIBLE_LICENSE', True) |
391 | if dont_want_license and not pn.endswith("-native") and not pn.endswith("-cross") and not pn.endswith("-cross-initial") and not pn.endswith("-cross-intermediate") and not pn.endswith("-crosssdk-intermediate") and not pn.endswith("-crosssdk") and not pn.endswith("-crosssdk-initial"): | 391 | if dont_want_license and not pn.endswith("-native") and not pn.endswith("-cross") and not pn.endswith("-cross-initial") and not pn.endswith("-cross-intermediate") and not pn.endswith("-crosssdk-intermediate") and not pn.endswith("-crosssdk") and not pn.endswith("-crosssdk-initial"): |
392 | hosttools_whitelist = (d.getVar('HOSTTOOLS_WHITELIST_%s' % dont_want_license, 1) or "").split() | 392 | hosttools_whitelist = (d.getVar('HOSTTOOLS_WHITELIST_%s' % dont_want_license, True) or "").split() |
393 | lgplv2_whitelist = (d.getVar('LGPLv2_WHITELIST_%s' % dont_want_license, 1) or "").split() | 393 | lgplv2_whitelist = (d.getVar('LGPLv2_WHITELIST_%s' % dont_want_license, True) or "").split() |
394 | dont_want_whitelist = (d.getVar('WHITELIST_%s' % dont_want_license, 1) or "").split() | 394 | dont_want_whitelist = (d.getVar('WHITELIST_%s' % dont_want_license, True) or "").split() |
395 | if pn not in hosttools_whitelist and pn not in lgplv2_whitelist and pn not in dont_want_whitelist: | 395 | if pn not in hosttools_whitelist and pn not in lgplv2_whitelist and pn not in dont_want_whitelist: |
396 | 396 | ||
397 | this_license = d.getVar('LICENSE', 1) | 397 | this_license = d.getVar('LICENSE', True) |
398 | if incompatible_license(d,dont_want_license): | 398 | if incompatible_license(d,dont_want_license): |
399 | bb.note("SKIPPING %s because it's %s" % (pn, this_license)) | 399 | bb.note("SKIPPING %s because it's %s" % (pn, this_license)) |
400 | raise bb.parse.SkipPackage("incompatible with license %s" % this_license) | 400 | raise bb.parse.SkipPackage("incompatible with license %s" % this_license) |
401 | 401 | ||
402 | srcuri = d.getVar('SRC_URI', 1) | 402 | srcuri = d.getVar('SRC_URI', True) |
403 | # Svn packages should DEPEND on subversion-native | 403 | # Svn packages should DEPEND on subversion-native |
404 | if "svn://" in srcuri: | 404 | if "svn://" in srcuri: |
405 | d.appendVarFlag('do_fetch', 'depends', ' subversion-native:do_populate_sysroot') | 405 | d.appendVarFlag('do_fetch', 'depends', ' subversion-native:do_populate_sysroot') |
@@ -426,8 +426,8 @@ python () { | |||
426 | d.appendVarFlag('do_unpack', 'depends', ' unzip-native:do_populate_sysroot') | 426 | d.appendVarFlag('do_unpack', 'depends', ' unzip-native:do_populate_sysroot') |
427 | 427 | ||
428 | # 'multimachine' handling | 428 | # 'multimachine' handling |
429 | mach_arch = d.getVar('MACHINE_ARCH', 1) | 429 | mach_arch = d.getVar('MACHINE_ARCH', True) |
430 | pkg_arch = d.getVar('PACKAGE_ARCH', 1) | 430 | pkg_arch = d.getVar('PACKAGE_ARCH', True) |
431 | 431 | ||
432 | if (pkg_arch == mach_arch): | 432 | if (pkg_arch == mach_arch): |
433 | # Already machine specific - nothing further to do | 433 | # Already machine specific - nothing further to do |
@@ -458,9 +458,9 @@ python () { | |||
458 | d.setVar('PACKAGE_ARCH', "${MACHINE_ARCH}") | 458 | d.setVar('PACKAGE_ARCH', "${MACHINE_ARCH}") |
459 | return | 459 | return |
460 | 460 | ||
461 | packages = d.getVar('PACKAGES', 1).split() | 461 | packages = d.getVar('PACKAGES', True).split() |
462 | for pkg in packages: | 462 | for pkg in packages: |
463 | pkgarch = d.getVar("PACKAGE_ARCH_%s" % pkg, 1) | 463 | pkgarch = d.getVar("PACKAGE_ARCH_%s" % pkg, True) |
464 | 464 | ||
465 | # We could look for != PACKAGE_ARCH here but how to choose | 465 | # We could look for != PACKAGE_ARCH here but how to choose |
466 | # if multiple differences are present? | 466 | # if multiple differences are present? |
diff --git a/meta/classes/copyleft_compliance.bbclass b/meta/classes/copyleft_compliance.bbclass index 2eb9dedd24..4082e7e15d 100644 --- a/meta/classes/copyleft_compliance.bbclass +++ b/meta/classes/copyleft_compliance.bbclass | |||
@@ -69,8 +69,8 @@ python do_prepare_copyleft_sources () { | |||
69 | else: | 69 | else: |
70 | bb.debug(1, 'copyleft: %s is included' % p) | 70 | bb.debug(1, 'copyleft: %s is included' % p) |
71 | 71 | ||
72 | sources_dir = d.getVar('COPYLEFT_SOURCES_DIR', 1) | 72 | sources_dir = d.getVar('COPYLEFT_SOURCES_DIR', True) |
73 | src_uri = d.getVar('SRC_URI', 1).split() | 73 | src_uri = d.getVar('SRC_URI', True).split() |
74 | fetch = bb.fetch2.Fetch(src_uri, d) | 74 | fetch = bb.fetch2.Fetch(src_uri, d) |
75 | ud = fetch.ud | 75 | ud = fetch.ud |
76 | 76 | ||
diff --git a/meta/classes/cpan-base.bbclass b/meta/classes/cpan-base.bbclass index 79582ca76c..6cb1fefc29 100644 --- a/meta/classes/cpan-base.bbclass +++ b/meta/classes/cpan-base.bbclass | |||
@@ -28,7 +28,7 @@ def get_perl_version(d): | |||
28 | 28 | ||
29 | # Determine where the library directories are | 29 | # Determine where the library directories are |
30 | def perl_get_libdirs(d): | 30 | def perl_get_libdirs(d): |
31 | libdir = d.getVar('libdir', 1) | 31 | libdir = d.getVar('libdir', True) |
32 | if is_target(d) == "no": | 32 | if is_target(d) == "no": |
33 | libdir += '/perl-native' | 33 | libdir += '/perl-native' |
34 | libdir += '/perl' | 34 | libdir += '/perl' |
diff --git a/meta/classes/cpan_build.bbclass b/meta/classes/cpan_build.bbclass index 981332c4fa..36ffc56b85 100644 --- a/meta/classes/cpan_build.bbclass +++ b/meta/classes/cpan_build.bbclass | |||
@@ -10,9 +10,9 @@ inherit cpan-base | |||
10 | # libmodule-build-perl) | 10 | # libmodule-build-perl) |
11 | # | 11 | # |
12 | def cpan_build_dep_prepend(d): | 12 | def cpan_build_dep_prepend(d): |
13 | if d.getVar('CPAN_BUILD_DEPS', 1): | 13 | if d.getVar('CPAN_BUILD_DEPS', True): |
14 | return '' | 14 | return '' |
15 | pn = d.getVar('PN', 1) | 15 | pn = d.getVar('PN', True) |
16 | if pn in ['libmodule-build-perl', 'libmodule-build-perl-native']: | 16 | if pn in ['libmodule-build-perl', 'libmodule-build-perl-native']: |
17 | return '' | 17 | return '' |
18 | return 'libmodule-build-perl-native ' | 18 | return 'libmodule-build-perl-native ' |
diff --git a/meta/classes/debian.bbclass b/meta/classes/debian.bbclass index 025abcfad0..3637e2ebe7 100644 --- a/meta/classes/debian.bbclass +++ b/meta/classes/debian.bbclass | |||
@@ -22,8 +22,8 @@ python () { | |||
22 | python debian_package_name_hook () { | 22 | python debian_package_name_hook () { |
23 | import glob, copy, stat, errno, re | 23 | import glob, copy, stat, errno, re |
24 | 24 | ||
25 | pkgdest = d.getVar('PKGDEST', 1) | 25 | pkgdest = d.getVar('PKGDEST', True) |
26 | packages = d.getVar('PACKAGES', 1) | 26 | packages = d.getVar('PACKAGES', True) |
27 | bin_re = re.compile(".*/s?" + os.path.basename(d.getVar("bindir", True)) + "$") | 27 | bin_re = re.compile(".*/s?" + os.path.basename(d.getVar("bindir", True)) + "$") |
28 | lib_re = re.compile(".*/" + os.path.basename(d.getVar("libdir", True)) + "$") | 28 | lib_re = re.compile(".*/" + os.path.basename(d.getVar("libdir", True)) + "$") |
29 | so_re = re.compile("lib.*\.so") | 29 | so_re = re.compile("lib.*\.so") |
@@ -60,7 +60,7 @@ python debian_package_name_hook () { | |||
60 | for f in files: | 60 | for f in files: |
61 | if so_re.match(f): | 61 | if so_re.match(f): |
62 | fp = os.path.join(root, f) | 62 | fp = os.path.join(root, f) |
63 | cmd = (d.getVar('BUILD_PREFIX', 1) or "") + "objdump -p " + fp + " 2>/dev/null" | 63 | cmd = (d.getVar('BUILD_PREFIX', True) or "") + "objdump -p " + fp + " 2>/dev/null" |
64 | fd = os.popen(cmd) | 64 | fd = os.popen(cmd) |
65 | lines = fd.readlines() | 65 | lines = fd.readlines() |
66 | fd.close() | 66 | fd.close() |
@@ -74,7 +74,7 @@ python debian_package_name_hook () { | |||
74 | if len(sonames) == 1: | 74 | if len(sonames) == 1: |
75 | soname = sonames[0] | 75 | soname = sonames[0] |
76 | elif len(sonames) > 1: | 76 | elif len(sonames) > 1: |
77 | lead = d.getVar('LEAD_SONAME', 1) | 77 | lead = d.getVar('LEAD_SONAME', True) |
78 | if lead: | 78 | if lead: |
79 | r = re.compile(lead) | 79 | r = re.compile(lead) |
80 | filtered = [] | 80 | filtered = [] |
@@ -117,7 +117,7 @@ python debian_package_name_hook () { | |||
117 | # and later | 117 | # and later |
118 | # DEBUG: LIBNAMES: pkgname libtic5 devname libtic pkg ncurses-libticw orig_pkg ncurses-libtic debian_pn None newpkg libticw | 118 | # DEBUG: LIBNAMES: pkgname libtic5 devname libtic pkg ncurses-libticw orig_pkg ncurses-libtic debian_pn None newpkg libticw |
119 | # so we need to handle ncurses-libticw->libticw5 before ncurses-libtic->libtic5 | 119 | # so we need to handle ncurses-libticw->libticw5 before ncurses-libtic->libtic5 |
120 | for pkg in sorted((d.getVar('AUTO_LIBNAME_PKGS', 1) or "").split(), reverse=True): | 120 | for pkg in sorted((d.getVar('AUTO_LIBNAME_PKGS', True) or "").split(), reverse=True): |
121 | auto_libname(packages, pkg) | 121 | auto_libname(packages, pkg) |
122 | } | 122 | } |
123 | 123 | ||
diff --git a/meta/classes/distrodata.bbclass b/meta/classes/distrodata.bbclass index ff5b836871..aba4bd7fa6 100644 --- a/meta/classes/distrodata.bbclass +++ b/meta/classes/distrodata.bbclass | |||
@@ -372,7 +372,7 @@ python do_checkpkg() { | |||
372 | 372 | ||
373 | f.close() | 373 | f.close() |
374 | if status != "ErrHostNoDir" and re.match("Err", status): | 374 | if status != "ErrHostNoDir" and re.match("Err", status): |
375 | logpath = d.getVar('LOG_DIR', 1) | 375 | logpath = d.getVar('LOG_DIR', True) |
376 | os.system("cp %s %s/" % (f.name, logpath)) | 376 | os.system("cp %s %s/" % (f.name, logpath)) |
377 | os.unlink(f.name) | 377 | os.unlink(f.name) |
378 | return status | 378 | return status |
diff --git a/meta/classes/distutils-base.bbclass b/meta/classes/distutils-base.bbclass index e7d0bb8071..6d18e08f14 100644 --- a/meta/classes/distutils-base.bbclass +++ b/meta/classes/distutils-base.bbclass | |||
@@ -1,4 +1,4 @@ | |||
1 | DEPENDS += "${@["python-native python", ""][(d.getVar('PACKAGES', 1) == '')]}" | 1 | DEPENDS += "${@["python-native python", ""][(d.getVar('PACKAGES', True) == '')]}" |
2 | RDEPENDS_${PN} += "${@['', 'python-core']['${PN}' == '${BPN}']}" | 2 | RDEPENDS_${PN} += "${@['', 'python-core']['${PN}' == '${BPN}']}" |
3 | 3 | ||
4 | inherit distutils-common-base | 4 | inherit distutils-common-base |
diff --git a/meta/classes/distutils-native-base.bbclass b/meta/classes/distutils-native-base.bbclass index 47367d796b..ceda512e39 100644 --- a/meta/classes/distutils-native-base.bbclass +++ b/meta/classes/distutils-native-base.bbclass | |||
@@ -1,3 +1,3 @@ | |||
1 | DEPENDS += "${@["python-native", ""][(d.getVar('PACKAGES', 1) == '')]}" | 1 | DEPENDS += "${@["python-native", ""][(d.getVar('PACKAGES', True) == '')]}" |
2 | 2 | ||
3 | inherit distutils-common-base | 3 | inherit distutils-common-base |
diff --git a/meta/classes/gconf.bbclass b/meta/classes/gconf.bbclass index 7bfa871bd2..095d04b1b8 100644 --- a/meta/classes/gconf.bbclass +++ b/meta/classes/gconf.bbclass | |||
@@ -32,8 +32,8 @@ done | |||
32 | 32 | ||
33 | python populate_packages_append () { | 33 | python populate_packages_append () { |
34 | import re | 34 | import re |
35 | packages = d.getVar('PACKAGES', 1).split() | 35 | packages = d.getVar('PACKAGES', True).split() |
36 | pkgdest = d.getVar('PKGDEST', 1) | 36 | pkgdest = d.getVar('PKGDEST', True) |
37 | 37 | ||
38 | for pkg in packages: | 38 | for pkg in packages: |
39 | schema_dir = '%s/%s/etc/gconf/schemas' % (pkgdest, pkg) | 39 | schema_dir = '%s/%s/etc/gconf/schemas' % (pkgdest, pkg) |
@@ -46,15 +46,15 @@ python populate_packages_append () { | |||
46 | if schemas != []: | 46 | if schemas != []: |
47 | bb.note("adding gconf postinst and prerm scripts to %s" % pkg) | 47 | bb.note("adding gconf postinst and prerm scripts to %s" % pkg) |
48 | d.setVar('SCHEMA_FILES', " ".join(schemas)) | 48 | d.setVar('SCHEMA_FILES', " ".join(schemas)) |
49 | postinst = d.getVar('pkg_postinst_%s' % pkg, 1) or d.getVar('pkg_postinst', 1) | 49 | postinst = d.getVar('pkg_postinst_%s' % pkg, True) or d.getVar('pkg_postinst', True) |
50 | if not postinst: | 50 | if not postinst: |
51 | postinst = '#!/bin/sh\n' | 51 | postinst = '#!/bin/sh\n' |
52 | postinst += d.getVar('gconf_postinst', 1) | 52 | postinst += d.getVar('gconf_postinst', True) |
53 | d.setVar('pkg_postinst_%s' % pkg, postinst) | 53 | d.setVar('pkg_postinst_%s' % pkg, postinst) |
54 | prerm = d.getVar('pkg_prerm_%s' % pkg, 1) or d.getVar('pkg_prerm', 1) | 54 | prerm = d.getVar('pkg_prerm_%s' % pkg, True) or d.getVar('pkg_prerm', True) |
55 | if not prerm: | 55 | if not prerm: |
56 | prerm = '#!/bin/sh\n' | 56 | prerm = '#!/bin/sh\n' |
57 | prerm += d.getVar('gconf_prerm', 1) | 57 | prerm += d.getVar('gconf_prerm', True) |
58 | d.setVar('pkg_prerm_%s' % pkg, prerm) | 58 | d.setVar('pkg_prerm_%s' % pkg, prerm) |
59 | rdepends = d.getVar("RDEPENDS_%s" % pkg, True) or "" | 59 | rdepends = d.getVar("RDEPENDS_%s" % pkg, True) or "" |
60 | rdepends += " gconf" | 60 | rdepends += " gconf" |
diff --git a/meta/classes/gtk-icon-cache.bbclass b/meta/classes/gtk-icon-cache.bbclass index 0204fd3fec..60e3401f4b 100644 --- a/meta/classes/gtk-icon-cache.bbclass +++ b/meta/classes/gtk-icon-cache.bbclass | |||
@@ -28,31 +28,31 @@ done | |||
28 | } | 28 | } |
29 | 29 | ||
30 | python populate_packages_append () { | 30 | python populate_packages_append () { |
31 | packages = d.getVar('PACKAGES', 1).split() | 31 | packages = d.getVar('PACKAGES', True).split() |
32 | pkgdest = d.getVar('PKGDEST', 1) | 32 | pkgdest = d.getVar('PKGDEST', True) |
33 | 33 | ||
34 | for pkg in packages: | 34 | for pkg in packages: |
35 | icon_dir = '%s/%s/%s/icons' % (pkgdest, pkg, d.getVar('datadir', 1)) | 35 | icon_dir = '%s/%s/%s/icons' % (pkgdest, pkg, d.getVar('datadir', True)) |
36 | if not os.path.exists(icon_dir): | 36 | if not os.path.exists(icon_dir): |
37 | continue | 37 | continue |
38 | 38 | ||
39 | bb.note("adding hicolor-icon-theme dependency to %s" % pkg) | 39 | bb.note("adding hicolor-icon-theme dependency to %s" % pkg) |
40 | rdepends = d.getVar('RDEPENDS_%s' % pkg, 1) | 40 | rdepends = d.getVar('RDEPENDS_%s' % pkg, True) |
41 | rdepends = rdepends + ' ' + d.getVar('MLPREFIX') + "hicolor-icon-theme" | 41 | rdepends = rdepends + ' ' + d.getVar('MLPREFIX') + "hicolor-icon-theme" |
42 | d.setVar('RDEPENDS_%s' % pkg, rdepends) | 42 | d.setVar('RDEPENDS_%s' % pkg, rdepends) |
43 | 43 | ||
44 | bb.note("adding gtk-icon-cache postinst and postrm scripts to %s" % pkg) | 44 | bb.note("adding gtk-icon-cache postinst and postrm scripts to %s" % pkg) |
45 | 45 | ||
46 | postinst = d.getVar('pkg_postinst_%s' % pkg, 1) or d.getVar('pkg_postinst', 1) | 46 | postinst = d.getVar('pkg_postinst_%s' % pkg, True) or d.getVar('pkg_postinst', True) |
47 | if not postinst: | 47 | if not postinst: |
48 | postinst = '#!/bin/sh\n' | 48 | postinst = '#!/bin/sh\n' |
49 | postinst += d.getVar('gtk_icon_cache_postinst', 1) | 49 | postinst += d.getVar('gtk_icon_cache_postinst', True) |
50 | d.setVar('pkg_postinst_%s' % pkg, postinst) | 50 | d.setVar('pkg_postinst_%s' % pkg, postinst) |
51 | 51 | ||
52 | postrm = d.getVar('pkg_postrm_%s' % pkg, 1) or d.getVar('pkg_postrm', 1) | 52 | postrm = d.getVar('pkg_postrm_%s' % pkg, True) or d.getVar('pkg_postrm', True) |
53 | if not postrm: | 53 | if not postrm: |
54 | postrm = '#!/bin/sh\n' | 54 | postrm = '#!/bin/sh\n' |
55 | postrm += d.getVar('gtk_icon_cache_postrm', 1) | 55 | postrm += d.getVar('gtk_icon_cache_postrm', True) |
56 | d.setVar('pkg_postrm_%s' % pkg, postrm) | 56 | d.setVar('pkg_postrm_%s' % pkg, postrm) |
57 | } | 57 | } |
58 | 58 | ||
diff --git a/meta/classes/image.bbclass b/meta/classes/image.bbclass index dfce381393..a62eb2cd57 100644 --- a/meta/classes/image.bbclass +++ b/meta/classes/image.bbclass | |||
@@ -98,15 +98,15 @@ python () { | |||
98 | # is searched for in the BBPATH (same as the old version.) | 98 | # is searched for in the BBPATH (same as the old version.) |
99 | # | 99 | # |
100 | def get_devtable_list(d): | 100 | def get_devtable_list(d): |
101 | devtable = d.getVar('IMAGE_DEVICE_TABLE', 1) | 101 | devtable = d.getVar('IMAGE_DEVICE_TABLE', True) |
102 | if devtable != None: | 102 | if devtable != None: |
103 | return devtable | 103 | return devtable |
104 | str = "" | 104 | str = "" |
105 | devtables = d.getVar('IMAGE_DEVICE_TABLES', 1) | 105 | devtables = d.getVar('IMAGE_DEVICE_TABLES', True) |
106 | if devtables == None: | 106 | if devtables == None: |
107 | devtables = 'files/device_table-minimal.txt' | 107 | devtables = 'files/device_table-minimal.txt' |
108 | for devtable in devtables.split(): | 108 | for devtable in devtables.split(): |
109 | str += " %s" % bb.which(d.getVar('BBPATH', 1), devtable) | 109 | str += " %s" % bb.which(d.getVar('BBPATH', True), devtable) |
110 | return str | 110 | return str |
111 | 111 | ||
112 | IMAGE_CLASSES ?= "image_types" | 112 | IMAGE_CLASSES ?= "image_types" |
@@ -119,7 +119,7 @@ ROOTFS_POSTPROCESS_COMMAND ?= "" | |||
119 | # some default locales | 119 | # some default locales |
120 | IMAGE_LINGUAS ?= "de-de fr-fr en-gb" | 120 | IMAGE_LINGUAS ?= "de-de fr-fr en-gb" |
121 | 121 | ||
122 | LINGUAS_INSTALL ?= "${@" ".join(map(lambda s: "locale-base-%s" % s, d.getVar('IMAGE_LINGUAS', 1).split()))}" | 122 | LINGUAS_INSTALL ?= "${@" ".join(map(lambda s: "locale-base-%s" % s, d.getVar('IMAGE_LINGUAS', True).split()))}" |
123 | 123 | ||
124 | PSEUDO_PASSWD = "${IMAGE_ROOTFS}" | 124 | PSEUDO_PASSWD = "${IMAGE_ROOTFS}" |
125 | 125 | ||
diff --git a/meta/classes/imagetest-qemu.bbclass b/meta/classes/imagetest-qemu.bbclass index d01d1f4979..d56b44b5c4 100644 --- a/meta/classes/imagetest-qemu.bbclass +++ b/meta/classes/imagetest-qemu.bbclass | |||
@@ -35,12 +35,12 @@ def qemuimagetest_main(d): | |||
35 | 35 | ||
36 | casestr = re.compile(r'(?P<scen>\w+\b):(?P<case>\S+$)') | 36 | casestr = re.compile(r'(?P<scen>\w+\b):(?P<case>\S+$)') |
37 | resultstr = re.compile(r'\s*(?P<case>\w+)\s*(?P<pass>\d+)\s*(?P<fail>\d+)\s*(?P<noresult>\d+)') | 37 | resultstr = re.compile(r'\s*(?P<case>\w+)\s*(?P<pass>\d+)\s*(?P<fail>\d+)\s*(?P<noresult>\d+)') |
38 | machine = d.getVar('MACHINE', 1) | 38 | machine = d.getVar('MACHINE', True) |
39 | pname = d.getVar('PN', 1) | 39 | pname = d.getVar('PN', True) |
40 | 40 | ||
41 | """function to save test cases running status""" | 41 | """function to save test cases running status""" |
42 | def teststatus(test, status, index, length): | 42 | def teststatus(test, status, index, length): |
43 | test_status = d.getVar('TEST_STATUS', 1) | 43 | test_status = d.getVar('TEST_STATUS', True) |
44 | if not os.path.exists(test_status): | 44 | if not os.path.exists(test_status): |
45 | raise bb.build.FuncFailed("No test status file existing under TEST_TMP") | 45 | raise bb.build.FuncFailed("No test status file existing under TEST_TMP") |
46 | 46 | ||
@@ -51,13 +51,13 @@ def qemuimagetest_main(d): | |||
51 | 51 | ||
52 | """funtion to run each case under scenario""" | 52 | """funtion to run each case under scenario""" |
53 | def runtest(scen, case, fulltestpath): | 53 | def runtest(scen, case, fulltestpath): |
54 | resultpath = d.getVar('TEST_RESULT', 1) | 54 | resultpath = d.getVar('TEST_RESULT', True) |
55 | tmppath = d.getVar('TEST_TMP', 1) | 55 | tmppath = d.getVar('TEST_TMP', True) |
56 | 56 | ||
57 | """initialize log file for testcase""" | 57 | """initialize log file for testcase""" |
58 | logpath = d.getVar('TEST_LOG', 1) | 58 | logpath = d.getVar('TEST_LOG', True) |
59 | bb.utils.mkdirhier("%s/%s" % (logpath, scen)) | 59 | bb.utils.mkdirhier("%s/%s" % (logpath, scen)) |
60 | caselog = os.path.join(logpath, "%s/log_%s.%s" % (scen, case, d.getVar('DATETIME', 1))) | 60 | caselog = os.path.join(logpath, "%s/log_%s.%s" % (scen, case, d.getVar('DATETIME', True))) |
61 | os.system("touch %s" % caselog) | 61 | os.system("touch %s" % caselog) |
62 | 62 | ||
63 | """export TEST_TMP, TEST_RESULT, DEPLOY_DIR and QEMUARCH""" | 63 | """export TEST_TMP, TEST_RESULT, DEPLOY_DIR and QEMUARCH""" |
@@ -141,7 +141,7 @@ def qemuimagetest_main(d): | |||
141 | 141 | ||
142 | """Clean tmp folder for testing""" | 142 | """Clean tmp folder for testing""" |
143 | def clean_tmp(): | 143 | def clean_tmp(): |
144 | tmppath = d.getVar('TEST_TMP', 1) | 144 | tmppath = d.getVar('TEST_TMP', True) |
145 | 145 | ||
146 | if os.path.isdir(tmppath): | 146 | if os.path.isdir(tmppath): |
147 | for f in os.listdir(tmppath): | 147 | for f in os.listdir(tmppath): |
@@ -155,28 +155,28 @@ def qemuimagetest_main(d): | |||
155 | clean_tmp() | 155 | clean_tmp() |
156 | 156 | ||
157 | """check testcase folder and create test log folder""" | 157 | """check testcase folder and create test log folder""" |
158 | testpath = d.getVar('TEST_DIR', 1) | 158 | testpath = d.getVar('TEST_DIR', True) |
159 | bb.utils.mkdirhier(testpath) | 159 | bb.utils.mkdirhier(testpath) |
160 | 160 | ||
161 | logpath = d.getVar('TEST_LOG', 1) | 161 | logpath = d.getVar('TEST_LOG', True) |
162 | bb.utils.mkdirhier(logpath) | 162 | bb.utils.mkdirhier(logpath) |
163 | 163 | ||
164 | tmppath = d.getVar('TEST_TMP', 1) | 164 | tmppath = d.getVar('TEST_TMP', True) |
165 | bb.utils.mkdirhier(tmppath) | 165 | bb.utils.mkdirhier(tmppath) |
166 | 166 | ||
167 | """initialize test status file""" | 167 | """initialize test status file""" |
168 | test_status = d.getVar('TEST_STATUS', 1) | 168 | test_status = d.getVar('TEST_STATUS', True) |
169 | if os.path.exists(test_status): | 169 | if os.path.exists(test_status): |
170 | os.remove(test_status) | 170 | os.remove(test_status) |
171 | os.system("touch %s" % test_status) | 171 | os.system("touch %s" % test_status) |
172 | 172 | ||
173 | """initialize result file""" | 173 | """initialize result file""" |
174 | resultpath = d.getVar('TEST_RESULT', 1) | 174 | resultpath = d.getVar('TEST_RESULT', True) |
175 | bb.utils.mkdirhier(resultpath) | 175 | bb.utils.mkdirhier(resultpath) |
176 | resultfile = os.path.join(resultpath, "testresult.%s" % d.getVar('DATETIME', 1)) | 176 | resultfile = os.path.join(resultpath, "testresult.%s" % d.getVar('DATETIME', True)) |
177 | sresultfile = os.path.join(resultpath, "testresult.log") | 177 | sresultfile = os.path.join(resultpath, "testresult.log") |
178 | 178 | ||
179 | machine = d.getVar('MACHINE', 1) | 179 | machine = d.getVar('MACHINE', True) |
180 | 180 | ||
181 | if os.path.exists(sresultfile): | 181 | if os.path.exists(sresultfile): |
182 | os.remove(sresultfile) | 182 | os.remove(sresultfile) |
@@ -188,7 +188,7 @@ def qemuimagetest_main(d): | |||
188 | f.close() | 188 | f.close() |
189 | 189 | ||
190 | """generate pre-defined testcase list""" | 190 | """generate pre-defined testcase list""" |
191 | testlist = d.getVar('TEST_SCEN', 1) | 191 | testlist = d.getVar('TEST_SCEN', True) |
192 | fulllist = generate_list(testlist) | 192 | fulllist = generate_list(testlist) |
193 | 193 | ||
194 | """Begin testing""" | 194 | """Begin testing""" |
diff --git a/meta/classes/kernel-arch.bbclass b/meta/classes/kernel-arch.bbclass index 8693395111..d37c1fb2ed 100644 --- a/meta/classes/kernel-arch.bbclass +++ b/meta/classes/kernel-arch.bbclass | |||
@@ -18,7 +18,7 @@ valid_archs = "alpha cris ia64 \ | |||
18 | def map_kernel_arch(a, d): | 18 | def map_kernel_arch(a, d): |
19 | import re | 19 | import re |
20 | 20 | ||
21 | valid_archs = d.getVar('valid_archs', 1).split() | 21 | valid_archs = d.getVar('valid_archs', True).split() |
22 | 22 | ||
23 | if re.match('(i.86|athlon|x86.64)$', a): return 'x86' | 23 | if re.match('(i.86|athlon|x86.64)$', a): return 'x86' |
24 | elif re.match('arm26$', a): return 'arm26' | 24 | elif re.match('arm26$', a): return 'arm26' |
@@ -32,7 +32,7 @@ def map_kernel_arch(a, d): | |||
32 | else: | 32 | else: |
33 | bb.error("cannot map '%s' to a linux kernel architecture" % a) | 33 | bb.error("cannot map '%s' to a linux kernel architecture" % a) |
34 | 34 | ||
35 | export ARCH = "${@map_kernel_arch(d.getVar('TARGET_ARCH', 1), d)}" | 35 | export ARCH = "${@map_kernel_arch(d.getVar('TARGET_ARCH', True), d)}" |
36 | 36 | ||
37 | def map_uboot_arch(a, d): | 37 | def map_uboot_arch(a, d): |
38 | import re | 38 | import re |
@@ -41,5 +41,5 @@ def map_uboot_arch(a, d): | |||
41 | elif re.match('i.86$', a): return 'x86' | 41 | elif re.match('i.86$', a): return 'x86' |
42 | return a | 42 | return a |
43 | 43 | ||
44 | export UBOOT_ARCH = "${@map_uboot_arch(d.getVar('ARCH', 1), d)}" | 44 | export UBOOT_ARCH = "${@map_uboot_arch(d.getVar('ARCH', True), d)}" |
45 | 45 | ||
diff --git a/meta/classes/kernel.bbclass b/meta/classes/kernel.bbclass index ec5d65e186..8fbec90ef1 100644 --- a/meta/classes/kernel.bbclass +++ b/meta/classes/kernel.bbclass | |||
@@ -11,9 +11,9 @@ INITRAMFS_IMAGE ?= "" | |||
11 | INITRAMFS_TASK ?= "" | 11 | INITRAMFS_TASK ?= "" |
12 | 12 | ||
13 | python __anonymous () { | 13 | python __anonymous () { |
14 | kerneltype = d.getVar('KERNEL_IMAGETYPE', 1) or '' | 14 | kerneltype = d.getVar('KERNEL_IMAGETYPE', True) or '' |
15 | if kerneltype == 'uImage': | 15 | if kerneltype == 'uImage': |
16 | depends = d.getVar("DEPENDS", 1) | 16 | depends = d.getVar("DEPENDS", True) |
17 | depends = "%s u-boot-mkimage-native" % depends | 17 | depends = "%s u-boot-mkimage-native" % depends |
18 | d.setVar("DEPENDS", depends) | 18 | d.setVar("DEPENDS", depends) |
19 | 19 | ||
@@ -75,7 +75,7 @@ EXTRA_OEMAKE = "" | |||
75 | 75 | ||
76 | KERNEL_ALT_IMAGETYPE ??= "" | 76 | KERNEL_ALT_IMAGETYPE ??= "" |
77 | 77 | ||
78 | KERNEL_IMAGETYPE_FOR_MAKE = "${@(lambda s: s[:-3] if s[-3:] == ".gz" else s)(d.getVar('KERNEL_IMAGETYPE', 1))}" | 78 | KERNEL_IMAGETYPE_FOR_MAKE = "${@(lambda s: s[:-3] if s[-3:] == ".gz" else s)(d.getVar('KERNEL_IMAGETYPE', True))}" |
79 | 79 | ||
80 | kernel_do_compile() { | 80 | kernel_do_compile() { |
81 | unset CFLAGS CPPFLAGS CXXFLAGS LDFLAGS MACHINE | 81 | unset CFLAGS CPPFLAGS CXXFLAGS LDFLAGS MACHINE |
@@ -306,10 +306,10 @@ module_conf_rfcomm = "alias bt-proto-3 rfcomm" | |||
306 | python populate_packages_prepend () { | 306 | python populate_packages_prepend () { |
307 | def extract_modinfo(file): | 307 | def extract_modinfo(file): |
308 | import tempfile, re | 308 | import tempfile, re |
309 | tempfile.tempdir = d.getVar("WORKDIR", 1) | 309 | tempfile.tempdir = d.getVar("WORKDIR", True) |
310 | tf = tempfile.mkstemp() | 310 | tf = tempfile.mkstemp() |
311 | tmpfile = tf[1] | 311 | tmpfile = tf[1] |
312 | cmd = "PATH=\"%s\" %sobjcopy -j .modinfo -O binary %s %s" % (d.getVar("PATH", 1), d.getVar("HOST_PREFIX", 1) or "", file, tmpfile) | 312 | cmd = "PATH=\"%s\" %sobjcopy -j .modinfo -O binary %s %s" % (d.getVar("PATH", True), d.getVar("HOST_PREFIX", True) or "", file, tmpfile) |
313 | os.system(cmd) | 313 | os.system(cmd) |
314 | f = open(tmpfile) | 314 | f = open(tmpfile) |
315 | l = f.read().split("\000") | 315 | l = f.read().split("\000") |
@@ -328,18 +328,18 @@ python populate_packages_prepend () { | |||
328 | def parse_depmod(): | 328 | def parse_depmod(): |
329 | import re | 329 | import re |
330 | 330 | ||
331 | dvar = d.getVar('PKGD', 1) | 331 | dvar = d.getVar('PKGD', True) |
332 | if not dvar: | 332 | if not dvar: |
333 | bb.error("PKGD not defined") | 333 | bb.error("PKGD not defined") |
334 | return | 334 | return |
335 | 335 | ||
336 | kernelver = d.getVar('KERNEL_VERSION', 1) | 336 | kernelver = d.getVar('KERNEL_VERSION', True) |
337 | kernelver_stripped = kernelver | 337 | kernelver_stripped = kernelver |
338 | m = re.match('^(.*-hh.*)[\.\+].*$', kernelver) | 338 | m = re.match('^(.*-hh.*)[\.\+].*$', kernelver) |
339 | if m: | 339 | if m: |
340 | kernelver_stripped = m.group(1) | 340 | kernelver_stripped = m.group(1) |
341 | path = d.getVar("PATH", 1) | 341 | path = d.getVar("PATH", True) |
342 | host_prefix = d.getVar("HOST_PREFIX", 1) or "" | 342 | host_prefix = d.getVar("HOST_PREFIX", True) or "" |
343 | 343 | ||
344 | cmd = "PATH=\"%s\" %sdepmod -n -a -r -b %s -F %s/boot/System.map-%s %s" % (path, host_prefix, dvar, dvar, kernelver, kernelver_stripped) | 344 | cmd = "PATH=\"%s\" %sdepmod -n -a -r -b %s -F %s/boot/System.map-%s %s" % (path, host_prefix, dvar, dvar, kernelver, kernelver_stripped) |
345 | f = os.popen(cmd, 'r') | 345 | f = os.popen(cmd, 'r') |
@@ -377,9 +377,9 @@ python populate_packages_prepend () { | |||
377 | 377 | ||
378 | def get_dependencies(file, pattern, format): | 378 | def get_dependencies(file, pattern, format): |
379 | # file no longer includes PKGD | 379 | # file no longer includes PKGD |
380 | file = file.replace(d.getVar('PKGD', 1) or '', '', 1) | 380 | file = file.replace(d.getVar('PKGD', True) or '', '', 1) |
381 | # instead is prefixed with /lib/modules/${KERNEL_VERSION} | 381 | # instead is prefixed with /lib/modules/${KERNEL_VERSION} |
382 | file = file.replace("/lib/modules/%s/" % d.getVar('KERNEL_VERSION', 1) or '', '', 1) | 382 | file = file.replace("/lib/modules/%s/" % d.getVar('KERNEL_VERSION', True) or '', '', 1) |
383 | 383 | ||
384 | if module_deps.has_key(file): | 384 | if module_deps.has_key(file): |
385 | import re | 385 | import re |
@@ -398,40 +398,40 @@ python populate_packages_prepend () { | |||
398 | import re | 398 | import re |
399 | vals = extract_modinfo(file) | 399 | vals = extract_modinfo(file) |
400 | 400 | ||
401 | dvar = d.getVar('PKGD', 1) | 401 | dvar = d.getVar('PKGD', True) |
402 | 402 | ||
403 | # If autoloading is requested, output /etc/modutils/<name> and append | 403 | # If autoloading is requested, output /etc/modutils/<name> and append |
404 | # appropriate modprobe commands to the postinst | 404 | # appropriate modprobe commands to the postinst |
405 | autoload = d.getVar('module_autoload_%s' % basename, 1) | 405 | autoload = d.getVar('module_autoload_%s' % basename, True) |
406 | if autoload: | 406 | if autoload: |
407 | name = '%s/etc/modutils/%s' % (dvar, basename) | 407 | name = '%s/etc/modutils/%s' % (dvar, basename) |
408 | f = open(name, 'w') | 408 | f = open(name, 'w') |
409 | for m in autoload.split(): | 409 | for m in autoload.split(): |
410 | f.write('%s\n' % m) | 410 | f.write('%s\n' % m) |
411 | f.close() | 411 | f.close() |
412 | postinst = d.getVar('pkg_postinst_%s' % pkg, 1) | 412 | postinst = d.getVar('pkg_postinst_%s' % pkg, True) |
413 | if not postinst: | 413 | if not postinst: |
414 | bb.fatal("pkg_postinst_%s not defined" % pkg) | 414 | bb.fatal("pkg_postinst_%s not defined" % pkg) |
415 | postinst += d.getVar('autoload_postinst_fragment', 1) % autoload | 415 | postinst += d.getVar('autoload_postinst_fragment', True) % autoload |
416 | d.setVar('pkg_postinst_%s' % pkg, postinst) | 416 | d.setVar('pkg_postinst_%s' % pkg, postinst) |
417 | 417 | ||
418 | # Write out any modconf fragment | 418 | # Write out any modconf fragment |
419 | modconf = d.getVar('module_conf_%s' % basename, 1) | 419 | modconf = d.getVar('module_conf_%s' % basename, True) |
420 | if modconf: | 420 | if modconf: |
421 | name = '%s/etc/modprobe.d/%s.conf' % (dvar, basename) | 421 | name = '%s/etc/modprobe.d/%s.conf' % (dvar, basename) |
422 | f = open(name, 'w') | 422 | f = open(name, 'w') |
423 | f.write("%s\n" % modconf) | 423 | f.write("%s\n" % modconf) |
424 | f.close() | 424 | f.close() |
425 | 425 | ||
426 | files = d.getVar('FILES_%s' % pkg, 1) | 426 | files = d.getVar('FILES_%s' % pkg, True) |
427 | files = "%s /etc/modutils/%s /etc/modutils/%s.conf /etc/modprobe.d/%s.conf" % (files, basename, basename, basename) | 427 | files = "%s /etc/modutils/%s /etc/modutils/%s.conf /etc/modprobe.d/%s.conf" % (files, basename, basename, basename) |
428 | d.setVar('FILES_%s' % pkg, files) | 428 | d.setVar('FILES_%s' % pkg, files) |
429 | 429 | ||
430 | if vals.has_key("description"): | 430 | if vals.has_key("description"): |
431 | old_desc = d.getVar('DESCRIPTION_' + pkg, 1) or "" | 431 | old_desc = d.getVar('DESCRIPTION_' + pkg, True) or "" |
432 | d.setVar('DESCRIPTION_' + pkg, old_desc + "; " + vals["description"]) | 432 | d.setVar('DESCRIPTION_' + pkg, old_desc + "; " + vals["description"]) |
433 | 433 | ||
434 | rdepends_str = d.getVar('RDEPENDS_' + pkg, 1) | 434 | rdepends_str = d.getVar('RDEPENDS_' + pkg, True) |
435 | if rdepends_str: | 435 | if rdepends_str: |
436 | rdepends = rdepends_str.split() | 436 | rdepends = rdepends_str.split() |
437 | else: | 437 | else: |
@@ -443,12 +443,12 @@ python populate_packages_prepend () { | |||
443 | module_regex = '^(.*)\.k?o$' | 443 | module_regex = '^(.*)\.k?o$' |
444 | module_pattern = 'kernel-module-%s' | 444 | module_pattern = 'kernel-module-%s' |
445 | 445 | ||
446 | postinst = d.getVar('pkg_postinst_modules', 1) | 446 | postinst = d.getVar('pkg_postinst_modules', True) |
447 | postrm = d.getVar('pkg_postrm_modules', 1) | 447 | postrm = d.getVar('pkg_postrm_modules', True) |
448 | do_split_packages(d, root='/lib/firmware', file_regex='^(.*)\.bin$', output_pattern='kernel-firmware-%s', description='Firmware for %s', recursive=True, extra_depends='') | 448 | do_split_packages(d, root='/lib/firmware', file_regex='^(.*)\.bin$', output_pattern='kernel-firmware-%s', description='Firmware for %s', recursive=True, extra_depends='') |
449 | do_split_packages(d, root='/lib/firmware', file_regex='^(.*)\.fw$', output_pattern='kernel-firmware-%s', description='Firmware for %s', recursive=True, extra_depends='') | 449 | do_split_packages(d, root='/lib/firmware', file_regex='^(.*)\.fw$', output_pattern='kernel-firmware-%s', description='Firmware for %s', recursive=True, extra_depends='') |
450 | do_split_packages(d, root='/lib/firmware', file_regex='^(.*)\.cis$', output_pattern='kernel-firmware-%s', description='Firmware for %s', recursive=True, extra_depends='') | 450 | do_split_packages(d, root='/lib/firmware', file_regex='^(.*)\.cis$', output_pattern='kernel-firmware-%s', description='Firmware for %s', recursive=True, extra_depends='') |
451 | do_split_packages(d, root='/lib/modules', file_regex=module_regex, output_pattern=module_pattern, description='%s kernel module', postinst=postinst, postrm=postrm, recursive=True, hook=frob_metadata, extra_depends='update-modules kernel-%s' % d.getVar("KERNEL_VERSION", 1)) | 451 | do_split_packages(d, root='/lib/modules', file_regex=module_regex, output_pattern=module_pattern, description='%s kernel module', postinst=postinst, postrm=postrm, recursive=True, hook=frob_metadata, extra_depends='update-modules kernel-%s' % d.getVar("KERNEL_VERSION", True)) |
452 | 452 | ||
453 | import re | 453 | import re |
454 | metapkg = "kernel-modules" | 454 | metapkg = "kernel-modules" |
@@ -460,7 +460,7 @@ python populate_packages_prepend () { | |||
460 | pkg = module_pattern % legitimize_package_name(re.match(module_regex, os.path.basename(i)).group(1)) | 460 | pkg = module_pattern % legitimize_package_name(re.match(module_regex, os.path.basename(i)).group(1)) |
461 | blacklist.append(pkg) | 461 | blacklist.append(pkg) |
462 | metapkg_rdepends = [] | 462 | metapkg_rdepends = [] |
463 | packages = d.getVar('PACKAGES', 1).split() | 463 | packages = d.getVar('PACKAGES', True).split() |
464 | for pkg in packages[1:]: | 464 | for pkg in packages[1:]: |
465 | if not pkg in blacklist and not pkg in metapkg_rdepends: | 465 | if not pkg in blacklist and not pkg in metapkg_rdepends: |
466 | metapkg_rdepends.append(pkg) | 466 | metapkg_rdepends.append(pkg) |
diff --git a/meta/classes/libc-common.bbclass b/meta/classes/libc-common.bbclass index ec33762a20..962f205f81 100644 --- a/meta/classes/libc-common.bbclass +++ b/meta/classes/libc-common.bbclass | |||
@@ -18,13 +18,13 @@ do_install() { | |||
18 | } | 18 | } |
19 | 19 | ||
20 | def get_libc_fpu_setting(bb, d): | 20 | def get_libc_fpu_setting(bb, d): |
21 | if d.getVar('TARGET_FPU', 1) in [ 'soft' ]: | 21 | if d.getVar('TARGET_FPU', True) in [ 'soft' ]: |
22 | return "--without-fp" | 22 | return "--without-fp" |
23 | return "" | 23 | return "" |
24 | 24 | ||
25 | python populate_packages_prepend () { | 25 | python populate_packages_prepend () { |
26 | if d.getVar('DEBIAN_NAMES', 1): | 26 | if d.getVar('DEBIAN_NAMES', True): |
27 | bpn = d.getVar('BPN', 1) | 27 | bpn = d.getVar('BPN', True) |
28 | d.setVar('PKG_'+bpn, 'libc6') | 28 | d.setVar('PKG_'+bpn, 'libc6') |
29 | d.setVar('PKG_'+bpn+'-dev', 'libc6-dev') | 29 | d.setVar('PKG_'+bpn+'-dev', 'libc6-dev') |
30 | } | 30 | } |
diff --git a/meta/classes/license.bbclass b/meta/classes/license.bbclass index cfc9eafb93..3de704f3f9 100644 --- a/meta/classes/license.bbclass +++ b/meta/classes/license.bbclass | |||
@@ -261,7 +261,7 @@ def incompatible_license(d,dont_want_license): | |||
261 | from fnmatch import fnmatchcase as fnmatch | 261 | from fnmatch import fnmatchcase as fnmatch |
262 | 262 | ||
263 | dont_want_licenses = [] | 263 | dont_want_licenses = [] |
264 | dont_want_licenses.append(d.getVar('INCOMPATIBLE_LICENSE', 1)) | 264 | dont_want_licenses.append(d.getVar('INCOMPATIBLE_LICENSE', True)) |
265 | if d.getVarFlag('SPDXLICENSEMAP', dont_want_license): | 265 | if d.getVarFlag('SPDXLICENSEMAP', dont_want_license): |
266 | dont_want_licenses.append(d.getVarFlag('SPDXLICENSEMAP', dont_want_license)) | 266 | dont_want_licenses.append(d.getVarFlag('SPDXLICENSEMAP', dont_want_license)) |
267 | 267 | ||
diff --git a/meta/classes/metadata_scm.bbclass b/meta/classes/metadata_scm.bbclass index 57609ef8cd..62650be675 100644 --- a/meta/classes/metadata_scm.bbclass +++ b/meta/classes/metadata_scm.bbclass | |||
@@ -27,7 +27,7 @@ def base_detect_branch(d): | |||
27 | return "<unknown>" | 27 | return "<unknown>" |
28 | 28 | ||
29 | def base_get_scmbasepath(d): | 29 | def base_get_scmbasepath(d): |
30 | return d.getVar( 'COREBASE', 1 ) | 30 | return d.getVar( 'COREBASE', True) |
31 | 31 | ||
32 | def base_get_metadata_monotone_branch(path, d): | 32 | def base_get_metadata_monotone_branch(path, d): |
33 | monotone_branch = "<unknown>" | 33 | monotone_branch = "<unknown>" |
diff --git a/meta/classes/package.bbclass b/meta/classes/package.bbclass index 4ed6972a7c..5c42619f3f 100644 --- a/meta/classes/package.bbclass +++ b/meta/classes/package.bbclass | |||
@@ -1067,7 +1067,7 @@ python emit_pkgdata() { | |||
1067 | return size | 1067 | return size |
1068 | 1068 | ||
1069 | packages = d.getVar('PACKAGES', True) | 1069 | packages = d.getVar('PACKAGES', True) |
1070 | pkgdest = d.getVar('PKGDEST', 1) | 1070 | pkgdest = d.getVar('PKGDEST', True) |
1071 | pkgdatadir = d.getVar('PKGDESTWORK', True) | 1071 | pkgdatadir = d.getVar('PKGDESTWORK', True) |
1072 | 1072 | ||
1073 | # Take shared lock since we're only reading, not writing | 1073 | # Take shared lock since we're only reading, not writing |
diff --git a/meta/classes/package_ipk.bbclass b/meta/classes/package_ipk.bbclass index e9d1ddcdbc..ff8b5b488a 100644 --- a/meta/classes/package_ipk.bbclass +++ b/meta/classes/package_ipk.bbclass | |||
@@ -15,12 +15,12 @@ python package_ipk_fn () { | |||
15 | } | 15 | } |
16 | 16 | ||
17 | python package_ipk_install () { | 17 | python package_ipk_install () { |
18 | pkg = d.getVar('PKG', 1) | 18 | pkg = d.getVar('PKG', True) |
19 | pkgfn = d.getVar('PKGFN', 1) | 19 | pkgfn = d.getVar('PKGFN', True) |
20 | rootfs = d.getVar('IMAGE_ROOTFS', 1) | 20 | rootfs = d.getVar('IMAGE_ROOTFS', True) |
21 | ipkdir = d.getVar('DEPLOY_DIR_IPK', 1) | 21 | ipkdir = d.getVar('DEPLOY_DIR_IPK', True) |
22 | stagingdir = d.getVar('STAGING_DIR', 1) | 22 | stagingdir = d.getVar('STAGING_DIR', True) |
23 | tmpdir = d.getVar('TMPDIR', 1) | 23 | tmpdir = d.getVar('TMPDIR', True) |
24 | 24 | ||
25 | if None in (pkg,pkgfn,rootfs): | 25 | if None in (pkg,pkgfn,rootfs): |
26 | raise bb.build.FuncFailed("missing variables (one or more of PKG, PKGFN, IMAGEROOTFS)") | 26 | raise bb.build.FuncFailed("missing variables (one or more of PKG, PKGFN, IMAGEROOTFS)") |
@@ -289,7 +289,7 @@ python do_package_ipk () { | |||
289 | 289 | ||
290 | localdata.setVar('ROOT', '') | 290 | localdata.setVar('ROOT', '') |
291 | localdata.setVar('ROOT_%s' % pkg, root) | 291 | localdata.setVar('ROOT_%s' % pkg, root) |
292 | pkgname = localdata.getVar('PKG_%s' % pkg, 1) | 292 | pkgname = localdata.getVar('PKG_%s' % pkg, True) |
293 | if not pkgname: | 293 | if not pkgname: |
294 | pkgname = pkg | 294 | pkgname = pkg |
295 | localdata.setVar('PKG', pkgname) | 295 | localdata.setVar('PKG', pkgname) |
@@ -298,7 +298,7 @@ python do_package_ipk () { | |||
298 | 298 | ||
299 | bb.data.update_data(localdata) | 299 | bb.data.update_data(localdata) |
300 | basedir = os.path.join(os.path.dirname(root)) | 300 | basedir = os.path.join(os.path.dirname(root)) |
301 | arch = localdata.getVar('PACKAGE_ARCH', 1) | 301 | arch = localdata.getVar('PACKAGE_ARCH', True) |
302 | pkgoutdir = "%s/%s" % (outdir, arch) | 302 | pkgoutdir = "%s/%s" % (outdir, arch) |
303 | bb.mkdirhier(pkgoutdir) | 303 | bb.mkdirhier(pkgoutdir) |
304 | os.chdir(root) | 304 | os.chdir(root) |
@@ -310,7 +310,7 @@ python do_package_ipk () { | |||
310 | except ValueError: | 310 | except ValueError: |
311 | pass | 311 | pass |
312 | if not g and localdata.getVar('ALLOW_EMPTY') != "1": | 312 | if not g and localdata.getVar('ALLOW_EMPTY') != "1": |
313 | bb.note("Not creating empty archive for %s-%s-%s" % (pkg, localdata.getVar('PKGV', 1), localdata.getVar('PKGR', 1))) | 313 | bb.note("Not creating empty archive for %s-%s-%s" % (pkg, localdata.getVar('PKGV', True), localdata.getVar('PKGR', True))) |
314 | bb.utils.unlockfile(lf) | 314 | bb.utils.unlockfile(lf) |
315 | continue | 315 | continue |
316 | 316 | ||
@@ -323,7 +323,7 @@ python do_package_ipk () { | |||
323 | raise bb.build.FuncFailed("unable to open control file for writing.") | 323 | raise bb.build.FuncFailed("unable to open control file for writing.") |
324 | 324 | ||
325 | fields = [] | 325 | fields = [] |
326 | pe = d.getVar('PKGE', 1) | 326 | pe = d.getVar('PKGE', True) |
327 | if pe and int(pe) > 0: | 327 | if pe and int(pe) > 0: |
328 | fields.append(["Version: %s:%s-%s\n", ['PKGE', 'PKGV', 'PKGR']]) | 328 | fields.append(["Version: %s:%s-%s\n", ['PKGE', 'PKGV', 'PKGR']]) |
329 | else: | 329 | else: |
@@ -340,7 +340,7 @@ python do_package_ipk () { | |||
340 | def pullData(l, d): | 340 | def pullData(l, d): |
341 | l2 = [] | 341 | l2 = [] |
342 | for i in l: | 342 | for i in l: |
343 | l2.append(d.getVar(i, 1)) | 343 | l2.append(d.getVar(i, True)) |
344 | return l2 | 344 | return l2 |
345 | 345 | ||
346 | ctrlfile.write("Package: %s\n" % pkgname) | 346 | ctrlfile.write("Package: %s\n" % pkgname) |
@@ -369,12 +369,12 @@ python do_package_ipk () { | |||
369 | 369 | ||
370 | bb.build.exec_func("mapping_rename_hook", localdata) | 370 | bb.build.exec_func("mapping_rename_hook", localdata) |
371 | 371 | ||
372 | rdepends = bb.utils.explode_dep_versions(localdata.getVar("RDEPENDS", 1) or "") | 372 | rdepends = bb.utils.explode_dep_versions(localdata.getVar("RDEPENDS", True) or "") |
373 | rrecommends = bb.utils.explode_dep_versions(localdata.getVar("RRECOMMENDS", 1) or "") | 373 | rrecommends = bb.utils.explode_dep_versions(localdata.getVar("RRECOMMENDS", True) or "") |
374 | rsuggests = bb.utils.explode_dep_versions(localdata.getVar("RSUGGESTS", 1) or "") | 374 | rsuggests = bb.utils.explode_dep_versions(localdata.getVar("RSUGGESTS", True) or "") |
375 | rprovides = bb.utils.explode_dep_versions(localdata.getVar("RPROVIDES", 1) or "") | 375 | rprovides = bb.utils.explode_dep_versions(localdata.getVar("RPROVIDES", True) or "") |
376 | rreplaces = bb.utils.explode_dep_versions(localdata.getVar("RREPLACES", 1) or "") | 376 | rreplaces = bb.utils.explode_dep_versions(localdata.getVar("RREPLACES", True) or "") |
377 | rconflicts = bb.utils.explode_dep_versions(localdata.getVar("RCONFLICTS", 1) or "") | 377 | rconflicts = bb.utils.explode_dep_versions(localdata.getVar("RCONFLICTS", True) or "") |
378 | 378 | ||
379 | if rdepends: | 379 | if rdepends: |
380 | ctrlfile.write("Depends: %s\n" % bb.utils.join_deps(rdepends)) | 380 | ctrlfile.write("Depends: %s\n" % bb.utils.join_deps(rdepends)) |
@@ -388,14 +388,14 @@ python do_package_ipk () { | |||
388 | ctrlfile.write("Replaces: %s\n" % bb.utils.join_deps(rreplaces)) | 388 | ctrlfile.write("Replaces: %s\n" % bb.utils.join_deps(rreplaces)) |
389 | if rconflicts: | 389 | if rconflicts: |
390 | ctrlfile.write("Conflicts: %s\n" % bb.utils.join_deps(rconflicts)) | 390 | ctrlfile.write("Conflicts: %s\n" % bb.utils.join_deps(rconflicts)) |
391 | src_uri = localdata.getVar("SRC_URI", 1) | 391 | src_uri = localdata.getVar("SRC_URI", True) |
392 | if src_uri: | 392 | if src_uri: |
393 | src_uri = re.sub("\s+", " ", src_uri) | 393 | src_uri = re.sub("\s+", " ", src_uri) |
394 | ctrlfile.write("Source: %s\n" % " ".join(src_uri.split())) | 394 | ctrlfile.write("Source: %s\n" % " ".join(src_uri.split())) |
395 | ctrlfile.close() | 395 | ctrlfile.close() |
396 | 396 | ||
397 | for script in ["preinst", "postinst", "prerm", "postrm"]: | 397 | for script in ["preinst", "postinst", "prerm", "postrm"]: |
398 | scriptvar = localdata.getVar('pkg_%s' % script, 1) | 398 | scriptvar = localdata.getVar('pkg_%s' % script, True) |
399 | if not scriptvar: | 399 | if not scriptvar: |
400 | continue | 400 | continue |
401 | try: | 401 | try: |
@@ -407,7 +407,7 @@ python do_package_ipk () { | |||
407 | scriptfile.close() | 407 | scriptfile.close() |
408 | os.chmod(os.path.join(controldir, script), 0755) | 408 | os.chmod(os.path.join(controldir, script), 0755) |
409 | 409 | ||
410 | conffiles_str = localdata.getVar("CONFFILES", 1) | 410 | conffiles_str = localdata.getVar("CONFFILES", True) |
411 | if conffiles_str: | 411 | if conffiles_str: |
412 | try: | 412 | try: |
413 | conffiles = file(os.path.join(controldir, 'conffiles'), 'w') | 413 | conffiles = file(os.path.join(controldir, 'conffiles'), 'w') |
@@ -419,7 +419,7 @@ python do_package_ipk () { | |||
419 | conffiles.close() | 419 | conffiles.close() |
420 | 420 | ||
421 | os.chdir(basedir) | 421 | os.chdir(basedir) |
422 | ret = os.system("PATH=\"%s\" %s %s %s" % (localdata.getVar("PATH", 1), | 422 | ret = os.system("PATH=\"%s\" %s %s %s" % (localdata.getVar("PATH", True), |
423 | d.getVar("OPKGBUILDCMD",1), pkg, pkgoutdir)) | 423 | d.getVar("OPKGBUILDCMD",1), pkg, pkgoutdir)) |
424 | if ret != 0: | 424 | if ret != 0: |
425 | bb.utils.unlockfile(lf) | 425 | bb.utils.unlockfile(lf) |
diff --git a/meta/classes/package_rpm.bbclass b/meta/classes/package_rpm.bbclass index a264712f9e..af8c63ed6f 100644 --- a/meta/classes/package_rpm.bbclass +++ b/meta/classes/package_rpm.bbclass | |||
@@ -619,7 +619,7 @@ python write_specfile () { | |||
619 | 619 | ||
620 | localdata.setVar('ROOT', '') | 620 | localdata.setVar('ROOT', '') |
621 | localdata.setVar('ROOT_%s' % pkg, root) | 621 | localdata.setVar('ROOT_%s' % pkg, root) |
622 | pkgname = localdata.getVar('PKG_%s' % pkg, 1) | 622 | pkgname = localdata.getVar('PKG_%s' % pkg, True) |
623 | if not pkgname: | 623 | if not pkgname: |
624 | pkgname = pkg | 624 | pkgname = pkg |
625 | localdata.setVar('PKG', pkgname) | 625 | localdata.setVar('PKG', pkgname) |
diff --git a/meta/classes/package_tar.bbclass b/meta/classes/package_tar.bbclass index 201bd91657..7590177e4b 100644 --- a/meta/classes/package_tar.bbclass +++ b/meta/classes/package_tar.bbclass | |||
@@ -9,9 +9,9 @@ python package_tar_fn () { | |||
9 | } | 9 | } |
10 | 10 | ||
11 | python package_tar_install () { | 11 | python package_tar_install () { |
12 | pkg = d.getVar('PKG', 1) | 12 | pkg = d.getVar('PKG', True) |
13 | pkgfn = d.getVar('PKGFN', 1) | 13 | pkgfn = d.getVar('PKGFN', True) |
14 | rootfs = d.getVar('IMAGE_ROOTFS', 1) | 14 | rootfs = d.getVar('IMAGE_ROOTFS', True) |
15 | 15 | ||
16 | if None in (pkg,pkgfn,rootfs): | 16 | if None in (pkg,pkgfn,rootfs): |
17 | bb.error("missing variables (one or more of PKG, PKGFN, IMAGEROOTFS)") | 17 | bb.error("missing variables (one or more of PKG, PKGFN, IMAGEROOTFS)") |
@@ -35,24 +35,24 @@ python package_tar_install () { | |||
35 | } | 35 | } |
36 | 36 | ||
37 | python do_package_tar () { | 37 | python do_package_tar () { |
38 | workdir = d.getVar('WORKDIR', 1) | 38 | workdir = d.getVar('WORKDIR', True) |
39 | if not workdir: | 39 | if not workdir: |
40 | bb.error("WORKDIR not defined, unable to package") | 40 | bb.error("WORKDIR not defined, unable to package") |
41 | return | 41 | return |
42 | 42 | ||
43 | outdir = d.getVar('DEPLOY_DIR_TAR', 1) | 43 | outdir = d.getVar('DEPLOY_DIR_TAR', True) |
44 | if not outdir: | 44 | if not outdir: |
45 | bb.error("DEPLOY_DIR_TAR not defined, unable to package") | 45 | bb.error("DEPLOY_DIR_TAR not defined, unable to package") |
46 | return | 46 | return |
47 | bb.mkdirhier(outdir) | 47 | bb.mkdirhier(outdir) |
48 | 48 | ||
49 | dvar = d.getVar('D', 1) | 49 | dvar = d.getVar('D', True) |
50 | if not dvar: | 50 | if not dvar: |
51 | bb.error("D not defined, unable to package") | 51 | bb.error("D not defined, unable to package") |
52 | return | 52 | return |
53 | bb.mkdirhier(dvar) | 53 | bb.mkdirhier(dvar) |
54 | 54 | ||
55 | packages = d.getVar('PACKAGES', 1) | 55 | packages = d.getVar('PACKAGES', True) |
56 | if not packages: | 56 | if not packages: |
57 | bb.debug(1, "PACKAGES not defined, nothing to package") | 57 | bb.debug(1, "PACKAGES not defined, nothing to package") |
58 | return | 58 | return |
@@ -79,11 +79,11 @@ python do_package_tar () { | |||
79 | pkgoutdir = outdir | 79 | pkgoutdir = outdir |
80 | bb.mkdirhier(pkgoutdir) | 80 | bb.mkdirhier(pkgoutdir) |
81 | bb.build.exec_func('package_tar_fn', localdata) | 81 | bb.build.exec_func('package_tar_fn', localdata) |
82 | tarfn = localdata.getVar('PKGFN', 1) | 82 | tarfn = localdata.getVar('PKGFN', True) |
83 | os.chdir(root) | 83 | os.chdir(root) |
84 | from glob import glob | 84 | from glob import glob |
85 | if not glob('*'): | 85 | if not glob('*'): |
86 | bb.note("Not creating empty archive for %s-%s-%s" % (pkg, localdata.getVar('PKGV', 1), localdata.getVar('PKGR', 1))) | 86 | bb.note("Not creating empty archive for %s-%s-%s" % (pkg, localdata.getVar('PKGV', True), localdata.getVar('PKGR', True))) |
87 | continue | 87 | continue |
88 | ret = os.system("tar -czf %s %s" % (tarfn, '.')) | 88 | ret = os.system("tar -czf %s %s" % (tarfn, '.')) |
89 | if ret != 0: | 89 | if ret != 0: |
diff --git a/meta/classes/packagedata.bbclass b/meta/classes/packagedata.bbclass index 9c7aede3bb..60f1aded0d 100644 --- a/meta/classes/packagedata.bbclass +++ b/meta/classes/packagedata.bbclass | |||
@@ -1,12 +1,12 @@ | |||
1 | python read_subpackage_metadata () { | 1 | python read_subpackage_metadata () { |
2 | import oe.packagedata | 2 | import oe.packagedata |
3 | 3 | ||
4 | data = oe.packagedata.read_pkgdata(d.getVar('PN', 1), d) | 4 | data = oe.packagedata.read_pkgdata(d.getVar('PN', True), d) |
5 | 5 | ||
6 | for key in data.keys(): | 6 | for key in data.keys(): |
7 | d.setVar(key, data[key]) | 7 | d.setVar(key, data[key]) |
8 | 8 | ||
9 | for pkg in d.getVar('PACKAGES', 1).split(): | 9 | for pkg in d.getVar('PACKAGES', True).split(): |
10 | sdata = oe.packagedata.read_subpkgdata(pkg, d) | 10 | sdata = oe.packagedata.read_subpkgdata(pkg, d) |
11 | for key in sdata.keys(): | 11 | for key in sdata.keys(): |
12 | d.setVar(key, sdata[key]) | 12 | d.setVar(key, sdata[key]) |
diff --git a/meta/classes/pkg_distribute.bbclass b/meta/classes/pkg_distribute.bbclass index 52643a2f90..9f249a0dfe 100644 --- a/meta/classes/pkg_distribute.bbclass +++ b/meta/classes/pkg_distribute.bbclass | |||
@@ -1,6 +1,6 @@ | |||
1 | PKG_DISTRIBUTECOMMAND[func] = "1" | 1 | PKG_DISTRIBUTECOMMAND[func] = "1" |
2 | python do_distribute_packages () { | 2 | python do_distribute_packages () { |
3 | cmd = d.getVar('PKG_DISTRIBUTECOMMAND', 1) | 3 | cmd = d.getVar('PKG_DISTRIBUTECOMMAND', True) |
4 | if not cmd: | 4 | if not cmd: |
5 | raise bb.build.FuncFailed("Unable to distribute packages, PKG_DISTRIBUTECOMMAND not defined") | 5 | raise bb.build.FuncFailed("Unable to distribute packages, PKG_DISTRIBUTECOMMAND not defined") |
6 | bb.build.exec_func('PKG_DISTRIBUTECOMMAND', d) | 6 | bb.build.exec_func('PKG_DISTRIBUTECOMMAND', d) |
diff --git a/meta/classes/pkg_metainfo.bbclass b/meta/classes/pkg_metainfo.bbclass index 1714a535c2..4b182690f2 100644 --- a/meta/classes/pkg_metainfo.bbclass +++ b/meta/classes/pkg_metainfo.bbclass | |||
@@ -1,5 +1,5 @@ | |||
1 | python do_pkg_write_metainfo () { | 1 | python do_pkg_write_metainfo () { |
2 | deploydir = d.getVar('DEPLOY_DIR', 1) | 2 | deploydir = d.getVar('DEPLOY_DIR', True) |
3 | if not deploydir: | 3 | if not deploydir: |
4 | bb.error("DEPLOY_DIR not defined, unable to write package info") | 4 | bb.error("DEPLOY_DIR not defined, unable to write package info") |
5 | return | 5 | return |
@@ -9,11 +9,11 @@ python do_pkg_write_metainfo () { | |||
9 | except OSError: | 9 | except OSError: |
10 | raise bb.build.FuncFailed("unable to open package-info file for writing.") | 10 | raise bb.build.FuncFailed("unable to open package-info file for writing.") |
11 | 11 | ||
12 | name = d.getVar('PN', 1) | 12 | name = d.getVar('PN', True) |
13 | version = d.getVar('PV', 1) | 13 | version = d.getVar('PV', True) |
14 | desc = d.getVar('DESCRIPTION', 1) | 14 | desc = d.getVar('DESCRIPTION', True) |
15 | page = d.getVar('HOMEPAGE', 1) | 15 | page = d.getVar('HOMEPAGE', True) |
16 | lic = d.getVar('LICENSE', 1) | 16 | lic = d.getVar('LICENSE', True) |
17 | 17 | ||
18 | infofile.write("|| "+ name +" || "+ version + " || "+ desc +" || "+ page +" || "+ lic + " ||\n" ) | 18 | infofile.write("|| "+ name +" || "+ version + " || "+ desc +" || "+ page +" || "+ lic + " ||\n" ) |
19 | infofile.close() | 19 | infofile.close() |
diff --git a/meta/classes/populate_sdk_deb.bbclass b/meta/classes/populate_sdk_deb.bbclass index fe3d849162..920c89a0f3 100644 --- a/meta/classes/populate_sdk_deb.bbclass +++ b/meta/classes/populate_sdk_deb.bbclass | |||
@@ -2,8 +2,8 @@ do_populate_sdk[depends] += "dpkg-native:do_populate_sysroot apt-native:do_popul | |||
2 | do_populate_sdk[recrdeptask] += "do_package_write_deb" | 2 | do_populate_sdk[recrdeptask] += "do_package_write_deb" |
3 | 3 | ||
4 | 4 | ||
5 | DEB_SDK_ARCH = "${@[d.getVar('SDK_ARCH', 1), "i386"]\ | 5 | DEB_SDK_ARCH = "${@[d.getVar('SDK_ARCH', True), "i386"]\ |
6 | [d.getVar('SDK_ARCH', 1) in \ | 6 | [d.getVar('SDK_ARCH', True) in \ |
7 | ["x86", "i486", "i586", "i686", "pentium"]]}" | 7 | ["x86", "i486", "i586", "i686", "pentium"]]}" |
8 | 8 | ||
9 | populate_sdk_post_deb () { | 9 | populate_sdk_post_deb () { |
diff --git a/meta/classes/qemu.bbclass b/meta/classes/qemu.bbclass index 1bdd209afe..22ebcfde40 100644 --- a/meta/classes/qemu.bbclass +++ b/meta/classes/qemu.bbclass | |||
@@ -6,7 +6,7 @@ | |||
6 | def qemu_target_binary(data): | 6 | def qemu_target_binary(data): |
7 | import bb | 7 | import bb |
8 | 8 | ||
9 | target_arch = data.getVar("TARGET_ARCH", 1) | 9 | target_arch = data.getVar("TARGET_ARCH", True) |
10 | if target_arch in ("i486", "i586", "i686"): | 10 | if target_arch in ("i486", "i586", "i686"): |
11 | target_arch = "i386" | 11 | target_arch = "i386" |
12 | elif target_arch == "powerpc": | 12 | elif target_arch == "powerpc": |
diff --git a/meta/classes/qt4e.bbclass b/meta/classes/qt4e.bbclass index f3d5caa455..05c24efaa1 100644 --- a/meta/classes/qt4e.bbclass +++ b/meta/classes/qt4e.bbclass | |||
@@ -1,4 +1,4 @@ | |||
1 | DEPENDS_prepend = "${@["qt4-embedded ", ""][(d.getVar('PN', 1)[:12] == 'qt4-embedded')]}" | 1 | DEPENDS_prepend = "${@["qt4-embedded ", ""][(d.getVar('PN', True)[:12] == 'qt4-embedded')]}" |
2 | 2 | ||
3 | inherit qmake2 | 3 | inherit qmake2 |
4 | 4 | ||
diff --git a/meta/classes/recipe_sanity.bbclass b/meta/classes/recipe_sanity.bbclass index bb60ffa00e..b3246599b9 100644 --- a/meta/classes/recipe_sanity.bbclass +++ b/meta/classes/recipe_sanity.bbclass | |||
@@ -1,5 +1,5 @@ | |||
1 | def __note(msg, d): | 1 | def __note(msg, d): |
2 | bb.note("%s: recipe_sanity: %s" % (d.getVar("P", 1), msg)) | 2 | bb.note("%s: recipe_sanity: %s" % (d.getVar("P", True), msg)) |
3 | 3 | ||
4 | __recipe_sanity_badruntimevars = "RDEPENDS RPROVIDES RRECOMMENDS RCONFLICTS" | 4 | __recipe_sanity_badruntimevars = "RDEPENDS RPROVIDES RRECOMMENDS RCONFLICTS" |
5 | def bad_runtime_vars(cfgdata, d): | 5 | def bad_runtime_vars(cfgdata, d): |
@@ -7,7 +7,7 @@ def bad_runtime_vars(cfgdata, d): | |||
7 | bb.data.inherits_class("cross", d): | 7 | bb.data.inherits_class("cross", d): |
8 | return | 8 | return |
9 | 9 | ||
10 | for var in d.getVar("__recipe_sanity_badruntimevars", 1).split(): | 10 | for var in d.getVar("__recipe_sanity_badruntimevars", True).split(): |
11 | val = d.getVar(var, 0) | 11 | val = d.getVar(var, 0) |
12 | if val and val != cfgdata.get(var): | 12 | if val and val != cfgdata.get(var): |
13 | __note("%s should be %s_${PN}" % (var, var), d) | 13 | __note("%s should be %s_${PN}" % (var, var), d) |
@@ -15,17 +15,17 @@ def bad_runtime_vars(cfgdata, d): | |||
15 | __recipe_sanity_reqvars = "DESCRIPTION" | 15 | __recipe_sanity_reqvars = "DESCRIPTION" |
16 | __recipe_sanity_reqdiffvars = "LICENSE" | 16 | __recipe_sanity_reqdiffvars = "LICENSE" |
17 | def req_vars(cfgdata, d): | 17 | def req_vars(cfgdata, d): |
18 | for var in d.getVar("__recipe_sanity_reqvars", 1).split(): | 18 | for var in d.getVar("__recipe_sanity_reqvars", True).split(): |
19 | if not d.getVar(var, 0): | 19 | if not d.getVar(var, 0): |
20 | __note("%s should be set" % var, d) | 20 | __note("%s should be set" % var, d) |
21 | 21 | ||
22 | for var in d.getVar("__recipe_sanity_reqdiffvars", 1).split(): | 22 | for var in d.getVar("__recipe_sanity_reqdiffvars", True).split(): |
23 | val = d.getVar(var, 0) | 23 | val = d.getVar(var, 0) |
24 | cfgval = cfgdata.get(var) | 24 | cfgval = cfgdata.get(var) |
25 | 25 | ||
26 | # Hardcoding is bad, but I'm lazy. We don't care about license being | 26 | # Hardcoding is bad, but I'm lazy. We don't care about license being |
27 | # unset if the recipe has no sources! | 27 | # unset if the recipe has no sources! |
28 | if var == "LICENSE" and d.getVar("SRC_URI", 1) == cfgdata.get("SRC_URI"): | 28 | if var == "LICENSE" and d.getVar("SRC_URI", True) == cfgdata.get("SRC_URI"): |
29 | continue | 29 | continue |
30 | 30 | ||
31 | if not val: | 31 | if not val: |
@@ -43,11 +43,11 @@ def var_renames_overwrite(cfgdata, d): | |||
43 | def incorrect_nonempty_PACKAGES(cfgdata, d): | 43 | def incorrect_nonempty_PACKAGES(cfgdata, d): |
44 | if bb.data.inherits_class("native", d) or \ | 44 | if bb.data.inherits_class("native", d) or \ |
45 | bb.data.inherits_class("cross", d): | 45 | bb.data.inherits_class("cross", d): |
46 | if d.getVar("PACKAGES", 1): | 46 | if d.getVar("PACKAGES", True): |
47 | return True | 47 | return True |
48 | 48 | ||
49 | def can_use_autotools_base(cfgdata, d): | 49 | def can_use_autotools_base(cfgdata, d): |
50 | cfg = d.getVar("do_configure", 1) | 50 | cfg = d.getVar("do_configure", True) |
51 | if not bb.data.inherits_class("autotools", d): | 51 | if not bb.data.inherits_class("autotools", d): |
52 | return False | 52 | return False |
53 | 53 | ||
@@ -65,10 +65,10 @@ def can_use_autotools_base(cfgdata, d): | |||
65 | 65 | ||
66 | def can_remove_FILESPATH(cfgdata, d): | 66 | def can_remove_FILESPATH(cfgdata, d): |
67 | expected = cfgdata.get("FILESPATH") | 67 | expected = cfgdata.get("FILESPATH") |
68 | #expected = "${@':'.join([os.path.normpath(os.path.join(fp, p, o)) for fp in d.getVar('FILESPATHBASE', 1).split(':') for p in d.getVar('FILESPATHPKG', 1).split(':') for o in (d.getVar('OVERRIDES', 1) + ':').split(':') if os.path.exists(os.path.join(fp, p, o))])}:${FILESDIR}" | 68 | #expected = "${@':'.join([os.path.normpath(os.path.join(fp, p, o)) for fp in d.getVar('FILESPATHBASE', True).split(':') for p in d.getVar('FILESPATHPKG', True).split(':') for o in (d.getVar('OVERRIDES', True) + ':').split(':') if os.path.exists(os.path.join(fp, p, o))])}:${FILESDIR}" |
69 | expectedpaths = bb.data.expand(expected, d) | 69 | expectedpaths = bb.data.expand(expected, d) |
70 | unexpanded = d.getVar("FILESPATH", 0) | 70 | unexpanded = d.getVar("FILESPATH", 0) |
71 | filespath = d.getVar("FILESPATH", 1).split(":") | 71 | filespath = d.getVar("FILESPATH", True).split(":") |
72 | filespath = [os.path.normpath(f) for f in filespath if os.path.exists(f)] | 72 | filespath = [os.path.normpath(f) for f in filespath if os.path.exists(f)] |
73 | for fp in filespath: | 73 | for fp in filespath: |
74 | if not fp in expectedpaths: | 74 | if not fp in expectedpaths: |
@@ -79,13 +79,13 @@ def can_remove_FILESPATH(cfgdata, d): | |||
79 | 79 | ||
80 | def can_remove_FILESDIR(cfgdata, d): | 80 | def can_remove_FILESDIR(cfgdata, d): |
81 | expected = cfgdata.get("FILESDIR") | 81 | expected = cfgdata.get("FILESDIR") |
82 | #expected = "${@bb.which(d.getVar('FILESPATH', 1), '.')}" | 82 | #expected = "${@bb.which(d.getVar('FILESPATH', True), '.')}" |
83 | unexpanded = d.getVar("FILESDIR", 0) | 83 | unexpanded = d.getVar("FILESDIR", 0) |
84 | if unexpanded is None: | 84 | if unexpanded is None: |
85 | return False | 85 | return False |
86 | 86 | ||
87 | expanded = os.path.normpath(d.getVar("FILESDIR", 1)) | 87 | expanded = os.path.normpath(d.getVar("FILESDIR", True)) |
88 | filespath = d.getVar("FILESPATH", 1).split(":") | 88 | filespath = d.getVar("FILESPATH", True).split(":") |
89 | filespath = [os.path.normpath(f) for f in filespath if os.path.exists(f)] | 89 | filespath = [os.path.normpath(f) for f in filespath if os.path.exists(f)] |
90 | 90 | ||
91 | return unexpanded != expected and \ | 91 | return unexpanded != expected and \ |
@@ -103,7 +103,7 @@ def can_remove_others(p, cfgdata, d): | |||
103 | continue | 103 | continue |
104 | 104 | ||
105 | try: | 105 | try: |
106 | expanded = d.getVar(k, 1) | 106 | expanded = d.getVar(k, True) |
107 | cfgexpanded = bb.data.expand(cfgunexpanded, d) | 107 | cfgexpanded = bb.data.expand(cfgunexpanded, d) |
108 | except bb.fetch.ParameterError: | 108 | except bb.fetch.ParameterError: |
109 | continue | 109 | continue |
@@ -115,8 +115,8 @@ def can_remove_others(p, cfgdata, d): | |||
115 | (p, cfgunexpanded, unexpanded, expanded)) | 115 | (p, cfgunexpanded, unexpanded, expanded)) |
116 | 116 | ||
117 | python do_recipe_sanity () { | 117 | python do_recipe_sanity () { |
118 | p = d.getVar("P", 1) | 118 | p = d.getVar("P", True) |
119 | p = "%s %s %s" % (d.getVar("PN", 1), d.getVar("PV", 1), d.getVar("PR", 1)) | 119 | p = "%s %s %s" % (d.getVar("PN", True), d.getVar("PV", True), d.getVar("PR", True)) |
120 | 120 | ||
121 | sanitychecks = [ | 121 | sanitychecks = [ |
122 | (can_remove_FILESDIR, "candidate for removal of FILESDIR"), | 122 | (can_remove_FILESDIR, "candidate for removal of FILESDIR"), |
diff --git a/meta/classes/siteconfig.bbclass b/meta/classes/siteconfig.bbclass index 880dcad1f3..ccbe5b99c9 100644 --- a/meta/classes/siteconfig.bbclass +++ b/meta/classes/siteconfig.bbclass | |||
@@ -2,7 +2,7 @@ python siteconfig_do_siteconfig () { | |||
2 | shared_state = sstate_state_fromvars(d) | 2 | shared_state = sstate_state_fromvars(d) |
3 | if shared_state['name'] != 'populate-sysroot': | 3 | if shared_state['name'] != 'populate-sysroot': |
4 | return | 4 | return |
5 | if not os.path.isdir(os.path.join(d.getVar('FILE_DIRNAME', 1), 'site_config')): | 5 | if not os.path.isdir(os.path.join(d.getVar('FILE_DIRNAME', True), 'site_config')): |
6 | bb.debug(1, "No site_config directory, skipping do_siteconfig") | 6 | bb.debug(1, "No site_config directory, skipping do_siteconfig") |
7 | return | 7 | return |
8 | bb.build.exec_func('do_siteconfig_gencache', d) | 8 | bb.build.exec_func('do_siteconfig_gencache', d) |
diff --git a/meta/classes/siteinfo.bbclass b/meta/classes/siteinfo.bbclass index bf6af2b838..8c256ceff3 100644 --- a/meta/classes/siteinfo.bbclass +++ b/meta/classes/siteinfo.bbclass | |||
@@ -130,7 +130,7 @@ def siteinfo_get_files(d, no_cache = False): | |||
130 | if no_cache: return sitefiles | 130 | if no_cache: return sitefiles |
131 | 131 | ||
132 | # Now check for siteconfig cache files | 132 | # Now check for siteconfig cache files |
133 | path_siteconfig = d.getVar('SITECONFIG_SYSROOTCACHE', 1) | 133 | path_siteconfig = d.getVar('SITECONFIG_SYSROOTCACHE', True) |
134 | if os.path.isdir(path_siteconfig): | 134 | if os.path.isdir(path_siteconfig): |
135 | for i in os.listdir(path_siteconfig): | 135 | for i in os.listdir(path_siteconfig): |
136 | filename = os.path.join(path_siteconfig, i) | 136 | filename = os.path.join(path_siteconfig, i) |
diff --git a/meta/classes/sourcepkg.bbclass b/meta/classes/sourcepkg.bbclass index 38edfe4e2e..2a78a90452 100644 --- a/meta/classes/sourcepkg.bbclass +++ b/meta/classes/sourcepkg.bbclass | |||
@@ -6,7 +6,7 @@ DISTRO ?= "openembedded" | |||
6 | 6 | ||
7 | def get_src_tree(d): | 7 | def get_src_tree(d): |
8 | 8 | ||
9 | workdir = d.getVar('WORKDIR', 1) | 9 | workdir = d.getVar('WORKDIR', True) |
10 | if not workdir: | 10 | if not workdir: |
11 | bb.error("WORKDIR not defined, unable to find source tree.") | 11 | bb.error("WORKDIR not defined, unable to find source tree.") |
12 | return | 12 | return |
@@ -55,8 +55,8 @@ sourcepkg_do_archive_bb() { | |||
55 | 55 | ||
56 | python sourcepkg_do_dumpdata() { | 56 | python sourcepkg_do_dumpdata() { |
57 | 57 | ||
58 | workdir = d.getVar('WORKDIR', 1) | 58 | workdir = d.getVar('WORKDIR', True) |
59 | distro = d.getVar('DISTRO', 1) | 59 | distro = d.getVar('DISTRO', True) |
60 | s_tree = get_src_tree(d) | 60 | s_tree = get_src_tree(d) |
61 | openembeddeddir = os.path.join(workdir, s_tree, distro) | 61 | openembeddeddir = os.path.join(workdir, s_tree, distro) |
62 | dumpfile = os.path.join(openembeddeddir, bb.data.expand("${P}-${PR}.showdata.dump",d)) | 62 | dumpfile = os.path.join(openembeddeddir, bb.data.expand("${P}-${PR}.showdata.dump",d)) |
@@ -74,7 +74,7 @@ python sourcepkg_do_dumpdata() { | |||
74 | # emit the metadata which isnt valid shell | 74 | # emit the metadata which isnt valid shell |
75 | for e in d.keys(): | 75 | for e in d.keys(): |
76 | if d.getVarFlag(e, 'python'): | 76 | if d.getVarFlag(e, 'python'): |
77 | f.write("\npython %s () {\n%s}\n" % (e, d.getVar(e, 1))) | 77 | f.write("\npython %s () {\n%s}\n" % (e, d.getVar(e, True))) |
78 | f.close() | 78 | f.close() |
79 | } | 79 | } |
80 | 80 | ||
diff --git a/meta/classes/src_distribute.bbclass b/meta/classes/src_distribute.bbclass index 651e492598..efa2720e23 100644 --- a/meta/classes/src_distribute.bbclass +++ b/meta/classes/src_distribute.bbclass | |||
@@ -3,12 +3,12 @@ python do_distribute_sources () { | |||
3 | l = bb.data.createCopy(d) | 3 | l = bb.data.createCopy(d) |
4 | bb.data.update_data(l) | 4 | bb.data.update_data(l) |
5 | 5 | ||
6 | sources_dir = d.getVar('SRC_DISTRIBUTEDIR', 1) | 6 | sources_dir = d.getVar('SRC_DISTRIBUTEDIR', True) |
7 | src_uri = d.getVar('SRC_URI', 1).split() | 7 | src_uri = d.getVar('SRC_URI', True).split() |
8 | fetcher = bb.fetch2.Fetch(src_uri, d) | 8 | fetcher = bb.fetch2.Fetch(src_uri, d) |
9 | ud = fetcher.ud | 9 | ud = fetcher.ud |
10 | 10 | ||
11 | licenses = d.getVar('LICENSE', 1).replace('&', '|') | 11 | licenses = d.getVar('LICENSE', True).replace('&', '|') |
12 | licenses = licenses.replace('(', '').replace(')', '') | 12 | licenses = licenses.replace('(', '').replace(')', '') |
13 | clean_licenses = "" | 13 | clean_licenses = "" |
14 | for x in licenses.split(): | 14 | for x in licenses.split(): |
@@ -20,7 +20,7 @@ python do_distribute_sources () { | |||
20 | 20 | ||
21 | for license in clean_licenses.split('|'): | 21 | for license in clean_licenses.split('|'): |
22 | for url in ud.values(): | 22 | for url in ud.values(): |
23 | cmd = d.getVar('SRC_DISTRIBUTECOMMAND', 1) | 23 | cmd = d.getVar('SRC_DISTRIBUTECOMMAND', True) |
24 | if not cmd: | 24 | if not cmd: |
25 | raise bb.build.FuncFailed("Unable to distribute sources, SRC_DISTRIBUTECOMMAND not defined") | 25 | raise bb.build.FuncFailed("Unable to distribute sources, SRC_DISTRIBUTECOMMAND not defined") |
26 | url.setup_localpath(d) | 26 | url.setup_localpath(d) |
@@ -29,9 +29,9 @@ python do_distribute_sources () { | |||
29 | if url.basename == '*': | 29 | if url.basename == '*': |
30 | import os.path | 30 | import os.path |
31 | dest_dir = os.path.basename(os.path.dirname(os.path.abspath(url.localpath))) | 31 | dest_dir = os.path.basename(os.path.dirname(os.path.abspath(url.localpath))) |
32 | d.setVar('DEST', "%s_%s/" % (d.getVar('PF', 1), dest_dir)) | 32 | d.setVar('DEST', "%s_%s/" % (d.getVar('PF', True), dest_dir)) |
33 | else: | 33 | else: |
34 | d.setVar('DEST', "%s_%s" % (d.getVar('PF', 1), url.basename)) | 34 | d.setVar('DEST', "%s_%s" % (d.getVar('PF', True), url.basename)) |
35 | else: | 35 | else: |
36 | d.setVar('DEST', '') | 36 | d.setVar('DEST', '') |
37 | 37 | ||
diff --git a/meta/classes/syslinux.bbclass b/meta/classes/syslinux.bbclass index ec93201581..b194fa69a3 100644 --- a/meta/classes/syslinux.bbclass +++ b/meta/classes/syslinux.bbclass | |||
@@ -57,12 +57,12 @@ python build_syslinux_menu () { | |||
57 | import copy | 57 | import copy |
58 | import sys | 58 | import sys |
59 | 59 | ||
60 | workdir = d.getVar('WORKDIR', 1) | 60 | workdir = d.getVar('WORKDIR', True) |
61 | if not workdir: | 61 | if not workdir: |
62 | bb.error("WORKDIR is not defined") | 62 | bb.error("WORKDIR is not defined") |
63 | return | 63 | return |
64 | 64 | ||
65 | labels = d.getVar('LABELS', 1) | 65 | labels = d.getVar('LABELS', True) |
66 | if not labels: | 66 | if not labels: |
67 | bb.debug(1, "LABELS not defined, nothing to do") | 67 | bb.debug(1, "LABELS not defined, nothing to do") |
68 | return | 68 | return |
@@ -71,7 +71,7 @@ python build_syslinux_menu () { | |||
71 | bb.debug(1, "No labels, nothing to do") | 71 | bb.debug(1, "No labels, nothing to do") |
72 | return | 72 | return |
73 | 73 | ||
74 | cfile = d.getVar('SYSLINUXMENU', 1) | 74 | cfile = d.getVar('SYSLINUXMENU', True) |
75 | if not cfile: | 75 | if not cfile: |
76 | raise bb.build.FuncFailed('Unable to read SYSLINUXMENU') | 76 | raise bb.build.FuncFailed('Unable to read SYSLINUXMENU') |
77 | 77 | ||
@@ -100,7 +100,7 @@ python build_syslinux_menu () { | |||
100 | localdata.setVar('OVERRIDES', label + ':' + overrides) | 100 | localdata.setVar('OVERRIDES', label + ':' + overrides) |
101 | bb.data.update_data(localdata) | 101 | bb.data.update_data(localdata) |
102 | 102 | ||
103 | usage = localdata.getVar('USAGE', 1) | 103 | usage = localdata.getVar('USAGE', True) |
104 | cfgfile.write(' \x0F\x30\x3E%16s\x0F\x30\x37: ' % (label)) | 104 | cfgfile.write(' \x0F\x30\x3E%16s\x0F\x30\x37: ' % (label)) |
105 | cfgfile.write('%s\n' % (usage)) | 105 | cfgfile.write('%s\n' % (usage)) |
106 | 106 | ||
@@ -114,12 +114,12 @@ python build_syslinux_cfg () { | |||
114 | import copy | 114 | import copy |
115 | import sys | 115 | import sys |
116 | 116 | ||
117 | workdir = d.getVar('WORKDIR', 1) | 117 | workdir = d.getVar('WORKDIR', True) |
118 | if not workdir: | 118 | if not workdir: |
119 | bb.error("WORKDIR not defined, unable to package") | 119 | bb.error("WORKDIR not defined, unable to package") |
120 | return | 120 | return |
121 | 121 | ||
122 | labels = d.getVar('LABELS', 1) | 122 | labels = d.getVar('LABELS', True) |
123 | if not labels: | 123 | if not labels: |
124 | bb.debug(1, "LABELS not defined, nothing to do") | 124 | bb.debug(1, "LABELS not defined, nothing to do") |
125 | return | 125 | return |
@@ -128,7 +128,7 @@ python build_syslinux_cfg () { | |||
128 | bb.debug(1, "No labels, nothing to do") | 128 | bb.debug(1, "No labels, nothing to do") |
129 | return | 129 | return |
130 | 130 | ||
131 | cfile = d.getVar('SYSLINUXCFG', 1) | 131 | cfile = d.getVar('SYSLINUXCFG', True) |
132 | if not cfile: | 132 | if not cfile: |
133 | raise bb.build.FuncFailed('Unable to read SYSLINUXCFG') | 133 | raise bb.build.FuncFailed('Unable to read SYSLINUXCFG') |
134 | 134 | ||
@@ -139,7 +139,7 @@ python build_syslinux_cfg () { | |||
139 | 139 | ||
140 | cfgfile.write('# Automatically created by OE\n') | 140 | cfgfile.write('# Automatically created by OE\n') |
141 | 141 | ||
142 | opts = d.getVar('SYSLINUX_OPTS', 1) | 142 | opts = d.getVar('SYSLINUX_OPTS', True) |
143 | 143 | ||
144 | if opts: | 144 | if opts: |
145 | for opt in opts.split(';'): | 145 | for opt in opts.split(';'): |
@@ -148,26 +148,26 @@ python build_syslinux_cfg () { | |||
148 | cfgfile.write('ALLOWOPTIONS 1\n'); | 148 | cfgfile.write('ALLOWOPTIONS 1\n'); |
149 | cfgfile.write('DEFAULT %s\n' % (labels.split()[0])) | 149 | cfgfile.write('DEFAULT %s\n' % (labels.split()[0])) |
150 | 150 | ||
151 | timeout = d.getVar('SYSLINUX_TIMEOUT', 1) | 151 | timeout = d.getVar('SYSLINUX_TIMEOUT', True) |
152 | 152 | ||
153 | if timeout: | 153 | if timeout: |
154 | cfgfile.write('TIMEOUT %s\n' % timeout) | 154 | cfgfile.write('TIMEOUT %s\n' % timeout) |
155 | else: | 155 | else: |
156 | cfgfile.write('TIMEOUT 50\n') | 156 | cfgfile.write('TIMEOUT 50\n') |
157 | 157 | ||
158 | prompt = d.getVar('SYSLINUX_PROMPT', 1) | 158 | prompt = d.getVar('SYSLINUX_PROMPT', True) |
159 | if prompt: | 159 | if prompt: |
160 | cfgfile.write('PROMPT %s\n' % prompt) | 160 | cfgfile.write('PROMPT %s\n' % prompt) |
161 | else: | 161 | else: |
162 | cfgfile.write('PROMPT 1\n') | 162 | cfgfile.write('PROMPT 1\n') |
163 | 163 | ||
164 | menu = d.getVar('AUTO_SYSLINUXMENU', 1) | 164 | menu = d.getVar('AUTO_SYSLINUXMENU', True) |
165 | 165 | ||
166 | # This is ugly. My bad. | 166 | # This is ugly. My bad. |
167 | 167 | ||
168 | if menu: | 168 | if menu: |
169 | bb.build.exec_func('build_syslinux_menu', d) | 169 | bb.build.exec_func('build_syslinux_menu', d) |
170 | mfile = d.getVar('SYSLINUXMENU', 1) | 170 | mfile = d.getVar('SYSLINUXMENU', True) |
171 | cfgfile.write('DISPLAY %s\n' % (mfile.split('/')[-1]) ) | 171 | cfgfile.write('DISPLAY %s\n' % (mfile.split('/')[-1]) ) |
172 | 172 | ||
173 | for label in labels.split(): | 173 | for label in labels.split(): |
@@ -182,8 +182,8 @@ python build_syslinux_cfg () { | |||
182 | 182 | ||
183 | cfgfile.write('LABEL %s\nKERNEL /vmlinuz\n' % (label)) | 183 | cfgfile.write('LABEL %s\nKERNEL /vmlinuz\n' % (label)) |
184 | 184 | ||
185 | append = localdata.getVar('APPEND', 1) | 185 | append = localdata.getVar('APPEND', True) |
186 | initrd = localdata.getVar('INITRD', 1) | 186 | initrd = localdata.getVar('INITRD', True) |
187 | 187 | ||
188 | if append: | 188 | if append: |
189 | cfgfile.write('APPEND ') | 189 | cfgfile.write('APPEND ') |
diff --git a/meta/classes/task.bbclass b/meta/classes/task.bbclass index 7891207a64..22c2fd3744 100644 --- a/meta/classes/task.bbclass +++ b/meta/classes/task.bbclass | |||
@@ -17,7 +17,7 @@ PACKAGE_ARCH = "all" | |||
17 | # to the list. Their dependencies (RRECOMMENDS) are handled as usual | 17 | # to the list. Their dependencies (RRECOMMENDS) are handled as usual |
18 | # by package_depchains in a following step. | 18 | # by package_depchains in a following step. |
19 | python () { | 19 | python () { |
20 | packages = d.getVar('PACKAGES', 1).split() | 20 | packages = d.getVar('PACKAGES', True).split() |
21 | genpackages = [] | 21 | genpackages = [] |
22 | for pkg in packages: | 22 | for pkg in packages: |
23 | for postfix in ['-dbg', '-dev']: | 23 | for postfix in ['-dbg', '-dev']: |
diff --git a/meta/classes/update-alternatives.bbclass b/meta/classes/update-alternatives.bbclass index 0e8e58bd03..ae58344d3d 100644 --- a/meta/classes/update-alternatives.bbclass +++ b/meta/classes/update-alternatives.bbclass | |||
@@ -94,22 +94,22 @@ python __anonymous() { | |||
94 | } | 94 | } |
95 | 95 | ||
96 | python populate_packages_prepend () { | 96 | python populate_packages_prepend () { |
97 | pkg = d.getVar('PN', 1) | 97 | pkg = d.getVar('PN', True) |
98 | bb.note('adding update-alternatives calls to postinst/postrm for %s' % pkg) | 98 | bb.note('adding update-alternatives calls to postinst/postrm for %s' % pkg) |
99 | postinst = d.getVar('pkg_postinst_%s' % pkg, 1) or d.getVar('pkg_postinst', 1) | 99 | postinst = d.getVar('pkg_postinst_%s' % pkg, True) or d.getVar('pkg_postinst', True) |
100 | if not postinst: | 100 | if not postinst: |
101 | postinst = '#!/bin/sh\n' | 101 | postinst = '#!/bin/sh\n' |
102 | if d.getVar('ALTERNATIVE_LINKS') != None: | 102 | if d.getVar('ALTERNATIVE_LINKS') != None: |
103 | postinst += d.getVar('update_alternatives_batch_postinst', 1) | 103 | postinst += d.getVar('update_alternatives_batch_postinst', True) |
104 | else: | 104 | else: |
105 | postinst += d.getVar('update_alternatives_postinst', 1) | 105 | postinst += d.getVar('update_alternatives_postinst', True) |
106 | d.setVar('pkg_postinst_%s' % pkg, postinst) | 106 | d.setVar('pkg_postinst_%s' % pkg, postinst) |
107 | postrm = d.getVar('pkg_postrm_%s' % pkg, 1) or d.getVar('pkg_postrm', 1) | 107 | postrm = d.getVar('pkg_postrm_%s' % pkg, True) or d.getVar('pkg_postrm', True) |
108 | if not postrm: | 108 | if not postrm: |
109 | postrm = '#!/bin/sh\n' | 109 | postrm = '#!/bin/sh\n' |
110 | if d.getVar('ALTERNATIVE_LINKS') != None: | 110 | if d.getVar('ALTERNATIVE_LINKS') != None: |
111 | postrm += d.getVar('update_alternatives_batch_postrm', 1) | 111 | postrm += d.getVar('update_alternatives_batch_postrm', True) |
112 | else: | 112 | else: |
113 | postrm += d.getVar('update_alternatives_postrm', 1) | 113 | postrm += d.getVar('update_alternatives_postrm', True) |
114 | d.setVar('pkg_postrm_%s' % pkg, postrm) | 114 | d.setVar('pkg_postrm_%s' % pkg, postrm) |
115 | } | 115 | } |
diff --git a/meta/classes/update-rc.d.bbclass b/meta/classes/update-rc.d.bbclass index db88a8e764..bddead4a25 100644 --- a/meta/classes/update-rc.d.bbclass +++ b/meta/classes/update-rc.d.bbclass | |||
@@ -47,7 +47,7 @@ python populate_packages_prepend () { | |||
47 | def update_rcd_package(pkg): | 47 | def update_rcd_package(pkg): |
48 | bb.debug(1, 'adding update-rc.d calls to postinst/postrm for %s' % pkg) | 48 | bb.debug(1, 'adding update-rc.d calls to postinst/postrm for %s' % pkg) |
49 | localdata = bb.data.createCopy(d) | 49 | localdata = bb.data.createCopy(d) |
50 | overrides = localdata.getVar("OVERRIDES", 1) | 50 | overrides = localdata.getVar("OVERRIDES", True) |
51 | localdata.setVar("OVERRIDES", "%s:%s" % (pkg, overrides)) | 51 | localdata.setVar("OVERRIDES", "%s:%s" % (pkg, overrides)) |
52 | bb.data.update_data(localdata) | 52 | bb.data.update_data(localdata) |
53 | 53 | ||
@@ -56,28 +56,28 @@ python populate_packages_prepend () { | |||
56 | execute on the target. Not doing so may cause update_rc.d postinst invoked | 56 | execute on the target. Not doing so may cause update_rc.d postinst invoked |
57 | twice to cause unwanted warnings. | 57 | twice to cause unwanted warnings. |
58 | """ | 58 | """ |
59 | postinst = localdata.getVar('pkg_postinst', 1) | 59 | postinst = localdata.getVar('pkg_postinst', True) |
60 | if not postinst: | 60 | if not postinst: |
61 | postinst = '#!/bin/sh\n' | 61 | postinst = '#!/bin/sh\n' |
62 | postinst += localdata.getVar('updatercd_postinst', 1) | 62 | postinst += localdata.getVar('updatercd_postinst', True) |
63 | d.setVar('pkg_postinst_%s' % pkg, postinst) | 63 | d.setVar('pkg_postinst_%s' % pkg, postinst) |
64 | 64 | ||
65 | prerm = localdata.getVar('pkg_prerm', 1) | 65 | prerm = localdata.getVar('pkg_prerm', True) |
66 | if not prerm: | 66 | if not prerm: |
67 | prerm = '#!/bin/sh\n' | 67 | prerm = '#!/bin/sh\n' |
68 | prerm += localdata.getVar('updatercd_prerm', 1) | 68 | prerm += localdata.getVar('updatercd_prerm', True) |
69 | d.setVar('pkg_prerm_%s' % pkg, prerm) | 69 | d.setVar('pkg_prerm_%s' % pkg, prerm) |
70 | 70 | ||
71 | postrm = localdata.getVar('pkg_postrm', 1) | 71 | postrm = localdata.getVar('pkg_postrm', True) |
72 | if not postrm: | 72 | if not postrm: |
73 | postrm = '#!/bin/sh\n' | 73 | postrm = '#!/bin/sh\n' |
74 | postrm += localdata.getVar('updatercd_postrm', 1) | 74 | postrm += localdata.getVar('updatercd_postrm', True) |
75 | d.setVar('pkg_postrm_%s' % pkg, postrm) | 75 | d.setVar('pkg_postrm_%s' % pkg, postrm) |
76 | 76 | ||
77 | pkgs = d.getVar('INITSCRIPT_PACKAGES', 1) | 77 | pkgs = d.getVar('INITSCRIPT_PACKAGES', True) |
78 | if pkgs == None: | 78 | if pkgs == None: |
79 | pkgs = d.getVar('UPDATERCPN', 1) | 79 | pkgs = d.getVar('UPDATERCPN', True) |
80 | packages = (d.getVar('PACKAGES', 1) or "").split() | 80 | packages = (d.getVar('PACKAGES', True) or "").split() |
81 | if not pkgs in packages and packages != []: | 81 | if not pkgs in packages and packages != []: |
82 | pkgs = packages[0] | 82 | pkgs = packages[0] |
83 | for pkg in pkgs.split(): | 83 | for pkg in pkgs.split(): |
diff --git a/meta/classes/utility-tasks.bbclass b/meta/classes/utility-tasks.bbclass index 009ef1fd04..bbdf6e159b 100644 --- a/meta/classes/utility-tasks.bbclass +++ b/meta/classes/utility-tasks.bbclass | |||
@@ -24,7 +24,7 @@ python do_clean() { | |||
24 | bb.note("Removing " + dir) | 24 | bb.note("Removing " + dir) |
25 | oe.path.remove(dir) | 25 | oe.path.remove(dir) |
26 | 26 | ||
27 | for f in (d.getVar('CLEANFUNCS', 1) or '').split(): | 27 | for f in (d.getVar('CLEANFUNCS', True) or '').split(): |
28 | bb.build.exec_func(f, d) | 28 | bb.build.exec_func(f, d) |
29 | } | 29 | } |
30 | 30 | ||
diff --git a/meta/classes/utils.bbclass b/meta/classes/utils.bbclass index 103fa9a546..3b5946308c 100644 --- a/meta/classes/utils.bbclass +++ b/meta/classes/utils.bbclass | |||
@@ -336,7 +336,7 @@ def base_set_filespath(path, d): | |||
336 | if extrapaths != "": | 336 | if extrapaths != "": |
337 | path = extrapaths.split(":") + path | 337 | path = extrapaths.split(":") + path |
338 | # The ":" ensures we have an 'empty' override | 338 | # The ":" ensures we have an 'empty' override |
339 | overrides = (d.getVar("OVERRIDES", 1) or "") + ":" | 339 | overrides = (d.getVar("OVERRIDES", True) or "") + ":" |
340 | for p in path: | 340 | for p in path: |
341 | if p != "": | 341 | if p != "": |
342 | for o in overrides.split(":"): | 342 | for o in overrides.split(":"): |