diff options
author | Richard Purdie <richard.purdie@linuxfoundation.org> | 2012-03-03 11:21:22 +0000 |
---|---|---|
committer | Richard Purdie <richard.purdie@linuxfoundation.org> | 2012-03-05 10:23:53 -0800 |
commit | b1aeaa8b0dcf27ef8dcd87279901c6f5817d9e94 (patch) | |
tree | 62b315fdc05e03677922a0161fdc7a39e2985466 /meta/classes | |
parent | 22f0df2aa737bda1ae3bce862ee5ceba5b2d204b (diff) | |
download | poky-b1aeaa8b0dcf27ef8dcd87279901c6f5817d9e94.tar.gz |
meta: Replace bb.data.expand(xxx, d) -> d.expand(xxx)
sed \
-e 's:bb.data.\(expand([^,()]*\), *\([^) ]*\) *):\2.\1):g' \
-i `grep -ril bb.data.expand *`
Signed-off-by: Richard Purdie <richard.purdie@linuxfoundation.org>
Diffstat (limited to 'meta/classes')
-rw-r--r-- | meta/classes/base.bbclass | 4 | ||||
-rw-r--r-- | meta/classes/buildstats.bbclass | 12 | ||||
-rw-r--r-- | meta/classes/cpan-base.bbclass | 2 | ||||
-rw-r--r-- | meta/classes/icecc.bbclass | 26 | ||||
-rw-r--r-- | meta/classes/image-swab.bbclass | 2 | ||||
-rw-r--r-- | meta/classes/insane.bbclass | 2 | ||||
-rw-r--r-- | meta/classes/kernel-yocto.bbclass | 2 | ||||
-rw-r--r-- | meta/classes/package.bbclass | 24 | ||||
-rw-r--r-- | meta/classes/package_deb.bbclass | 2 | ||||
-rw-r--r-- | meta/classes/package_rpm.bbclass | 6 | ||||
-rw-r--r-- | meta/classes/package_tar.bbclass | 4 | ||||
-rw-r--r-- | meta/classes/recipe_sanity.bbclass | 6 | ||||
-rw-r--r-- | meta/classes/relocatable.bbclass | 8 | ||||
-rw-r--r-- | meta/classes/sanity.bbclass | 18 | ||||
-rw-r--r-- | meta/classes/sourcepkg.bbclass | 4 | ||||
-rw-r--r-- | meta/classes/sstate.bbclass | 30 | ||||
-rw-r--r-- | meta/classes/syslinux.bbclass | 2 | ||||
-rw-r--r-- | meta/classes/utility-tasks.bbclass | 2 |
18 files changed, 78 insertions, 78 deletions
diff --git a/meta/classes/base.bbclass b/meta/classes/base.bbclass index 48e4a28d83..82dcda9aec 100644 --- a/meta/classes/base.bbclass +++ b/meta/classes/base.bbclass | |||
@@ -122,7 +122,7 @@ def generate_git_config(e): | |||
122 | gitconfig_path = e.data.getVar('GIT_CONFIG', True) | 122 | gitconfig_path = e.data.getVar('GIT_CONFIG', True) |
123 | proxy_command = " gitProxy = %s\n" % data.getVar('OE_GIT_PROXY_COMMAND', e.data, True) | 123 | proxy_command = " gitProxy = %s\n" % data.getVar('OE_GIT_PROXY_COMMAND', e.data, True) |
124 | 124 | ||
125 | bb.mkdirhier(bb.data.expand("${GIT_CONFIG_PATH}", e.data)) | 125 | bb.mkdirhier(e.data.expand("${GIT_CONFIG_PATH}")) |
126 | if (os.path.exists(gitconfig_path)): | 126 | if (os.path.exists(gitconfig_path)): |
127 | os.remove(gitconfig_path) | 127 | os.remove(gitconfig_path) |
128 | 128 | ||
@@ -307,7 +307,7 @@ python () { | |||
307 | def appendVar(varname, appends): | 307 | def appendVar(varname, appends): |
308 | if not appends: | 308 | if not appends: |
309 | return | 309 | return |
310 | varname = bb.data.expand(varname, d) | 310 | varname = d.expand(varname) |
311 | d.appendVar(varname, " " + " ".join(appends)) | 311 | d.appendVar(varname, " " + " ".join(appends)) |
312 | 312 | ||
313 | extradeps = [] | 313 | extradeps = [] |
diff --git a/meta/classes/buildstats.bbclass b/meta/classes/buildstats.bbclass index efd3426c16..dc9afb101b 100644 --- a/meta/classes/buildstats.bbclass +++ b/meta/classes/buildstats.bbclass | |||
@@ -132,7 +132,7 @@ def get_timedata(var, data): | |||
132 | def write_task_data(status, logfile, dev, e): | 132 | def write_task_data(status, logfile, dev, e): |
133 | bn = get_bn(e) | 133 | bn = get_bn(e) |
134 | bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn) | 134 | bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn) |
135 | taskdir = os.path.join(bsdir, bb.data.expand("${PF}", e.data)) | 135 | taskdir = os.path.join(bsdir, e.data.expand("${PF}")) |
136 | file = open(os.path.join(logfile), "a") | 136 | file = open(os.path.join(logfile), "a") |
137 | timedata = get_timedata("__timedata_task", e.data) | 137 | timedata = get_timedata("__timedata_task", e.data) |
138 | if timedata: | 138 | if timedata: |
@@ -205,7 +205,7 @@ python run_buildstats () { | |||
205 | bn = get_bn(e) | 205 | bn = get_bn(e) |
206 | device = get_device(e) | 206 | device = get_device(e) |
207 | bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn) | 207 | bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn) |
208 | taskdir = os.path.join(bsdir, bb.data.expand("${PF}", e.data)) | 208 | taskdir = os.path.join(bsdir, e.data.expand("${PF}")) |
209 | build_time = os.path.join(bsdir, "build_stats") | 209 | build_time = os.path.join(bsdir, "build_stats") |
210 | file = open(build_time, "a") | 210 | file = open(build_time, "a") |
211 | ######################################################################## | 211 | ######################################################################## |
@@ -230,7 +230,7 @@ python run_buildstats () { | |||
230 | bn = get_bn(e) | 230 | bn = get_bn(e) |
231 | device = get_device(e) | 231 | device = get_device(e) |
232 | bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn) | 232 | bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn) |
233 | taskdir = os.path.join(bsdir, bb.data.expand("${PF}", e.data)) | 233 | taskdir = os.path.join(bsdir, e.data.expand("${PF}")) |
234 | if device != "NoLogicalDevice": | 234 | if device != "NoLogicalDevice": |
235 | set_diskdata("__diskdata_task", device, e.data) | 235 | set_diskdata("__diskdata_task", device, e.data) |
236 | set_timedata("__timedata_task", e.data) | 236 | set_timedata("__timedata_task", e.data) |
@@ -248,7 +248,7 @@ python run_buildstats () { | |||
248 | bn = get_bn(e) | 248 | bn = get_bn(e) |
249 | device = get_device(e) | 249 | device = get_device(e) |
250 | bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn) | 250 | bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn) |
251 | taskdir = os.path.join(bsdir, bb.data.expand("${PF}", e.data)) | 251 | taskdir = os.path.join(bsdir, e.data.expand("${PF}")) |
252 | write_task_data("passed", os.path.join(taskdir, e.task), device, e) | 252 | write_task_data("passed", os.path.join(taskdir, e.task), device, e) |
253 | if e.task == "do_rootfs": | 253 | if e.task == "do_rootfs": |
254 | bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn) | 254 | bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn) |
@@ -263,7 +263,7 @@ python run_buildstats () { | |||
263 | bn = get_bn(e) | 263 | bn = get_bn(e) |
264 | device = get_device(e) | 264 | device = get_device(e) |
265 | bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn) | 265 | bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn) |
266 | taskdir = os.path.join(bsdir, bb.data.expand("${PF}", e.data)) | 266 | taskdir = os.path.join(bsdir, e.data.expand("${PF}")) |
267 | write_task_data("failed", os.path.join(taskdir, e.task), device, e) | 267 | write_task_data("failed", os.path.join(taskdir, e.task), device, e) |
268 | ######################################################################## | 268 | ######################################################################## |
269 | # Lets make things easier and tell people where the build failed in | 269 | # Lets make things easier and tell people where the build failed in |
@@ -272,7 +272,7 @@ python run_buildstats () { | |||
272 | ######################################################################## | 272 | ######################################################################## |
273 | build_status = os.path.join(bsdir, "build_stats") | 273 | build_status = os.path.join(bsdir, "build_stats") |
274 | file = open(build_status,"a") | 274 | file = open(build_status,"a") |
275 | file.write(bb.data.expand("Failed at: ${PF} at task: %s \n" % e.task, e.data)) | 275 | file.write(e.data.expand("Failed at: ${PF} at task: %s \n" % e.task)) |
276 | file.close() | 276 | file.close() |
277 | 277 | ||
278 | } | 278 | } |
diff --git a/meta/classes/cpan-base.bbclass b/meta/classes/cpan-base.bbclass index 6cb1fefc29..b4b7b81d8d 100644 --- a/meta/classes/cpan-base.bbclass +++ b/meta/classes/cpan-base.bbclass | |||
@@ -12,7 +12,7 @@ PERL_OWN_DIR = "${@["", "/perl-native"][(bb.data.inherits_class('native', d))]}" | |||
12 | # Determine the staged version of perl from the perl configuration file | 12 | # Determine the staged version of perl from the perl configuration file |
13 | def get_perl_version(d): | 13 | def get_perl_version(d): |
14 | import re | 14 | import re |
15 | cfg = bb.data.expand('${STAGING_LIBDIR}${PERL_OWN_DIR}/perl/config.sh', d) | 15 | cfg = d.expand('${STAGING_LIBDIR}${PERL_OWN_DIR}/perl/config.sh') |
16 | try: | 16 | try: |
17 | f = open(cfg, 'r') | 17 | f = open(cfg, 'r') |
18 | except IOError: | 18 | except IOError: |
diff --git a/meta/classes/icecc.bbclass b/meta/classes/icecc.bbclass index a14e02db0e..45570739f2 100644 --- a/meta/classes/icecc.bbclass +++ b/meta/classes/icecc.bbclass | |||
@@ -39,7 +39,7 @@ def icecc_dep_prepend(d): | |||
39 | DEPENDS_prepend += "${@icecc_dep_prepend(d)} " | 39 | DEPENDS_prepend += "${@icecc_dep_prepend(d)} " |
40 | 40 | ||
41 | def get_cross_kernel_cc(bb,d): | 41 | def get_cross_kernel_cc(bb,d): |
42 | kernel_cc = bb.data.expand('${KERNEL_CC}', d) | 42 | kernel_cc = d.expand('${KERNEL_CC}') |
43 | kernel_cc = kernel_cc.replace('ccache', '').strip() | 43 | kernel_cc = kernel_cc.replace('ccache', '').strip() |
44 | kernel_cc = kernel_cc.split(' ')[0] | 44 | kernel_cc = kernel_cc.split(' ')[0] |
45 | kernel_cc = kernel_cc.strip() | 45 | kernel_cc = kernel_cc.strip() |
@@ -49,7 +49,7 @@ def create_path(compilers, bb, d): | |||
49 | """ | 49 | """ |
50 | Create Symlinks for the icecc in the staging directory | 50 | Create Symlinks for the icecc in the staging directory |
51 | """ | 51 | """ |
52 | staging = os.path.join(bb.data.expand('${STAGING_BINDIR}', d), "ice") | 52 | staging = os.path.join(d.expand('${STAGING_BINDIR}'), "ice") |
53 | if icc_is_kernel(bb, d): | 53 | if icc_is_kernel(bb, d): |
54 | staging += "-kernel" | 54 | staging += "-kernel" |
55 | 55 | ||
@@ -78,7 +78,7 @@ def create_path(compilers, bb, d): | |||
78 | return staging | 78 | return staging |
79 | 79 | ||
80 | def use_icc(bb,d): | 80 | def use_icc(bb,d): |
81 | package_tmp = bb.data.expand('${PN}', d) | 81 | package_tmp = d.expand('${PN}') |
82 | 82 | ||
83 | system_class_blacklist = [ "none" ] | 83 | system_class_blacklist = [ "none" ] |
84 | user_class_blacklist = (d.getVar('ICECC_USER_CLASS_BL') or "none").split() | 84 | user_class_blacklist = (d.getVar('ICECC_USER_CLASS_BL') or "none").split() |
@@ -101,7 +101,7 @@ def use_icc(bb,d): | |||
101 | return "no" | 101 | return "no" |
102 | 102 | ||
103 | if d.getVar('PARALLEL_MAKE') == "": | 103 | if d.getVar('PARALLEL_MAKE') == "": |
104 | bb.note(package_tmp, " ", bb.data.expand('${PV}', d), " has empty PARALLEL_MAKE, disable icecc") | 104 | bb.note(package_tmp, " ", d.expand('${PV}'), " has empty PARALLEL_MAKE, disable icecc") |
105 | return "no" | 105 | return "no" |
106 | 106 | ||
107 | return "yes" | 107 | return "yes" |
@@ -124,19 +124,19 @@ def icc_version(bb, d): | |||
124 | 124 | ||
125 | if icc_is_native(bb, d): | 125 | if icc_is_native(bb, d): |
126 | archive_name = "local-host-env" | 126 | archive_name = "local-host-env" |
127 | elif bb.data.expand('${HOST_PREFIX}', d) == "": | 127 | elif d.expand('${HOST_PREFIX}') == "": |
128 | bb.fatal(bb.data.expand("${PN}", d), " NULL prefix") | 128 | bb.fatal(d.expand("${PN}"), " NULL prefix") |
129 | else: | 129 | else: |
130 | prefix = bb.data.expand('${HOST_PREFIX}' , d) | 130 | prefix = d.expand('${HOST_PREFIX}' ) |
131 | distro = bb.data.expand('${DISTRO}', d) | 131 | distro = d.expand('${DISTRO}') |
132 | target_sys = bb.data.expand('${TARGET_SYS}', d) | 132 | target_sys = d.expand('${TARGET_SYS}') |
133 | float = d.getVar('TARGET_FPU') or "hard" | 133 | float = d.getVar('TARGET_FPU') or "hard" |
134 | archive_name = prefix + distro + "-" + target_sys + "-" + float | 134 | archive_name = prefix + distro + "-" + target_sys + "-" + float |
135 | if icc_is_kernel(bb, d): | 135 | if icc_is_kernel(bb, d): |
136 | archive_name += "-kernel" | 136 | archive_name += "-kernel" |
137 | 137 | ||
138 | import socket | 138 | import socket |
139 | ice_dir = bb.data.expand('${STAGING_DIR_NATIVE}${prefix_native}', d) | 139 | ice_dir = d.expand('${STAGING_DIR_NATIVE}${prefix_native}') |
140 | tar_file = os.path.join(ice_dir, 'ice', archive_name + "-@VERSION@-" + socket.gethostname() + '.tar.gz') | 140 | tar_file = os.path.join(ice_dir, 'ice', archive_name + "-@VERSION@-" + socket.gethostname() + '.tar.gz') |
141 | 141 | ||
142 | return tar_file | 142 | return tar_file |
@@ -146,7 +146,7 @@ def icc_path(bb,d): | |||
146 | return create_path( [get_cross_kernel_cc(bb,d), ], bb, d) | 146 | return create_path( [get_cross_kernel_cc(bb,d), ], bb, d) |
147 | 147 | ||
148 | else: | 148 | else: |
149 | prefix = bb.data.expand('${HOST_PREFIX}', d) | 149 | prefix = d.expand('${HOST_PREFIX}') |
150 | return create_path( [prefix+"gcc", prefix+"g++"], bb, d) | 150 | return create_path( [prefix+"gcc", prefix+"g++"], bb, d) |
151 | 151 | ||
152 | def icc_get_tool(bb, d, tool): | 152 | def icc_get_tool(bb, d, tool): |
@@ -155,8 +155,8 @@ def icc_get_tool(bb, d, tool): | |||
155 | elif icc_is_kernel(bb, d): | 155 | elif icc_is_kernel(bb, d): |
156 | return os.popen("which %s" % get_cross_kernel_cc(bb, d)).read()[:-1] | 156 | return os.popen("which %s" % get_cross_kernel_cc(bb, d)).read()[:-1] |
157 | else: | 157 | else: |
158 | ice_dir = bb.data.expand('${STAGING_BINDIR_TOOLCHAIN}', d) | 158 | ice_dir = d.expand('${STAGING_BINDIR_TOOLCHAIN}') |
159 | target_sys = bb.data.expand('${TARGET_SYS}', d) | 159 | target_sys = d.expand('${TARGET_SYS}') |
160 | return os.path.join(ice_dir, "%s-%s" % (target_sys, tool)) | 160 | return os.path.join(ice_dir, "%s-%s" % (target_sys, tool)) |
161 | 161 | ||
162 | set_icecc_env() { | 162 | set_icecc_env() { |
diff --git a/meta/classes/image-swab.bbclass b/meta/classes/image-swab.bbclass index 5aace0f5c8..0414653f72 100644 --- a/meta/classes/image-swab.bbclass +++ b/meta/classes/image-swab.bbclass | |||
@@ -54,7 +54,7 @@ python() { | |||
54 | deps = (d.getVarFlag('do_setscene', 'depends') or "").split() | 54 | deps = (d.getVarFlag('do_setscene', 'depends') or "").split() |
55 | deps.append('strace-native:do_populate_sysroot') | 55 | deps.append('strace-native:do_populate_sysroot') |
56 | d.setVarFlag('do_setscene', 'depends', " ".join(deps)) | 56 | d.setVarFlag('do_setscene', 'depends', " ".join(deps)) |
57 | logdir = bb.data.expand("${TRACE_LOGDIR}", d) | 57 | logdir = d.expand("${TRACE_LOGDIR}") |
58 | bb.utils.mkdirhier(logdir) | 58 | bb.utils.mkdirhier(logdir) |
59 | else: | 59 | else: |
60 | d.setVar('STRACEFUNC', '') | 60 | d.setVar('STRACEFUNC', '') |
diff --git a/meta/classes/insane.bbclass b/meta/classes/insane.bbclass index 7a84465ca6..6298136c39 100644 --- a/meta/classes/insane.bbclass +++ b/meta/classes/insane.bbclass | |||
@@ -673,7 +673,7 @@ python do_package_qa () { | |||
673 | python do_qa_staging() { | 673 | python do_qa_staging() { |
674 | bb.note("QA checking staging") | 674 | bb.note("QA checking staging") |
675 | 675 | ||
676 | if not package_qa_check_staged(bb.data.expand('${SYSROOT_DESTDIR}/${STAGING_LIBDIR}',d), d): | 676 | if not package_qa_check_staged(d.expand('${SYSROOT_DESTDIR}/${STAGING_LIBDIR}'), d): |
677 | bb.fatal("QA staging was broken by the package built above") | 677 | bb.fatal("QA staging was broken by the package built above") |
678 | } | 678 | } |
679 | 679 | ||
diff --git a/meta/classes/kernel-yocto.bbclass b/meta/classes/kernel-yocto.bbclass index 187e3cc07c..aabca789cf 100644 --- a/meta/classes/kernel-yocto.bbclass +++ b/meta/classes/kernel-yocto.bbclass | |||
@@ -244,7 +244,7 @@ python do_kernel_configcheck() { | |||
244 | bb.plain("NOTE: validating kernel configuration") | 244 | bb.plain("NOTE: validating kernel configuration") |
245 | 245 | ||
246 | pathprefix = "export PATH=%s:%s; " % (d.getVar('PATH', True), "${S}/scripts/util/") | 246 | pathprefix = "export PATH=%s:%s; " % (d.getVar('PATH', True), "${S}/scripts/util/") |
247 | cmd = bb.data.expand("cd ${B}/..; kconf_check -config- ${B} ${S} ${B} ${KBRANCH}",d ) | 247 | cmd = d.expand("cd ${B}/..; kconf_check -config- ${B} ${S} ${B} ${KBRANCH}") |
248 | ret, result = commands.getstatusoutput("%s%s" % (pathprefix, cmd)) | 248 | ret, result = commands.getstatusoutput("%s%s" % (pathprefix, cmd)) |
249 | 249 | ||
250 | bb.plain( "%s" % result ) | 250 | bb.plain( "%s" % result ) |
diff --git a/meta/classes/package.bbclass b/meta/classes/package.bbclass index 5c42619f3f..e8c449711f 100644 --- a/meta/classes/package.bbclass +++ b/meta/classes/package.bbclass | |||
@@ -195,10 +195,10 @@ def splitfile(file, debugfile, debugsrcdir, d): | |||
195 | dvar = d.getVar('PKGD', True) | 195 | dvar = d.getVar('PKGD', True) |
196 | pathprefix = "export PATH=%s; " % d.getVar('PATH', True) | 196 | pathprefix = "export PATH=%s; " % d.getVar('PATH', True) |
197 | objcopy = d.getVar("OBJCOPY", True) | 197 | objcopy = d.getVar("OBJCOPY", True) |
198 | debugedit = bb.data.expand("${STAGING_LIBDIR_NATIVE}/rpm/bin/debugedit", d) | 198 | debugedit = d.expand("${STAGING_LIBDIR_NATIVE}/rpm/bin/debugedit") |
199 | workdir = d.getVar("WORKDIR", True) | 199 | workdir = d.getVar("WORKDIR", True) |
200 | workparentdir = os.path.dirname(workdir) | 200 | workparentdir = os.path.dirname(workdir) |
201 | sourcefile = bb.data.expand("${WORKDIR}/debugsources.list", d) | 201 | sourcefile = d.expand("${WORKDIR}/debugsources.list") |
202 | 202 | ||
203 | # We ignore kernel modules, we don't generate debug info files. | 203 | # We ignore kernel modules, we don't generate debug info files. |
204 | if file.find("/lib/modules/") != -1 and file.endswith(".ko"): | 204 | if file.find("/lib/modules/") != -1 and file.endswith(".ko"): |
@@ -238,11 +238,11 @@ def splitfile2(debugsrcdir, d): | |||
238 | pathprefix = "export PATH=%s; " % d.getVar('PATH', True) | 238 | pathprefix = "export PATH=%s; " % d.getVar('PATH', True) |
239 | strip = d.getVar("STRIP", True) | 239 | strip = d.getVar("STRIP", True) |
240 | objcopy = d.getVar("OBJCOPY", True) | 240 | objcopy = d.getVar("OBJCOPY", True) |
241 | debugedit = bb.data.expand("${STAGING_LIBDIR_NATIVE}/rpm/bin/debugedit", d) | 241 | debugedit = d.expand("${STAGING_LIBDIR_NATIVE}/rpm/bin/debugedit") |
242 | workdir = d.getVar("WORKDIR", True) | 242 | workdir = d.getVar("WORKDIR", True) |
243 | workparentdir = os.path.dirname(workdir) | 243 | workparentdir = os.path.dirname(workdir) |
244 | workbasedir = os.path.basename(workdir) | 244 | workbasedir = os.path.basename(workdir) |
245 | sourcefile = bb.data.expand("${WORKDIR}/debugsources.list", d) | 245 | sourcefile = d.expand("${WORKDIR}/debugsources.list") |
246 | 246 | ||
247 | if debugsrcdir: | 247 | if debugsrcdir: |
248 | nosuchdir = [] | 248 | nosuchdir = [] |
@@ -624,7 +624,7 @@ python fixup_perms () { | |||
624 | if len(lsplit) != 8 and not (len(lsplit) == 3 and lsplit[1].lower() == "link"): | 624 | if len(lsplit) != 8 and not (len(lsplit) == 3 and lsplit[1].lower() == "link"): |
625 | bb.error("Fixup perms: %s invalid line: %s" % (conf, line)) | 625 | bb.error("Fixup perms: %s invalid line: %s" % (conf, line)) |
626 | continue | 626 | continue |
627 | entry = fs_perms_entry(bb.data.expand(line, d)) | 627 | entry = fs_perms_entry(d.expand(line)) |
628 | if entry and entry.path: | 628 | if entry and entry.path: |
629 | fs_perms_table[entry.path] = entry | 629 | fs_perms_table[entry.path] = entry |
630 | f.close() | 630 | f.close() |
@@ -1071,9 +1071,9 @@ python emit_pkgdata() { | |||
1071 | pkgdatadir = d.getVar('PKGDESTWORK', True) | 1071 | pkgdatadir = d.getVar('PKGDESTWORK', True) |
1072 | 1072 | ||
1073 | # Take shared lock since we're only reading, not writing | 1073 | # Take shared lock since we're only reading, not writing |
1074 | lf = bb.utils.lockfile(bb.data.expand("${PACKAGELOCK}", d), True) | 1074 | lf = bb.utils.lockfile(d.expand("${PACKAGELOCK}"), True) |
1075 | 1075 | ||
1076 | data_file = pkgdatadir + bb.data.expand("/${PN}" , d) | 1076 | data_file = pkgdatadir + d.expand("/${PN}" ) |
1077 | f = open(data_file, 'w') | 1077 | f = open(data_file, 'w') |
1078 | f.write("PACKAGES: %s\n" % packages) | 1078 | f.write("PACKAGES: %s\n" % packages) |
1079 | f.close() | 1079 | f.close() |
@@ -1154,7 +1154,7 @@ python package_do_filedeps() { | |||
1154 | pkgdest = d.getVar('PKGDEST', True) | 1154 | pkgdest = d.getVar('PKGDEST', True) |
1155 | packages = d.getVar('PACKAGES', True) | 1155 | packages = d.getVar('PACKAGES', True) |
1156 | 1156 | ||
1157 | rpmdeps = bb.data.expand("${RPMDEPS}", d) | 1157 | rpmdeps = d.expand("${RPMDEPS}") |
1158 | r = re.compile(r'[<>=]+ +[^ ]*') | 1158 | r = re.compile(r'[<>=]+ +[^ ]*') |
1159 | 1159 | ||
1160 | # Quick routine to process the results of the rpmdeps call... | 1160 | # Quick routine to process the results of the rpmdeps call... |
@@ -1253,7 +1253,7 @@ python package_do_shlibs() { | |||
1253 | shlibswork_dir = d.getVar('SHLIBSWORKDIR', True) | 1253 | shlibswork_dir = d.getVar('SHLIBSWORKDIR', True) |
1254 | 1254 | ||
1255 | # Take shared lock since we're only reading, not writing | 1255 | # Take shared lock since we're only reading, not writing |
1256 | lf = bb.utils.lockfile(bb.data.expand("${PACKAGELOCK}", d)) | 1256 | lf = bb.utils.lockfile(d.expand("${PACKAGELOCK}")) |
1257 | 1257 | ||
1258 | def linux_so(root, path, file): | 1258 | def linux_so(root, path, file): |
1259 | cmd = d.getVar('OBJDUMP', True) + " -p " + pipes.quote(os.path.join(root, file)) + " 2>/dev/null" | 1259 | cmd = d.getVar('OBJDUMP', True) + " -p " + pipes.quote(os.path.join(root, file)) + " 2>/dev/null" |
@@ -1499,7 +1499,7 @@ python package_do_pkgconfig () { | |||
1499 | if m: | 1499 | if m: |
1500 | name = m.group(1) | 1500 | name = m.group(1) |
1501 | val = m.group(2) | 1501 | val = m.group(2) |
1502 | pd.setVar(name, bb.data.expand(val, pd)) | 1502 | pd.setVar(name, pd.expand(val)) |
1503 | continue | 1503 | continue |
1504 | m = field_re.match(l) | 1504 | m = field_re.match(l) |
1505 | if m: | 1505 | if m: |
@@ -1509,7 +1509,7 @@ python package_do_pkgconfig () { | |||
1509 | pkgconfig_needed[pkg] += exp.replace(',', ' ').split() | 1509 | pkgconfig_needed[pkg] += exp.replace(',', ' ').split() |
1510 | 1510 | ||
1511 | # Take shared lock since we're only reading, not writing | 1511 | # Take shared lock since we're only reading, not writing |
1512 | lf = bb.utils.lockfile(bb.data.expand("${PACKAGELOCK}", d)) | 1512 | lf = bb.utils.lockfile(d.expand("${PACKAGELOCK}")) |
1513 | 1513 | ||
1514 | for pkg in packages.split(): | 1514 | for pkg in packages.split(): |
1515 | pkgs_file = os.path.join(shlibswork_dir, pkg + ".pclist") | 1515 | pkgs_file = os.path.join(shlibswork_dir, pkg + ".pclist") |
@@ -1560,7 +1560,7 @@ python read_shlibdeps () { | |||
1560 | rdepends = bb.utils.explode_dep_versions(d.getVar('RDEPENDS_' + pkg, False) or d.getVar('RDEPENDS', False) or "") | 1560 | rdepends = bb.utils.explode_dep_versions(d.getVar('RDEPENDS_' + pkg, False) or d.getVar('RDEPENDS', False) or "") |
1561 | 1561 | ||
1562 | for extension in ".shlibdeps", ".pcdeps", ".clilibdeps": | 1562 | for extension in ".shlibdeps", ".pcdeps", ".clilibdeps": |
1563 | depsfile = bb.data.expand("${PKGDEST}/" + pkg + extension, d) | 1563 | depsfile = d.expand("${PKGDEST}/" + pkg + extension) |
1564 | if os.access(depsfile, os.R_OK): | 1564 | if os.access(depsfile, os.R_OK): |
1565 | fd = file(depsfile) | 1565 | fd = file(depsfile) |
1566 | lines = fd.readlines() | 1566 | lines = fd.readlines() |
diff --git a/meta/classes/package_deb.bbclass b/meta/classes/package_deb.bbclass index 8721fd2121..1f7ec9ca3e 100644 --- a/meta/classes/package_deb.bbclass +++ b/meta/classes/package_deb.bbclass | |||
@@ -20,7 +20,7 @@ python do_package_deb_install () { | |||
20 | pkgfn = d.getVar('PKGFN', True) | 20 | pkgfn = d.getVar('PKGFN', True) |
21 | rootfs = d.getVar('IMAGE_ROOTFS', True) | 21 | rootfs = d.getVar('IMAGE_ROOTFS', True) |
22 | debdir = d.getVar('DEPLOY_DIR_DEB', True) | 22 | debdir = d.getVar('DEPLOY_DIR_DEB', True) |
23 | apt_config = bb.data.expand('${STAGING_ETCDIR_NATIVE}/apt/apt.conf', d) | 23 | apt_config = d.expand('${STAGING_ETCDIR_NATIVE}/apt/apt.conf') |
24 | stagingbindir = d.getVar('STAGING_BINDIR_NATIVE', True) | 24 | stagingbindir = d.getVar('STAGING_BINDIR_NATIVE', True) |
25 | tmpdir = d.getVar('TMPDIR', True) | 25 | tmpdir = d.getVar('TMPDIR', True) |
26 | 26 | ||
diff --git a/meta/classes/package_rpm.bbclass b/meta/classes/package_rpm.bbclass index af8c63ed6f..68313eccc8 100644 --- a/meta/classes/package_rpm.bbclass +++ b/meta/classes/package_rpm.bbclass | |||
@@ -998,9 +998,9 @@ python do_package_rpm () { | |||
998 | d.setVar('PACKAGE_ARCH_EXTEND', ml_prefix + package_arch) | 998 | d.setVar('PACKAGE_ARCH_EXTEND', ml_prefix + package_arch) |
999 | else: | 999 | else: |
1000 | d.setVar('PACKAGE_ARCH_EXTEND', package_arch) | 1000 | d.setVar('PACKAGE_ARCH_EXTEND', package_arch) |
1001 | pkgwritedir = bb.data.expand('${PKGWRITEDIRRPM}/${PACKAGE_ARCH_EXTEND}', d) | 1001 | pkgwritedir = d.expand('${PKGWRITEDIRRPM}/${PACKAGE_ARCH_EXTEND}') |
1002 | pkgarch = bb.data.expand('${PACKAGE_ARCH_EXTEND}${TARGET_VENDOR}-${TARGET_OS}', d) | 1002 | pkgarch = d.expand('${PACKAGE_ARCH_EXTEND}${TARGET_VENDOR}-${TARGET_OS}') |
1003 | magicfile = bb.data.expand('${STAGING_DIR_NATIVE}${datadir_native}/misc/magic.mgc', d) | 1003 | magicfile = d.expand('${STAGING_DIR_NATIVE}${datadir_native}/misc/magic.mgc') |
1004 | bb.mkdirhier(pkgwritedir) | 1004 | bb.mkdirhier(pkgwritedir) |
1005 | os.chmod(pkgwritedir, 0755) | 1005 | os.chmod(pkgwritedir, 0755) |
1006 | 1006 | ||
diff --git a/meta/classes/package_tar.bbclass b/meta/classes/package_tar.bbclass index 7590177e4b..68b1bf0fed 100644 --- a/meta/classes/package_tar.bbclass +++ b/meta/classes/package_tar.bbclass | |||
@@ -4,7 +4,7 @@ IMAGE_PKGTYPE ?= "tar" | |||
4 | 4 | ||
5 | python package_tar_fn () { | 5 | python package_tar_fn () { |
6 | fn = os.path.join(d.getVar('DEPLOY_DIR_TAR'), "%s-%s-%s.tar.gz" % (d.getVar('PKG'), d.getVar('PKGV'), d.getVar('PKGR'))) | 6 | fn = os.path.join(d.getVar('DEPLOY_DIR_TAR'), "%s-%s-%s.tar.gz" % (d.getVar('PKG'), d.getVar('PKGV'), d.getVar('PKGR'))) |
7 | fn = bb.data.expand(fn, d) | 7 | fn = d.expand(fn) |
8 | d.setVar('PKGFN', fn) | 8 | d.setVar('PKGFN', fn) |
9 | } | 9 | } |
10 | 10 | ||
@@ -68,7 +68,7 @@ python do_package_tar () { | |||
68 | overrides = localdata.getVar('OVERRIDES') | 68 | overrides = localdata.getVar('OVERRIDES') |
69 | if not overrides: | 69 | if not overrides: |
70 | raise bb.build.FuncFailed('OVERRIDES not defined') | 70 | raise bb.build.FuncFailed('OVERRIDES not defined') |
71 | overrides = bb.data.expand(overrides, localdata) | 71 | overrides = localdata.expand(overrides) |
72 | localdata.setVar('OVERRIDES', '%s:%s' % (overrides, pkg)) | 72 | localdata.setVar('OVERRIDES', '%s:%s' % (overrides, pkg)) |
73 | 73 | ||
74 | bb.data.update_data(localdata) | 74 | bb.data.update_data(localdata) |
diff --git a/meta/classes/recipe_sanity.bbclass b/meta/classes/recipe_sanity.bbclass index b3246599b9..da8ad76c96 100644 --- a/meta/classes/recipe_sanity.bbclass +++ b/meta/classes/recipe_sanity.bbclass | |||
@@ -66,7 +66,7 @@ def can_use_autotools_base(cfgdata, d): | |||
66 | def can_remove_FILESPATH(cfgdata, d): | 66 | def can_remove_FILESPATH(cfgdata, d): |
67 | expected = cfgdata.get("FILESPATH") | 67 | expected = cfgdata.get("FILESPATH") |
68 | #expected = "${@':'.join([os.path.normpath(os.path.join(fp, p, o)) for fp in d.getVar('FILESPATHBASE', True).split(':') for p in d.getVar('FILESPATHPKG', True).split(':') for o in (d.getVar('OVERRIDES', True) + ':').split(':') if os.path.exists(os.path.join(fp, p, o))])}:${FILESDIR}" | 68 | #expected = "${@':'.join([os.path.normpath(os.path.join(fp, p, o)) for fp in d.getVar('FILESPATHBASE', True).split(':') for p in d.getVar('FILESPATHPKG', True).split(':') for o in (d.getVar('OVERRIDES', True) + ':').split(':') if os.path.exists(os.path.join(fp, p, o))])}:${FILESDIR}" |
69 | expectedpaths = bb.data.expand(expected, d) | 69 | expectedpaths = d.expand(expected) |
70 | unexpanded = d.getVar("FILESPATH", 0) | 70 | unexpanded = d.getVar("FILESPATH", 0) |
71 | filespath = d.getVar("FILESPATH", True).split(":") | 71 | filespath = d.getVar("FILESPATH", True).split(":") |
72 | filespath = [os.path.normpath(f) for f in filespath if os.path.exists(f)] | 72 | filespath = [os.path.normpath(f) for f in filespath if os.path.exists(f)] |
@@ -91,7 +91,7 @@ def can_remove_FILESDIR(cfgdata, d): | |||
91 | return unexpanded != expected and \ | 91 | return unexpanded != expected and \ |
92 | os.path.exists(expanded) and \ | 92 | os.path.exists(expanded) and \ |
93 | (expanded in filespath or | 93 | (expanded in filespath or |
94 | expanded == bb.data.expand(expected, d)) | 94 | expanded == d.expand(expected)) |
95 | 95 | ||
96 | def can_remove_others(p, cfgdata, d): | 96 | def can_remove_others(p, cfgdata, d): |
97 | for k in ["S", "PV", "PN", "DESCRIPTION", "LICENSE", "DEPENDS", | 97 | for k in ["S", "PV", "PN", "DESCRIPTION", "LICENSE", "DEPENDS", |
@@ -104,7 +104,7 @@ def can_remove_others(p, cfgdata, d): | |||
104 | 104 | ||
105 | try: | 105 | try: |
106 | expanded = d.getVar(k, True) | 106 | expanded = d.getVar(k, True) |
107 | cfgexpanded = bb.data.expand(cfgunexpanded, d) | 107 | cfgexpanded = d.expand(cfgunexpanded) |
108 | except bb.fetch.ParameterError: | 108 | except bb.fetch.ParameterError: |
109 | continue | 109 | continue |
110 | 110 | ||
diff --git a/meta/classes/relocatable.bbclass b/meta/classes/relocatable.bbclass index 54227a91ca..072f533f4f 100644 --- a/meta/classes/relocatable.bbclass +++ b/meta/classes/relocatable.bbclass | |||
@@ -7,9 +7,9 @@ def process_dir (directory, d): | |||
7 | import subprocess as sub | 7 | import subprocess as sub |
8 | import stat | 8 | import stat |
9 | 9 | ||
10 | cmd = bb.data.expand('${CHRPATH_BIN}', d) | 10 | cmd = d.expand('${CHRPATH_BIN}') |
11 | tmpdir = d.getVar('TMPDIR') | 11 | tmpdir = d.getVar('TMPDIR') |
12 | basedir = bb.data.expand('${base_prefix}', d) | 12 | basedir = d.expand('${base_prefix}') |
13 | 13 | ||
14 | #bb.debug("Checking %s for binaries to process" % directory) | 14 | #bb.debug("Checking %s for binaries to process" % directory) |
15 | if not os.path.exists(directory): | 15 | if not os.path.exists(directory): |
@@ -82,7 +82,7 @@ def process_dir (directory, d): | |||
82 | os.chmod(fpath, perms) | 82 | os.chmod(fpath, perms) |
83 | 83 | ||
84 | def rpath_replace (path, d): | 84 | def rpath_replace (path, d): |
85 | bindirs = bb.data.expand("${bindir} ${sbindir} ${base_sbindir} ${base_bindir} ${libdir} ${base_libdir} ${libexecdir} ${PREPROCESS_RELOCATE_DIRS}", d).split() | 85 | bindirs = d.expand("${bindir} ${sbindir} ${base_sbindir} ${base_bindir} ${libdir} ${base_libdir} ${libexecdir} ${PREPROCESS_RELOCATE_DIRS}").split() |
86 | 86 | ||
87 | for bindir in bindirs: | 87 | for bindir in bindirs: |
88 | #bb.note ("Processing directory " + bindir) | 88 | #bb.note ("Processing directory " + bindir) |
@@ -90,5 +90,5 @@ def rpath_replace (path, d): | |||
90 | process_dir (directory, d) | 90 | process_dir (directory, d) |
91 | 91 | ||
92 | python relocatable_binaries_preprocess() { | 92 | python relocatable_binaries_preprocess() { |
93 | rpath_replace(bb.data.expand('${SYSROOT_DESTDIR}', d), d) | 93 | rpath_replace(d.expand('${SYSROOT_DESTDIR}'), d) |
94 | } | 94 | } |
diff --git a/meta/classes/sanity.bbclass b/meta/classes/sanity.bbclass index d8835dae64..1b941ac58c 100644 --- a/meta/classes/sanity.bbclass +++ b/meta/classes/sanity.bbclass | |||
@@ -13,12 +13,12 @@ def raise_sanity_error(msg): | |||
13 | 13 | ||
14 | def check_conf_exists(fn, data): | 14 | def check_conf_exists(fn, data): |
15 | bbpath = [] | 15 | bbpath = [] |
16 | fn = bb.data.expand(fn, data) | 16 | fn = data.expand(fn) |
17 | vbbpath = data.getVar("BBPATH") | 17 | vbbpath = data.getVar("BBPATH") |
18 | if vbbpath: | 18 | if vbbpath: |
19 | bbpath += vbbpath.split(":") | 19 | bbpath += vbbpath.split(":") |
20 | for p in bbpath: | 20 | for p in bbpath: |
21 | currname = os.path.join(bb.data.expand(p, data), fn) | 21 | currname = os.path.join(data.expand(p), fn) |
22 | if os.access(currname, os.R_OK): | 22 | if os.access(currname, os.R_OK): |
23 | return True | 23 | return True |
24 | return False | 24 | return False |
@@ -411,16 +411,16 @@ def check_sanity(e): | |||
411 | f.write(current_abi) | 411 | f.write(current_abi) |
412 | elif abi == "2" and current_abi == "3": | 412 | elif abi == "2" and current_abi == "3": |
413 | bb.note("Converting staging from layout version 2 to layout version 3") | 413 | bb.note("Converting staging from layout version 2 to layout version 3") |
414 | os.system(bb.data.expand("mv ${TMPDIR}/staging ${TMPDIR}/sysroots", e.data)) | 414 | os.system(e.data.expand("mv ${TMPDIR}/staging ${TMPDIR}/sysroots")) |
415 | os.system(bb.data.expand("ln -s sysroots ${TMPDIR}/staging", e.data)) | 415 | os.system(e.data.expand("ln -s sysroots ${TMPDIR}/staging")) |
416 | os.system(bb.data.expand("cd ${TMPDIR}/stamps; for i in */*do_populate_staging; do new=`echo $i | sed -e 's/do_populate_staging/do_populate_sysroot/'`; mv $i $new; done", e.data)) | 416 | os.system(e.data.expand("cd ${TMPDIR}/stamps; for i in */*do_populate_staging; do new=`echo $i | sed -e 's/do_populate_staging/do_populate_sysroot/'`; mv $i $new; done")) |
417 | f = file(abifile, "w") | 417 | f = file(abifile, "w") |
418 | f.write(current_abi) | 418 | f.write(current_abi) |
419 | elif abi == "3" and current_abi == "4": | 419 | elif abi == "3" and current_abi == "4": |
420 | bb.note("Converting staging layout from version 3 to layout version 4") | 420 | bb.note("Converting staging layout from version 3 to layout version 4") |
421 | if os.path.exists(bb.data.expand("${STAGING_DIR_NATIVE}${bindir_native}/${MULTIMACH_HOST_SYS}", e.data)): | 421 | if os.path.exists(e.data.expand("${STAGING_DIR_NATIVE}${bindir_native}/${MULTIMACH_HOST_SYS}")): |
422 | os.system(bb.data.expand("mv ${STAGING_DIR_NATIVE}${bindir_native}/${MULTIMACH_HOST_SYS} ${STAGING_BINDIR_CROSS}", e.data)) | 422 | os.system(e.data.expand("mv ${STAGING_DIR_NATIVE}${bindir_native}/${MULTIMACH_HOST_SYS} ${STAGING_BINDIR_CROSS}")) |
423 | os.system(bb.data.expand("ln -s ${STAGING_BINDIR_CROSS} ${STAGING_DIR_NATIVE}${bindir_native}/${MULTIMACH_HOST_SYS}", e.data)) | 423 | os.system(e.data.expand("ln -s ${STAGING_BINDIR_CROSS} ${STAGING_DIR_NATIVE}${bindir_native}/${MULTIMACH_HOST_SYS}")) |
424 | 424 | ||
425 | f = file(abifile, "w") | 425 | f = file(abifile, "w") |
426 | f.write(current_abi) | 426 | f.write(current_abi) |
@@ -428,7 +428,7 @@ def check_sanity(e): | |||
428 | messages = messages + "Staging layout has changed. The cross directory has been deprecated and cross packages are now built under the native sysroot.\nThis requires a rebuild.\n" | 428 | messages = messages + "Staging layout has changed. The cross directory has been deprecated and cross packages are now built under the native sysroot.\nThis requires a rebuild.\n" |
429 | elif abi == "5" and current_abi == "6": | 429 | elif abi == "5" and current_abi == "6": |
430 | bb.note("Converting staging layout from version 5 to layout version 6") | 430 | bb.note("Converting staging layout from version 5 to layout version 6") |
431 | os.system(bb.data.expand("mv ${TMPDIR}/pstagelogs ${SSTATE_MANIFESTS}", e.data)) | 431 | os.system(e.data.expand("mv ${TMPDIR}/pstagelogs ${SSTATE_MANIFESTS}")) |
432 | f = file(abifile, "w") | 432 | f = file(abifile, "w") |
433 | f.write(current_abi) | 433 | f.write(current_abi) |
434 | elif abi == "7" and current_abi == "8": | 434 | elif abi == "7" and current_abi == "8": |
diff --git a/meta/classes/sourcepkg.bbclass b/meta/classes/sourcepkg.bbclass index 2a78a90452..102c109324 100644 --- a/meta/classes/sourcepkg.bbclass +++ b/meta/classes/sourcepkg.bbclass | |||
@@ -17,7 +17,7 @@ def get_src_tree(d): | |||
17 | return | 17 | return |
18 | 18 | ||
19 | s_tree_raw = s.split('/')[1] | 19 | s_tree_raw = s.split('/')[1] |
20 | s_tree = bb.data.expand(s_tree_raw, d) | 20 | s_tree = d.expand(s_tree_raw) |
21 | 21 | ||
22 | src_tree_path = os.path.join(workdir, s_tree) | 22 | src_tree_path = os.path.join(workdir, s_tree) |
23 | try: | 23 | try: |
@@ -59,7 +59,7 @@ python sourcepkg_do_dumpdata() { | |||
59 | distro = d.getVar('DISTRO', True) | 59 | distro = d.getVar('DISTRO', True) |
60 | s_tree = get_src_tree(d) | 60 | s_tree = get_src_tree(d) |
61 | openembeddeddir = os.path.join(workdir, s_tree, distro) | 61 | openembeddeddir = os.path.join(workdir, s_tree, distro) |
62 | dumpfile = os.path.join(openembeddeddir, bb.data.expand("${P}-${PR}.showdata.dump",d)) | 62 | dumpfile = os.path.join(openembeddeddir, d.expand("${P}-${PR}.showdata.dump")) |
63 | 63 | ||
64 | try: | 64 | try: |
65 | os.mkdir(openembeddeddir) | 65 | os.mkdir(openembeddeddir) |
diff --git a/meta/classes/sstate.bbclass b/meta/classes/sstate.bbclass index e4338e0f78..0d16d118d9 100644 --- a/meta/classes/sstate.bbclass +++ b/meta/classes/sstate.bbclass | |||
@@ -24,16 +24,16 @@ python () { | |||
24 | if bb.data.inherits_class('native', d): | 24 | if bb.data.inherits_class('native', d): |
25 | d.setVar('SSTATE_PKGARCH', d.getVar('BUILD_ARCH')) | 25 | d.setVar('SSTATE_PKGARCH', d.getVar('BUILD_ARCH')) |
26 | elif bb.data.inherits_class('cross', d): | 26 | elif bb.data.inherits_class('cross', d): |
27 | d.setVar('SSTATE_PKGARCH', bb.data.expand("${BUILD_ARCH}_${TUNE_PKGARCH}", d)) | 27 | d.setVar('SSTATE_PKGARCH', d.expand("${BUILD_ARCH}_${TUNE_PKGARCH}")) |
28 | d.setVar('SSTATE_MANMACH', bb.data.expand("${BUILD_ARCH}_${MACHINE}", d)) | 28 | d.setVar('SSTATE_MANMACH', d.expand("${BUILD_ARCH}_${MACHINE}")) |
29 | elif bb.data.inherits_class('crosssdk', d): | 29 | elif bb.data.inherits_class('crosssdk', d): |
30 | d.setVar('SSTATE_PKGARCH', bb.data.expand("${BUILD_ARCH}_${PACKAGE_ARCH}", d)) | 30 | d.setVar('SSTATE_PKGARCH', d.expand("${BUILD_ARCH}_${PACKAGE_ARCH}")) |
31 | elif bb.data.inherits_class('nativesdk', d): | 31 | elif bb.data.inherits_class('nativesdk', d): |
32 | d.setVar('SSTATE_PKGARCH', bb.data.expand("${SDK_ARCH}", d)) | 32 | d.setVar('SSTATE_PKGARCH', d.expand("${SDK_ARCH}")) |
33 | elif bb.data.inherits_class('cross-canadian', d): | 33 | elif bb.data.inherits_class('cross-canadian', d): |
34 | d.setVar('SSTATE_PKGARCH', bb.data.expand("${SDK_ARCH}_${PACKAGE_ARCH}", d)) | 34 | d.setVar('SSTATE_PKGARCH', d.expand("${SDK_ARCH}_${PACKAGE_ARCH}")) |
35 | else: | 35 | else: |
36 | d.setVar('SSTATE_MANMACH', bb.data.expand("${MACHINE}", d)) | 36 | d.setVar('SSTATE_MANMACH', d.expand("${MACHINE}")) |
37 | 37 | ||
38 | # These classes encode staging paths into their scripts data so can only be | 38 | # These classes encode staging paths into their scripts data so can only be |
39 | # reused if we manipulate the paths | 39 | # reused if we manipulate the paths |
@@ -97,8 +97,8 @@ def sstate_install(ss, d): | |||
97 | 97 | ||
98 | sharedfiles = [] | 98 | sharedfiles = [] |
99 | shareddirs = [] | 99 | shareddirs = [] |
100 | bb.mkdirhier(bb.data.expand("${SSTATE_MANIFESTS}", d)) | 100 | bb.mkdirhier(d.expand("${SSTATE_MANIFESTS}")) |
101 | manifest = bb.data.expand("${SSTATE_MANFILEPREFIX}.%s" % ss['name'], d) | 101 | manifest = d.expand("${SSTATE_MANFILEPREFIX}.%s" % ss['name']) |
102 | 102 | ||
103 | if os.access(manifest, os.R_OK): | 103 | if os.access(manifest, os.R_OK): |
104 | bb.fatal("Package already staged (%s)?!" % manifest) | 104 | bb.fatal("Package already staged (%s)?!" % manifest) |
@@ -153,7 +153,7 @@ def sstate_installpkg(ss, d): | |||
153 | bb.mkdirhier(dir) | 153 | bb.mkdirhier(dir) |
154 | oe.path.remove(dir) | 154 | oe.path.remove(dir) |
155 | 155 | ||
156 | sstateinst = bb.data.expand("${WORKDIR}/sstate-install-%s/" % ss['name'], d) | 156 | sstateinst = d.expand("${WORKDIR}/sstate-install-%s/" % ss['name']) |
157 | sstatepkg = d.getVar('SSTATE_PKG', True) + '_' + ss['name'] + ".tgz" | 157 | sstatepkg = d.getVar('SSTATE_PKG', True) + '_' + ss['name'] + ".tgz" |
158 | 158 | ||
159 | if not os.path.exists(sstatepkg): | 159 | if not os.path.exists(sstatepkg): |
@@ -246,7 +246,7 @@ def sstate_clean_manifest(manifest, d): | |||
246 | def sstate_clean(ss, d): | 246 | def sstate_clean(ss, d): |
247 | import oe.path | 247 | import oe.path |
248 | 248 | ||
249 | manifest = bb.data.expand("${SSTATE_MANFILEPREFIX}.%s" % ss['name'], d) | 249 | manifest = d.expand("${SSTATE_MANFILEPREFIX}.%s" % ss['name']) |
250 | 250 | ||
251 | if os.path.exists(manifest): | 251 | if os.path.exists(manifest): |
252 | locks = [] | 252 | locks = [] |
@@ -351,7 +351,7 @@ def sstate_package(ss, d): | |||
351 | 351 | ||
352 | tmpdir = d.getVar('TMPDIR', True) | 352 | tmpdir = d.getVar('TMPDIR', True) |
353 | 353 | ||
354 | sstatebuild = bb.data.expand("${WORKDIR}/sstate-build-%s/" % ss['name'], d) | 354 | sstatebuild = d.expand("${WORKDIR}/sstate-build-%s/" % ss['name']) |
355 | sstatepkg = d.getVar('SSTATE_PKG', True) + '_'+ ss['name'] + ".tgz" | 355 | sstatepkg = d.getVar('SSTATE_PKG', True) + '_'+ ss['name'] + ".tgz" |
356 | bb.mkdirhier(sstatebuild) | 356 | bb.mkdirhier(sstatebuild) |
357 | bb.mkdirhier(os.path.dirname(sstatepkg)) | 357 | bb.mkdirhier(os.path.dirname(sstatepkg)) |
@@ -397,7 +397,7 @@ def pstaging_fetch(sstatepkg, d): | |||
397 | localdata = bb.data.createCopy(d) | 397 | localdata = bb.data.createCopy(d) |
398 | bb.data.update_data(localdata) | 398 | bb.data.update_data(localdata) |
399 | 399 | ||
400 | dldir = bb.data.expand("${SSTATE_DIR}", localdata) | 400 | dldir = localdata.expand("${SSTATE_DIR}") |
401 | srcuri = "file://" + os.path.basename(sstatepkg) | 401 | srcuri = "file://" + os.path.basename(sstatepkg) |
402 | 402 | ||
403 | bb.mkdirhier(dldir) | 403 | bb.mkdirhier(dldir) |
@@ -484,7 +484,7 @@ def sstate_checkhashes(sq_fn, sq_task, sq_hash, sq_hashfn, d): | |||
484 | } | 484 | } |
485 | 485 | ||
486 | for task in range(len(sq_fn)): | 486 | for task in range(len(sq_fn)): |
487 | sstatefile = bb.data.expand("${SSTATE_DIR}/" + sq_hashfn[task] + "_" + mapping[sq_task[task]] + ".tgz", d) | 487 | sstatefile = d.expand("${SSTATE_DIR}/" + sq_hashfn[task] + "_" + mapping[sq_task[task]] + ".tgz") |
488 | sstatefile = sstatefile.replace("${BB_TASKHASH}", sq_hash[task]) | 488 | sstatefile = sstatefile.replace("${BB_TASKHASH}", sq_hash[task]) |
489 | if os.path.exists(sstatefile): | 489 | if os.path.exists(sstatefile): |
490 | bb.debug(2, "SState: Found valid sstate file %s" % sstatefile) | 490 | bb.debug(2, "SState: Found valid sstate file %s" % sstatefile) |
@@ -499,7 +499,7 @@ def sstate_checkhashes(sq_fn, sq_task, sq_hash, sq_hashfn, d): | |||
499 | localdata = bb.data.createCopy(d) | 499 | localdata = bb.data.createCopy(d) |
500 | bb.data.update_data(localdata) | 500 | bb.data.update_data(localdata) |
501 | 501 | ||
502 | dldir = bb.data.expand("${SSTATE_DIR}", localdata) | 502 | dldir = localdata.expand("${SSTATE_DIR}") |
503 | localdata.setVar('DL_DIR', dldir) | 503 | localdata.setVar('DL_DIR', dldir) |
504 | localdata.setVar('PREMIRRORS', mirrors) | 504 | localdata.setVar('PREMIRRORS', mirrors) |
505 | 505 | ||
@@ -509,7 +509,7 @@ def sstate_checkhashes(sq_fn, sq_task, sq_hash, sq_hashfn, d): | |||
509 | if task in ret: | 509 | if task in ret: |
510 | continue | 510 | continue |
511 | 511 | ||
512 | sstatefile = bb.data.expand("${SSTATE_DIR}/" + sq_hashfn[task] + "_" + mapping[sq_task[task]] + ".tgz", d) | 512 | sstatefile = d.expand("${SSTATE_DIR}/" + sq_hashfn[task] + "_" + mapping[sq_task[task]] + ".tgz") |
513 | sstatefile = sstatefile.replace("${BB_TASKHASH}", sq_hash[task]) | 513 | sstatefile = sstatefile.replace("${BB_TASKHASH}", sq_hash[task]) |
514 | 514 | ||
515 | srcuri = "file://" + os.path.basename(sstatefile) | 515 | srcuri = "file://" + os.path.basename(sstatefile) |
diff --git a/meta/classes/syslinux.bbclass b/meta/classes/syslinux.bbclass index b194fa69a3..700ea53911 100644 --- a/meta/classes/syslinux.bbclass +++ b/meta/classes/syslinux.bbclass | |||
@@ -95,7 +95,7 @@ python build_syslinux_menu () { | |||
95 | overrides = localdata.getVar('OVERRIDES') | 95 | overrides = localdata.getVar('OVERRIDES') |
96 | if not overrides: | 96 | if not overrides: |
97 | raise bb.build.FuncFailed('OVERRIDES not defined') | 97 | raise bb.build.FuncFailed('OVERRIDES not defined') |
98 | overrides = bb.data.expand(overrides, localdata) | 98 | overrides = localdata.expand(overrides) |
99 | 99 | ||
100 | localdata.setVar('OVERRIDES', label + ':' + overrides) | 100 | localdata.setVar('OVERRIDES', label + ':' + overrides) |
101 | bb.data.update_data(localdata) | 101 | bb.data.update_data(localdata) |
diff --git a/meta/classes/utility-tasks.bbclass b/meta/classes/utility-tasks.bbclass index bbdf6e159b..cbb000a1e3 100644 --- a/meta/classes/utility-tasks.bbclass +++ b/meta/classes/utility-tasks.bbclass | |||
@@ -16,7 +16,7 @@ addtask clean | |||
16 | do_clean[nostamp] = "1" | 16 | do_clean[nostamp] = "1" |
17 | python do_clean() { | 17 | python do_clean() { |
18 | """clear the build and temp directories""" | 18 | """clear the build and temp directories""" |
19 | dir = bb.data.expand("${WORKDIR}", d) | 19 | dir = d.expand("${WORKDIR}") |
20 | bb.note("Removing " + dir) | 20 | bb.note("Removing " + dir) |
21 | oe.path.remove(dir) | 21 | oe.path.remove(dir) |
22 | 22 | ||