summaryrefslogtreecommitdiff
path: root/meta/classes
diff options
context:
space:
mode:
Diffstat (limited to 'meta/classes')
-rw-r--r--meta/classes/autotools.bbclass6
-rw-r--r--meta/classes/base.bbclass114
-rw-r--r--meta/classes/bugzilla.bbclass30
-rw-r--r--meta/classes/buildstats.bbclass36
-rw-r--r--meta/classes/cpan-base.bbclass2
-rw-r--r--meta/classes/cpan_build.bbclass4
-rw-r--r--meta/classes/cross-canadian.bbclass2
-rw-r--r--meta/classes/debian.bbclass18
-rw-r--r--meta/classes/distrodata.bbclass162
-rw-r--r--meta/classes/distutils-base.bbclass2
-rw-r--r--meta/classes/distutils-native-base.bbclass2
-rw-r--r--meta/classes/gconf.bbclass16
-rw-r--r--meta/classes/gtk-icon-cache.bbclass24
-rw-r--r--meta/classes/icecc.bbclass16
-rw-r--r--meta/classes/image-swab.bbclass4
-rw-r--r--meta/classes/image.bbclass20
-rw-r--r--meta/classes/image_types.bbclass4
-rw-r--r--meta/classes/imagetest-qemu.bbclass56
-rw-r--r--meta/classes/insane.bbclass78
-rw-r--r--meta/classes/kernel-arch.bbclass6
-rw-r--r--meta/classes/kernel-yocto.bbclass2
-rw-r--r--meta/classes/kernel.bbclass68
-rw-r--r--meta/classes/libc-common.bbclass10
-rw-r--r--meta/classes/libc-package.bbclass66
-rw-r--r--meta/classes/license.bbclass18
-rw-r--r--meta/classes/metadata_scm.bbclass2
-rw-r--r--meta/classes/native.bbclass12
-rw-r--r--meta/classes/nativesdk.bbclass14
-rw-r--r--meta/classes/package.bbclass232
-rw-r--r--meta/classes/package_deb.bbclass82
-rw-r--r--meta/classes/package_ipk.bbclass82
-rw-r--r--meta/classes/package_rpm.bbclass136
-rw-r--r--meta/classes/package_tar.bbclass38
-rw-r--r--meta/classes/packagedata.bbclass8
-rw-r--r--meta/classes/packagehistory.bbclass8
-rw-r--r--meta/classes/patch.bbclass28
-rw-r--r--meta/classes/pkg_distribute.bbclass2
-rw-r--r--meta/classes/pkg_metainfo.bbclass12
-rw-r--r--meta/classes/populate_sdk_deb.bbclass4
-rw-r--r--meta/classes/populate_sdk_rpm.bbclass2
-rw-r--r--meta/classes/qemu.bbclass2
-rw-r--r--meta/classes/qt4e.bbclass2
-rw-r--r--meta/classes/qt4x11.bbclass2
-rw-r--r--meta/classes/relocatable.bbclass2
-rw-r--r--meta/classes/rootfs_ipk.bbclass10
-rw-r--r--meta/classes/rootfs_rpm.bbclass14
-rw-r--r--meta/classes/sanity.bbclass10
-rw-r--r--meta/classes/siteconfig.bbclass2
-rw-r--r--meta/classes/siteinfo.bbclass2
-rw-r--r--meta/classes/sourcepkg.bbclass12
-rw-r--r--meta/classes/src_distribute.bbclass16
-rw-r--r--meta/classes/sstate.bbclass94
-rw-r--r--meta/classes/staging.bbclass6
-rw-r--r--meta/classes/syslinux.bbclass34
-rw-r--r--meta/classes/task.bbclass2
-rw-r--r--meta/classes/toolchain-scripts.bbclass6
-rw-r--r--meta/classes/update-alternatives.bbclass38
-rw-r--r--meta/classes/update-rc.d.bbclass36
-rw-r--r--meta/classes/useradd.bbclass8
-rw-r--r--meta/classes/utility-tasks.bbclass8
-rw-r--r--meta/classes/utils.bbclass4
61 files changed, 869 insertions, 869 deletions
diff --git a/meta/classes/autotools.bbclass b/meta/classes/autotools.bbclass
index 451c7fcc1..37e7d4b48 100644
--- a/meta/classes/autotools.bbclass
+++ b/meta/classes/autotools.bbclass
@@ -1,8 +1,8 @@
def autotools_dep_prepend(d):
- if bb.data.getVar('INHIBIT_AUTOTOOLS_DEPS', d, 1):
+ if d.getVar('INHIBIT_AUTOTOOLS_DEPS', 1):
return ''
- pn = bb.data.getVar('PN', d, 1)
+ pn = d.getVar('PN', 1)
deps = ''
if pn in ['autoconf-native', 'automake-native', 'help2man-native']:
@@ -13,7 +13,7 @@ def autotools_dep_prepend(d):
deps += 'libtool-native '
if not bb.data.inherits_class('native', d) \
and not bb.data.inherits_class('cross', d) \
- and not bb.data.getVar('INHIBIT_DEFAULT_DEPS', d, 1):
+ and not d.getVar('INHIBIT_DEFAULT_DEPS', 1):
deps += 'libtool-cross '
return deps + 'gnu-config-native '
diff --git a/meta/classes/base.bbclass b/meta/classes/base.bbclass
index f5397446d..f3f798f9b 100644
--- a/meta/classes/base.bbclass
+++ b/meta/classes/base.bbclass
@@ -60,9 +60,9 @@ def base_dep_prepend(d):
# INHIBIT_DEFAULT_DEPS doesn't apply to the patch command. Whether or not
# we need that built is the responsibility of the patch function / class, not
# the application.
- if not bb.data.getVar('INHIBIT_DEFAULT_DEPS', d):
- if (bb.data.getVar('HOST_SYS', d, 1) !=
- bb.data.getVar('BUILD_SYS', d, 1)):
+ if not d.getVar('INHIBIT_DEFAULT_DEPS'):
+ if (d.getVar('HOST_SYS', 1) !=
+ d.getVar('BUILD_SYS', 1)):
deps += " virtual/${TARGET_PREFIX}gcc virtual/${TARGET_PREFIX}compilerlibs virtual/libc "
return deps
@@ -73,13 +73,13 @@ DEPENDS_prepend="${BASEDEPENDS} "
FILESPATH = "${@base_set_filespath([ "${FILE_DIRNAME}/${PF}", "${FILE_DIRNAME}/${P}", "${FILE_DIRNAME}/${PN}", "${FILE_DIRNAME}/${BP}", "${FILE_DIRNAME}/${BPN}", "${FILE_DIRNAME}/files", "${FILE_DIRNAME}" ], d)}"
# THISDIR only works properly with imediate expansion as it has to run
# in the context of the location its used (:=)
-THISDIR = "${@os.path.dirname(bb.data.getVar('FILE', d, True))}"
+THISDIR = "${@os.path.dirname(d.getVar('FILE', True))}"
addtask fetch
do_fetch[dirs] = "${DL_DIR}"
python base_do_fetch() {
- src_uri = (bb.data.getVar('SRC_URI', d, True) or "").split()
+ src_uri = (d.getVar('SRC_URI', True) or "").split()
if len(src_uri) == 0:
return
@@ -96,14 +96,14 @@ python base_do_fetch() {
addtask unpack after do_fetch
do_unpack[dirs] = "${WORKDIR}"
python base_do_unpack() {
- src_uri = (bb.data.getVar('SRC_URI', d, True) or "").split()
+ src_uri = (d.getVar('SRC_URI', True) or "").split()
if len(src_uri) == 0:
return
localdata = bb.data.createCopy(d)
bb.data.update_data(localdata)
- rootdir = bb.data.getVar('WORKDIR', localdata, True)
+ rootdir = localdata.getVar('WORKDIR', True)
try:
fetcher = bb.fetch2.Fetch(src_uri, localdata)
@@ -118,7 +118,7 @@ def generate_git_config(e):
from bb import data
if data.getVar('GIT_CORE_CONFIG', e.data, True):
- gitconfig_path = bb.data.getVar('GIT_CONFIG', e.data, True)
+ gitconfig_path = e.data.getVar('GIT_CONFIG', True)
proxy_command = " gitproxy = %s\n" % data.getVar('GIT_PROXY_COMMAND', e.data, True)
bb.mkdirhier(bb.data.expand("${STAGING_DIR_NATIVE}/usr/etc/", e.data))
@@ -207,11 +207,11 @@ python base_eventhandler() {
name = getName(e)
if name.startswith("BuildStarted"):
- bb.data.setVar( 'BB_VERSION', bb.__version__, e.data )
+ e.data.setVar( 'BB_VERSION', bb.__version__)
statusvars = ['BB_VERSION', 'TARGET_ARCH', 'TARGET_OS', 'MACHINE', 'DISTRO', 'DISTRO_VERSION','TUNE_FEATURES', 'TARGET_FPU']
- statuslines = ["%-17s = \"%s\"" % (i, bb.data.getVar(i, e.data, 1) or '') for i in statusvars]
+ statuslines = ["%-17s = \"%s\"" % (i, e.data.getVar(i, 1) or '') for i in statusvars]
- layers = (bb.data.getVar("BBLAYERS", e.data, 1) or "").split()
+ layers = (e.data.getVar("BBLAYERS", 1) or "").split()
layers_branch_rev = ["%-17s = \"%s:%s\"" % (os.path.basename(i), \
base_get_metadata_git_branch(i, None).strip(), \
base_get_metadata_git_revision(i, None)) \
@@ -237,7 +237,7 @@ python base_eventhandler() {
needed_vars = [ "TARGET_ARCH", "TARGET_OS" ]
pesteruser = []
for v in needed_vars:
- val = bb.data.getVar(v, e.data, 1)
+ val = e.data.getVar(v, 1)
if not val or val == 'INVALID':
pesteruser.append(v)
if pesteruser:
@@ -330,23 +330,23 @@ python () {
appendVar('EXTRA_OECONF', extraconf)
# If PRINC is set, try and increase the PR value by the amount specified
- princ = bb.data.getVar('PRINC', d, True)
+ princ = d.getVar('PRINC', True)
if princ:
- pr = bb.data.getVar('PR', d, True)
+ pr = d.getVar('PR', True)
pr_prefix = re.search("\D+",pr)
prval = re.search("\d+",pr)
if pr_prefix is None or prval is None:
bb.error("Unable to analyse format of PR variable: %s" % pr)
nval = int(prval.group(0)) + int(princ)
pr = pr_prefix.group(0) + str(nval) + pr[prval.end():]
- bb.data.setVar('PR', pr, d)
+ d.setVar('PR', pr)
- pn = bb.data.getVar('PN', d, 1)
- license = bb.data.getVar('LICENSE', d, True)
+ pn = d.getVar('PN', 1)
+ license = d.getVar('LICENSE', True)
if license == "INVALID":
bb.fatal('This recipe does not have the LICENSE field set (%s)' % pn)
- commercial_license = " %s " % bb.data.getVar('COMMERCIAL_LICENSE', d, 1)
+ commercial_license = " %s " % d.getVar('COMMERCIAL_LICENSE', 1)
import re
pnr = "[ \t]%s[ \t]" % pn.replace('+', "\+")
if commercial_license and re.search(pnr, commercial_license):
@@ -356,86 +356,86 @@ python () {
# If we're building a target package we need to use fakeroot (pseudo)
# in order to capture permissions, owners, groups and special files
if not bb.data.inherits_class('native', d) and not bb.data.inherits_class('cross', d):
- bb.data.setVarFlag('do_configure', 'umask', 022, d)
- bb.data.setVarFlag('do_compile', 'umask', 022, d)
- deps = (bb.data.getVarFlag('do_install', 'depends', d) or "").split()
+ d.setVarFlag('do_configure', 'umask', 022)
+ d.setVarFlag('do_compile', 'umask', 022)
+ deps = (d.getVarFlag('do_install', 'depends') or "").split()
deps.append('virtual/fakeroot-native:do_populate_sysroot')
bb.data.setVarFlag('do_install', 'depends', " ".join(deps),d)
- bb.data.setVarFlag('do_install', 'fakeroot', 1, d)
- bb.data.setVarFlag('do_install', 'umask', 022, d)
- deps = (bb.data.getVarFlag('do_package', 'depends', d) or "").split()
+ d.setVarFlag('do_install', 'fakeroot', 1)
+ d.setVarFlag('do_install', 'umask', 022)
+ deps = (d.getVarFlag('do_package', 'depends') or "").split()
deps.append('virtual/fakeroot-native:do_populate_sysroot')
bb.data.setVarFlag('do_package', 'depends', " ".join(deps),d)
- bb.data.setVarFlag('do_package', 'fakeroot', 1, d)
- bb.data.setVarFlag('do_package', 'umask', 022, d)
- bb.data.setVarFlag('do_package_setscene', 'fakeroot', 1, d)
- source_mirror_fetch = bb.data.getVar('SOURCE_MIRROR_FETCH', d, 0)
+ d.setVarFlag('do_package', 'fakeroot', 1)
+ d.setVarFlag('do_package', 'umask', 022)
+ d.setVarFlag('do_package_setscene', 'fakeroot', 1)
+ source_mirror_fetch = d.getVar('SOURCE_MIRROR_FETCH', 0)
if not source_mirror_fetch:
- need_host = bb.data.getVar('COMPATIBLE_HOST', d, 1)
+ need_host = d.getVar('COMPATIBLE_HOST', 1)
if need_host:
import re
- this_host = bb.data.getVar('HOST_SYS', d, 1)
+ this_host = d.getVar('HOST_SYS', 1)
if not re.match(need_host, this_host):
raise bb.parse.SkipPackage("incompatible with host %s" % this_host)
- need_machine = bb.data.getVar('COMPATIBLE_MACHINE', d, 1)
+ need_machine = d.getVar('COMPATIBLE_MACHINE', 1)
if need_machine:
import re
- this_machine = bb.data.getVar('MACHINE', d, 1)
+ this_machine = d.getVar('MACHINE', 1)
if this_machine and not re.match(need_machine, this_machine):
- this_soc_family = bb.data.getVar('SOC_FAMILY', d, 1)
+ this_soc_family = d.getVar('SOC_FAMILY', 1)
if (this_soc_family and not re.match(need_machine, this_soc_family)) or not this_soc_family:
raise bb.parse.SkipPackage("incompatible with machine %s" % this_machine)
- dont_want_license = bb.data.getVar('INCOMPATIBLE_LICENSE', d, 1)
+ dont_want_license = d.getVar('INCOMPATIBLE_LICENSE', 1)
if dont_want_license and not pn.endswith("-native") and not pn.endswith("-cross") and not pn.endswith("-cross-initial") and not pn.endswith("-cross-intermediate") and not pn.endswith("-crosssdk-intermediate") and not pn.endswith("-crosssdk") and not pn.endswith("-crosssdk-initial"):
- hosttools_whitelist = (bb.data.getVar('HOSTTOOLS_WHITELIST_%s' % dont_want_license, d, 1) or "").split()
- lgplv2_whitelist = (bb.data.getVar('LGPLv2_WHITELIST_%s' % dont_want_license, d, 1) or "").split()
- dont_want_whitelist = (bb.data.getVar('WHITELIST_%s' % dont_want_license, d, 1) or "").split()
+ hosttools_whitelist = (d.getVar('HOSTTOOLS_WHITELIST_%s' % dont_want_license, 1) or "").split()
+ lgplv2_whitelist = (d.getVar('LGPLv2_WHITELIST_%s' % dont_want_license, 1) or "").split()
+ dont_want_whitelist = (d.getVar('WHITELIST_%s' % dont_want_license, 1) or "").split()
if pn not in hosttools_whitelist and pn not in lgplv2_whitelist and pn not in dont_want_whitelist:
import re
- this_license = bb.data.getVar('LICENSE', d, 1)
+ this_license = d.getVar('LICENSE', 1)
if this_license and re.search(dont_want_license, this_license):
bb.note("SKIPPING %s because it's %s" % (pn, this_license))
raise bb.parse.SkipPackage("incompatible with license %s" % this_license)
# Git packages should DEPEND on git-native
- srcuri = bb.data.getVar('SRC_URI', d, 1)
+ srcuri = d.getVar('SRC_URI', 1)
if "git://" in srcuri:
- depends = bb.data.getVarFlag('do_fetch', 'depends', d) or ""
+ depends = d.getVarFlag('do_fetch', 'depends') or ""
depends = depends + " git-native:do_populate_sysroot"
- bb.data.setVarFlag('do_fetch', 'depends', depends, d)
+ d.setVarFlag('do_fetch', 'depends', depends)
# Mercurial packages should DEPEND on mercurial-native
elif "hg://" in srcuri:
- depends = bb.data.getVarFlag('do_fetch', 'depends', d) or ""
+ depends = d.getVarFlag('do_fetch', 'depends') or ""
depends = depends + " mercurial-native:do_populate_sysroot"
- bb.data.setVarFlag('do_fetch', 'depends', depends, d)
+ d.setVarFlag('do_fetch', 'depends', depends)
# OSC packages should DEPEND on osc-native
elif "osc://" in srcuri:
- depends = bb.data.getVarFlag('do_fetch', 'depends', d) or ""
+ depends = d.getVarFlag('do_fetch', 'depends') or ""
depends = depends + " osc-native:do_populate_sysroot"
- bb.data.setVarFlag('do_fetch', 'depends', depends, d)
+ d.setVarFlag('do_fetch', 'depends', depends)
# *.xz should depends on xz-native for unpacking
# Not endswith because of "*.patch.xz;patch=1". Need bb.decodeurl in future
if '.xz' in srcuri:
- depends = bb.data.getVarFlag('do_unpack', 'depends', d) or ""
+ depends = d.getVarFlag('do_unpack', 'depends') or ""
depends = depends + " xz-native:do_populate_sysroot"
- bb.data.setVarFlag('do_unpack', 'depends', depends, d)
+ d.setVarFlag('do_unpack', 'depends', depends)
# unzip-native should already be staged before unpacking ZIP recipes
if ".zip" in srcuri:
- depends = bb.data.getVarFlag('do_unpack', 'depends', d) or ""
+ depends = d.getVarFlag('do_unpack', 'depends') or ""
depends = depends + " unzip-native:do_populate_sysroot"
- bb.data.setVarFlag('do_unpack', 'depends', depends, d)
+ d.setVarFlag('do_unpack', 'depends', depends)
# 'multimachine' handling
- mach_arch = bb.data.getVar('MACHINE_ARCH', d, 1)
- pkg_arch = bb.data.getVar('PACKAGE_ARCH', d, 1)
+ mach_arch = d.getVar('MACHINE_ARCH', 1)
+ pkg_arch = d.getVar('PACKAGE_ARCH', 1)
if (pkg_arch == mach_arch):
# Already machine specific - nothing further to do
@@ -445,7 +445,7 @@ python () {
# We always try to scan SRC_URI for urls with machine overrides
# unless the package sets SRC_URI_OVERRIDES_PACKAGE_ARCH=0
#
- override = bb.data.getVar('SRC_URI_OVERRIDES_PACKAGE_ARCH', d, 1)
+ override = d.getVar('SRC_URI_OVERRIDES_PACKAGE_ARCH', 1)
if override != '0':
paths = []
for p in [ "${PF}", "${P}", "${PN}", "files", "" ]:
@@ -461,18 +461,18 @@ python () {
for mp in paths:
if local.startswith(mp):
#bb.note("overriding PACKAGE_ARCH from %s to %s" % (pkg_arch, mach_arch))
- bb.data.setVar('PACKAGE_ARCH', "${MACHINE_ARCH}", d)
+ d.setVar('PACKAGE_ARCH', "${MACHINE_ARCH}")
return
- packages = bb.data.getVar('PACKAGES', d, 1).split()
+ packages = d.getVar('PACKAGES', 1).split()
for pkg in packages:
- pkgarch = bb.data.getVar("PACKAGE_ARCH_%s" % pkg, d, 1)
+ pkgarch = d.getVar("PACKAGE_ARCH_%s" % pkg, 1)
# We could look for != PACKAGE_ARCH here but how to choose
# if multiple differences are present?
# Look through PACKAGE_ARCHS for the priority order?
if pkgarch and pkgarch == mach_arch:
- bb.data.setVar('PACKAGE_ARCH', "${MACHINE_ARCH}", d)
+ d.setVar('PACKAGE_ARCH', "${MACHINE_ARCH}")
bb.warn("Recipe %s is marked as only being architecture specific but seems to have machine specific packages?! The recipe may as well mark itself as machine specific directly." % d.getVar("PN", True))
}
@@ -483,7 +483,7 @@ python do_cleansstate() {
addtask cleanall after do_cleansstate
python do_cleanall() {
- src_uri = (bb.data.getVar('SRC_URI', d, True) or "").split()
+ src_uri = (d.getVar('SRC_URI', True) or "").split()
if len(src_uri) == 0:
return
diff --git a/meta/classes/bugzilla.bbclass b/meta/classes/bugzilla.bbclass
index 801bd66d2..4028d261c 100644
--- a/meta/classes/bugzilla.bbclass
+++ b/meta/classes/bugzilla.bbclass
@@ -109,14 +109,14 @@ python bugzilla_eventhandler() {
return
if name == "TaskFailed":
- xmlrpc = bb.data.getVar("BUGZILLA_XMLRPC", data, True)
- user = bb.data.getVar("BUGZILLA_USER", data, True)
- passw = bb.data.getVar("BUGZILLA_PASS", data, True)
- product = bb.data.getVar("BUGZILLA_PRODUCT", data, True)
- compon = bb.data.getVar("BUGZILLA_COMPONENT", data, True)
- version = bb.data.getVar("BUGZILLA_VERSION", data, True)
-
- proxy = bb.data.getVar('http_proxy', data, True )
+ xmlrpc = data.getVar("BUGZILLA_XMLRPC", True)
+ user = data.getVar("BUGZILLA_USER", True)
+ passw = data.getVar("BUGZILLA_PASS", True)
+ product = data.getVar("BUGZILLA_PRODUCT", True)
+ compon = data.getVar("BUGZILLA_COMPONENT", True)
+ version = data.getVar("BUGZILLA_VERSION", True)
+
+ proxy = data.getVar('http_proxy', True )
if (proxy):
import urllib2
s, u, p, hostport = urllib2._parse_proxy(proxy)
@@ -132,14 +132,14 @@ python bugzilla_eventhandler() {
'component': compon}
# evil hack to figure out what is going on
- debug_file = open(os.path.join(bb.data.getVar("TMPDIR", data, True),"..","bugzilla-log"),"a")
+ debug_file = open(os.path.join(data.getVar("TMPDIR", True),"..","bugzilla-log"),"a")
file = None
- bugname = "%(package)s-%(pv)s-autobuild" % { "package" : bb.data.getVar("PN", data, True),
- "pv" : bb.data.getVar("PV", data, True),
+ bugname = "%(package)s-%(pv)s-autobuild" % { "package" : data.getVar("PN", True),
+ "pv" : data.getVar("PV", True),
}
- log_file = glob.glob("%s/log.%s.*" % (bb.data.getVar('T', event.data, True), event.task))
- text = "The %s step in %s failed at %s for machine %s" % (e.task, bb.data.getVar("PN", data, True), bb.data.getVar('DATETIME', data, True), bb.data.getVar( 'MACHINE', data, True ) )
+ log_file = glob.glob("%s/log.%s.*" % (event.data.getVar('T', True), event.task))
+ text = "The %s step in %s failed at %s for machine %s" % (e.task, data.getVar("PN", True), data.getVar('DATETIME', True), data.getVar( 'MACHINE', True ) )
if len(log_file) != 0:
print >> debug_file, "Adding log file %s" % log_file[0]
file = open(log_file[0], 'r')
@@ -167,7 +167,7 @@ python bugzilla_eventhandler() {
if bug_number and log:
print >> debug_file, "The bug is known as '%s'" % bug_number
- desc = "Build log for machine %s" % (bb.data.getVar('MACHINE', data, True))
+ desc = "Build log for machine %s" % (data.getVar('MACHINE', True))
if not bugzilla_create_attachment(debug_file, server, args.copy(), bug_number, text, log_file[0], log, desc):
print >> debug_file, "Failed to attach the build log for bug #%s" % bug_number
else:
@@ -181,6 +181,6 @@ python bugzilla_eventhandler() {
# store bug number for oestats-client
if bug_number:
- bb.data.setVar('OESTATS_BUG_NUMBER', bug_number, data)
+ data.setVar('OESTATS_BUG_NUMBER', bug_number)
}
diff --git a/meta/classes/buildstats.bbclass b/meta/classes/buildstats.bbclass
index 96c98d409..4cd8fe693 100644
--- a/meta/classes/buildstats.bbclass
+++ b/meta/classes/buildstats.bbclass
@@ -21,25 +21,25 @@ def get_cputime():
return sum(int(field) for field in fields)
def set_bn(e):
- bn = e.getPkgs()[0] + "-" + bb.data.getVar('MACHINE',e.data, True)
+ bn = e.getPkgs()[0] + "-" + e.data.getVar('MACHINE', True)
try:
- os.remove(bb.data.getVar('BNFILE', e.data, True))
+ os.remove(e.data.getVar('BNFILE', True))
except:
pass
- file = open(bb.data.getVar('BNFILE', e.data, True), "w")
- file.write(os.path.join(bn, bb.data.getVar('BUILDNAME', e.data, True)))
+ file = open(e.data.getVar('BNFILE', True), "w")
+ file.write(os.path.join(bn, e.data.getVar('BUILDNAME', True)))
file.close()
def get_bn(e):
- file = open(bb.data.getVar('BNFILE', e.data, True))
+ file = open(e.data.getVar('BNFILE', True))
bn = file.readline()
file.close()
return bn
def set_device(e):
- tmpdir = bb.data.getVar('TMPDIR', e.data, True)
+ tmpdir = e.data.getVar('TMPDIR', True)
try:
- os.remove(bb.data.getVar('DEVFILE', e.data, True))
+ os.remove(e.data.getVar('DEVFILE', True))
except:
pass
############################################################################
@@ -66,12 +66,12 @@ def set_device(e):
rdev=line.split()[2]
else:
rdev="NoLogicalDevice"
- file = open(bb.data.getVar('DEVFILE', e.data, True), "w")
+ file = open(e.data.getVar('DEVFILE', True), "w")
file.write(rdev)
file.close()
def get_device(e):
- file = open(bb.data.getVar('DEVFILE', e.data, True))
+ file = open(e.data.getVar('DEVFILE', True))
device = file.readline()
file.close()
return device
@@ -126,7 +126,7 @@ def get_timedata(var, data):
def write_task_data(status, logfile, dev, e):
bn = get_bn(e)
- bsdir = os.path.join(bb.data.getVar('BUILDSTATS_BASE', e.data, True), bn)
+ bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn)
taskdir = os.path.join(bsdir, bb.data.expand("${PF}", e.data))
file = open(os.path.join(logfile), "a")
timedata = get_timedata("__timedata_task", e.data)
@@ -168,7 +168,7 @@ python run_buildstats () {
# set the buildname
########################################################################
try:
- bb.mkdirhier(bb.data.getVar('BUILDSTATS_BASE', e.data, True))
+ bb.mkdirhier(e.data.getVar('BUILDSTATS_BASE', True))
except:
pass
set_bn(e)
@@ -176,7 +176,7 @@ python run_buildstats () {
set_device(e)
device = get_device(e)
- bsdir = os.path.join(bb.data.getVar('BUILDSTATS_BASE', e.data, True), bn)
+ bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn)
try:
bb.mkdirhier(bsdir)
except:
@@ -199,7 +199,7 @@ python run_buildstats () {
elif isinstance(e, bb.event.BuildCompleted):
bn = get_bn(e)
device = get_device(e)
- bsdir = os.path.join(bb.data.getVar('BUILDSTATS_BASE', e.data, True), bn)
+ bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn)
taskdir = os.path.join(bsdir, bb.data.expand("${PF}", e.data))
build_time = os.path.join(bsdir, "build_stats")
file = open(build_time, "a")
@@ -224,7 +224,7 @@ python run_buildstats () {
if isinstance(e, bb.build.TaskStarted):
bn = get_bn(e)
device = get_device(e)
- bsdir = os.path.join(bb.data.getVar('BUILDSTATS_BASE', e.data, True), bn)
+ bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn)
taskdir = os.path.join(bsdir, bb.data.expand("${PF}", e.data))
if device != "NoLogicalDevice":
set_diskdata("__diskdata_task", device, e.data)
@@ -242,14 +242,14 @@ python run_buildstats () {
elif isinstance(e, bb.build.TaskSucceeded):
bn = get_bn(e)
device = get_device(e)
- bsdir = os.path.join(bb.data.getVar('BUILDSTATS_BASE', e.data, True), bn)
+ bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn)
taskdir = os.path.join(bsdir, bb.data.expand("${PF}", e.data))
write_task_data("passed", os.path.join(taskdir, e.task), device, e)
if e.task == "do_rootfs":
- bsdir = os.path.join(bb.data.getVar('BUILDSTATS_BASE', e.data, True), bn)
+ bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn)
bs=os.path.join(bsdir, "build_stats")
file = open(bs,"a")
- rootfs = bb.data.getVar('IMAGE_ROOTFS', e.data, True)
+ rootfs = e.data.getVar('IMAGE_ROOTFS', True)
rootfs_size = subprocess.Popen(["du", "-sh", rootfs], stdout=subprocess.PIPE).stdout.read()
file.write("Uncompressed Rootfs size: %s" % rootfs_size)
file.close()
@@ -257,7 +257,7 @@ python run_buildstats () {
elif isinstance(e, bb.build.TaskFailed):
bn = get_bn(e)
device = get_device(e)
- bsdir = os.path.join(bb.data.getVar('BUILDSTATS_BASE', e.data, True), bn)
+ bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn)
taskdir = os.path.join(bsdir, bb.data.expand("${PF}", e.data))
write_task_data("failed", os.path.join(taskdir, e.task), device, e)
########################################################################
diff --git a/meta/classes/cpan-base.bbclass b/meta/classes/cpan-base.bbclass
index b5dbdaea8..79582ca76 100644
--- a/meta/classes/cpan-base.bbclass
+++ b/meta/classes/cpan-base.bbclass
@@ -28,7 +28,7 @@ def get_perl_version(d):
# Determine where the library directories are
def perl_get_libdirs(d):
- libdir = bb.data.getVar('libdir', d, 1)
+ libdir = d.getVar('libdir', 1)
if is_target(d) == "no":
libdir += '/perl-native'
libdir += '/perl'
diff --git a/meta/classes/cpan_build.bbclass b/meta/classes/cpan_build.bbclass
index cc503a424..981332c4f 100644
--- a/meta/classes/cpan_build.bbclass
+++ b/meta/classes/cpan_build.bbclass
@@ -10,9 +10,9 @@ inherit cpan-base
# libmodule-build-perl)
#
def cpan_build_dep_prepend(d):
- if bb.data.getVar('CPAN_BUILD_DEPS', d, 1):
+ if d.getVar('CPAN_BUILD_DEPS', 1):
return ''
- pn = bb.data.getVar('PN', d, 1)
+ pn = d.getVar('PN', 1)
if pn in ['libmodule-build-perl', 'libmodule-build-perl-native']:
return ''
return 'libmodule-build-perl-native '
diff --git a/meta/classes/cross-canadian.bbclass b/meta/classes/cross-canadian.bbclass
index 601175db4..6f5bcd0ad 100644
--- a/meta/classes/cross-canadian.bbclass
+++ b/meta/classes/cross-canadian.bbclass
@@ -16,7 +16,7 @@ STAGING_BINDIR_TOOLCHAIN = "${STAGING_DIR_NATIVE}${bindir_native}/${SDK_ARCH}${S
#
PACKAGE_ARCH = "${SDK_ARCH}-nativesdk"
python () {
- archs = bb.data.getVar('PACKAGE_ARCHS', d, True).split()
+ archs = d.getVar('PACKAGE_ARCHS', True).split()
sdkarchs = []
for arch in archs:
sdkarchs.append(arch + '-nativesdk')
diff --git a/meta/classes/debian.bbclass b/meta/classes/debian.bbclass
index 554525dc8..025abcfad 100644
--- a/meta/classes/debian.bbclass
+++ b/meta/classes/debian.bbclass
@@ -22,8 +22,8 @@ python () {
python debian_package_name_hook () {
import glob, copy, stat, errno, re
- pkgdest = bb.data.getVar('PKGDEST', d, 1)
- packages = bb.data.getVar('PACKAGES', d, 1)
+ pkgdest = d.getVar('PKGDEST', 1)
+ packages = d.getVar('PACKAGES', 1)
bin_re = re.compile(".*/s?" + os.path.basename(d.getVar("bindir", True)) + "$")
lib_re = re.compile(".*/" + os.path.basename(d.getVar("libdir", True)) + "$")
so_re = re.compile("lib.*\.so")
@@ -60,7 +60,7 @@ python debian_package_name_hook () {
for f in files:
if so_re.match(f):
fp = os.path.join(root, f)
- cmd = (bb.data.getVar('BUILD_PREFIX', d, 1) or "") + "objdump -p " + fp + " 2>/dev/null"
+ cmd = (d.getVar('BUILD_PREFIX', 1) or "") + "objdump -p " + fp + " 2>/dev/null"
fd = os.popen(cmd)
lines = fd.readlines()
fd.close()
@@ -74,7 +74,7 @@ python debian_package_name_hook () {
if len(sonames) == 1:
soname = sonames[0]
elif len(sonames) > 1:
- lead = bb.data.getVar('LEAD_SONAME', d, 1)
+ lead = d.getVar('LEAD_SONAME', 1)
if lead:
r = re.compile(lead)
filtered = []
@@ -95,21 +95,21 @@ python debian_package_name_hook () {
if soname_result:
(pkgname, devname) = soname_result
for pkg in packages.split():
- if (bb.data.getVar('PKG_' + pkg, d) or bb.data.getVar('DEBIAN_NOAUTONAME_' + pkg, d)):
+ if (d.getVar('PKG_' + pkg) or d.getVar('DEBIAN_NOAUTONAME_' + pkg)):
continue
- debian_pn = bb.data.getVar('DEBIANNAME_' + pkg, d)
+ debian_pn = d.getVar('DEBIANNAME_' + pkg)
if debian_pn:
newpkg = debian_pn
elif pkg == orig_pkg:
newpkg = pkgname
else:
newpkg = pkg.replace(orig_pkg, devname, 1)
- mlpre=bb.data.getVar('MLPREFIX', d, True)
+ mlpre=d.getVar('MLPREFIX', True)
if mlpre:
if not newpkg.find(mlpre) == 0:
newpkg = mlpre + newpkg
if newpkg != pkg:
- bb.data.setVar('PKG_' + pkg, newpkg, d)
+ d.setVar('PKG_' + pkg, newpkg)
# reversed sort is needed when some package is substring of another
# ie in ncurses we get without reverse sort:
@@ -117,7 +117,7 @@ python debian_package_name_hook () {
# and later
# DEBUG: LIBNAMES: pkgname libtic5 devname libtic pkg ncurses-libticw orig_pkg ncurses-libtic debian_pn None newpkg libticw
# so we need to handle ncurses-libticw->libticw5 before ncurses-libtic->libtic5
- for pkg in sorted((bb.data.getVar('AUTO_LIBNAME_PKGS', d, 1) or "").split(), reverse=True):
+ for pkg in sorted((d.getVar('AUTO_LIBNAME_PKGS', 1) or "").split(), reverse=True):
auto_libname(packages, pkg)
}
diff --git a/meta/classes/distrodata.bbclass b/meta/classes/distrodata.bbclass
index ce7b931b1..687247a64 100644
--- a/meta/classes/distrodata.bbclass
+++ b/meta/classes/distrodata.bbclass
@@ -19,87 +19,87 @@ addtask distrodata_np
do_distrodata_np[nostamp] = "1"
python do_distrodata_np() {
localdata = bb.data.createCopy(d)
- pn = bb.data.getVar("PN", d, True)
+ pn = d.getVar("PN", True)
bb.note("Package Name: %s" % pn)
import oe.distro_check as dist_check
- tmpdir = bb.data.getVar('TMPDIR', d, True)
+ tmpdir = d.getVar('TMPDIR', True)
distro_check_dir = os.path.join(tmpdir, "distro_check")
- datetime = bb.data.getVar('DATETIME', localdata, True)
+ datetime = localdata.getVar('DATETIME', True)
dist_check.update_distro_data(distro_check_dir, datetime)
if pn.find("-native") != -1:
pnstripped = pn.split("-native")
bb.note("Native Split: %s" % pnstripped)
- bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+ bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
bb.data.update_data(localdata)
if pn.find("-nativesdk") != -1:
pnstripped = pn.split("-nativesdk")
bb.note("Native Split: %s" % pnstripped)
- bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+ bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
bb.data.update_data(localdata)
if pn.find("-cross") != -1:
pnstripped = pn.split("-cross")
bb.note("cross Split: %s" % pnstripped)
- bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+ bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
bb.data.update_data(localdata)
if pn.find("-crosssdk") != -1:
pnstripped = pn.split("-crosssdk")
bb.note("cross Split: %s" % pnstripped)
- bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+ bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
bb.data.update_data(localdata)
if pn.find("-initial") != -1:
pnstripped = pn.split("-initial")
bb.note("initial Split: %s" % pnstripped)
- bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+ bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
bb.data.update_data(localdata)
"""generate package information from .bb file"""
- pname = bb.data.getVar('PN', localdata, True)
- pcurver = bb.data.getVar('PV', localdata, True)
- pdesc = bb.data.getVar('DESCRIPTION', localdata, True)
+ pname = localdata.getVar('PN', True)
+ pcurver = localdata.getVar('PV', True)
+ pdesc = localdata.getVar('DESCRIPTION', True)
if pdesc is not None:
pdesc = pdesc.replace(',','')
pdesc = pdesc.replace('\n','')
- pgrp = bb.data.getVar('SECTION', localdata, True)
- plicense = bb.data.getVar('LICENSE', localdata, True).replace(',','_')
- if bb.data.getVar('LIC_FILES_CHKSUM', localdata, True):
+ pgrp = localdata.getVar('SECTION', True)
+ plicense = localdata.getVar('LICENSE', True).replace(',','_')
+ if localdata.getVar('LIC_FILES_CHKSUM', True):
pchksum="1"
else:
pchksum="0"
- if bb.data.getVar('RECIPE_STATUS', localdata, True):
+ if localdata.getVar('RECIPE_STATUS', True):
hasrstatus="1"
else:
hasrstatus="0"
- rstatus = bb.data.getVar('RECIPE_STATUS', localdata, True)
+ rstatus = localdata.getVar('RECIPE_STATUS', True)
if rstatus is not None:
rstatus = rstatus.replace(',','')
- pupver = bb.data.getVar('RECIPE_LATEST_VERSION', localdata, True)
+ pupver = localdata.getVar('RECIPE_LATEST_VERSION', True)
if pcurver == pupver:
vermatch="1"
else:
vermatch="0"
- noupdate_reason = bb.data.getVar('RECIPE_NO_UPDATE_REASON', localdata, True)
+ noupdate_reason = localdata.getVar('RECIPE_NO_UPDATE_REASON', True)
if noupdate_reason is None:
noupdate="0"
else:
noupdate="1"
noupdate_reason = noupdate_reason.replace(',','')
- ris = bb.data.getVar('RECIPE_INTEL_SECTION', localdata, True)
- maintainer = bb.data.getVar('RECIPE_MAINTAINER', localdata, True)
- rttr = bb.data.getVar('RECIPE_TIME_BETWEEN_LAST_TWO_RELEASES', localdata, True)
- rlrd = bb.data.getVar('RECIPE_LATEST_RELEASE_DATE', localdata, True)
- dc = bb.data.getVar('DEPENDENCY_CHECK', localdata, True)
- rc = bb.data.getVar('RECIPE_COMMENTS', localdata, True)
+ ris = localdata.getVar('RECIPE_INTEL_SECTION', True)
+ maintainer = localdata.getVar('RECIPE_MAINTAINER', True)
+ rttr = localdata.getVar('RECIPE_TIME_BETWEEN_LAST_TWO_RELEASES', True)
+ rlrd = localdata.getVar('RECIPE_LATEST_RELEASE_DATE', True)
+ dc = localdata.getVar('DEPENDENCY_CHECK', True)
+ rc = localdata.getVar('RECIPE_COMMENTS', True)
result = dist_check.compare_in_distro_packages_list(distro_check_dir, localdata)
bb.note("DISTRO: %s,%s,%s,%s,%s,%s,%s,%s,%s, %s, %s, %s\n" % \
@@ -113,81 +113,81 @@ python do_distrodata_np() {
addtask distrodata
do_distrodata[nostamp] = "1"
python do_distrodata() {
- logpath = bb.data.getVar('LOG_DIR', d, True)
+ logpath = d.getVar('LOG_DIR', True)
bb.utils.mkdirhier(logpath)
logfile = os.path.join(logpath, "distrodata.csv")
import oe.distro_check as dist_check
localdata = bb.data.createCopy(d)
- tmpdir = bb.data.getVar('TMPDIR', d, True)
+ tmpdir = d.getVar('TMPDIR', True)
distro_check_dir = os.path.join(tmpdir, "distro_check")
- datetime = bb.data.getVar('DATETIME', localdata, True)
+ datetime = localdata.getVar('DATETIME', True)
dist_check.update_distro_data(distro_check_dir, datetime)
- pn = bb.data.getVar("PN", d, True)
+ pn = d.getVar("PN", True)
bb.note("Package Name: %s" % pn)
if pn.find("-native") != -1:
pnstripped = pn.split("-native")
bb.note("Native Split: %s" % pnstripped)
- bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+ bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
bb.data.update_data(localdata)
if pn.find("-cross") != -1:
pnstripped = pn.split("-cross")
bb.note("cross Split: %s" % pnstripped)
- bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+ bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
bb.data.update_data(localdata)
if pn.find("-initial") != -1:
pnstripped = pn.split("-initial")
bb.note("initial Split: %s" % pnstripped)
- bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+ bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
bb.data.update_data(localdata)
"""generate package information from .bb file"""
- pname = bb.data.getVar('PN', localdata, True)
- pcurver = bb.data.getVar('PV', localdata, True)
- pdesc = bb.data.getVar('DESCRIPTION', localdata, True)
+ pname = localdata.getVar('PN', True)
+ pcurver = localdata.getVar('PV', True)
+ pdesc = localdata.getVar('DESCRIPTION', True)
if pdesc is not None:
pdesc = pdesc.replace(',','')
pdesc = pdesc.replace('\n','')
- pgrp = bb.data.getVar('SECTION', localdata, True)
- plicense = bb.data.getVar('LICENSE', localdata, True).replace(',','_')
- if bb.data.getVar('LIC_FILES_CHKSUM', localdata, True):
+ pgrp = localdata.getVar('SECTION', True)
+ plicense = localdata.getVar('LICENSE', True).replace(',','_')
+ if localdata.getVar('LIC_FILES_CHKSUM', True):
pchksum="1"
else:
pchksum="0"
- if bb.data.getVar('RECIPE_STATUS', localdata, True):
+ if localdata.getVar('RECIPE_STATUS', True):
hasrstatus="1"
else:
hasrstatus="0"
- rstatus = bb.data.getVar('RECIPE_STATUS', localdata, True)
+ rstatus = localdata.getVar('RECIPE_STATUS', True)
if rstatus is not None:
rstatus = rstatus.replace(',','')
- pupver = bb.data.getVar('RECIPE_LATEST_VERSION', localdata, True)
+ pupver = localdata.getVar('RECIPE_LATEST_VERSION', True)
if pcurver == pupver:
vermatch="1"
else:
vermatch="0"
- noupdate_reason = bb.data.getVar('RECIPE_NO_UPDATE_REASON', localdata, True)
+ noupdate_reason = localdata.getVar('RECIPE_NO_UPDATE_REASON', True)
if noupdate_reason is None:
noupdate="0"
else:
noupdate="1"
noupdate_reason = noupdate_reason.replace(',','')
- ris = bb.data.getVar('RECIPE_INTEL_SECTION', localdata, True)
- maintainer = bb.data.getVar('RECIPE_MAINTAINER', localdata, True)
- rttr = bb.data.getVar('RECIPE_TIME_BETWEEN_LAST_TWO_RELEASES', localdata, True)
- rlrd = bb.data.getVar('RECIPE_LATEST_RELEASE_DATE', localdata, True)
- dc = bb.data.getVar('DEPENDENCY_CHECK', localdata, True)
- rc = bb.data.getVar('RECIPE_COMMENTS', localdata, True)
+ ris = localdata.getVar('RECIPE_INTEL_SECTION', True)
+ maintainer = localdata.getVar('RECIPE_MAINTAINER', True)
+ rttr = localdata.getVar('RECIPE_TIME_BETWEEN_LAST_TWO_RELEASES', True)
+ rlrd = localdata.getVar('RECIPE_LATEST_RELEASE_DATE', True)
+ dc = localdata.getVar('DEPENDENCY_CHECK', True)
+ rc = localdata.getVar('RECIPE_COMMENTS', True)
# do the comparison
result = dist_check.compare_in_distro_packages_list(distro_check_dir, localdata)
@@ -298,7 +298,7 @@ python do_checkpkg() {
Clear internal url cache as it's a temporary check. Not doing so will have
bitbake check url multiple times when looping through a single url
"""
- fn = bb.data.getVar('FILE', d, True)
+ fn = d.getVar('FILE', True)
bb.fetch2.urldata_cache[fn] = {}
"""
@@ -329,7 +329,7 @@ python do_checkpkg() {
Return new version if success, or else error in "Errxxxx" style
"""
def check_new_dir(url, curver, d):
- pn = bb.data.getVar('PN', d, True)
+ pn = d.getVar('PN', True)
f = tempfile.NamedTemporaryFile(delete=False, prefix="%s-1-" % pn)
status = internal_fetch_wget(url, d, f)
fhtml = f.read()
@@ -372,7 +372,7 @@ python do_checkpkg() {
f.close()
if status != "ErrHostNoDir" and re.match("Err", status):
- logpath = bb.data.getVar('LOG_DIR', d, 1)
+ logpath = d.getVar('LOG_DIR', 1)
os.system("cp %s %s/" % (f.name, logpath))
os.unlink(f.name)
return status
@@ -388,7 +388,7 @@ python do_checkpkg() {
"""possible to have no version in pkg name, such as spectrum-fw"""
if not re.search("\d+", curname):
return pcurver
- pn = bb.data.getVar('PN', d, True)
+ pn = d.getVar('PN', True)
f = tempfile.NamedTemporaryFile(delete=False, prefix="%s-2-" % pn)
status = internal_fetch_wget(url, d, f)
fhtml = f.read()
@@ -431,55 +431,55 @@ python do_checkpkg() {
f.close()
"""if host hasn't directory information, no need to save tmp file"""
if status != "ErrHostNoDir" and re.match("Err", status):
- logpath = bb.data.getVar('LOG_DIR', d, True)
+ logpath = d.getVar('LOG_DIR', True)
os.system("cp %s %s/" % (f.name, logpath))
os.unlink(f.name)
return status
"""first check whether a uri is provided"""
- src_uri = bb.data.getVar('SRC_URI', d, True)
+ src_uri = d.getVar('SRC_URI', True)
if not src_uri:
return
"""initialize log files."""
- logpath = bb.data.getVar('LOG_DIR', d, True)
+ logpath = d.getVar('LOG_DIR', True)
bb.utils.mkdirhier(logpath)
logfile = os.path.join(logpath, "checkpkg.csv")
"""generate package information from .bb file"""
- pname = bb.data.getVar('PN', d, True)
+ pname = d.getVar('PN', True)
if pname.find("-native") != -1:
pnstripped = pname.split("-native")
bb.note("Native Split: %s" % pnstripped)
- bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+ bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
bb.data.update_data(localdata)
if pname.find("-cross") != -1:
pnstripped = pname.split("-cross")
bb.note("cross Split: %s" % pnstripped)
- bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+ bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
bb.data.update_data(localdata)
if pname.find("-initial") != -1:
pnstripped = pname.split("-initial")
bb.note("initial Split: %s" % pnstripped)
- bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+ bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
bb.data.update_data(localdata)
- pdesc = bb.data.getVar('DESCRIPTION', localdata, True)
- pgrp = bb.data.getVar('SECTION', localdata, True)
- pversion = bb.data.getVar('PV', localdata, True)
- plicense = bb.data.getVar('LICENSE', localdata, True)
- psection = bb.data.getVar('SECTION', localdata, True)
- phome = bb.data.getVar('HOMEPAGE', localdata, True)
- prelease = bb.data.getVar('PR', localdata, True)
- ppriority = bb.data.getVar('PRIORITY', localdata, True)
- pdepends = bb.data.getVar('DEPENDS', localdata, True)
- pbugtracker = bb.data.getVar('BUGTRACKER', localdata, True)
- ppe = bb.data.getVar('PE', localdata, True)
- psrcuri = bb.data.getVar('SRC_URI', localdata, True)
- maintainer = bb.data.getVar('RECIPE_MAINTAINER', localdata, True)
+ pdesc = localdata.getVar('DESCRIPTION', True)
+ pgrp = localdata.getVar('SECTION', True)
+ pversion = localdata.getVar('PV', True)
+ plicense = localdata.getVar('LICENSE', True)
+ psection = localdata.getVar('SECTION', True)
+ phome = localdata.getVar('HOMEPAGE', True)
+ prelease = localdata.getVar('PR', True)
+ ppriority = localdata.getVar('PRIORITY', True)
+ pdepends = localdata.getVar('DEPENDS', True)
+ pbugtracker = localdata.getVar('BUGTRACKER', True)
+ ppe = localdata.getVar('PE', True)
+ psrcuri = localdata.getVar('SRC_URI', True)
+ maintainer = localdata.getVar('RECIPE_MAINTAINER', True)
found = 0
for uri in src_uri.split():
@@ -497,9 +497,9 @@ python do_checkpkg() {
(type, host, path, user, pswd, parm) = bb.decodeurl(uri)
if type in ['http', 'https', 'ftp']:
- pcurver = bb.data.getVar('PV', d, True)
+ pcurver = d.getVar('PV', True)
else:
- pcurver = bb.data.getVar("SRCREV", d, True)
+ pcurver = d.getVar("SRCREV", True)
if type in ['http', 'https', 'ftp']:
newver = pcurver
@@ -639,7 +639,7 @@ python do_checkpkg() {
pstatus += ":%s%s" % (host, path)
"""Read from manual distro tracking fields as alternative"""
- pmver = bb.data.getVar("RECIPE_LATEST_VERSION", d, True)
+ pmver = d.getVar("RECIPE_LATEST_VERSION", True)
if not pmver:
pmver = "N/A"
pmstatus = "ErrNoRecipeData"
@@ -688,12 +688,12 @@ python do_distro_check() {
localdata = bb.data.createCopy(d)
bb.data.update_data(localdata)
- tmpdir = bb.data.getVar('TMPDIR', d, True)
+ tmpdir = d.getVar('TMPDIR', True)
distro_check_dir = os.path.join(tmpdir, "distro_check")
- logpath = bb.data.getVar('LOG_DIR', d, True)
+ logpath = d.getVar('LOG_DIR', True)
bb.utils.mkdirhier(logpath)
result_file = os.path.join(logpath, "distrocheck.csv")
- datetime = bb.data.getVar('DATETIME', localdata, True)
+ datetime = localdata.getVar('DATETIME', True)
dc.update_distro_data(distro_check_dir, datetime)
# do the comparison
@@ -734,12 +734,12 @@ python do_checklicense() {
import os
import bb
import shutil
- logpath = bb.data.getVar('LOG_DIR', d, True)
+ logpath = d.getVar('LOG_DIR', True)
bb.utils.mkdirhier(logpath)
- pn = bb.data.getVar('PN', d, True)
+ pn = d.getVar('PN', True)
logfile = os.path.join(logpath, "missinglicense.csv")
- generic_directory = bb.data.getVar('COMMON_LICENSE_DIR', d, True)
- license_types = bb.data.getVar('LICENSE', d, True)
+ generic_directory = d.getVar('COMMON_LICENSE_DIR', True)
+ license_types = d.getVar('LICENSE', True)
for license_type in ((license_types.replace('+', '').replace('|', '&')
.replace('(', '').replace(')', '').replace(';', '')
.replace(',', '').replace(" ", "").split("&"))):
diff --git a/meta/classes/distutils-base.bbclass b/meta/classes/distutils-base.bbclass
index e84b0fcc5..e7d0bb807 100644
--- a/meta/classes/distutils-base.bbclass
+++ b/meta/classes/distutils-base.bbclass
@@ -1,4 +1,4 @@
-DEPENDS += "${@["python-native python", ""][(bb.data.getVar('PACKAGES', d, 1) == '')]}"
+DEPENDS += "${@["python-native python", ""][(d.getVar('PACKAGES', 1) == '')]}"
RDEPENDS_${PN} += "${@['', 'python-core']['${PN}' == '${BPN}']}"
inherit distutils-common-base
diff --git a/meta/classes/distutils-native-base.bbclass b/meta/classes/distutils-native-base.bbclass
index 2703fe074..47367d796 100644
--- a/meta/classes/distutils-native-base.bbclass
+++ b/meta/classes/distutils-native-base.bbclass
@@ -1,3 +1,3 @@
-DEPENDS += "${@["python-native", ""][(bb.data.getVar('PACKAGES', d, 1) == '')]}"
+DEPENDS += "${@["python-native", ""][(d.getVar('PACKAGES', 1) == '')]}"
inherit distutils-common-base
diff --git a/meta/classes/gconf.bbclass b/meta/classes/gconf.bbclass
index 67986787d..bffc92ea7 100644
--- a/meta/classes/gconf.bbclass
+++ b/meta/classes/gconf.bbclass
@@ -27,8 +27,8 @@ done
python populate_packages_append () {
import re
- packages = bb.data.getVar('PACKAGES', d, 1).split()
- pkgdest = bb.data.getVar('PKGDEST', d, 1)
+ packages = d.getVar('PACKAGES', 1).split()
+ pkgdest = d.getVar('PKGDEST', 1)
for pkg in packages:
schema_dir = '%s/%s/etc/gconf/schemas' % (pkgdest, pkg)
@@ -41,15 +41,15 @@ python populate_packages_append () {
if schemas != []:
bb.note("adding gconf postinst and prerm scripts to %s" % pkg)
bb.data.setVar('SCHEMA_FILES', " ".join(schemas), d)
- postinst = bb.data.getVar('pkg_postinst_%s' % pkg, d, 1) or bb.data.getVar('pkg_postinst', d, 1)
+ postinst = d.getVar('pkg_postinst_%s' % pkg, 1) or d.getVar('pkg_postinst', 1)
if not postinst:
postinst = '#!/bin/sh\n'
- postinst += bb.data.getVar('gconf_postinst', d, 1)
- bb.data.setVar('pkg_postinst_%s' % pkg, postinst, d)
- prerm = bb.data.getVar('pkg_prerm_%s' % pkg, d, 1) or bb.data.getVar('pkg_prerm', d, 1)
+ postinst += d.getVar('gconf_postinst', 1)
+ d.setVar('pkg_postinst_%s' % pkg, postinst)
+ prerm = d.getVar('pkg_prerm_%s' % pkg, 1) or d.getVar('pkg_prerm', 1)
if not prerm:
prerm = '#!/bin/sh\n'
- prerm += bb.data.getVar('gconf_prerm', d, 1)
- bb.data.setVar('pkg_prerm_%s' % pkg, prerm, d)
+ prerm += d.getVar('gconf_prerm', 1)
+ d.setVar('pkg_prerm_%s' % pkg, prerm)
}
diff --git a/meta/classes/gtk-icon-cache.bbclass b/meta/classes/gtk-icon-cache.bbclass
index d0840d59b..eac3061b0 100644
--- a/meta/classes/gtk-icon-cache.bbclass
+++ b/meta/classes/gtk-icon-cache.bbclass
@@ -28,31 +28,31 @@ done
}
python populate_packages_append () {
- packages = bb.data.getVar('PACKAGES', d, 1).split()
- pkgdest = bb.data.getVar('PKGDEST', d, 1)
+ packages = d.getVar('PACKAGES', 1).split()
+ pkgdest = d.getVar('PKGDEST', 1)
for pkg in packages:
- icon_dir = '%s/%s/%s/icons' % (pkgdest, pkg, bb.data.getVar('datadir', d, 1))
+ icon_dir = '%s/%s/%s/icons' % (pkgdest, pkg, d.getVar('datadir', 1))
if not os.path.exists(icon_dir):
continue
bb.note("adding hicolor-icon-theme dependency to %s" % pkg)
- rdepends = bb.data.getVar('RDEPENDS_%s' % pkg, d, 1)
- rdepends = rdepends + ' ' + bb.data.getVar('MLPREFIX', d) + "hicolor-icon-theme"
- bb.data.setVar('RDEPENDS_%s' % pkg, rdepends, d)
+ rdepends = d.getVar('RDEPENDS_%s' % pkg, 1)
+ rdepends = rdepends + ' ' + d.getVar('MLPREFIX') + "hicolor-icon-theme"
+ d.setVar('RDEPENDS_%s' % pkg, rdepends)
bb.note("adding gtk-icon-cache postinst and postrm scripts to %s" % pkg)
- postinst = bb.data.getVar('pkg_postinst_%s' % pkg, d, 1) or bb.data.getVar('pkg_postinst', d, 1)
+ postinst = d.getVar('pkg_postinst_%s' % pkg, 1) or d.getVar('pkg_postinst', 1)
if not postinst:
postinst = '#!/bin/sh\n'
- postinst += bb.data.getVar('gtk_icon_cache_postinst', d, 1)
- bb.data.setVar('pkg_postinst_%s' % pkg, postinst, d)
+ postinst += d.getVar('gtk_icon_cache_postinst', 1)
+ d.setVar('pkg_postinst_%s' % pkg, postinst)
- postrm = bb.data.getVar('pkg_postrm_%s' % pkg, d, 1) or bb.data.getVar('pkg_postrm', d, 1)
+ postrm = d.getVar('pkg_postrm_%s' % pkg, 1) or d.getVar('pkg_postrm', 1)
if not postrm:
postrm = '#!/bin/sh\n'
- postrm += bb.data.getVar('gtk_icon_cache_postrm', d, 1)
- bb.data.setVar('pkg_postrm_%s' % pkg, postrm, d)
+ postrm += d.getVar('gtk_icon_cache_postrm', 1)
+ d.setVar('pkg_postrm_%s' % pkg, postrm)
}
diff --git a/meta/classes/icecc.bbclass b/meta/classes/icecc.bbclass
index f8e9d8859..7e3676af3 100644
--- a/meta/classes/icecc.bbclass
+++ b/meta/classes/icecc.bbclass
@@ -32,7 +32,7 @@ def icecc_dep_prepend(d):
# INHIBIT_DEFAULT_DEPS doesn't apply to the patch command. Whether or not
# we need that built is the responsibility of the patch function / class, not
# the application.
- if not bb.data.getVar('INHIBIT_DEFAULT_DEPS', d):
+ if not d.getVar('INHIBIT_DEFAULT_DEPS'):
return "icecc-create-env-native"
return ""
@@ -54,7 +54,7 @@ def create_path(compilers, bb, d):
staging += "-kernel"
#check if the icecc path is set by the user
- icecc = bb.data.getVar('ICECC_PATH', d) or os.popen("which icecc").read()[:-1]
+ icecc = d.getVar('ICECC_PATH') or os.popen("which icecc").read()[:-1]
# Create the dir if necessary
try:
@@ -81,7 +81,7 @@ def use_icc(bb,d):
package_tmp = bb.data.expand('${PN}', d)
system_class_blacklist = [ "none" ]
- user_class_blacklist = (bb.data.getVar('ICECC_USER_CLASS_BL', d) or "none").split()
+ user_class_blacklist = (d.getVar('ICECC_USER_CLASS_BL') or "none").split()
package_class_blacklist = system_class_blacklist + user_class_blacklist
for black in package_class_blacklist:
@@ -92,7 +92,7 @@ def use_icc(bb,d):
#"system" package blacklist contains a list of packages that can not distribute compile tasks
#for one reason or the other
system_package_blacklist = [ "uclibc", "glibc", "gcc", "bind", "u-boot", "dhcp-forwarder", "enchant", "connman", "orbit2" ]
- user_package_blacklist = (bb.data.getVar('ICECC_USER_PACKAGE_BL', d) or "").split()
+ user_package_blacklist = (d.getVar('ICECC_USER_PACKAGE_BL') or "").split()
package_blacklist = system_package_blacklist + user_package_blacklist
for black in package_blacklist:
@@ -100,7 +100,7 @@ def use_icc(bb,d):
#bb.note(package_tmp, ' found in blacklist, disable icecc')
return "no"
- if bb.data.getVar('PARALLEL_MAKE', d) == "":
+ if d.getVar('PARALLEL_MAKE') == "":
bb.note(package_tmp, " ", bb.data.expand('${PV}', d), " has empty PARALLEL_MAKE, disable icecc")
return "no"
@@ -119,8 +119,8 @@ def icc_version(bb, d):
if use_icc(bb, d) == "no":
return ""
- parallel = bb.data.getVar('ICECC_PARALLEL_MAKE', d) or ""
- bb.data.setVar("PARALLEL_MAKE", parallel, d)
+ parallel = d.getVar('ICECC_PARALLEL_MAKE') or ""
+ d.setVar("PARALLEL_MAKE", parallel)
if icc_is_native(bb, d):
archive_name = "local-host-env"
@@ -130,7 +130,7 @@ def icc_version(bb, d):
prefix = bb.data.expand('${HOST_PREFIX}' , d)
distro = bb.data.expand('${DISTRO}', d)
target_sys = bb.data.expand('${TARGET_SYS}', d)
- float = bb.data.getVar('TARGET_FPU', d) or "hard"
+ float = d.getVar('TARGET_FPU') or "hard"
archive_name = prefix + distro + "-" + target_sys + "-" + float
if icc_is_kernel(bb, d):
archive_name += "-kernel"
diff --git a/meta/classes/image-swab.bbclass b/meta/classes/image-swab.bbclass
index b939ec4b4..23183b3ec 100644
--- a/meta/classes/image-swab.bbclass
+++ b/meta/classes/image-swab.bbclass
@@ -51,13 +51,13 @@ python() {
# and cross packages which aren't swabber-native or one of its dependencies
# I have ignored them for now...
if not bb.data.inherits_class('native', d) and not bb.data.inherits_class('nativesdk', d) and not bb.data.inherits_class('cross', d):
- deps = (bb.data.getVarFlag('do_setscene', 'depends', d) or "").split()
+ deps = (d.getVarFlag('do_setscene', 'depends') or "").split()
deps.append('strace-native:do_populate_sysroot')
bb.data.setVarFlag('do_setscene', 'depends', " ".join(deps), d)
logdir = bb.data.expand("${TRACE_LOGDIR}", d)
bb.utils.mkdirhier(logdir)
else:
- bb.data.setVar('STRACEFUNC', '', d)
+ d.setVar('STRACEFUNC', '')
}
STRACEPID = "${@os.getpid()}"
diff --git a/meta/classes/image.bbclass b/meta/classes/image.bbclass
index 14726d253..4642fa63e 100644
--- a/meta/classes/image.bbclass
+++ b/meta/classes/image.bbclass
@@ -74,17 +74,17 @@ IMAGE_TYPE = ${@base_contains("IMAGE_FSTYPES", "live", "live", "empty", d)}
inherit image-${IMAGE_TYPE}
python () {
- deps = bb.data.getVarFlag('do_rootfs', 'depends', d) or ""
- for type in (bb.data.getVar('IMAGE_FSTYPES', d, True) or "").split():
- for dep in ((bb.data.getVar('IMAGE_DEPENDS_%s' % type, d) or "").split() or []):
+ deps = d.getVarFlag('do_rootfs', 'depends') or ""
+ for type in (d.getVar('IMAGE_FSTYPES', True) or "").split():
+ for dep in ((d.getVar('IMAGE_DEPENDS_%s' % type) or "").split() or []):
deps += " %s:do_populate_sysroot" % dep
- for dep in (bb.data.getVar('EXTRA_IMAGEDEPENDS', d, True) or "").split():
+ for dep in (d.getVar('EXTRA_IMAGEDEPENDS', True) or "").split():
deps += " %s:do_populate_sysroot" % dep
- bb.data.setVarFlag('do_rootfs', 'depends', deps, d)
+ d.setVarFlag('do_rootfs', 'depends', deps)
# If we don't do this we try and run the mapping hooks while parsing which is slow
# bitbake should really provide something to let us know this...
- if bb.data.getVar('BB_WORKERCONTEXT', d, True) is not None:
+ if d.getVar('BB_WORKERCONTEXT', True) is not None:
runtime_mapping_rename("PACKAGE_INSTALL", d)
runtime_mapping_rename("PACKAGE_INSTALL_ATTEMPTONLY", d)
}
@@ -98,15 +98,15 @@ python () {
# is searched for in the BBPATH (same as the old version.)
#
def get_devtable_list(d):
- devtable = bb.data.getVar('IMAGE_DEVICE_TABLE', d, 1)
+ devtable = d.getVar('IMAGE_DEVICE_TABLE', 1)
if devtable != None:
return devtable
str = ""
- devtables = bb.data.getVar('IMAGE_DEVICE_TABLES', d, 1)
+ devtables = d.getVar('IMAGE_DEVICE_TABLES', 1)
if devtables == None:
devtables = 'files/device_table-minimal.txt'
for devtable in devtables.split():
- str += " %s" % bb.which(bb.data.getVar('BBPATH', d, 1), devtable)
+ str += " %s" % bb.which(d.getVar('BBPATH', 1), devtable)
return str
IMAGE_CLASSES ?= "image_types"
@@ -119,7 +119,7 @@ ROOTFS_POSTPROCESS_COMMAND ?= ""
# some default locales
IMAGE_LINGUAS ?= "de-de fr-fr en-gb"
-LINGUAS_INSTALL = "${@" ".join(map(lambda s: "locale-base-%s" % s, bb.data.getVar('IMAGE_LINGUAS', d, 1).split()))}"
+LINGUAS_INSTALL = "${@" ".join(map(lambda s: "locale-base-%s" % s, d.getVar('IMAGE_LINGUAS', 1).split()))}"
do_rootfs[nostamp] = "1"
do_rootfs[dirs] = "${TOPDIR}"
diff --git a/meta/classes/image_types.bbclass b/meta/classes/image_types.bbclass
index 9549a9e3e..ea0d9a56d 100644
--- a/meta/classes/image_types.bbclass
+++ b/meta/classes/image_types.bbclass
@@ -1,8 +1,8 @@
def get_imagecmds(d):
cmds = "\n"
- old_overrides = bb.data.getVar('OVERRIDES', d, 0)
+ old_overrides = d.getVar('OVERRIDES', 0)
- types = bb.data.getVar('IMAGE_FSTYPES', d, True).split()
+ types = d.getVar('IMAGE_FSTYPES', True).split()
# Live images will be processed via inheriting bbclass and
# does not get processed here.
# live images also depend on ext3 so ensure its present
diff --git a/meta/classes/imagetest-qemu.bbclass b/meta/classes/imagetest-qemu.bbclass
index 4ea86c04f..de142bc0f 100644
--- a/meta/classes/imagetest-qemu.bbclass
+++ b/meta/classes/imagetest-qemu.bbclass
@@ -35,12 +35,12 @@ def qemuimagetest_main(d):
casestr = re.compile(r'(?P<scen>\w+\b):(?P<case>\S+$)')
resultstr = re.compile(r'\s*(?P<case>\w+)\s*(?P<pass>\d+)\s*(?P<fail>\d+)\s*(?P<noresult>\d+)')
- machine = bb.data.getVar('MACHINE', d, 1)
- pname = bb.data.getVar('PN', d, 1)
+ machine = d.getVar('MACHINE', 1)
+ pname = d.getVar('PN', 1)
"""function to save test cases running status"""
def teststatus(test, status, index, length):
- test_status = bb.data.getVar('TEST_STATUS', d, 1)
+ test_status = d.getVar('TEST_STATUS', 1)
if not os.path.exists(test_status):
raise bb.build.FuncFailed("No test status file existing under TEST_TMP")
@@ -51,30 +51,30 @@ def qemuimagetest_main(d):
"""funtion to run each case under scenario"""
def runtest(scen, case, fulltestpath):
- resultpath = bb.data.getVar('TEST_RESULT', d, 1)
- tmppath = bb.data.getVar('TEST_TMP', d, 1)
+ resultpath = d.getVar('TEST_RESULT', 1)
+ tmppath = d.getVar('TEST_TMP', 1)
"""initialize log file for testcase"""
- logpath = bb.data.getVar('TEST_LOG', d, 1)
+ logpath = d.getVar('TEST_LOG', 1)
bb.utils.mkdirhier("%s/%s" % (logpath, scen))
- caselog = os.path.join(logpath, "%s/log_%s.%s" % (scen, case, bb.data.getVar('DATETIME', d, 1)))
+ caselog = os.path.join(logpath, "%s/log_%s.%s" % (scen, case, d.getVar('DATETIME', 1)))
os.system("touch %s" % caselog)
"""export TEST_TMP, TEST_RESULT, DEPLOY_DIR and QEMUARCH"""
- os.environ["PATH"] = bb.data.getVar("PATH", d, True)
+ os.environ["PATH"] = d.getVar("PATH", True)
os.environ["TEST_TMP"] = tmppath
os.environ["TEST_RESULT"] = resultpath
- os.environ["DEPLOY_DIR"] = bb.data.getVar("DEPLOY_DIR", d, True)
+ os.environ["DEPLOY_DIR"] = d.getVar("DEPLOY_DIR", True)
os.environ["QEMUARCH"] = machine
os.environ["QEMUTARGET"] = pname
- os.environ["DISPLAY"] = bb.data.getVar("DISPLAY", d, True)
- os.environ["COREBASE"] = bb.data.getVar("COREBASE", d, True)
- os.environ["TOPDIR"] = bb.data.getVar("TOPDIR", d, True)
- os.environ["OE_TMPDIR"] = bb.data.getVar("TMPDIR", d, True)
- os.environ["TEST_STATUS"] = bb.data.getVar("TEST_STATUS", d, True)
- os.environ["TARGET_IPSAVE"] = bb.data.getVar("TARGET_IPSAVE", d, True)
- os.environ["TEST_SERIALIZE"] = bb.data.getVar("TEST_SERIALIZE", d, True)
- os.environ["SDK_NAME"] = bb.data.getVar("SDK_NAME", d, True)
+ os.environ["DISPLAY"] = d.getVar("DISPLAY", True)
+ os.environ["COREBASE"] = d.getVar("COREBASE", True)
+ os.environ["TOPDIR"] = d.getVar("TOPDIR", True)
+ os.environ["OE_TMPDIR"] = d.getVar("TMPDIR", True)
+ os.environ["TEST_STATUS"] = d.getVar("TEST_STATUS", True)
+ os.environ["TARGET_IPSAVE"] = d.getVar("TARGET_IPSAVE", True)
+ os.environ["TEST_SERIALIZE"] = d.getVar("TEST_SERIALIZE", True)
+ os.environ["SDK_NAME"] = d.getVar("SDK_NAME", True)
"""run Test Case"""
bb.note("Run %s test in scenario %s" % (case, scen))
@@ -92,13 +92,13 @@ def qemuimagetest_main(d):
if n:
item = n.group('scen')
casefile = n.group('case')
- for dir in bb.data.getVar("QEMUIMAGETESTS", d, True).split():
+ for dir in d.getVar("QEMUIMAGETESTS", True).split():
fulltestcase = os.path.join(dir, item, casefile)
if not os.path.isfile(fulltestcase):
raise bb.build.FuncFailed("Testcase %s not found" % fulltestcase)
list.append((item, casefile, fulltestcase))
else:
- for dir in bb.data.getVar("QEMUIMAGETESTS", d, True).split():
+ for dir in d.getVar("QEMUIMAGETESTS", True).split():
scenlist = os.path.join(dir, "scenario", machine, pname)
if not os.path.isfile(scenlist):
raise bb.build.FuncFailed("No scenario list file named %s found" % scenlist)
@@ -118,7 +118,7 @@ def qemuimagetest_main(d):
"""Clean tmp folder for testing"""
def clean_tmp():
- tmppath = bb.data.getVar('TEST_TMP', d, 1)
+ tmppath = d.getVar('TEST_TMP', 1)
if os.path.isdir(tmppath):
for f in os.listdir(tmppath):
@@ -132,28 +132,28 @@ def qemuimagetest_main(d):
clean_tmp()
"""check testcase folder and create test log folder"""
- testpath = bb.data.getVar('TEST_DIR', d, 1)
+ testpath = d.getVar('TEST_DIR', 1)
bb.utils.mkdirhier(testpath)
- logpath = bb.data.getVar('TEST_LOG', d, 1)
+ logpath = d.getVar('TEST_LOG', 1)
bb.utils.mkdirhier(logpath)
- tmppath = bb.data.getVar('TEST_TMP', d, 1)
+ tmppath = d.getVar('TEST_TMP', 1)
bb.utils.mkdirhier(tmppath)
"""initialize test status file"""
- test_status = bb.data.getVar('TEST_STATUS', d, 1)
+ test_status = d.getVar('TEST_STATUS', 1)
if os.path.exists(test_status):
os.remove(test_status)
os.system("touch %s" % test_status)
"""initialize result file"""
- resultpath = bb.data.getVar('TEST_RESULT', d, 1)
+ resultpath = d.getVar('TEST_RESULT', 1)
bb.utils.mkdirhier(resultpath)
- resultfile = os.path.join(resultpath, "testresult.%s" % bb.data.getVar('DATETIME', d, 1))
+ resultfile = os.path.join(resultpath, "testresult.%s" % d.getVar('DATETIME', 1))
sresultfile = os.path.join(resultpath, "testresult.log")
- machine = bb.data.getVar('MACHINE', d, 1)
+ machine = d.getVar('MACHINE', 1)
if os.path.exists(sresultfile):
os.remove(sresultfile)
@@ -165,7 +165,7 @@ def qemuimagetest_main(d):
f.close()
"""generate pre-defined testcase list"""
- testlist = bb.data.getVar('TEST_SCEN', d, 1)
+ testlist = d.getVar('TEST_SCEN', 1)
fulllist = generate_list(testlist)
"""Begin testing"""
diff --git a/meta/classes/insane.bbclass b/meta/classes/insane.bbclass
index 017f7bedc..a65f3ee86 100644
--- a/meta/classes/insane.bbclass
+++ b/meta/classes/insane.bbclass
@@ -105,7 +105,7 @@ ERROR_QA ?= "dev-so debug-deps dev-deps debug-files arch la2 pkgconfig la perms"
def package_qa_clean_path(path,d):
""" Remove the common prefix from the path. In this case it is the TMPDIR"""
- return path.replace(bb.data.getVar('TMPDIR',d,True),"")
+ return path.replace(d.getVar('TMPDIR',True),"")
def package_qa_write_error(error, d):
logfile = d.getVar('QA_LOGFILE', True)
@@ -132,13 +132,13 @@ def package_qa_check_rpath(file,name, d, elf, messages):
if not elf:
return
- scanelf = os.path.join(bb.data.getVar('STAGING_BINDIR_NATIVE',d,True),'scanelf')
- bad_dirs = [bb.data.getVar('TMPDIR', d, True) + "/work", bb.data.getVar('STAGING_DIR_TARGET', d, True)]
- bad_dir_test = bb.data.getVar('TMPDIR', d, True)
+ scanelf = os.path.join(d.getVar('STAGING_BINDIR_NATIVE',True),'scanelf')
+ bad_dirs = [d.getVar('TMPDIR', True) + "/work", d.getVar('STAGING_DIR_TARGET', True)]
+ bad_dir_test = d.getVar('TMPDIR', True)
if not os.path.exists(scanelf):
bb.fatal("Can not check RPATH, scanelf (part of pax-utils-native) not found")
- if not bad_dirs[0] in bb.data.getVar('WORKDIR', d, True):
+ if not bad_dirs[0] in d.getVar('WORKDIR', True):
bb.fatal("This class assumed that WORKDIR is ${TMPDIR}/work... Not doing any check")
output = os.popen("%s -B -F%%r#F '%s'" % (scanelf,file))
@@ -156,11 +156,11 @@ def package_qa_check_useless_rpaths(file, name, d, elf, messages):
if not elf:
return
- objdump = bb.data.getVar('OBJDUMP', d, True)
- env_path = bb.data.getVar('PATH', d, True)
+ objdump = d.getVar('OBJDUMP', True)
+ env_path = d.getVar('PATH', True)
- libdir = bb.data.getVar("libdir", d, True)
- base_libdir = bb.data.getVar("base_libdir", d, True)
+ libdir = d.getVar("libdir", True)
+ base_libdir = d.getVar("base_libdir", True)
import re
rpath_re = re.compile("\s+RPATH\s+(.*)")
@@ -209,8 +209,8 @@ def package_qa_check_arch(path,name,d, elf, messages):
if not elf:
return
- target_os = bb.data.getVar('TARGET_OS', d, True)
- target_arch = bb.data.getVar('TARGET_ARCH', d, True)
+ target_os = d.getVar('TARGET_OS', True)
+ target_arch = d.getVar('TARGET_ARCH', True)
# FIXME: Cross package confuse this check, so just skip them
for s in ['cross', 'nativesdk', 'cross-canadian']:
@@ -243,7 +243,7 @@ def package_qa_check_desktop(path, name, d, elf, messages):
Run all desktop files through desktop-file-validate.
"""
if path.endswith(".desktop"):
- desktop_file_validate = os.path.join(bb.data.getVar('STAGING_BINDIR_NATIVE',d,True),'desktop-file-validate')
+ desktop_file_validate = os.path.join(d.getVar('STAGING_BINDIR_NATIVE',True),'desktop-file-validate')
output = os.popen("%s %s" % (desktop_file_validate, path))
# This only produces output on errors
for l in output:
@@ -261,14 +261,14 @@ def package_qa_hash_style(path, name, d, elf, messages):
if os.path.islink(path):
return
- gnu_hash = "--hash-style=gnu" in bb.data.getVar('LDFLAGS', d, True)
+ gnu_hash = "--hash-style=gnu" in d.getVar('LDFLAGS', True)
if not gnu_hash:
- gnu_hash = "--hash-style=both" in bb.data.getVar('LDFLAGS', d, True)
+ gnu_hash = "--hash-style=both" in d.getVar('LDFLAGS', True)
if not gnu_hash:
return
- objdump = bb.data.getVar('OBJDUMP', d, True)
- env_path = bb.data.getVar('PATH', d, True)
+ objdump = d.getVar('OBJDUMP', True)
+ env_path = d.getVar('PATH', True)
sane = False
has_syms = False
@@ -299,7 +299,7 @@ def package_qa_check_buildpaths(path, name, d, elf, messages):
if os.path.islink(path):
return
- tmpdir = bb.data.getVar('TMPDIR', d, True)
+ tmpdir = d.getVar('TMPDIR', True)
file_content = open(path).read()
if tmpdir in file_content:
messages.append("File %s in package contained reference to tmpdir" % package_qa_clean_path(path,d))
@@ -311,9 +311,9 @@ def package_qa_check_license(workdir, d):
import tempfile
sane = True
- lic_files = bb.data.getVar('LIC_FILES_CHKSUM', d, True)
- lic = bb.data.getVar('LICENSE', d, True)
- pn = bb.data.getVar('PN', d, True)
+ lic_files = d.getVar('LIC_FILES_CHKSUM', True)
+ lic = d.getVar('LICENSE', True)
+ pn = d.getVar('PN', True)
if lic == "CLOSED":
return True
@@ -324,7 +324,7 @@ def package_qa_check_license(workdir, d):
bb.error(pn + ": Recipe file does not have license file information (LIC_FILES_CHKSUM)")
return False
- srcdir = bb.data.getVar('S', d, True)
+ srcdir = d.getVar('S', True)
for url in lic_files.split():
(type, host, path, user, pswd, parm) = bb.decodeurl(url)
@@ -384,7 +384,7 @@ def package_qa_check_staged(path,d):
"""
sane = True
- tmpdir = bb.data.getVar('TMPDIR', d, True)
+ tmpdir = d.getVar('TMPDIR', True)
workdir = os.path.join(tmpdir, "work")
installed = "installed=yes"
@@ -417,8 +417,8 @@ def package_qa_walk(path, warnfuncs, errorfuncs, skip, package, d):
import oe.qa
#if this will throw an exception, then fix the dict above
- target_os = bb.data.getVar('TARGET_OS', d, True)
- target_arch = bb.data.getVar('TARGET_ARCH', d, True)
+ target_os = d.getVar('TARGET_OS', True)
+ target_arch = d.getVar('TARGET_ARCH', True)
warnings = []
errors = []
@@ -457,19 +457,19 @@ def package_qa_check_rdepends(pkg, pkgdest, skip, d):
localdata = bb.data.createCopy(d)
root = "%s/%s" % (pkgdest, pkg)
- bb.data.setVar('ROOT', '', localdata)
- bb.data.setVar('ROOT_%s' % pkg, root, localdata)
- pkgname = bb.data.getVar('PKG_%s' % pkg, localdata, True)
+ localdata.setVar('ROOT', '')
+ localdata.setVar('ROOT_%s' % pkg, root)
+ pkgname = localdata.getVar('PKG_%s' % pkg, True)
if not pkgname:
pkgname = pkg
- bb.data.setVar('PKG', pkgname, localdata)
+ localdata.setVar('PKG', pkgname)
- bb.data.setVar('OVERRIDES', pkg, localdata)
+ localdata.setVar('OVERRIDES', pkg)
bb.data.update_data(localdata)
# Now check the RDEPENDS
- rdepends = bb.utils.explode_deps(bb.data.getVar('RDEPENDS', localdata, True) or "")
+ rdepends = bb.utils.explode_deps(localdata.getVar('RDEPENDS', True) or "")
# Now do the sanity check!!!
@@ -487,8 +487,8 @@ def package_qa_check_rdepends(pkg, pkgdest, skip, d):
python do_package_qa () {
bb.note("DO PACKAGE QA")
- logdir = bb.data.getVar('T', d, True)
- pkg = bb.data.getVar('PN', d, True)
+ logdir = d.getVar('T', True)
+ pkg = d.getVar('PN', True)
# Check the compile log for host contamination
compilelog = os.path.join(logdir,"log.do_compile")
@@ -508,8 +508,8 @@ python do_package_qa () {
(pkg, installlog))
# Scan the packages...
- pkgdest = bb.data.getVar('PKGDEST', d, True)
- packages = bb.data.getVar('PACKAGES',d, True)
+ pkgdest = d.getVar('PKGDEST', True)
+ packages = d.getVar('PACKAGES', True)
# no packages should be scanned
if not packages:
@@ -521,7 +521,7 @@ python do_package_qa () {
walk_sane = True
rdepends_sane = True
for package in packages.split():
- skip = (bb.data.getVar('INSANE_SKIP_' + package, d, True) or "").split()
+ skip = (d.getVar('INSANE_SKIP_' + package, True) or "").split()
if skip:
bb.note("Package %s skipping QA tests: %s" % (package, str(skip)))
warnchecks = []
@@ -560,7 +560,7 @@ python do_qa_staging() {
python do_qa_configure() {
configs = []
- workdir = bb.data.getVar('WORKDIR', d, True)
+ workdir = d.getVar('WORKDIR', True)
bb.note("Checking autotools environment for common misconfiguration")
for root, dirs, files in os.walk(workdir):
statement = "grep -e 'CROSS COMPILE Badness:' -e 'is unsafe for cross-compilation' %s > /dev/null" % \
@@ -575,8 +575,8 @@ Rerun configure task after fixing this. The path was '%s'""" % root)
if "configure.in" in files:
configs.append(os.path.join(root, "configure.in"))
- cnf = bb.data.getVar('EXTRA_OECONF', d, True) or ""
- if "gettext" not in bb.data.getVar('P', d, True) and "gcc-runtime" not in bb.data.getVar('P', d, True) and "--disable-nls" not in cnf:
+ cnf = d.getVar('EXTRA_OECONF', True) or ""
+ if "gettext" not in d.getVar('P', True) and "gcc-runtime" not in d.getVar('P', True) and "--disable-nls" not in cnf:
ml = d.getVar("MLPREFIX", True) or ""
if bb.data.inherits_class('native', d) or bb.data.inherits_class('cross', d) or bb.data.inherits_class('crosssdk', d) or bb.data.inherits_class('nativesdk', d):
gt = "gettext-native"
@@ -584,7 +584,7 @@ Rerun configure task after fixing this. The path was '%s'""" % root)
gt = "gettext-nativesdk"
else:
gt = "virtual/" + ml + "gettext"
- deps = bb.utils.explode_deps(bb.data.getVar('DEPENDS', d, True) or "")
+ deps = bb.utils.explode_deps(d.getVar('DEPENDS', True) or "")
if gt not in deps:
for config in configs:
gnu = "grep \"^[[:space:]]*AM_GNU_GETTEXT\" %s >/dev/null" % config
diff --git a/meta/classes/kernel-arch.bbclass b/meta/classes/kernel-arch.bbclass
index 225d5994f..869339511 100644
--- a/meta/classes/kernel-arch.bbclass
+++ b/meta/classes/kernel-arch.bbclass
@@ -18,7 +18,7 @@ valid_archs = "alpha cris ia64 \
def map_kernel_arch(a, d):
import re
- valid_archs = bb.data.getVar('valid_archs', d, 1).split()
+ valid_archs = d.getVar('valid_archs', 1).split()
if re.match('(i.86|athlon|x86.64)$', a): return 'x86'
elif re.match('arm26$', a): return 'arm26'
@@ -32,7 +32,7 @@ def map_kernel_arch(a, d):
else:
bb.error("cannot map '%s' to a linux kernel architecture" % a)
-export ARCH = "${@map_kernel_arch(bb.data.getVar('TARGET_ARCH', d, 1), d)}"
+export ARCH = "${@map_kernel_arch(d.getVar('TARGET_ARCH', 1), d)}"
def map_uboot_arch(a, d):
import re
@@ -41,5 +41,5 @@ def map_uboot_arch(a, d):
elif re.match('i.86$', a): return 'x86'
return a
-export UBOOT_ARCH = "${@map_uboot_arch(bb.data.getVar('ARCH', d, 1), d)}"
+export UBOOT_ARCH = "${@map_uboot_arch(d.getVar('ARCH', 1), d)}"
diff --git a/meta/classes/kernel-yocto.bbclass b/meta/classes/kernel-yocto.bbclass
index 2205686dd..d555fc04b 100644
--- a/meta/classes/kernel-yocto.bbclass
+++ b/meta/classes/kernel-yocto.bbclass
@@ -104,7 +104,7 @@ python do_kernel_configcheck() {
bb.plain("NOTE: validating kernel configuration")
- pathprefix = "export PATH=%s; " % bb.data.getVar('PATH', d, True)
+ pathprefix = "export PATH=%s; " % d.getVar('PATH', True)
cmd = bb.data.expand("cd ${B}/..; kconf_check -config- ${B} ${S} ${B} ${KBRANCH}",d )
ret, result = commands.getstatusoutput("%s%s" % (pathprefix, cmd))
diff --git a/meta/classes/kernel.bbclass b/meta/classes/kernel.bbclass
index dc711f287..12e0b83b9 100644
--- a/meta/classes/kernel.bbclass
+++ b/meta/classes/kernel.bbclass
@@ -11,15 +11,15 @@ INITRAMFS_IMAGE ?= ""
INITRAMFS_TASK ?= ""
python __anonymous () {
- kerneltype = bb.data.getVar('KERNEL_IMAGETYPE', d, 1) or ''
+ kerneltype = d.getVar('KERNEL_IMAGETYPE', 1) or ''
if kerneltype == 'uImage':
- depends = bb.data.getVar("DEPENDS", d, 1)
+ depends = d.getVar("DEPENDS", 1)
depends = "%s u-boot-mkimage-native" % depends
- bb.data.setVar("DEPENDS", depends, d)
+ d.setVar("DEPENDS", depends)
- image = bb.data.getVar('INITRAMFS_IMAGE', d, True)
+ image = d.getVar('INITRAMFS_IMAGE', True)
if image:
- bb.data.setVar('INITRAMFS_TASK', '${INITRAMFS_IMAGE}:do_rootfs', d)
+ d.setVar('INITRAMFS_TASK', '${INITRAMFS_IMAGE}:do_rootfs')
}
inherit kernel-arch deploy
@@ -31,7 +31,7 @@ PACKAGES_DYNAMIC += "kernel-firmware-*"
export OS = "${TARGET_OS}"
export CROSS_COMPILE = "${TARGET_PREFIX}"
-KERNEL_PRIORITY = "${@bb.data.getVar('PV',d,1).split('-')[0].split('.')[-1]}"
+KERNEL_PRIORITY = "${@d.getVar('PV',1).split('-')[0].split('.')[-1]}"
KERNEL_RELEASE ?= "${KERNEL_VERSION}"
@@ -56,7 +56,7 @@ KERNEL_IMAGEDEST = "boot"
#
# configuration
#
-export CMDLINE_CONSOLE = "console=${@bb.data.getVar("KERNEL_CONSOLE",d,1) or "ttyS0"}"
+export CMDLINE_CONSOLE = "console=${@d.getVar("KERNEL_CONSOLE",1) or "ttyS0"}"
KERNEL_VERSION = "${@get_kernelversion('${B}')}"
@@ -75,7 +75,7 @@ EXTRA_OEMAKE = ""
KERNEL_ALT_IMAGETYPE ??= ""
-KERNEL_IMAGETYPE_FOR_MAKE = "${@(lambda s: s[:-3] if s[-3:] == ".gz" else s)(bb.data.getVar('KERNEL_IMAGETYPE', d, 1))}"
+KERNEL_IMAGETYPE_FOR_MAKE = "${@(lambda s: s[:-3] if s[-3:] == ".gz" else s)(d.getVar('KERNEL_IMAGETYPE', 1))}"
kernel_do_compile() {
unset CFLAGS CPPFLAGS CXXFLAGS LDFLAGS MACHINE
@@ -306,10 +306,10 @@ module_conf_rfcomm = "alias bt-proto-3 rfcomm"
python populate_packages_prepend () {
def extract_modinfo(file):
import tempfile, re
- tempfile.tempdir = bb.data.getVar("WORKDIR", d, 1)
+ tempfile.tempdir = d.getVar("WORKDIR", 1)
tf = tempfile.mkstemp()
tmpfile = tf[1]
- cmd = "PATH=\"%s\" %sobjcopy -j .modinfo -O binary %s %s" % (bb.data.getVar("PATH", d, 1), bb.data.getVar("HOST_PREFIX", d, 1) or "", file, tmpfile)
+ cmd = "PATH=\"%s\" %sobjcopy -j .modinfo -O binary %s %s" % (d.getVar("PATH", 1), d.getVar("HOST_PREFIX", 1) or "", file, tmpfile)
os.system(cmd)
f = open(tmpfile)
l = f.read().split("\000")
@@ -328,18 +328,18 @@ python populate_packages_prepend () {
def parse_depmod():
import re
- dvar = bb.data.getVar('PKGD', d, 1)
+ dvar = d.getVar('PKGD', 1)
if not dvar:
bb.error("PKGD not defined")
return
- kernelver = bb.data.getVar('KERNEL_VERSION', d, 1)
+ kernelver = d.getVar('KERNEL_VERSION', 1)
kernelver_stripped = kernelver
m = re.match('^(.*-hh.*)[\.\+].*$', kernelver)
if m:
kernelver_stripped = m.group(1)
- path = bb.data.getVar("PATH", d, 1)
- host_prefix = bb.data.getVar("HOST_PREFIX", d, 1) or ""
+ path = d.getVar("PATH", 1)
+ host_prefix = d.getVar("HOST_PREFIX", 1) or ""
cmd = "PATH=\"%s\" %sdepmod -n -a -r -b %s -F %s/boot/System.map-%s %s" % (path, host_prefix, dvar, dvar, kernelver, kernelver_stripped)
f = os.popen(cmd, 'r')
@@ -377,9 +377,9 @@ python populate_packages_prepend () {
def get_dependencies(file, pattern, format):
# file no longer includes PKGD
- file = file.replace(bb.data.getVar('PKGD', d, 1) or '', '', 1)
+ file = file.replace(d.getVar('PKGD', 1) or '', '', 1)
# instead is prefixed with /lib/modules/${KERNEL_VERSION}
- file = file.replace("/lib/modules/%s/" % bb.data.getVar('KERNEL_VERSION', d, 1) or '', '', 1)
+ file = file.replace("/lib/modules/%s/" % d.getVar('KERNEL_VERSION', 1) or '', '', 1)
if module_deps.has_key(file):
import re
@@ -398,40 +398,40 @@ python populate_packages_prepend () {
import re
vals = extract_modinfo(file)
- dvar = bb.data.getVar('PKGD', d, 1)
+ dvar = d.getVar('PKGD', 1)
# If autoloading is requested, output /etc/modutils/<name> and append
# appropriate modprobe commands to the postinst
- autoload = bb.data.getVar('module_autoload_%s' % basename, d, 1)
+ autoload = d.getVar('module_autoload_%s' % basename, 1)
if autoload:
name = '%s/etc/modutils/%s' % (dvar, basename)
f = open(name, 'w')
for m in autoload.split():
f.write('%s\n' % m)
f.close()
- postinst = bb.data.getVar('pkg_postinst_%s' % pkg, d, 1)
+ postinst = d.getVar('pkg_postinst_%s' % pkg, 1)
if not postinst:
bb.fatal("pkg_postinst_%s not defined" % pkg)
- postinst += bb.data.getVar('autoload_postinst_fragment', d, 1) % autoload
- bb.data.setVar('pkg_postinst_%s' % pkg, postinst, d)
+ postinst += d.getVar('autoload_postinst_fragment', 1) % autoload
+ d.setVar('pkg_postinst_%s' % pkg, postinst)
# Write out any modconf fragment
- modconf = bb.data.getVar('module_conf_%s' % basename, d, 1)
+ modconf = d.getVar('module_conf_%s' % basename, 1)
if modconf:
name = '%s/etc/modprobe.d/%s.conf' % (dvar, basename)
f = open(name, 'w')
f.write("%s\n" % modconf)
f.close()
- files = bb.data.getVar('FILES_%s' % pkg, d, 1)
+ files = d.getVar('FILES_%s' % pkg, 1)
files = "%s /etc/modutils/%s /etc/modutils/%s.conf /etc/modprobe.d/%s.conf" % (files, basename, basename, basename)
- bb.data.setVar('FILES_%s' % pkg, files, d)
+ d.setVar('FILES_%s' % pkg, files)
if vals.has_key("description"):
- old_desc = bb.data.getVar('DESCRIPTION_' + pkg, d, 1) or ""
- bb.data.setVar('DESCRIPTION_' + pkg, old_desc + "; " + vals["description"], d)
+ old_desc = d.getVar('DESCRIPTION_' + pkg, 1) or ""
+ d.setVar('DESCRIPTION_' + pkg, old_desc + "; " + vals["description"])
- rdepends_str = bb.data.getVar('RDEPENDS_' + pkg, d, 1)
+ rdepends_str = d.getVar('RDEPENDS_' + pkg, 1)
if rdepends_str:
rdepends = rdepends_str.split()
else:
@@ -443,29 +443,29 @@ python populate_packages_prepend () {
module_regex = '^(.*)\.k?o$'
module_pattern = 'kernel-module-%s'
- postinst = bb.data.getVar('pkg_postinst_modules', d, 1)
- postrm = bb.data.getVar('pkg_postrm_modules', d, 1)
+ postinst = d.getVar('pkg_postinst_modules', 1)
+ postrm = d.getVar('pkg_postrm_modules', 1)
do_split_packages(d, root='/lib/firmware', file_regex='^(.*)\.bin$', output_pattern='kernel-firmware-%s', description='Firmware for %s', recursive=True, extra_depends='')
do_split_packages(d, root='/lib/firmware', file_regex='^(.*)\.fw$', output_pattern='kernel-firmware-%s', description='Firmware for %s', recursive=True, extra_depends='')
do_split_packages(d, root='/lib/firmware', file_regex='^(.*)\.cis$', output_pattern='kernel-firmware-%s', description='Firmware for %s', recursive=True, extra_depends='')
- do_split_packages(d, root='/lib/modules', file_regex=module_regex, output_pattern=module_pattern, description='%s kernel module', postinst=postinst, postrm=postrm, recursive=True, hook=frob_metadata, extra_depends='update-modules kernel-%s' % bb.data.getVar("KERNEL_VERSION", d, 1))
+ do_split_packages(d, root='/lib/modules', file_regex=module_regex, output_pattern=module_pattern, description='%s kernel module', postinst=postinst, postrm=postrm, recursive=True, hook=frob_metadata, extra_depends='update-modules kernel-%s' % d.getVar("KERNEL_VERSION", 1))
import re
metapkg = "kernel-modules"
- bb.data.setVar('ALLOW_EMPTY_' + metapkg, "1", d)
- bb.data.setVar('FILES_' + metapkg, "", d)
+ d.setVar('ALLOW_EMPTY_' + metapkg, "1")
+ d.setVar('FILES_' + metapkg, "")
blacklist = [ 'kernel-dev', 'kernel-image', 'kernel-base', 'kernel-vmlinux', 'perf', 'perf-dbg', 'kernel-misc' ]
for l in module_deps.values():
for i in l:
pkg = module_pattern % legitimize_package_name(re.match(module_regex, os.path.basename(i)).group(1))
blacklist.append(pkg)
metapkg_rdepends = []
- packages = bb.data.getVar('PACKAGES', d, 1).split()
+ packages = d.getVar('PACKAGES', 1).split()
for pkg in packages[1:]:
if not pkg in blacklist and not pkg in metapkg_rdepends:
metapkg_rdepends.append(pkg)
bb.data.setVar('RDEPENDS_' + metapkg, ' '.join(metapkg_rdepends), d)
- bb.data.setVar('DESCRIPTION_' + metapkg, 'Kernel modules meta package', d)
+ d.setVar('DESCRIPTION_' + metapkg, 'Kernel modules meta package')
packages.append(metapkg)
bb.data.setVar('PACKAGES', ' '.join(packages), d)
}
diff --git a/meta/classes/libc-common.bbclass b/meta/classes/libc-common.bbclass
index 0ee9a55b0..ec33762a2 100644
--- a/meta/classes/libc-common.bbclass
+++ b/meta/classes/libc-common.bbclass
@@ -18,13 +18,13 @@ do_install() {
}
def get_libc_fpu_setting(bb, d):
- if bb.data.getVar('TARGET_FPU', d, 1) in [ 'soft' ]:
+ if d.getVar('TARGET_FPU', 1) in [ 'soft' ]:
return "--without-fp"
return ""
python populate_packages_prepend () {
- if bb.data.getVar('DEBIAN_NAMES', d, 1):
- bpn = bb.data.getVar('BPN', d, 1)
- bb.data.setVar('PKG_'+bpn, 'libc6', d)
- bb.data.setVar('PKG_'+bpn+'-dev', 'libc6-dev', d)
+ if d.getVar('DEBIAN_NAMES', 1):
+ bpn = d.getVar('BPN', 1)
+ d.setVar('PKG_'+bpn, 'libc6')
+ d.setVar('PKG_'+bpn+'-dev', 'libc6-dev')
}
diff --git a/meta/classes/libc-package.bbclass b/meta/classes/libc-package.bbclass
index 6ef2f9725..fc1a5794d 100644
--- a/meta/classes/libc-package.bbclass
+++ b/meta/classes/libc-package.bbclass
@@ -10,7 +10,7 @@
GLIBC_INTERNAL_USE_BINARY_LOCALE ?= "ondevice"
python __anonymous () {
- enabled = bb.data.getVar("ENABLE_BINARY_LOCALE_GENERATION", d, 1)
+ enabled = d.getVar("ENABLE_BINARY_LOCALE_GENERATION", 1)
pn = d.getVar("PN", True)
if pn.endswith("-initial"):
@@ -19,21 +19,21 @@ python __anonymous () {
if enabled and int(enabled):
import re
- target_arch = bb.data.getVar("TARGET_ARCH", d, 1)
- binary_arches = bb.data.getVar("BINARY_LOCALE_ARCHES", d, 1) or ""
- use_cross_localedef = bb.data.getVar("LOCALE_GENERATION_WITH_CROSS-LOCALEDEF", d, 1) or ""
+ target_arch = d.getVar("TARGET_ARCH", 1)
+ binary_arches = d.getVar("BINARY_LOCALE_ARCHES", 1) or ""
+ use_cross_localedef = d.getVar("LOCALE_GENERATION_WITH_CROSS-LOCALEDEF", 1) or ""
for regexp in binary_arches.split(" "):
r = re.compile(regexp)
if r.match(target_arch):
- depends = bb.data.getVar("DEPENDS", d, 1)
+ depends = d.getVar("DEPENDS", 1)
if use_cross_localedef == "1" :
depends = "%s cross-localedef-native" % depends
else:
depends = "%s qemu-native" % depends
- bb.data.setVar("DEPENDS", depends, d)
- bb.data.setVar("GLIBC_INTERNAL_USE_BINARY_LOCALE", "compile", d)
+ d.setVar("DEPENDS", depends)
+ d.setVar("GLIBC_INTERNAL_USE_BINARY_LOCALE", "compile")
break
}
@@ -109,19 +109,19 @@ inherit qemu
python package_do_split_gconvs () {
import os, re
- if (bb.data.getVar('PACKAGE_NO_GCONV', d, 1) == '1'):
+ if (d.getVar('PACKAGE_NO_GCONV', 1) == '1'):
bb.note("package requested not splitting gconvs")
return
- if not bb.data.getVar('PACKAGES', d, 1):
+ if not d.getVar('PACKAGES', 1):
return
- bpn = bb.data.getVar('BPN', d, 1)
- libdir = bb.data.getVar('libdir', d, 1)
+ bpn = d.getVar('BPN', 1)
+ libdir = d.getVar('libdir', 1)
if not libdir:
bb.error("libdir not defined")
return
- datadir = bb.data.getVar('datadir', d, 1)
+ datadir = d.getVar('datadir', 1)
if not datadir:
bb.error("datadir not defined")
return
@@ -191,17 +191,17 @@ python package_do_split_gconvs () {
do_split_packages(d, locales_dir, file_regex='(.*)', output_pattern=bpn+'-localedata-%s', \
description='locale definition for %s', hook=calc_locale_deps, extra_depends='')
- bb.data.setVar('PACKAGES', bb.data.getVar('PACKAGES', d) + ' ' + bb.data.getVar('MLPREFIX', d) + bpn + '-gconv', d)
+ bb.data.setVar('PACKAGES', d.getVar('PACKAGES') + ' ' + d.getVar('MLPREFIX') + bpn + '-gconv', d)
- use_bin = bb.data.getVar("GLIBC_INTERNAL_USE_BINARY_LOCALE", d, 1)
+ use_bin = d.getVar("GLIBC_INTERNAL_USE_BINARY_LOCALE", 1)
dot_re = re.compile("(.*)\.(.*)")
#GLIBC_GENERATE_LOCALES var specifies which locales to be supported, empty or "all" means all locales
if use_bin != "precompiled":
- supported = bb.data.getVar('GLIBC_GENERATE_LOCALES', d, 1)
+ supported = d.getVar('GLIBC_GENERATE_LOCALES', 1)
if not supported or supported == "all":
- f = open(base_path_join(bb.data.getVar('WORKDIR', d, 1), "SUPPORTED"), "r")
+ f = open(base_path_join(d.getVar('WORKDIR', 1), "SUPPORTED"), "r")
supported = f.readlines()
f.close()
else:
@@ -209,7 +209,7 @@ python package_do_split_gconvs () {
supported = map(lambda s:s.replace(".", " ") + "\n", supported)
else:
supported = []
- full_bin_path = bb.data.getVar('PKGD', d, True) + binary_locales_dir
+ full_bin_path = d.getVar('PKGD', True) + binary_locales_dir
for dir in os.listdir(full_bin_path):
dbase = dir.split(".")
d2 = " "
@@ -218,7 +218,7 @@ python package_do_split_gconvs () {
supported.append(dbase[0] + d2)
# Collate the locales by base and encoding
- utf8_only = int(bb.data.getVar('LOCALE_UTF8_ONLY', d, 1) or 0)
+ utf8_only = int(d.getVar('LOCALE_UTF8_ONLY', 1) or 0)
encodings = {}
for l in supported:
l = l[:-1]
@@ -235,9 +235,9 @@ python package_do_split_gconvs () {
def output_locale_source(name, pkgname, locale, encoding):
bb.data.setVar('RDEPENDS_%s' % pkgname, 'localedef %s-localedata-%s %s-charmap-%s' % \
(bpn, legitimize_package_name(locale), bpn, legitimize_package_name(encoding)), d)
- bb.data.setVar('pkg_postinst_%s' % pkgname, bb.data.getVar('locale_base_postinst', d, 1) \
+ bb.data.setVar('pkg_postinst_%s' % pkgname, d.getVar('locale_base_postinst', 1) \
% (locale, encoding, locale), d)
- bb.data.setVar('pkg_postrm_%s' % pkgname, bb.data.getVar('locale_base_postrm', d, 1) % \
+ bb.data.setVar('pkg_postrm_%s' % pkgname, d.getVar('locale_base_postrm', 1) % \
(locale, encoding, locale), d)
def output_locale_binary_rdepends(name, pkgname, locale, encoding):
@@ -248,23 +248,23 @@ python package_do_split_gconvs () {
libc_name = name
bb.data.setVar('RDEPENDS_%s' % pkgname, legitimize_package_name('%s-binary-localedata-%s' \
% (bpn, libc_name)), d)
- rprovides = (bb.data.getVar('RPROVIDES_%s' % pkgname, d, True) or "").split()
+ rprovides = (d.getVar('RPROVIDES_%s' % pkgname, True) or "").split()
rprovides.append(legitimize_package_name('%s-binary-localedata-%s' % (bpn, libc_name)))
bb.data.setVar('RPROVIDES_%s' % pkgname, " ".join(rprovides), d)
commands = {}
def output_locale_binary(name, pkgname, locale, encoding):
- treedir = base_path_join(bb.data.getVar("WORKDIR", d, 1), "locale-tree")
- ldlibdir = base_path_join(treedir, bb.data.getVar("base_libdir", d, 1))
- path = bb.data.getVar("PATH", d, 1)
+ treedir = base_path_join(d.getVar("WORKDIR", 1), "locale-tree")
+ ldlibdir = base_path_join(treedir, d.getVar("base_libdir", 1))
+ path = d.getVar("PATH", 1)
i18npath = base_path_join(treedir, datadir, "i18n")
gconvpath = base_path_join(treedir, "iconvdata")
outputpath = base_path_join(treedir, libdir, "locale")
- use_cross_localedef = bb.data.getVar("LOCALE_GENERATION_WITH_CROSS-LOCALEDEF", d, 1) or "0"
+ use_cross_localedef = d.getVar("LOCALE_GENERATION_WITH_CROSS-LOCALEDEF", 1) or "0"
if use_cross_localedef == "1":
- target_arch = bb.data.getVar('TARGET_ARCH', d, True)
+ target_arch = d.getVar('TARGET_ARCH', True)
locale_arch_options = { \
"arm": " --uint32-align=4 --little-endian ", \
"powerpc": " --uint32-align=4 --big-endian ", \
@@ -292,9 +292,9 @@ python package_do_split_gconvs () {
--inputfile=%s/i18n/locales/%s --charmap=%s %s" \
% (treedir, datadir, locale, encoding, name)
- qemu_options = bb.data.getVar("QEMU_OPTIONS_%s" % bb.data.getVar('PACKAGE_ARCH', d, 1), d, 1)
+ qemu_options = bb.data.getVar("QEMU_OPTIONS_%s" % d.getVar('PACKAGE_ARCH', 1), d, 1)
if not qemu_options:
- qemu_options = bb.data.getVar('QEMU_OPTIONS', d, 1)
+ qemu_options = d.getVar('QEMU_OPTIONS', 1)
cmd = "PSEUDO_RELOADED=YES PATH=\"%s\" I18NPATH=\"%s\" %s -L %s \
-E LD_LIBRARY_PATH=%s %s %s/bin/localedef %s" % \
@@ -305,14 +305,14 @@ python package_do_split_gconvs () {
bb.note("generating locale %s (%s)" % (locale, encoding))
def output_locale(name, locale, encoding):
- pkgname = bb.data.getVar('MLPREFIX', d) + 'locale-base-' + legitimize_package_name(name)
- bb.data.setVar('ALLOW_EMPTY_%s' % pkgname, '1', d)
- bb.data.setVar('PACKAGES', '%s %s' % (pkgname, bb.data.getVar('PACKAGES', d, 1)), d)
+ pkgname = d.getVar('MLPREFIX') + 'locale-base-' + legitimize_package_name(name)
+ d.setVar('ALLOW_EMPTY_%s' % pkgname, '1')
+ bb.data.setVar('PACKAGES', '%s %s' % (pkgname, d.getVar('PACKAGES', 1)), d)
rprovides = ' virtual-locale-%s' % legitimize_package_name(name)
m = re.match("(.*)_(.*)", name)
if m:
rprovides += ' virtual-locale-%s' % m.group(1)
- bb.data.setVar('RPROVIDES_%s' % pkgname, rprovides, d)
+ d.setVar('RPROVIDES_%s' % pkgname, rprovides)
if use_bin == "compile":
output_locale_binary_rdepends(name, pkgname, locale, encoding)
@@ -347,7 +347,7 @@ python package_do_split_gconvs () {
bb.note(" " + " ".join(non_utf8))
if use_bin == "compile":
- makefile = base_path_join(bb.data.getVar("WORKDIR", d, 1), "locale-tree", "Makefile")
+ makefile = base_path_join(d.getVar("WORKDIR", 1), "locale-tree", "Makefile")
m = open(makefile, "w")
m.write("all: %s\n\n" % " ".join(commands.keys()))
for cmd in commands:
diff --git a/meta/classes/license.bbclass b/meta/classes/license.bbclass
index baf35f00c..4d036b171 100644
--- a/meta/classes/license.bbclass
+++ b/meta/classes/license.bbclass
@@ -104,10 +104,10 @@ python do_populate_lic() {
# If the generic does not exist we need to check to see if there is an SPDX mapping to it
if not os.path.isfile(os.path.join(generic_directory, license_type)):
- if bb.data.getVarFlag('SPDXLICENSEMAP', license_type, d) != None:
+ if d.getVarFlag('SPDXLICENSEMAP', license_type) != None:
# Great, there is an SPDXLICENSEMAP. We can copy!
bb.note("We need to use a SPDXLICENSEMAP for %s" % (license_type))
- spdx_generic = bb.data.getVarFlag('SPDXLICENSEMAP', license_type, d)
+ spdx_generic = d.getVarFlag('SPDXLICENSEMAP', license_type)
copy_license(generic_directory, gen_lic_dest, spdx_generic)
link_license(gen_lic_dest, destdir, spdx_generic)
else:
@@ -120,16 +120,16 @@ python do_populate_lic() {
link_license(gen_lic_dest, destdir, license_type)
# All the license types for the package
- license_types = bb.data.getVar('LICENSE', d, True)
+ license_types = d.getVar('LICENSE', True)
# All the license files for the package
- lic_files = bb.data.getVar('LIC_FILES_CHKSUM', d, True)
- pn = bb.data.getVar('PN', d, True)
+ lic_files = d.getVar('LIC_FILES_CHKSUM', True)
+ pn = d.getVar('PN', True)
# The base directory we wrangle licenses to
- destdir = os.path.join(bb.data.getVar('LICSSTATEDIR', d, True), pn)
+ destdir = os.path.join(d.getVar('LICSSTATEDIR', True), pn)
# The license files are located in S/LIC_FILE_CHECKSUM.
- srcdir = bb.data.getVar('S', d, True)
+ srcdir = d.getVar('S', True)
# Directory we store the generic licenses as set in the distro configuration
- generic_directory = bb.data.getVar('COMMON_LICENSE_DIR', d, True)
+ generic_directory = d.getVar('COMMON_LICENSE_DIR', True)
try:
bb.mkdirhier(destdir)
@@ -154,7 +154,7 @@ python do_populate_lic() {
if ret is False or ret == 0:
bb.warn("%s could not be copied for some reason. It may not exist. WARN for now." % srclicfile)
- gen_lic_dest = os.path.join(bb.data.getVar('LICENSE_DIRECTORY', d, True), "common-licenses")
+ gen_lic_dest = os.path.join(d.getVar('LICENSE_DIRECTORY', True), "common-licenses")
clean_licenses = ""
diff --git a/meta/classes/metadata_scm.bbclass b/meta/classes/metadata_scm.bbclass
index 99ac05c18..57609ef8c 100644
--- a/meta/classes/metadata_scm.bbclass
+++ b/meta/classes/metadata_scm.bbclass
@@ -27,7 +27,7 @@ def base_detect_branch(d):
return "<unknown>"
def base_get_scmbasepath(d):
- return bb.data.getVar( 'COREBASE', d, 1 )
+ return d.getVar( 'COREBASE', 1 )
def base_get_metadata_monotone_branch(path, d):
monotone_branch = "<unknown>"
diff --git a/meta/classes/native.bbclass b/meta/classes/native.bbclass
index ba8b0bf25..5f25bdd2c 100644
--- a/meta/classes/native.bbclass
+++ b/meta/classes/native.bbclass
@@ -98,18 +98,18 @@ python native_virtclass_handler () {
if not isinstance(e, bb.event.RecipePreFinalise):
return
- classextend = bb.data.getVar('BBCLASSEXTEND', e.data, True) or ""
+ classextend = e.data.getVar('BBCLASSEXTEND', True) or ""
if "native" not in classextend:
return
- pn = bb.data.getVar("PN", e.data, True)
+ pn = e.data.getVar("PN", True)
if not pn.endswith("-native"):
return
def map_dependencies(varname, d, suffix = ""):
if suffix:
varname = varname + "_" + suffix
- deps = bb.data.getVar(varname, d, True)
+ deps = d.getVar(varname, True)
if not deps:
return
deps = bb.utils.explode_deps(deps)
@@ -131,15 +131,15 @@ python native_virtclass_handler () {
map_dependencies("RPROVIDES", e.data, pkg)
map_dependencies("RREPLACES", e.data, pkg)
- provides = bb.data.getVar("PROVIDES", e.data, True)
+ provides = e.data.getVar("PROVIDES", True)
for prov in provides.split():
if prov.find(pn) != -1:
continue
if not prov.endswith("-native"):
provides = provides.replace(prov, prov + "-native")
- bb.data.setVar("PROVIDES", provides, e.data)
+ e.data.setVar("PROVIDES", provides)
- bb.data.setVar("OVERRIDES", bb.data.getVar("OVERRIDES", e.data, False) + ":virtclass-native", e.data)
+ bb.data.setVar("OVERRIDES", e.data.getVar("OVERRIDES", False) + ":virtclass-native", e.data)
}
addhandler native_virtclass_handler
diff --git a/meta/classes/nativesdk.bbclass b/meta/classes/nativesdk.bbclass
index bb59ac57a..ca24efaa7 100644
--- a/meta/classes/nativesdk.bbclass
+++ b/meta/classes/nativesdk.bbclass
@@ -11,7 +11,7 @@ STAGING_BINDIR_TOOLCHAIN = "${STAGING_DIR_NATIVE}${bindir_native}/${SDK_ARCH}${S
#
PACKAGE_ARCH = "${SDK_ARCH}-nativesdk"
python () {
- archs = bb.data.getVar('PACKAGE_ARCHS', d, True).split()
+ archs = d.getVar('PACKAGE_ARCHS', True).split()
sdkarchs = []
for arch in archs:
sdkarchs.append(arch + '-nativesdk')
@@ -62,22 +62,22 @@ python nativesdk_virtclass_handler () {
if not isinstance(e, bb.event.RecipePreFinalise):
return
- pn = bb.data.getVar("PN", e.data, True)
+ pn = e.data.getVar("PN", True)
if not pn.endswith("-nativesdk"):
return
- bb.data.setVar("OVERRIDES", bb.data.getVar("OVERRIDES", e.data, False) + ":virtclass-nativesdk", e.data)
+ bb.data.setVar("OVERRIDES", e.data.getVar("OVERRIDES", False) + ":virtclass-nativesdk", e.data)
}
python () {
- pn = bb.data.getVar("PN", d, True)
+ pn = d.getVar("PN", True)
if not pn.endswith("-nativesdk"):
return
def map_dependencies(varname, d, suffix = ""):
if suffix:
varname = varname + "_" + suffix
- deps = bb.data.getVar(varname, d, True)
+ deps = d.getVar(varname, True)
if not deps:
return
deps = bb.utils.explode_deps(deps)
@@ -101,13 +101,13 @@ python () {
# map_dependencies("RPROVIDES", d, pkg)
# map_dependencies("RREPLACES", d, pkg)
- provides = bb.data.getVar("PROVIDES", d, True)
+ provides = d.getVar("PROVIDES", True)
for prov in provides.split():
if prov.find(pn) != -1:
continue
if not prov.endswith("-nativesdk"):
provides = provides.replace(prov, prov + "-nativesdk")
- bb.data.setVar("PROVIDES", provides, d)
+ d.setVar("PROVIDES", provides)
}
addhandler nativesdk_virtclass_handler
diff --git a/meta/classes/package.bbclass b/meta/classes/package.bbclass
index 0e1d8dbfc..d01275f33 100644
--- a/meta/classes/package.bbclass
+++ b/meta/classes/package.bbclass
@@ -88,9 +88,9 @@ def do_split_packages(d, root, file_regex, output_pattern, description, postinst
if newdeps:
extra_depends = " ".join(newdeps)
- dvar = bb.data.getVar('PKGD', d, True)
+ dvar = d.getVar('PKGD', True)
- packages = bb.data.getVar('PACKAGES', d, True).split()
+ packages = d.getVar('PACKAGES', True).split()
if postinst:
postinst = '#!/bin/sh\n' + postinst + '\n'
@@ -136,7 +136,7 @@ def do_split_packages(d, root, file_regex, output_pattern, description, postinst
packages = [pkg] + packages
else:
packages.append(pkg)
- oldfiles = bb.data.getVar('FILES_' + pkg, d, True)
+ oldfiles = d.getVar('FILES_' + pkg, True)
if not oldfiles:
the_files = [os.path.join(root, o)]
if aux_files_pattern:
@@ -153,17 +153,17 @@ def do_split_packages(d, root, file_regex, output_pattern, description, postinst
the_files.append(aux_files_pattern_verbatim % m.group(1))
bb.data.setVar('FILES_' + pkg, " ".join(the_files), d)
if extra_depends != '':
- the_depends = bb.data.getVar('RDEPENDS_' + pkg, d, True)
+ the_depends = d.getVar('RDEPENDS_' + pkg, True)
if the_depends:
the_depends = '%s %s' % (the_depends, extra_depends)
else:
the_depends = extra_depends
- bb.data.setVar('RDEPENDS_' + pkg, the_depends, d)
- bb.data.setVar('DESCRIPTION_' + pkg, description % on, d)
+ d.setVar('RDEPENDS_' + pkg, the_depends)
+ d.setVar('DESCRIPTION_' + pkg, description % on)
if postinst:
- bb.data.setVar('pkg_postinst_' + pkg, postinst, d)
+ d.setVar('pkg_postinst_' + pkg, postinst)
if postrm:
- bb.data.setVar('pkg_postrm_' + pkg, postrm, d)
+ d.setVar('pkg_postrm_' + pkg, postrm)
else:
bb.data.setVar('FILES_' + pkg, oldfiles + " " + os.path.join(root, o), d)
if callable(hook):
@@ -174,13 +174,13 @@ def do_split_packages(d, root, file_regex, output_pattern, description, postinst
PACKAGE_DEPENDS += "file-native"
python () {
- if bb.data.getVar('PACKAGES', d, True) != '':
- deps = bb.data.getVarFlag('do_package', 'depends', d) or ""
- for dep in (bb.data.getVar('PACKAGE_DEPENDS', d, True) or "").split():
+ if d.getVar('PACKAGES', True) != '':
+ deps = d.getVarFlag('do_package', 'depends') or ""
+ for dep in (d.getVar('PACKAGE_DEPENDS', True) or "").split():
deps += " %s:do_populate_sysroot" % dep
- bb.data.setVarFlag('do_package', 'depends', deps, d)
+ d.setVarFlag('do_package', 'depends', deps)
- deps = (bb.data.getVarFlag('do_package', 'deptask', d) or "").split()
+ deps = (d.getVarFlag('do_package', 'deptask') or "").split()
# shlibs requires any DEPENDS to have already packaged for the *.list files
deps.append("do_package")
bb.data.setVarFlag('do_package', 'deptask', " ".join(deps), d)
@@ -198,9 +198,9 @@ def splitfile(file, debugfile, debugsrcdir, d):
import commands, stat
- dvar = bb.data.getVar('PKGD', d, True)
- pathprefix = "export PATH=%s; " % bb.data.getVar('PATH', d, True)
- objcopy = bb.data.getVar("OBJCOPY", d, True)
+ dvar = d.getVar('PKGD', True)
+ pathprefix = "export PATH=%s; " % d.getVar('PATH', True)
+ objcopy = d.getVar("OBJCOPY", True)
debugedit = bb.data.expand("${STAGING_LIBDIR_NATIVE}/rpm/bin/debugedit", d)
workdir = bb.data.expand("${WORKDIR}", d)
workparentdir = os.path.dirname(workdir)
@@ -240,10 +240,10 @@ def splitfile2(debugsrcdir, d):
import commands, stat
- dvar = bb.data.getVar('PKGD', d, True)
- pathprefix = "export PATH=%s; " % bb.data.getVar('PATH', d, True)
- strip = bb.data.getVar("STRIP", d, True)
- objcopy = bb.data.getVar("OBJCOPY", d, True)
+ dvar = d.getVar('PKGD', True)
+ pathprefix = "export PATH=%s; " % d.getVar('PATH', True)
+ strip = d.getVar("STRIP", True)
+ objcopy = d.getVar("OBJCOPY", True)
debugedit = bb.data.expand("${STAGING_LIBDIR_NATIVE}/rpm/bin/debugedit", d)
workdir = bb.data.expand("${WORKDIR}", d)
workparentdir = os.path.dirname(workdir)
@@ -279,8 +279,8 @@ def runstrip(file, elftype, d):
import commands, stat
- pathprefix = "export PATH=%s; " % bb.data.getVar('PATH', d, True)
- strip = bb.data.getVar("STRIP", d, True)
+ pathprefix = "export PATH=%s; " % d.getVar('PATH', True)
+ strip = d.getVar("STRIP", True)
# Handle kernel modules specifically - .debug directories here are pointless
if file.find("/lib/modules/") != -1 and file.endswith(".ko"):
@@ -329,10 +329,10 @@ def get_package_mapping (pkg, d):
return pkg
def runtime_mapping_rename (varname, d):
- #bb.note("%s before: %s" % (varname, bb.data.getVar(varname, d, True)))
+ #bb.note("%s before: %s" % (varname, d.getVar(varname, True)))
new_depends = []
- deps = bb.utils.explode_dep_versions(bb.data.getVar(varname, d, True) or "")
+ deps = bb.utils.explode_dep_versions(d.getVar(varname, True) or "")
for depend in deps:
# Have to be careful with any version component of the depend
new_depend = get_package_mapping(depend, d)
@@ -343,7 +343,7 @@ def runtime_mapping_rename (varname, d):
bb.data.setVar(varname, " ".join(new_depends) or None, d)
- #bb.note("%s after: %s" % (varname, bb.data.getVar(varname, d, True)))
+ #bb.note("%s after: %s" % (varname, d.getVar(varname, True)))
#
# Package functions suitable for inclusion in PACKAGEFUNCS
@@ -359,19 +359,19 @@ python package_get_auto_pr() {
}
python package_do_split_locales() {
- if (bb.data.getVar('PACKAGE_NO_LOCALE', d, True) == '1'):
+ if (d.getVar('PACKAGE_NO_LOCALE', True) == '1'):
bb.debug(1, "package requested not splitting locales")
return
- packages = (bb.data.getVar('PACKAGES', d, True) or "").split()
+ packages = (d.getVar('PACKAGES', True) or "").split()
- datadir = bb.data.getVar('datadir', d, True)
+ datadir = d.getVar('datadir', True)
if not datadir:
bb.note("datadir not defined")
return
- dvar = bb.data.getVar('PKGD', d, True)
- pn = bb.data.getVar('PN', d, True)
+ dvar = d.getVar('PKGD', True)
+ pn = d.getVar('PN', True)
if pn + '-locale' in packages:
packages.remove(pn + '-locale')
@@ -392,9 +392,9 @@ python package_do_split_locales() {
if mainpkg.find('-dev'):
mainpkg = mainpkg.replace('-dev', '')
- summary = bb.data.getVar('SUMMARY', d, True) or pn
- description = bb.data.getVar('DESCRIPTION', d, True) or ""
- locale_section = bb.data.getVar('LOCALE_SECTION', d, True)
+ summary = d.getVar('SUMMARY', True) or pn
+ description = d.getVar('DESCRIPTION', True) or ""
+ locale_section = d.getVar('LOCALE_SECTION', True)
for l in locales:
ln = legitimize_package_name(l)
pkg = pn + '-locale-' + ln
@@ -405,7 +405,7 @@ python package_do_split_locales() {
bb.data.setVar('SUMMARY_' + pkg, '%s - %s translations' % (summary, l), d)
bb.data.setVar('DESCRIPTION_' + pkg, '%s This package contains language translation files for the %s locale.' % (description, l), d)
if locale_section:
- bb.data.setVar('SECTION_' + pkg, locale_section, d)
+ d.setVar('SECTION_' + pkg, locale_section)
bb.data.setVar('PACKAGES', ' '.join(packages), d)
@@ -415,14 +415,14 @@ python package_do_split_locales() {
# glibc-localedata-translit* won't install as a dependency
# for some other package which breaks meta-toolchain
# Probably breaks since virtual-locale- isn't provided anywhere
- #rdep = (bb.data.getVar('RDEPENDS_%s' % mainpkg, d, True) or bb.data.getVar('RDEPENDS', d, True) or "").split()
+ #rdep = (d.getVar('RDEPENDS_%s' % mainpkg, True) or d.getVar('RDEPENDS', True) or "").split()
#rdep.append('%s-locale*' % pn)
#bb.data.setVar('RDEPENDS_%s' % mainpkg, ' '.join(rdep), d)
}
python perform_packagecopy () {
- dest = bb.data.getVar('D', d, True)
- dvar = bb.data.getVar('PKGD', d, True)
+ dest = d.getVar('D', True)
+ dvar = d.getVar('PKGD', True)
bb.mkdirhier(dvar)
@@ -549,16 +549,16 @@ python fixup_perms () {
# paths are resolved via BBPATH
def get_fs_perms_list(d):
str = ""
- fs_perms_tables = bb.data.getVar('FILESYSTEM_PERMS_TABLES', d, True)
+ fs_perms_tables = d.getVar('FILESYSTEM_PERMS_TABLES', True)
if not fs_perms_tables:
fs_perms_tables = 'files/fs-perms.txt'
for conf_file in fs_perms_tables.split():
- str += " %s" % bb.which(bb.data.getVar('BBPATH', d, True), conf_file)
+ str += " %s" % bb.which(d.getVar('BBPATH', True), conf_file)
return str
- dvar = bb.data.getVar('PKGD', d, True)
+ dvar = d.getVar('PKGD', True)
fs_perms_table = {}
@@ -586,7 +586,7 @@ python fixup_perms () {
'oldincludedir' ]
for path in target_path_vars:
- dir = bb.data.getVar(path, d, True) or ""
+ dir = d.getVar(path, True) or ""
if dir == "":
continue
fs_perms_table[dir] = fs_perms_entry(bb.data.expand("%s 0755 root root false - - -" % (dir), d))
@@ -664,11 +664,11 @@ python fixup_perms () {
python split_and_strip_files () {
import commands, stat, errno
- dvar = bb.data.getVar('PKGD', d, True)
- pn = bb.data.getVar('PN', d, True)
+ dvar = d.getVar('PKGD', True)
+ pn = d.getVar('PN', True)
# We default to '.debug' style
- if bb.data.getVar('PACKAGE_DEBUG_SPLIT_STYLE', d, True) == 'debug-file-directory':
+ if d.getVar('PACKAGE_DEBUG_SPLIT_STYLE', True) == 'debug-file-directory':
# Single debug-file-directory style debug info
debugappend = ".debug"
debugdir = ""
@@ -691,7 +691,7 @@ python split_and_strip_files () {
# 8 - shared library
def isELF(path):
type = 0
- pathprefix = "export PATH=%s; " % bb.data.getVar('PATH', d, True)
+ pathprefix = "export PATH=%s; " % d.getVar('PATH', True)
ret, result = commands.getstatusoutput("%sfile '%s'" % (pathprefix, path))
if ret:
@@ -715,8 +715,8 @@ python split_and_strip_files () {
#
file_list = {}
file_links = {}
- if (bb.data.getVar('INHIBIT_PACKAGE_DEBUG_SPLIT', d, True) != '1') and \
- (bb.data.getVar('INHIBIT_PACKAGE_STRIP', d, True) != '1'):
+ if (d.getVar('INHIBIT_PACKAGE_DEBUG_SPLIT', True) != '1') and \
+ (d.getVar('INHIBIT_PACKAGE_STRIP', True) != '1'):
for root, dirs, files in os.walk(dvar):
for f in files:
file = os.path.join(root, f)
@@ -764,7 +764,7 @@ python split_and_strip_files () {
#
# First lets process debug splitting
#
- if (bb.data.getVar('INHIBIT_PACKAGE_DEBUG_SPLIT', d, True) != '1'):
+ if (d.getVar('INHIBIT_PACKAGE_DEBUG_SPLIT', True) != '1'):
for file in file_list:
src = file[len(dvar):]
dest = debuglibdir + os.path.dirname(src) + debugdir + "/" + os.path.basename(src) + debugappend
@@ -842,7 +842,7 @@ python split_and_strip_files () {
#
# Now lets go back over things and strip them
#
- if (bb.data.getVar('INHIBIT_PACKAGE_STRIP', d, True) != '1'):
+ if (d.getVar('INHIBIT_PACKAGE_STRIP', True) != '1'):
for file in file_list:
if file_list[file].startswith("ELF: "):
elf_file = int(file_list[file][5:])
@@ -856,11 +856,11 @@ python split_and_strip_files () {
python populate_packages () {
import glob, stat, errno, re
- workdir = bb.data.getVar('WORKDIR', d, True)
- outdir = bb.data.getVar('DEPLOY_DIR', d, True)
- dvar = bb.data.getVar('PKGD', d, True)
- packages = bb.data.getVar('PACKAGES', d, True)
- pn = bb.data.getVar('PN', d, True)
+ workdir = d.getVar('WORKDIR', True)
+ outdir = d.getVar('DEPLOY_DIR', True)
+ dvar = d.getVar('PKGD', True)
+ packages = d.getVar('PACKAGES', True)
+ pn = d.getVar('PN', True)
bb.mkdirhier(outdir)
os.chdir(dvar)
@@ -877,7 +877,7 @@ python populate_packages () {
else:
package_list.append(pkg)
- pkgdest = bb.data.getVar('PKGDEST', d, True)
+ pkgdest = d.getVar('PKGDEST', True)
os.system('rm -rf %s' % pkgdest)
seen = []
@@ -887,14 +887,14 @@ python populate_packages () {
root = os.path.join(pkgdest, pkg)
bb.mkdirhier(root)
- bb.data.setVar('PKG', pkg, localdata)
- overrides = bb.data.getVar('OVERRIDES', localdata, True)
+ localdata.setVar('PKG', pkg)
+ overrides = localdata.getVar('OVERRIDES', True)
if not overrides:
raise bb.build.FuncFailed('OVERRIDES not defined')
- bb.data.setVar('OVERRIDES', overrides + ':' + pkg, localdata)
+ localdata.setVar('OVERRIDES', overrides + ':' + pkg)
bb.data.update_data(localdata)
- filesvar = bb.data.getVar('FILES', localdata, True) or ""
+ filesvar = localdata.getVar('FILES', True) or ""
files = filesvar.split()
file_links = {}
for file in files:
@@ -973,9 +973,9 @@ python populate_packages () {
bb.build.exec_func("package_name_hook", d)
for pkg in package_list:
- pkgname = bb.data.getVar('PKG_%s' % pkg, d, True)
+ pkgname = d.getVar('PKG_%s' % pkg, True)
if pkgname is None:
- bb.data.setVar('PKG_%s' % pkg, pkg, d)
+ d.setVar('PKG_%s' % pkg, pkg)
dangling_links = {}
pkg_files = {}
@@ -999,7 +999,7 @@ python populate_packages () {
dangling_links[pkg].append(os.path.normpath(target))
for pkg in package_list:
- rdepends = bb.utils.explode_dep_versions(bb.data.getVar('RDEPENDS_' + pkg, d, True) or bb.data.getVar('RDEPENDS', d, True) or "")
+ rdepends = bb.utils.explode_dep_versions(d.getVar('RDEPENDS_' + pkg, True) or d.getVar('RDEPENDS', True) or "")
for l in dangling_links[pkg]:
found = False
@@ -1040,9 +1040,9 @@ python emit_pkgdata() {
f.write('%s: %s\n' % (var, encode(val)))
return
- packages = bb.data.getVar('PACKAGES', d, True)
- pkgdest = bb.data.getVar('PKGDEST', d, 1)
- pkgdatadir = bb.data.getVar('PKGDESTWORK', d, True)
+ packages = d.getVar('PACKAGES', True)
+ pkgdest = d.getVar('PKGDEST', 1)
+ pkgdatadir = d.getVar('PKGDESTWORK', True)
# Take shared lock since we're only reading, not writing
lf = bb.utils.lockfile(bb.data.expand("${PACKAGELOCK}", d), True)
@@ -1052,7 +1052,7 @@ python emit_pkgdata() {
f.write("PACKAGES: %s\n" % packages)
f.close()
- workdir = bb.data.getVar('WORKDIR', d, True)
+ workdir = d.getVar('WORKDIR', True)
for pkg in packages.split():
subdata_file = pkgdatadir + "/runtime/%s" % pkg
@@ -1080,19 +1080,19 @@ python emit_pkgdata() {
write_if_exists(sf, pkg, 'pkg_preinst')
write_if_exists(sf, pkg, 'pkg_prerm')
write_if_exists(sf, pkg, 'FILERPROVIDESFLIST')
- for dfile in (bb.data.getVar('FILERPROVIDESFLIST_' + pkg, d, True) or "").split():
+ for dfile in (d.getVar('FILERPROVIDESFLIST_' + pkg, True) or "").split():
write_if_exists(sf, pkg, 'FILERPROVIDES_' + dfile)
write_if_exists(sf, pkg, 'FILERDEPENDSFLIST')
- for dfile in (bb.data.getVar('FILERDEPENDSFLIST_' + pkg, d, True) or "").split():
+ for dfile in (d.getVar('FILERDEPENDSFLIST_' + pkg, True) or "").split():
write_if_exists(sf, pkg, 'FILERDEPENDS_' + dfile)
sf.close()
- allow_empty = bb.data.getVar('ALLOW_EMPTY_%s' % pkg, d, True)
+ allow_empty = d.getVar('ALLOW_EMPTY_%s' % pkg, True)
if not allow_empty:
- allow_empty = bb.data.getVar('ALLOW_EMPTY', d, True)
+ allow_empty = d.getVar('ALLOW_EMPTY', True)
root = "%s/%s" % (pkgdest, pkg)
os.chdir(root)
g = glob('*')
@@ -1123,8 +1123,8 @@ RPMDEPS = "${STAGING_LIBDIR_NATIVE}/rpm/bin/rpmdeps"
python package_do_filedeps() {
import os, re
- pkgdest = bb.data.getVar('PKGDEST', d, True)
- packages = bb.data.getVar('PACKAGES', d, True)
+ pkgdest = d.getVar('PKGDEST', True)
+ packages = d.getVar('PACKAGES', True)
rpmdeps = bb.data.expand("${RPMDEPS}", d)
r = re.compile(r'[<>=]+ +[^ ]*')
@@ -1189,7 +1189,7 @@ SHLIBSWORKDIR = "${WORKDIR}/shlibs"
python package_do_shlibs() {
import re
- exclude_shlibs = bb.data.getVar('EXCLUDE_FROM_SHLIBS', d, 0)
+ exclude_shlibs = d.getVar('EXCLUDE_FROM_SHLIBS', 0)
if exclude_shlibs:
bb.note("not generating shlibs")
return
@@ -1197,27 +1197,27 @@ python package_do_shlibs() {
lib_re = re.compile("^.*\.so")
libdir_re = re.compile(".*/lib$")
- packages = bb.data.getVar('PACKAGES', d, True)
- targetos = bb.data.getVar('TARGET_OS', d, True)
+ packages = d.getVar('PACKAGES', True)
+ targetos = d.getVar('TARGET_OS', True)
- workdir = bb.data.getVar('WORKDIR', d, True)
+ workdir = d.getVar('WORKDIR', True)
- ver = bb.data.getVar('PKGV', d, True)
+ ver = d.getVar('PKGV', True)
if not ver:
bb.error("PKGV not defined")
return
- pkgdest = bb.data.getVar('PKGDEST', d, True)
+ pkgdest = d.getVar('PKGDEST', True)
- shlibs_dir = bb.data.getVar('SHLIBSDIR', d, True)
- shlibswork_dir = bb.data.getVar('SHLIBSWORKDIR', d, True)
+ shlibs_dir = d.getVar('SHLIBSDIR', True)
+ shlibswork_dir = d.getVar('SHLIBSWORKDIR', True)
# Take shared lock since we're only reading, not writing
lf = bb.utils.lockfile(bb.data.expand("${PACKAGELOCK}", d))
def linux_so(root, path, file):
- cmd = bb.data.getVar('OBJDUMP', d, True) + " -p " + os.path.join(root, file) + " 2>/dev/null"
- cmd = "PATH=\"%s\" %s" % (bb.data.getVar('PATH', d, True), cmd)
+ cmd = d.getVar('OBJDUMP', True) + " -p " + os.path.join(root, file) + " 2>/dev/null"
+ cmd = "PATH=\"%s\" %s" % (d.getVar('PATH', True), cmd)
fd = os.popen(cmd)
lines = fd.readlines()
fd.close()
@@ -1264,7 +1264,7 @@ python package_do_shlibs() {
if not combo in sonames:
sonames.append(combo)
if file.endswith('.dylib') or file.endswith('.so'):
- lafile = fullpath.replace(os.path.join(pkgdest, pkg), bb.data.getVar('PKGD', d, True))
+ lafile = fullpath.replace(os.path.join(pkgdest, pkg), d.getVar('PKGD', True))
# Drop suffix
lafile = lafile.rsplit(".",1)[0]
lapath = os.path.dirname(lafile)
@@ -1299,26 +1299,26 @@ python package_do_shlibs() {
needed[pkg].append(name)
#bb.note("Adding %s for %s" % (name, pkg))
- if bb.data.getVar('PACKAGE_SNAP_LIB_SYMLINKS', d, True) == "1":
+ if d.getVar('PACKAGE_SNAP_LIB_SYMLINKS', True) == "1":
snap_symlinks = True
else:
snap_symlinks = False
- if (bb.data.getVar('USE_LDCONFIG', d, True) or "1") == "1":
+ if (d.getVar('USE_LDCONFIG', True) or "1") == "1":
use_ldconfig = True
else:
use_ldconfig = False
needed = {}
shlib_provider = {}
- private_libs = bb.data.getVar('PRIVATE_LIBS', d, True)
+ private_libs = d.getVar('PRIVATE_LIBS', True)
for pkg in packages.split():
needs_ldconfig = False
bb.debug(2, "calculating shlib provides for %s" % pkg)
- pkgver = bb.data.getVar('PKGV_' + pkg, d, True)
+ pkgver = d.getVar('PKGV_' + pkg, True)
if not pkgver:
- pkgver = bb.data.getVar('PV_' + pkg, d, True)
+ pkgver = d.getVar('PV_' + pkg, True)
if not pkgver:
pkgver = ver
@@ -1352,11 +1352,11 @@ python package_do_shlibs() {
fd.close()
if needs_ldconfig and use_ldconfig:
bb.debug(1, 'adding ldconfig call to postinst for %s' % pkg)
- postinst = bb.data.getVar('pkg_postinst_%s' % pkg, d, True) or bb.data.getVar('pkg_postinst', d, True)
+ postinst = d.getVar('pkg_postinst_%s' % pkg, True) or d.getVar('pkg_postinst', True)
if not postinst:
postinst = '#!/bin/sh\n'
- postinst += bb.data.getVar('ldconfig_postinst_fragment', d, True)
- bb.data.setVar('pkg_postinst_%s' % pkg, postinst, d)
+ postinst += d.getVar('ldconfig_postinst_fragment', True)
+ d.setVar('pkg_postinst_%s' % pkg, postinst)
list_re = re.compile('^(.*)\.list$')
for dir in [shlibs_dir]:
@@ -1380,7 +1380,7 @@ python package_do_shlibs() {
bb.utils.unlockfile(lf)
- assumed_libs = bb.data.getVar('ASSUME_SHLIBS', d, True)
+ assumed_libs = d.getVar('ASSUME_SHLIBS', True)
if assumed_libs:
for e in assumed_libs.split():
l, dep_pkg = e.split(":")
@@ -1424,12 +1424,12 @@ python package_do_shlibs() {
python package_do_pkgconfig () {
import re
- packages = bb.data.getVar('PACKAGES', d, True)
- workdir = bb.data.getVar('WORKDIR', d, True)
- pkgdest = bb.data.getVar('PKGDEST', d, True)
+ packages = d.getVar('PACKAGES', True)
+ workdir = d.getVar('WORKDIR', True)
+ pkgdest = d.getVar('PKGDEST', True)
- shlibs_dir = bb.data.getVar('SHLIBSDIR', d, True)
- shlibswork_dir = bb.data.getVar('SHLIBSWORKDIR', d, True)
+ shlibs_dir = d.getVar('SHLIBSDIR', True)
+ shlibswork_dir = d.getVar('SHLIBSWORKDIR', True)
pc_re = re.compile('(.*)\.pc$')
var_re = re.compile('(.*)=(.*)')
@@ -1515,9 +1515,9 @@ python package_do_pkgconfig () {
}
python read_shlibdeps () {
- packages = bb.data.getVar('PACKAGES', d, True).split()
+ packages = d.getVar('PACKAGES', True).split()
for pkg in packages:
- rdepends = bb.utils.explode_dep_versions(bb.data.getVar('RDEPENDS_' + pkg, d, 0) or bb.data.getVar('RDEPENDS', d, 0) or "")
+ rdepends = bb.utils.explode_dep_versions(d.getVar('RDEPENDS_' + pkg, 0) or d.getVar('RDEPENDS', 0) or "")
for extension in ".shlibdeps", ".pcdeps", ".clilibdeps":
depsfile = bb.data.expand("${PKGDEST}/" + pkg + extension, d)
@@ -1544,14 +1544,14 @@ python package_depchains() {
package.
"""
- packages = bb.data.getVar('PACKAGES', d, True)
- postfixes = (bb.data.getVar('DEPCHAIN_POST', d, True) or '').split()
- prefixes = (bb.data.getVar('DEPCHAIN_PRE', d, True) or '').split()
+ packages = d.getVar('PACKAGES', True)
+ postfixes = (d.getVar('DEPCHAIN_POST', True) or '').split()
+ prefixes = (d.getVar('DEPCHAIN_PRE', True) or '').split()
def pkg_adddeprrecs(pkg, base, suffix, getname, depends, d):
#bb.note('depends for %s is %s' % (base, depends))
- rreclist = bb.utils.explode_dep_versions(bb.data.getVar('RRECOMMENDS_' + pkg, d, True) or bb.data.getVar('RRECOMMENDS', d, True) or "")
+ rreclist = bb.utils.explode_dep_versions(d.getVar('RRECOMMENDS_' + pkg, True) or d.getVar('RRECOMMENDS', True) or "")
for depend in depends:
if depend.find('-native') != -1 or depend.find('-cross') != -1 or depend.startswith('virtual/'):
@@ -1572,7 +1572,7 @@ python package_depchains() {
def pkg_addrrecs(pkg, base, suffix, getname, rdepends, d):
#bb.note('rdepends for %s is %s' % (base, rdepends))
- rreclist = bb.utils.explode_dep_versions(bb.data.getVar('RRECOMMENDS_' + pkg, d, True) or bb.data.getVar('RRECOMMENDS', d, True) or "")
+ rreclist = bb.utils.explode_dep_versions(d.getVar('RRECOMMENDS_' + pkg, True) or d.getVar('RRECOMMENDS', True) or "")
for depend in rdepends:
if depend.find('virtual-locale-') != -1:
@@ -1596,15 +1596,15 @@ python package_depchains() {
list.append(dep)
depends = []
- for dep in bb.utils.explode_deps(bb.data.getVar('DEPENDS', d, True) or ""):
+ for dep in bb.utils.explode_deps(d.getVar('DEPENDS', True) or ""):
add_dep(depends, dep)
rdepends = []
- for dep in bb.utils.explode_deps(bb.data.getVar('RDEPENDS', d, True) or ""):
+ for dep in bb.utils.explode_deps(d.getVar('RDEPENDS', True) or ""):
add_dep(rdepends, dep)
for pkg in packages.split():
- for dep in bb.utils.explode_deps(bb.data.getVar('RDEPENDS_' + pkg, d, True) or ""):
+ for dep in bb.utils.explode_deps(d.getVar('RDEPENDS_' + pkg, True) or ""):
add_dep(rdepends, dep)
#bb.note('rdepends is %s' % rdepends)
@@ -1630,7 +1630,7 @@ python package_depchains() {
for suffix in pkgs:
for pkg in pkgs[suffix]:
- if bb.data.getVarFlag('RRECOMMENDS_' + pkg, 'nodeprrecs', d):
+ if d.getVarFlag('RRECOMMENDS_' + pkg, 'nodeprrecs'):
continue
(base, func) = pkgs[suffix][pkg]
if suffix == "-dev":
@@ -1639,7 +1639,7 @@ python package_depchains() {
pkg_addrrecs(pkg, base, suffix, func, rdepends, d)
else:
rdeps = []
- for dep in bb.utils.explode_deps(bb.data.getVar('RDEPENDS_' + base, d, True) or bb.data.getVar('RDEPENDS', d, True) or ""):
+ for dep in bb.utils.explode_deps(d.getVar('RDEPENDS_' + base, True) or d.getVar('RDEPENDS', True) or ""):
add_dep(rdeps, dep)
pkg_addrrecs(pkg, base, suffix, func, rdeps, d)
}
@@ -1679,22 +1679,22 @@ python do_package () {
# as any change to rpmdeps requires this to be rerun.
# PACKAGE_BBCLASS_VERSION = "1"
- packages = (bb.data.getVar('PACKAGES', d, True) or "").split()
+ packages = (d.getVar('PACKAGES', True) or "").split()
if len(packages) < 1:
bb.debug(1, "No packages to build, skipping do_package")
return
- workdir = bb.data.getVar('WORKDIR', d, True)
- outdir = bb.data.getVar('DEPLOY_DIR', d, True)
- dest = bb.data.getVar('D', d, True)
- dvar = bb.data.getVar('PKGD', d, True)
- pn = bb.data.getVar('PN', d, True)
+ workdir = d.getVar('WORKDIR', True)
+ outdir = d.getVar('DEPLOY_DIR', True)
+ dest = d.getVar('D', True)
+ dvar = d.getVar('PKGD', True)
+ pn = d.getVar('PN', True)
if not workdir or not outdir or not dest or not dvar or not pn or not packages:
bb.error("WORKDIR, DEPLOY_DIR, D, PN and PKGD all must be defined, unable to package")
return
- for f in (bb.data.getVar('PACKAGEFUNCS', d, True) or '').split():
+ for f in (d.getVar('PACKAGEFUNCS', True) or '').split():
bb.build.exec_func(f, d)
}
diff --git a/meta/classes/package_deb.bbclass b/meta/classes/package_deb.bbclass
index 6733e6453..71e46a8c8 100644
--- a/meta/classes/package_deb.bbclass
+++ b/meta/classes/package_deb.bbclass
@@ -11,18 +11,18 @@ DPKG_ARCH ?= "${TARGET_ARCH}"
PKGWRITEDIRDEB = "${WORKDIR}/deploy-debs"
python package_deb_fn () {
- bb.data.setVar('PKGFN', bb.data.getVar('PKG',d), d)
+ bb.data.setVar('PKGFN', d.getVar('PKG'), d)
}
addtask package_deb_install
python do_package_deb_install () {
- pkg = bb.data.getVar('PKG', d, True)
- pkgfn = bb.data.getVar('PKGFN', d, True)
- rootfs = bb.data.getVar('IMAGE_ROOTFS', d, True)
- debdir = bb.data.getVar('DEPLOY_DIR_DEB', d, True)
+ pkg = d.getVar('PKG', True)
+ pkgfn = d.getVar('PKGFN', True)
+ rootfs = d.getVar('IMAGE_ROOTFS', True)
+ debdir = d.getVar('DEPLOY_DIR_DEB', True)
apt_config = bb.data.expand('${STAGING_ETCDIR_NATIVE}/apt/apt.conf', d)
- stagingbindir = bb.data.getVar('STAGING_BINDIR_NATIVE', d, True)
- tmpdir = bb.data.getVar('TMPDIR', d, True)
+ stagingbindir = d.getVar('STAGING_BINDIR_NATIVE', True)
+ tmpdir = d.getVar('TMPDIR', True)
if None in (pkg,pkgfn,rootfs):
raise bb.build.FuncFailed("missing variables (one or more of PKG, PKGFN, IMAGE_ROOTFS)")
@@ -206,22 +206,22 @@ python do_package_deb () {
import re, copy
import textwrap
- workdir = bb.data.getVar('WORKDIR', d, True)
+ workdir = d.getVar('WORKDIR', True)
if not workdir:
bb.error("WORKDIR not defined, unable to package")
return
- outdir = bb.data.getVar('PKGWRITEDIRDEB', d, True)
+ outdir = d.getVar('PKGWRITEDIRDEB', True)
if not outdir:
bb.error("PKGWRITEDIRDEB not defined, unable to package")
return
- packages = bb.data.getVar('PACKAGES', d, True)
+ packages = d.getVar('PACKAGES', True)
if not packages:
bb.debug(1, "PACKAGES not defined, nothing to package")
return
- tmpdir = bb.data.getVar('TMPDIR', d, True)
+ tmpdir = d.getVar('TMPDIR', True)
if os.access(os.path.join(tmpdir, "stamps", "DEB_PACKAGE_INDEX_CLEAN"),os.R_OK):
os.unlink(os.path.join(tmpdir, "stamps", "DEB_PACKAGE_INDEX_CLEAN"))
@@ -230,7 +230,7 @@ python do_package_deb () {
bb.debug(1, "No packages; nothing to do")
return
- pkgdest = bb.data.getVar('PKGDEST', d, True)
+ pkgdest = d.getVar('PKGDEST', True)
for pkg in packages.split():
localdata = bb.data.createCopy(d)
@@ -238,19 +238,19 @@ python do_package_deb () {
lf = bb.utils.lockfile(root + ".lock")
- bb.data.setVar('ROOT', '', localdata)
- bb.data.setVar('ROOT_%s' % pkg, root, localdata)
- pkgname = bb.data.getVar('PKG_%s' % pkg, localdata, True)
+ localdata.setVar('ROOT', '')
+ localdata.setVar('ROOT_%s' % pkg, root)
+ pkgname = localdata.getVar('PKG_%s' % pkg, True)
if not pkgname:
pkgname = pkg
- bb.data.setVar('PKG', pkgname, localdata)
+ localdata.setVar('PKG', pkgname)
- bb.data.setVar('OVERRIDES', pkg, localdata)
+ localdata.setVar('OVERRIDES', pkg)
bb.data.update_data(localdata)
basedir = os.path.join(os.path.dirname(root))
- pkgoutdir = os.path.join(outdir, bb.data.getVar('PACKAGE_ARCH', localdata, True))
+ pkgoutdir = os.path.join(outdir, localdata.getVar('PACKAGE_ARCH', True))
bb.mkdirhier(pkgoutdir)
os.chdir(root)
@@ -261,8 +261,8 @@ python do_package_deb () {
del g[g.index('./DEBIAN')]
except ValueError:
pass
- if not g and bb.data.getVar('ALLOW_EMPTY', localdata) != "1":
- bb.note("Not creating empty archive for %s-%s-%s" % (pkg, bb.data.getVar('PKGV', localdata, True), bb.data.getVar('PKGR', localdata, True)))
+ if not g and localdata.getVar('ALLOW_EMPTY') != "1":
+ bb.note("Not creating empty archive for %s-%s-%s" % (pkg, localdata.getVar('PKGV', True), localdata.getVar('PKGR', True)))
bb.utils.unlockfile(lf)
continue
@@ -278,7 +278,7 @@ python do_package_deb () {
raise bb.build.FuncFailed("unable to open control file for writing.")
fields = []
- pe = bb.data.getVar('PKGE', d, True)
+ pe = d.getVar('PKGE', True)
if pe and int(pe) > 0:
fields.append(["Version: %s:%s-%s\n", ['PKGE', 'PKGV', 'PKGR']])
else:
@@ -298,10 +298,10 @@ python do_package_deb () {
def pullData(l, d):
l2 = []
for i in l:
- data = bb.data.getVar(i, d, True)
+ data = d.getVar(i, True)
if data is None:
raise KeyError(f)
- if i == 'DPKG_ARCH' and bb.data.getVar('PACKAGE_ARCH', d, True) == 'all':
+ if i == 'DPKG_ARCH' and d.getVar('PACKAGE_ARCH', True) == 'all':
data = 'all'
l2.append(data)
return l2
@@ -311,12 +311,12 @@ python do_package_deb () {
try:
for (c, fs) in fields:
for f in fs:
- if bb.data.getVar(f, localdata) is None:
+ if localdata.getVar(f) is None:
raise KeyError(f)
# Special behavior for description...
if 'DESCRIPTION' in fs:
- summary = bb.data.getVar('SUMMARY', localdata, True) or bb.data.getVar('DESCRIPTION', localdata, True) or "."
- description = bb.data.getVar('DESCRIPTION', localdata, True) or "."
+ summary = localdata.getVar('SUMMARY', True) or localdata.getVar('DESCRIPTION', True) or "."
+ description = localdata.getVar('DESCRIPTION', True) or "."
description = textwrap.dedent(description).strip()
ctrlfile.write('Description: %s\n' % unicode(summary))
ctrlfile.write('%s\n' % unicode(textwrap.fill(description, width=74, initial_indent=' ', subsequent_indent=' ')))
@@ -332,18 +332,18 @@ python do_package_deb () {
bb.build.exec_func("mapping_rename_hook", localdata)
- rdepends = bb.utils.explode_dep_versions(bb.data.getVar("RDEPENDS", localdata, True) or "")
+ rdepends = bb.utils.explode_dep_versions(localdata.getVar("RDEPENDS", True) or "")
for dep in rdepends:
if '*' in dep:
del rdepends[dep]
- rrecommends = bb.utils.explode_dep_versions(bb.data.getVar("RRECOMMENDS", localdata, True) or "")
+ rrecommends = bb.utils.explode_dep_versions(localdata.getVar("RRECOMMENDS", True) or "")
for dep in rrecommends:
if '*' in dep:
del rrecommends[dep]
- rsuggests = bb.utils.explode_dep_versions(bb.data.getVar("RSUGGESTS", localdata, True) or "")
- rprovides = bb.utils.explode_dep_versions(bb.data.getVar("RPROVIDES", localdata, True) or "")
- rreplaces = bb.utils.explode_dep_versions(bb.data.getVar("RREPLACES", localdata, True) or "")
- rconflicts = bb.utils.explode_dep_versions(bb.data.getVar("RCONFLICTS", localdata, True) or "")
+ rsuggests = bb.utils.explode_dep_versions(localdata.getVar("RSUGGESTS", True) or "")
+ rprovides = bb.utils.explode_dep_versions(localdata.getVar("RPROVIDES", True) or "")
+ rreplaces = bb.utils.explode_dep_versions(localdata.getVar("RREPLACES", True) or "")
+ rconflicts = bb.utils.explode_dep_versions(localdata.getVar("RCONFLICTS", True) or "")
if rdepends:
ctrlfile.write("Depends: %s\n" % unicode(bb.utils.join_deps(rdepends)))
if rsuggests:
@@ -359,7 +359,7 @@ python do_package_deb () {
ctrlfile.close()
for script in ["preinst", "postinst", "prerm", "postrm"]:
- scriptvar = bb.data.getVar('pkg_%s' % script, localdata, True)
+ scriptvar = localdata.getVar('pkg_%s' % script, True)
if not scriptvar:
continue
try:
@@ -372,7 +372,7 @@ python do_package_deb () {
scriptfile.close()
os.chmod(os.path.join(controldir, script), 0755)
- conffiles_str = bb.data.getVar("CONFFILES", localdata, True)
+ conffiles_str = localdata.getVar("CONFFILES", True)
if conffiles_str:
try:
conffiles = file(os.path.join(controldir, 'conffiles'), 'w')
@@ -384,7 +384,7 @@ python do_package_deb () {
conffiles.close()
os.chdir(basedir)
- ret = os.system("PATH=\"%s\" dpkg-deb -b %s %s" % (bb.data.getVar("PATH", localdata, True), root, pkgoutdir))
+ ret = os.system("PATH=\"%s\" dpkg-deb -b %s %s" % (localdata.getVar("PATH", True), root, pkgoutdir))
if ret != 0:
bb.utils.prunedir(controldir)
bb.utils.unlockfile(lf)
@@ -405,17 +405,17 @@ python do_package_write_deb_setscene () {
addtask do_package_write_deb_setscene
python () {
- if bb.data.getVar('PACKAGES', d, True) != '':
- deps = (bb.data.getVarFlag('do_package_write_deb', 'depends', d) or "").split()
+ if d.getVar('PACKAGES', True) != '':
+ deps = (d.getVarFlag('do_package_write_deb', 'depends') or "").split()
deps.append('dpkg-native:do_populate_sysroot')
deps.append('virtual/fakeroot-native:do_populate_sysroot')
bb.data.setVarFlag('do_package_write_deb', 'depends', " ".join(deps), d)
- bb.data.setVarFlag('do_package_write_deb', 'fakeroot', "1", d)
- bb.data.setVarFlag('do_package_write_deb_setscene', 'fakeroot', "1", d)
+ d.setVarFlag('do_package_write_deb', 'fakeroot', "1")
+ d.setVarFlag('do_package_write_deb_setscene', 'fakeroot', "1")
# Map TARGET_ARCH to Debian's ideas about architectures
- if bb.data.getVar('DPKG_ARCH', d, True) in ["x86", "i486", "i586", "i686", "pentium"]:
- bb.data.setVar('DPKG_ARCH', 'i386', d)
+ if d.getVar('DPKG_ARCH', True) in ["x86", "i486", "i586", "i686", "pentium"]:
+ d.setVar('DPKG_ARCH', 'i386')
}
python do_package_write_deb () {
diff --git a/meta/classes/package_ipk.bbclass b/meta/classes/package_ipk.bbclass
index d41b40d2c..df608fc0e 100644
--- a/meta/classes/package_ipk.bbclass
+++ b/meta/classes/package_ipk.bbclass
@@ -11,16 +11,16 @@ PKGWRITEDIRIPK = "${WORKDIR}/deploy-ipks"
OPKGBUILDCMD ??= "opkg-build"
python package_ipk_fn () {
- bb.data.setVar('PKGFN', bb.data.getVar('PKG',d), d)
+ bb.data.setVar('PKGFN', d.getVar('PKG'), d)
}
python package_ipk_install () {
- pkg = bb.data.getVar('PKG', d, 1)
- pkgfn = bb.data.getVar('PKGFN', d, 1)
- rootfs = bb.data.getVar('IMAGE_ROOTFS', d, 1)
- ipkdir = bb.data.getVar('DEPLOY_DIR_IPK', d, 1)
- stagingdir = bb.data.getVar('STAGING_DIR', d, 1)
- tmpdir = bb.data.getVar('TMPDIR', d, 1)
+ pkg = d.getVar('PKG', 1)
+ pkgfn = d.getVar('PKGFN', 1)
+ rootfs = d.getVar('IMAGE_ROOTFS', 1)
+ ipkdir = d.getVar('DEPLOY_DIR_IPK', 1)
+ stagingdir = d.getVar('STAGING_DIR', 1)
+ tmpdir = d.getVar('TMPDIR', 1)
if None in (pkg,pkgfn,rootfs):
raise bb.build.FuncFailed("missing variables (one or more of PKG, PKGFN, IMAGEROOTFS)")
@@ -36,7 +36,7 @@ python package_ipk_install () {
# Generate ipk.conf if it or the stamp doesnt exist
conffile = os.path.join(stagingdir,"ipkg.conf")
if not os.access(conffile, os.R_OK):
- ipkg_archs = bb.data.getVar('PACKAGE_ARCHS',d)
+ ipkg_archs = d.getVar('PACKAGE_ARCHS')
if ipkg_archs is None:
bb.error("PACKAGE_ARCHS missing")
raise FuncFailed
@@ -259,15 +259,15 @@ python do_package_ipk () {
import re, copy
import textwrap
- workdir = bb.data.getVar('WORKDIR', d, True)
- outdir = bb.data.getVar('PKGWRITEDIRIPK', d, True)
- tmpdir = bb.data.getVar('TMPDIR', d, True)
- pkgdest = bb.data.getVar('PKGDEST', d, True)
+ workdir = d.getVar('WORKDIR', True)
+ outdir = d.getVar('PKGWRITEDIRIPK', True)
+ tmpdir = d.getVar('TMPDIR', True)
+ pkgdest = d.getVar('PKGDEST', True)
if not workdir or not outdir or not tmpdir:
bb.error("Variables incorrectly set, unable to package")
return
- packages = bb.data.getVar('PACKAGES', d, True)
+ packages = d.getVar('PACKAGES', True)
if not packages or packages == '':
bb.debug(1, "No packages; nothing to do")
return
@@ -283,18 +283,18 @@ python do_package_ipk () {
lf = bb.utils.lockfile(root + ".lock")
- bb.data.setVar('ROOT', '', localdata)
- bb.data.setVar('ROOT_%s' % pkg, root, localdata)
- pkgname = bb.data.getVar('PKG_%s' % pkg, localdata, 1)
+ localdata.setVar('ROOT', '')
+ localdata.setVar('ROOT_%s' % pkg, root)
+ pkgname = localdata.getVar('PKG_%s' % pkg, 1)
if not pkgname:
pkgname = pkg
- bb.data.setVar('PKG', pkgname, localdata)
+ localdata.setVar('PKG', pkgname)
- bb.data.setVar('OVERRIDES', pkg, localdata)
+ localdata.setVar('OVERRIDES', pkg)
bb.data.update_data(localdata)
basedir = os.path.join(os.path.dirname(root))
- arch = bb.data.getVar('PACKAGE_ARCH', localdata, 1)
+ arch = localdata.getVar('PACKAGE_ARCH', 1)
pkgoutdir = "%s/%s" % (outdir, arch)
bb.mkdirhier(pkgoutdir)
os.chdir(root)
@@ -305,8 +305,8 @@ python do_package_ipk () {
del g[g.index('./CONTROL')]
except ValueError:
pass
- if not g and bb.data.getVar('ALLOW_EMPTY', localdata) != "1":
- bb.note("Not creating empty archive for %s-%s-%s" % (pkg, bb.data.getVar('PKGV', localdata, 1), bb.data.getVar('PKGR', localdata, 1)))
+ if not g and localdata.getVar('ALLOW_EMPTY') != "1":
+ bb.note("Not creating empty archive for %s-%s-%s" % (pkg, localdata.getVar('PKGV', 1), localdata.getVar('PKGR', 1)))
bb.utils.unlockfile(lf)
continue
@@ -319,7 +319,7 @@ python do_package_ipk () {
raise bb.build.FuncFailed("unable to open control file for writing.")
fields = []
- pe = bb.data.getVar('PKGE', d, 1)
+ pe = d.getVar('PKGE', 1)
if pe and int(pe) > 0:
fields.append(["Version: %s:%s-%s\n", ['PKGE', 'PKGV', 'PKGR']])
else:
@@ -336,7 +336,7 @@ python do_package_ipk () {
def pullData(l, d):
l2 = []
for i in l:
- l2.append(bb.data.getVar(i, d, 1))
+ l2.append(d.getVar(i, 1))
return l2
ctrlfile.write("Package: %s\n" % pkgname)
@@ -344,12 +344,12 @@ python do_package_ipk () {
try:
for (c, fs) in fields:
for f in fs:
- if bb.data.getVar(f, localdata) is None:
+ if localdata.getVar(f) is None:
raise KeyError(f)
# Special behavior for description...
if 'DESCRIPTION' in fs:
- summary = bb.data.getVar('SUMMARY', localdata, True) or bb.data.getVar('DESCRIPTION', localdata, True) or "."
- description = bb.data.getVar('DESCRIPTION', localdata, True) or "."
+ summary = localdata.getVar('SUMMARY', True) or localdata.getVar('DESCRIPTION', True) or "."
+ description = localdata.getVar('DESCRIPTION', True) or "."
description = textwrap.dedent(description).strip()
ctrlfile.write('Description: %s\n' % summary)
ctrlfile.write('%s\n' % textwrap.fill(description, width=74, initial_indent=' ', subsequent_indent=' '))
@@ -365,12 +365,12 @@ python do_package_ipk () {
bb.build.exec_func("mapping_rename_hook", localdata)
- rdepends = bb.utils.explode_dep_versions(bb.data.getVar("RDEPENDS", localdata, 1) or "")
- rrecommends = bb.utils.explode_dep_versions(bb.data.getVar("RRECOMMENDS", localdata, 1) or "")
- rsuggests = bb.utils.explode_dep_versions(bb.data.getVar("RSUGGESTS", localdata, 1) or "")
- rprovides = bb.utils.explode_dep_versions(bb.data.getVar("RPROVIDES", localdata, 1) or "")
- rreplaces = bb.utils.explode_dep_versions(bb.data.getVar("RREPLACES", localdata, 1) or "")
- rconflicts = bb.utils.explode_dep_versions(bb.data.getVar("RCONFLICTS", localdata, 1) or "")
+ rdepends = bb.utils.explode_dep_versions(localdata.getVar("RDEPENDS", 1) or "")
+ rrecommends = bb.utils.explode_dep_versions(localdata.getVar("RRECOMMENDS", 1) or "")
+ rsuggests = bb.utils.explode_dep_versions(localdata.getVar("RSUGGESTS", 1) or "")
+ rprovides = bb.utils.explode_dep_versions(localdata.getVar("RPROVIDES", 1) or "")
+ rreplaces = bb.utils.explode_dep_versions(localdata.getVar("RREPLACES", 1) or "")
+ rconflicts = bb.utils.explode_dep_versions(localdata.getVar("RCONFLICTS", 1) or "")
if rdepends:
ctrlfile.write("Depends: %s\n" % bb.utils.join_deps(rdepends))
@@ -384,14 +384,14 @@ python do_package_ipk () {
ctrlfile.write("Replaces: %s\n" % bb.utils.join_deps(rreplaces))
if rconflicts:
ctrlfile.write("Conflicts: %s\n" % bb.utils.join_deps(rconflicts))
- src_uri = bb.data.getVar("SRC_URI", localdata, 1)
+ src_uri = localdata.getVar("SRC_URI", 1)
if src_uri:
src_uri = re.sub("\s+", " ", src_uri)
ctrlfile.write("Source: %s\n" % " ".join(src_uri.split()))
ctrlfile.close()
for script in ["preinst", "postinst", "prerm", "postrm"]:
- scriptvar = bb.data.getVar('pkg_%s' % script, localdata, 1)
+ scriptvar = localdata.getVar('pkg_%s' % script, 1)
if not scriptvar:
continue
try:
@@ -403,7 +403,7 @@ python do_package_ipk () {
scriptfile.close()
os.chmod(os.path.join(controldir, script), 0755)
- conffiles_str = bb.data.getVar("CONFFILES", localdata, 1)
+ conffiles_str = localdata.getVar("CONFFILES", 1)
if conffiles_str:
try:
conffiles = file(os.path.join(controldir, 'conffiles'), 'w')
@@ -415,8 +415,8 @@ python do_package_ipk () {
conffiles.close()
os.chdir(basedir)
- ret = os.system("PATH=\"%s\" %s %s %s" % (bb.data.getVar("PATH", localdata, 1),
- bb.data.getVar("OPKGBUILDCMD",d,1), pkg, pkgoutdir))
+ ret = os.system("PATH=\"%s\" %s %s %s" % (localdata.getVar("PATH", 1),
+ d.getVar("OPKGBUILDCMD",1), pkg, pkgoutdir))
if ret != 0:
bb.utils.unlockfile(lf)
raise bb.build.FuncFailed("opkg-build execution failed")
@@ -437,13 +437,13 @@ python do_package_write_ipk_setscene () {
addtask do_package_write_ipk_setscene
python () {
- if bb.data.getVar('PACKAGES', d, True) != '':
- deps = (bb.data.getVarFlag('do_package_write_ipk', 'depends', d) or "").split()
+ if d.getVar('PACKAGES', True) != '':
+ deps = (d.getVarFlag('do_package_write_ipk', 'depends') or "").split()
deps.append('opkg-utils-native:do_populate_sysroot')
deps.append('virtual/fakeroot-native:do_populate_sysroot')
bb.data.setVarFlag('do_package_write_ipk', 'depends', " ".join(deps), d)
- bb.data.setVarFlag('do_package_write_ipk', 'fakeroot', "1", d)
- bb.data.setVarFlag('do_package_write_ipk_setscene', 'fakeroot', "1", d)
+ d.setVarFlag('do_package_write_ipk', 'fakeroot', "1")
+ d.setVarFlag('do_package_write_ipk_setscene', 'fakeroot', "1")
}
python do_package_write_ipk () {
diff --git a/meta/classes/package_rpm.bbclass b/meta/classes/package_rpm.bbclass
index f804a0fc1..2c5545c11 100644
--- a/meta/classes/package_rpm.bbclass
+++ b/meta/classes/package_rpm.bbclass
@@ -8,7 +8,7 @@ RPMBUILD="rpmbuild"
PKGWRITEDIRRPM = "${WORKDIR}/deploy-rpms"
python package_rpm_fn () {
- bb.data.setVar('PKGFN', bb.data.getVar('PKG',d), d)
+ bb.data.setVar('PKGFN', d.getVar('PKG'), d)
}
python package_rpm_install () {
@@ -406,7 +406,7 @@ python write_specfile () {
name = "".join(name.split(eext[1] + '-'))
return name
-# ml = bb.data.getVar("MLPREFIX", d, True)
+# ml = d.getVar("MLPREFIX", True)
# if ml and name and len(ml) != 0 and name.find(ml) == 0:
# return ml.join(name.split(ml, 1)[1:])
# return name
@@ -426,7 +426,7 @@ python write_specfile () {
# after renaming we cannot look up the dependencies in the packagedata
# store.
def translate_vers(varname, d):
- depends = bb.data.getVar(varname, d, True)
+ depends = d.getVar(varname, True)
if depends:
depends_dict = bb.utils.explode_dep_versions(depends)
newdeps_dict = {}
@@ -481,34 +481,34 @@ python write_specfile () {
scr = scr[:pos] + 'if [ "$1" = "0" ] ; then\n' + scr[pos:] + '\nfi'
return scr
- packages = bb.data.getVar('PACKAGES', d, True)
+ packages = d.getVar('PACKAGES', True)
if not packages or packages == '':
bb.debug(1, "No packages; nothing to do")
return
- pkgdest = bb.data.getVar('PKGDEST', d, True)
+ pkgdest = d.getVar('PKGDEST', True)
if not pkgdest:
bb.fatal("No PKGDEST")
return
- outspecfile = bb.data.getVar('OUTSPECFILE', d, True)
+ outspecfile = d.getVar('OUTSPECFILE', True)
if not outspecfile:
bb.fatal("No OUTSPECFILE")
return
# Construct the SPEC file...
- srcname = strip_multilib(bb.data.getVar('PN', d, True), d)
- srcsummary = (bb.data.getVar('SUMMARY', d, True) or bb.data.getVar('DESCRIPTION', d, True) or ".")
- srcversion = bb.data.getVar('PKGV', d, True).replace('-', '+')
- srcrelease = bb.data.getVar('PKGR', d, True)
- srcepoch = (bb.data.getVar('PKGE', d, True) or "")
- srclicense = bb.data.getVar('LICENSE', d, True)
- srcsection = bb.data.getVar('SECTION', d, True)
- srcmaintainer = bb.data.getVar('MAINTAINER', d, True)
- srchomepage = bb.data.getVar('HOMEPAGE', d, True)
- srcdescription = bb.data.getVar('DESCRIPTION', d, True) or "."
-
- srcdepends = strip_multilib(bb.data.getVar('DEPENDS', d, True), d)
+ srcname = strip_multilib(d.getVar('PN', True), d)
+ srcsummary = (d.getVar('SUMMARY', True) or d.getVar('DESCRIPTION', True) or ".")
+ srcversion = d.getVar('PKGV', True).replace('-', '+')
+ srcrelease = d.getVar('PKGR', True)
+ srcepoch = (d.getVar('PKGE', True) or "")
+ srclicense = d.getVar('LICENSE', True)
+ srcsection = d.getVar('SECTION', True)
+ srcmaintainer = d.getVar('MAINTAINER', True)
+ srchomepage = d.getVar('HOMEPAGE', True)
+ srcdescription = d.getVar('DESCRIPTION', True) or "."
+
+ srcdepends = strip_multilib(d.getVar('DEPENDS', True), d)
srcrdepends = []
srcrrecommends = []
srcrsuggests = []
@@ -538,28 +538,28 @@ python write_specfile () {
lf = bb.utils.lockfile(root + ".lock")
- bb.data.setVar('ROOT', '', localdata)
- bb.data.setVar('ROOT_%s' % pkg, root, localdata)
- pkgname = bb.data.getVar('PKG_%s' % pkg, localdata, 1)
+ localdata.setVar('ROOT', '')
+ localdata.setVar('ROOT_%s' % pkg, root)
+ pkgname = localdata.getVar('PKG_%s' % pkg, 1)
if not pkgname:
pkgname = pkg
- bb.data.setVar('PKG', pkgname, localdata)
+ localdata.setVar('PKG', pkgname)
- bb.data.setVar('OVERRIDES', pkg, localdata)
+ localdata.setVar('OVERRIDES', pkg)
bb.data.update_data(localdata)
- conffiles = (bb.data.getVar('CONFFILES', localdata, True) or "").split()
+ conffiles = (localdata.getVar('CONFFILES', True) or "").split()
splitname = strip_multilib(pkgname, d)
- splitsummary = (bb.data.getVar('SUMMARY', localdata, True) or bb.data.getVar('DESCRIPTION', localdata, True) or ".")
- splitversion = (bb.data.getVar('PKGV', localdata, True) or "").replace('-', '+')
- splitrelease = (bb.data.getVar('PKGR', localdata, True) or "")
- splitepoch = (bb.data.getVar('PKGE', localdata, True) or "")
- splitlicense = (bb.data.getVar('LICENSE', localdata, True) or "")
- splitsection = (bb.data.getVar('SECTION', localdata, True) or "")
- splitdescription = (bb.data.getVar('DESCRIPTION', localdata, True) or ".")
+ splitsummary = (localdata.getVar('SUMMARY', True) or localdata.getVar('DESCRIPTION', True) or ".")
+ splitversion = (localdata.getVar('PKGV', True) or "").replace('-', '+')
+ splitrelease = (localdata.getVar('PKGR', True) or "")
+ splitepoch = (localdata.getVar('PKGE', True) or "")
+ splitlicense = (localdata.getVar('LICENSE', True) or "")
+ splitsection = (localdata.getVar('SECTION', True) or "")
+ splitdescription = (localdata.getVar('DESCRIPTION', True) or ".")
translate_vers('RDEPENDS', localdata)
translate_vers('RRECOMMENDS', localdata)
@@ -571,12 +571,12 @@ python write_specfile () {
# Map the dependencies into their final form
bb.build.exec_func("mapping_rename_hook", localdata)
- splitrdepends = strip_multilib(bb.data.getVar('RDEPENDS', localdata, True), d) or ""
- splitrrecommends = strip_multilib(bb.data.getVar('RRECOMMENDS', localdata, True), d) or ""
- splitrsuggests = strip_multilib(bb.data.getVar('RSUGGESTS', localdata, True), d) or ""
- splitrprovides = strip_multilib(bb.data.getVar('RPROVIDES', localdata, True), d) or ""
- splitrreplaces = strip_multilib(bb.data.getVar('RREPLACES', localdata, True), d) or ""
- splitrconflicts = strip_multilib(bb.data.getVar('RCONFLICTS', localdata, True), d) or ""
+ splitrdepends = strip_multilib(localdata.getVar('RDEPENDS', True), d) or ""
+ splitrrecommends = strip_multilib(localdata.getVar('RRECOMMENDS', True), d) or ""
+ splitrsuggests = strip_multilib(localdata.getVar('RSUGGESTS', True), d) or ""
+ splitrprovides = strip_multilib(localdata.getVar('RPROVIDES', True), d) or ""
+ splitrreplaces = strip_multilib(localdata.getVar('RREPLACES', True), d) or ""
+ splitrconflicts = strip_multilib(localdata.getVar('RCONFLICTS', True), d) or ""
splitrobsoletes = []
# For now we need to manually supplement RPROVIDES with any update-alternatives links
@@ -592,14 +592,14 @@ python write_specfile () {
srcrreplaces = splitrreplaces
srcrconflicts = splitrconflicts
- srcpreinst = bb.data.getVar('pkg_preinst', localdata, True)
- srcpostinst = bb.data.getVar('pkg_postinst', localdata, True)
- srcprerm = bb.data.getVar('pkg_prerm', localdata, True)
- srcpostrm = bb.data.getVar('pkg_postrm', localdata, True)
+ srcpreinst = localdata.getVar('pkg_preinst', True)
+ srcpostinst = localdata.getVar('pkg_postinst', True)
+ srcprerm = localdata.getVar('pkg_prerm', True)
+ srcpostrm = localdata.getVar('pkg_postrm', True)
file_list = []
walk_files(root, file_list, conffiles)
- if not file_list and bb.data.getVar('ALLOW_EMPTY', localdata) != "1":
+ if not file_list and localdata.getVar('ALLOW_EMPTY') != "1":
bb.note("Not creating empty RPM package for %s" % splitname)
else:
bb.note("Creating RPM package for %s" % splitname)
@@ -672,7 +672,7 @@ python write_specfile () {
# Now process scriptlets
for script in ["preinst", "postinst", "prerm", "postrm"]:
- scriptvar = bb.data.getVar('pkg_%s' % script, localdata, True)
+ scriptvar = localdata.getVar('pkg_%s' % script, True)
if not scriptvar:
continue
if script == 'preinst':
@@ -691,7 +691,7 @@ python write_specfile () {
# Now process files
file_list = []
walk_files(root, file_list, conffiles)
- if not file_list and bb.data.getVar('ALLOW_EMPTY', localdata) != "1":
+ if not file_list and localdata.getVar('ALLOW_EMPTY') != "1":
bb.note("Not creating empty RPM package for %s" % splitname)
else:
spec_files_bottom.append('%%files -n %s' % splitname)
@@ -813,29 +813,29 @@ python do_package_rpm () {
# We need a simple way to remove the MLPREFIX from the package name,
# and dependency information...
def strip_multilib(name, d):
- ml = bb.data.getVar("MLPREFIX", d, True)
+ ml = d.getVar("MLPREFIX", True)
if ml and name and len(ml) != 0 and name.find(ml) >= 0:
return "".join(name.split(ml))
return name
- workdir = bb.data.getVar('WORKDIR', d, True)
- outdir = bb.data.getVar('DEPLOY_DIR_IPK', d, True)
- tmpdir = bb.data.getVar('TMPDIR', d, True)
- pkgd = bb.data.getVar('PKGD', d, True)
- pkgdest = bb.data.getVar('PKGDEST', d, True)
+ workdir = d.getVar('WORKDIR', True)
+ outdir = d.getVar('DEPLOY_DIR_IPK', True)
+ tmpdir = d.getVar('TMPDIR', True)
+ pkgd = d.getVar('PKGD', True)
+ pkgdest = d.getVar('PKGDEST', True)
if not workdir or not outdir or not pkgd or not tmpdir:
bb.error("Variables incorrectly set, unable to package")
return
- packages = bb.data.getVar('PACKAGES', d, True)
+ packages = d.getVar('PACKAGES', True)
if not packages or packages == '':
bb.debug(1, "No packages; nothing to do")
return
# Construct the spec file...
- srcname = strip_multilib(bb.data.getVar('PN', d, True), d)
+ srcname = strip_multilib(d.getVar('PN', True), d)
outspecfile = workdir + "/" + srcname + ".spec"
- bb.data.setVar('OUTSPECFILE', outspecfile, d)
+ d.setVar('OUTSPECFILE', outspecfile)
bb.build.exec_func('write_specfile', d)
# Construct per file dependencies file
@@ -844,10 +844,10 @@ python do_package_rpm () {
outfile.write("\n# Dependency table\n")
for pkg in packages.split():
dependsflist_key = 'FILE' + varname + 'FLIST' + "_" + pkg
- dependsflist = (bb.data.getVar(dependsflist_key, d, True) or "")
+ dependsflist = (d.getVar(dependsflist_key, True) or "")
for dfile in dependsflist.split():
key = "FILE" + varname + "_" + dfile + "_" + pkg
- depends_dict = bb.utils.explode_dep_versions(bb.data.getVar(key, d, True) or "")
+ depends_dict = bb.utils.explode_dep_versions(d.getVar(key, True) or "")
file = dfile.replace("@underscore@", "_")
file = file.replace("@closebrace@", "]")
file = file.replace("@openbrace@", "[")
@@ -899,15 +899,15 @@ python do_package_rpm () {
os.chmod(outprovides, 0755)
# Setup the rpmbuild arguments...
- rpmbuild = bb.data.getVar('RPMBUILD', d, True)
- targetsys = bb.data.getVar('TARGET_SYS', d, True)
- targetvendor = bb.data.getVar('TARGET_VENDOR', d, True)
- package_arch = bb.data.getVar('PACKAGE_ARCH', d, True) or ""
+ rpmbuild = d.getVar('RPMBUILD', True)
+ targetsys = d.getVar('TARGET_SYS', True)
+ targetvendor = d.getVar('TARGET_VENDOR', True)
+ package_arch = d.getVar('PACKAGE_ARCH', True) or ""
if package_arch not in "all any noarch".split():
- ml_prefix = (bb.data.getVar('MLPREFIX', d, True) or "").replace("-", "_")
- bb.data.setVar('PACKAGE_ARCH_EXTEND', ml_prefix + package_arch, d)
+ ml_prefix = (d.getVar('MLPREFIX', True) or "").replace("-", "_")
+ d.setVar('PACKAGE_ARCH_EXTEND', ml_prefix + package_arch)
else:
- bb.data.setVar('PACKAGE_ARCH_EXTEND', package_arch, d)
+ d.setVar('PACKAGE_ARCH_EXTEND', package_arch)
pkgwritedir = bb.data.expand('${PKGWRITEDIRRPM}/${PACKAGE_ARCH_EXTEND}', d)
pkgarch = bb.data.expand('${PACKAGE_ARCH_EXTEND}${TARGET_VENDOR}-${TARGET_OS}', d)
magicfile = bb.data.expand('${STAGING_DIR_NATIVE}/usr/share/misc/magic.mgc', d)
@@ -927,19 +927,19 @@ python do_package_rpm () {
cmd = cmd + " -bb " + outspecfile
# Build the rpm package!
- bb.data.setVar('BUILDSPEC', cmd + "\n", d)
- bb.data.setVarFlag('BUILDSPEC', 'func', '1', d)
+ d.setVar('BUILDSPEC', cmd + "\n")
+ d.setVarFlag('BUILDSPEC', 'func', '1')
bb.build.exec_func('BUILDSPEC', d)
}
python () {
- if bb.data.getVar('PACKAGES', d, True) != '':
- deps = (bb.data.getVarFlag('do_package_write_rpm', 'depends', d) or "").split()
+ if d.getVar('PACKAGES', True) != '':
+ deps = (d.getVarFlag('do_package_write_rpm', 'depends') or "").split()
deps.append('rpm-native:do_populate_sysroot')
deps.append('virtual/fakeroot-native:do_populate_sysroot')
bb.data.setVarFlag('do_package_write_rpm', 'depends', " ".join(deps), d)
- bb.data.setVarFlag('do_package_write_rpm', 'fakeroot', 1, d)
- bb.data.setVarFlag('do_package_write_rpm_setscene', 'fakeroot', 1, d)
+ d.setVarFlag('do_package_write_rpm', 'fakeroot', 1)
+ d.setVarFlag('do_package_write_rpm_setscene', 'fakeroot', 1)
}
SSTATETASKS += "do_package_write_rpm"
diff --git a/meta/classes/package_tar.bbclass b/meta/classes/package_tar.bbclass
index a806e4514..f26a2c000 100644
--- a/meta/classes/package_tar.bbclass
+++ b/meta/classes/package_tar.bbclass
@@ -3,15 +3,15 @@ inherit package
IMAGE_PKGTYPE ?= "tar"
python package_tar_fn () {
- fn = os.path.join(bb.data.getVar('DEPLOY_DIR_TAR', d), "%s-%s-%s.tar.gz" % (bb.data.getVar('PKG', d), bb.data.getVar('PKGV', d), bb.data.getVar('PKGR', d)))
+ fn = os.path.join(d.getVar('DEPLOY_DIR_TAR'), "%s-%s-%s.tar.gz" % (d.getVar('PKG'), d.getVar('PKGV'), d.getVar('PKGR')))
fn = bb.data.expand(fn, d)
- bb.data.setVar('PKGFN', fn, d)
+ d.setVar('PKGFN', fn)
}
python package_tar_install () {
- pkg = bb.data.getVar('PKG', d, 1)
- pkgfn = bb.data.getVar('PKGFN', d, 1)
- rootfs = bb.data.getVar('IMAGE_ROOTFS', d, 1)
+ pkg = d.getVar('PKG', 1)
+ pkgfn = d.getVar('PKGFN', 1)
+ rootfs = d.getVar('IMAGE_ROOTFS', 1)
if None in (pkg,pkgfn,rootfs):
bb.error("missing variables (one or more of PKG, PKGFN, IMAGEROOTFS)")
@@ -35,24 +35,24 @@ python package_tar_install () {
}
python do_package_tar () {
- workdir = bb.data.getVar('WORKDIR', d, 1)
+ workdir = d.getVar('WORKDIR', 1)
if not workdir:
bb.error("WORKDIR not defined, unable to package")
return
- outdir = bb.data.getVar('DEPLOY_DIR_TAR', d, 1)
+ outdir = d.getVar('DEPLOY_DIR_TAR', 1)
if not outdir:
bb.error("DEPLOY_DIR_TAR not defined, unable to package")
return
bb.mkdirhier(outdir)
- dvar = bb.data.getVar('D', d, 1)
+ dvar = d.getVar('D', 1)
if not dvar:
bb.error("D not defined, unable to package")
return
bb.mkdirhier(dvar)
- packages = bb.data.getVar('PACKAGES', d, 1)
+ packages = d.getVar('PACKAGES', 1)
if not packages:
bb.debug(1, "PACKAGES not defined, nothing to package")
return
@@ -61,11 +61,11 @@ python do_package_tar () {
localdata = bb.data.createCopy(d)
root = "%s/install/%s" % (workdir, pkg)
- bb.data.setVar('ROOT', '', localdata)
- bb.data.setVar('ROOT_%s' % pkg, root, localdata)
- bb.data.setVar('PKG', pkg, localdata)
+ localdata.setVar('ROOT', '')
+ localdata.setVar('ROOT_%s' % pkg, root)
+ localdata.setVar('PKG', pkg)
- overrides = bb.data.getVar('OVERRIDES', localdata)
+ overrides = localdata.getVar('OVERRIDES')
if not overrides:
raise bb.build.FuncFailed('OVERRIDES not defined')
overrides = bb.data.expand(overrides, localdata)
@@ -73,17 +73,17 @@ python do_package_tar () {
bb.data.update_data(localdata)
- root = bb.data.getVar('ROOT', localdata)
+ root = localdata.getVar('ROOT')
bb.mkdirhier(root)
basedir = os.path.dirname(root)
pkgoutdir = outdir
bb.mkdirhier(pkgoutdir)
bb.build.exec_func('package_tar_fn', localdata)
- tarfn = bb.data.getVar('PKGFN', localdata, 1)
+ tarfn = localdata.getVar('PKGFN', 1)
os.chdir(root)
from glob import glob
if not glob('*'):
- bb.note("Not creating empty archive for %s-%s-%s" % (pkg, bb.data.getVar('PKGV', localdata, 1), bb.data.getVar('PKGR', localdata, 1)))
+ bb.note("Not creating empty archive for %s-%s-%s" % (pkg, localdata.getVar('PKGV', 1), localdata.getVar('PKGR', 1)))
continue
ret = os.system("tar -czf %s %s" % (tarfn, '.'))
if ret != 0:
@@ -91,12 +91,12 @@ python do_package_tar () {
}
python () {
- if bb.data.getVar('PACKAGES', d, True) != '':
- deps = (bb.data.getVarFlag('do_package_write_tar', 'depends', d) or "").split()
+ if d.getVar('PACKAGES', True) != '':
+ deps = (d.getVarFlag('do_package_write_tar', 'depends') or "").split()
deps.append('tar-native:do_populate_sysroot')
deps.append('virtual/fakeroot-native:do_populate_sysroot')
bb.data.setVarFlag('do_package_write_tar', 'depends', " ".join(deps), d)
- bb.data.setVarFlag('do_package_write_ipk', 'fakeroot', "1", d)
+ d.setVarFlag('do_package_write_ipk', 'fakeroot', "1")
}
diff --git a/meta/classes/packagedata.bbclass b/meta/classes/packagedata.bbclass
index bf051feea..9c7aede3b 100644
--- a/meta/classes/packagedata.bbclass
+++ b/meta/classes/packagedata.bbclass
@@ -1,13 +1,13 @@
python read_subpackage_metadata () {
import oe.packagedata
- data = oe.packagedata.read_pkgdata(bb.data.getVar('PN', d, 1), d)
+ data = oe.packagedata.read_pkgdata(d.getVar('PN', 1), d)
for key in data.keys():
- bb.data.setVar(key, data[key], d)
+ d.setVar(key, data[key])
- for pkg in bb.data.getVar('PACKAGES', d, 1).split():
+ for pkg in d.getVar('PACKAGES', 1).split():
sdata = oe.packagedata.read_subpkgdata(pkg, d)
for key in sdata.keys():
- bb.data.setVar(key, sdata[key], d)
+ d.setVar(key, sdata[key])
}
diff --git a/meta/classes/packagehistory.bbclass b/meta/classes/packagehistory.bbclass
index 492bbac21..2cdf9d8a7 100644
--- a/meta/classes/packagehistory.bbclass
+++ b/meta/classes/packagehistory.bbclass
@@ -10,8 +10,8 @@ PKGHIST_DIR = "${TMPDIR}/pkghistory/${BASEPKG_TARGET_SYS}/"
# for comparision when writing future packages
#
python emit_pkghistory() {
- packages = bb.data.getVar('PACKAGES', d, True)
- pkghistdir = bb.data.getVar('PKGHIST_DIR', d, True)
+ packages = d.getVar('PACKAGES', True)
+ pkghistdir = d.getVar('PKGHIST_DIR', True)
# Should check PACKAGES here to see if anything removed
@@ -72,14 +72,14 @@ def check_pkghistory(pkg, pe, pv, pr, lastversion):
def write_pkghistory(pkg, pe, pv, pr, d):
bb.debug(2, "Writing package history")
- pkghistdir = bb.data.getVar('PKGHIST_DIR', d, True)
+ pkghistdir = d.getVar('PKGHIST_DIR', True)
verpath = os.path.join(pkghistdir, pkg, pe, pv, pr)
if not os.path.exists(verpath):
os.makedirs(verpath)
def write_latestlink(pkg, pe, pv, pr, d):
- pkghistdir = bb.data.getVar('PKGHIST_DIR', d, True)
+ pkghistdir = d.getVar('PKGHIST_DIR', True)
def rm_link(path):
try:
diff --git a/meta/classes/patch.bbclass b/meta/classes/patch.bbclass
index 86046e1ff..b2b6d976c 100644
--- a/meta/classes/patch.bbclass
+++ b/meta/classes/patch.bbclass
@@ -10,7 +10,7 @@ inherit terminal
python patch_do_patch() {
import oe.patch
- src_uri = (bb.data.getVar('SRC_URI', d, 1) or '').split()
+ src_uri = (d.getVar('SRC_URI', 1) or '').split()
if not src_uri:
return
@@ -20,23 +20,23 @@ python patch_do_patch() {
"git": oe.patch.GitApplyTree,
}
- cls = patchsetmap[bb.data.getVar('PATCHTOOL', d, 1) or 'quilt']
+ cls = patchsetmap[d.getVar('PATCHTOOL', 1) or 'quilt']
resolvermap = {
"noop": oe.patch.NOOPResolver,
"user": oe.patch.UserResolver,
}
- rcls = resolvermap[bb.data.getVar('PATCHRESOLVE', d, 1) or 'user']
+ rcls = resolvermap[d.getVar('PATCHRESOLVE', 1) or 'user']
- s = bb.data.getVar('S', d, 1)
+ s = d.getVar('S', 1)
path = os.getenv('PATH')
- os.putenv('PATH', bb.data.getVar('PATH', d, 1))
+ os.putenv('PATH', d.getVar('PATH', 1))
classes = {}
- workdir = bb.data.getVar('WORKDIR', d, 1)
+ workdir = d.getVar('WORKDIR', 1)
for url in src_uri:
(type, host, path, user, pswd, parm) = bb.decodeurl(url)
@@ -76,13 +76,13 @@ python patch_do_patch() {
pname = os.path.basename(local)
if "mindate" in parm or "maxdate" in parm:
- pn = bb.data.getVar('PN', d, 1)
- srcdate = bb.data.getVar('SRCDATE_%s' % pn, d, 1)
+ pn = d.getVar('PN', 1)
+ srcdate = d.getVar('SRCDATE_%s' % pn, 1)
if not srcdate:
- srcdate = bb.data.getVar('SRCDATE', d, 1)
+ srcdate = d.getVar('SRCDATE', 1)
if srcdate == "now":
- srcdate = bb.data.getVar('DATE', d, 1)
+ srcdate = d.getVar('DATE', 1)
if "maxdate" in parm and parm["maxdate"] < srcdate:
bb.note("Patch '%s' is outdated" % pname)
@@ -94,25 +94,25 @@ python patch_do_patch() {
if "minrev" in parm:
- srcrev = bb.data.getVar('SRCREV', d, 1)
+ srcrev = d.getVar('SRCREV', 1)
if srcrev and srcrev < parm["minrev"]:
bb.note("Patch '%s' applies to later revisions" % pname)
continue
if "maxrev" in parm:
- srcrev = bb.data.getVar('SRCREV', d, 1)
+ srcrev = d.getVar('SRCREV', 1)
if srcrev and srcrev > parm["maxrev"]:
bb.note("Patch '%s' applies to earlier revisions" % pname)
continue
if "rev" in parm:
- srcrev = bb.data.getVar('SRCREV', d, 1)
+ srcrev = d.getVar('SRCREV', 1)
if srcrev and parm["rev"] not in srcrev:
bb.note("Patch '%s' doesn't apply to revision" % pname)
continue
if "notrev" in parm:
- srcrev = bb.data.getVar('SRCREV', d, 1)
+ srcrev = d.getVar('SRCREV', 1)
if srcrev and parm["notrev"] in srcrev:
bb.note("Patch '%s' doesn't apply to revision" % pname)
continue
diff --git a/meta/classes/pkg_distribute.bbclass b/meta/classes/pkg_distribute.bbclass
index 81978e3e3..52643a2f9 100644
--- a/meta/classes/pkg_distribute.bbclass
+++ b/meta/classes/pkg_distribute.bbclass
@@ -1,6 +1,6 @@
PKG_DISTRIBUTECOMMAND[func] = "1"
python do_distribute_packages () {
- cmd = bb.data.getVar('PKG_DISTRIBUTECOMMAND', d, 1)
+ cmd = d.getVar('PKG_DISTRIBUTECOMMAND', 1)
if not cmd:
raise bb.build.FuncFailed("Unable to distribute packages, PKG_DISTRIBUTECOMMAND not defined")
bb.build.exec_func('PKG_DISTRIBUTECOMMAND', d)
diff --git a/meta/classes/pkg_metainfo.bbclass b/meta/classes/pkg_metainfo.bbclass
index ac4f73c77..1714a535c 100644
--- a/meta/classes/pkg_metainfo.bbclass
+++ b/meta/classes/pkg_metainfo.bbclass
@@ -1,5 +1,5 @@
python do_pkg_write_metainfo () {
- deploydir = bb.data.getVar('DEPLOY_DIR', d, 1)
+ deploydir = d.getVar('DEPLOY_DIR', 1)
if not deploydir:
bb.error("DEPLOY_DIR not defined, unable to write package info")
return
@@ -9,11 +9,11 @@ python do_pkg_write_metainfo () {
except OSError:
raise bb.build.FuncFailed("unable to open package-info file for writing.")
- name = bb.data.getVar('PN', d, 1)
- version = bb.data.getVar('PV', d, 1)
- desc = bb.data.getVar('DESCRIPTION', d, 1)
- page = bb.data.getVar('HOMEPAGE', d, 1)
- lic = bb.data.getVar('LICENSE', d, 1)
+ name = d.getVar('PN', 1)
+ version = d.getVar('PV', 1)
+ desc = d.getVar('DESCRIPTION', 1)
+ page = d.getVar('HOMEPAGE', 1)
+ lic = d.getVar('LICENSE', 1)
infofile.write("|| "+ name +" || "+ version + " || "+ desc +" || "+ page +" || "+ lic + " ||\n" )
infofile.close()
diff --git a/meta/classes/populate_sdk_deb.bbclass b/meta/classes/populate_sdk_deb.bbclass
index be7b5520c..eedc80b26 100644
--- a/meta/classes/populate_sdk_deb.bbclass
+++ b/meta/classes/populate_sdk_deb.bbclass
@@ -2,8 +2,8 @@ do_populate_sdk[depends] += "dpkg-native:do_populate_sysroot apt-native:do_popul
do_populate_sdk[recrdeptask] += "do_package_write_deb"
-DEB_SDK_ARCH = "${@[bb.data.getVar('SDK_ARCH', d, 1), "i386"]\
- [bb.data.getVar('SDK_ARCH', d, 1) in \
+DEB_SDK_ARCH = "${@[d.getVar('SDK_ARCH', 1), "i386"]\
+ [d.getVar('SDK_ARCH', 1) in \
["x86", "i486", "i586", "i686", "pentium"]]}"
populate_sdk_post_deb () {
diff --git a/meta/classes/populate_sdk_rpm.bbclass b/meta/classes/populate_sdk_rpm.bbclass
index 9989d0abf..829d83a8c 100644
--- a/meta/classes/populate_sdk_rpm.bbclass
+++ b/meta/classes/populate_sdk_rpm.bbclass
@@ -127,6 +127,6 @@ python () {
localdata.setVar("DEFAULTTUNE", localdata.getVar("DEFAULTTUNE_virtclass-multilib-" + eext[1], False) or "")
ml_package_archs += localdata.getVar("PACKAGE_ARCHS", True) or ""
#bb.note("ML_PACKAGE_ARCHS %s %s %s" % (eext[1], localdata.getVar("PACKAGE_ARCHS", True) or "(none)", overrides))
- bb.data.setVar('MULTILIB_PACKAGE_ARCHS', ml_package_archs, d)
+ d.setVar('MULTILIB_PACKAGE_ARCHS', ml_package_archs)
}
diff --git a/meta/classes/qemu.bbclass b/meta/classes/qemu.bbclass
index 66dfb2b0d..1bdd209af 100644
--- a/meta/classes/qemu.bbclass
+++ b/meta/classes/qemu.bbclass
@@ -6,7 +6,7 @@
def qemu_target_binary(data):
import bb
- target_arch = bb.data.getVar("TARGET_ARCH", data, 1)
+ target_arch = data.getVar("TARGET_ARCH", 1)
if target_arch in ("i486", "i586", "i686"):
target_arch = "i386"
elif target_arch == "powerpc":
diff --git a/meta/classes/qt4e.bbclass b/meta/classes/qt4e.bbclass
index 670605ba4..d955aca5d 100644
--- a/meta/classes/qt4e.bbclass
+++ b/meta/classes/qt4e.bbclass
@@ -1,4 +1,4 @@
-DEPENDS_prepend = "${@["qt4-embedded ", ""][(bb.data.getVar('PN', d, 1)[:12] == 'qt4-embedded')]}"
+DEPENDS_prepend = "${@["qt4-embedded ", ""][(d.getVar('PN', 1)[:12] == 'qt4-embedded')]}"
inherit qmake2
diff --git a/meta/classes/qt4x11.bbclass b/meta/classes/qt4x11.bbclass
index ee2cdca3a..3f955d48d 100644
--- a/meta/classes/qt4x11.bbclass
+++ b/meta/classes/qt4x11.bbclass
@@ -1,4 +1,4 @@
-DEPENDS_prepend = "${@["qt4-x11-free ", ""][(bb.data.getVar('BPN', d, True)[:12] == 'qt4-x11-free')]}"
+DEPENDS_prepend = "${@["qt4-x11-free ", ""][(d.getVar('BPN', True)[:12] == 'qt4-x11-free')]}"
inherit qmake2
diff --git a/meta/classes/relocatable.bbclass b/meta/classes/relocatable.bbclass
index e665e317c..54227a91c 100644
--- a/meta/classes/relocatable.bbclass
+++ b/meta/classes/relocatable.bbclass
@@ -8,7 +8,7 @@ def process_dir (directory, d):
import stat
cmd = bb.data.expand('${CHRPATH_BIN}', d)
- tmpdir = bb.data.getVar('TMPDIR', d)
+ tmpdir = d.getVar('TMPDIR')
basedir = bb.data.expand('${base_prefix}', d)
#bb.debug("Checking %s for binaries to process" % directory)
diff --git a/meta/classes/rootfs_ipk.bbclass b/meta/classes/rootfs_ipk.bbclass
index ffc4a72c0..3b4c392ce 100644
--- a/meta/classes/rootfs_ipk.bbclass
+++ b/meta/classes/rootfs_ipk.bbclass
@@ -185,13 +185,13 @@ ipk_insert_feed_uris () {
python () {
- if bb.data.getVar('BUILD_IMAGES_FROM_FEEDS', d, True):
- flags = bb.data.getVarFlag('do_rootfs', 'recrdeptask', d)
+ if d.getVar('BUILD_IMAGES_FROM_FEEDS', True):
+ flags = d.getVarFlag('do_rootfs', 'recrdeptask')
flags = flags.replace("do_package_write_ipk", "")
flags = flags.replace("do_deploy", "")
flags = flags.replace("do_populate_sysroot", "")
- bb.data.setVarFlag('do_rootfs', 'recrdeptask', flags, d)
- bb.data.setVar('OPKG_PREPROCESS_COMMANDS', "package_generate_archlist\nipk_insert_feed_uris", d)
- bb.data.setVar('OPKG_POSTPROCESS_COMMANDS', '', d)
+ d.setVarFlag('do_rootfs', 'recrdeptask', flags)
+ d.setVar('OPKG_PREPROCESS_COMMANDS', "package_generate_archlist\nipk_insert_feed_uris")
+ d.setVar('OPKG_POSTPROCESS_COMMANDS', '')
}
diff --git a/meta/classes/rootfs_rpm.bbclass b/meta/classes/rootfs_rpm.bbclass
index 9f5076488..95e9455e5 100644
--- a/meta/classes/rootfs_rpm.bbclass
+++ b/meta/classes/rootfs_rpm.bbclass
@@ -200,14 +200,14 @@ install_all_locales() {
}
python () {
- if bb.data.getVar('BUILD_IMAGES_FROM_FEEDS', d, True):
- flags = bb.data.getVarFlag('do_rootfs', 'recrdeptask', d)
+ if d.getVar('BUILD_IMAGES_FROM_FEEDS', True):
+ flags = d.getVarFlag('do_rootfs', 'recrdeptask')
flags = flags.replace("do_package_write_rpm", "")
flags = flags.replace("do_deploy", "")
flags = flags.replace("do_populate_sysroot", "")
- bb.data.setVarFlag('do_rootfs', 'recrdeptask', flags, d)
- bb.data.setVar('RPM_PREPROCESS_COMMANDS', '', d)
- bb.data.setVar('RPM_POSTPROCESS_COMMANDS', '', d)
+ d.setVarFlag('do_rootfs', 'recrdeptask', flags)
+ d.setVar('RPM_PREPROCESS_COMMANDS', '')
+ d.setVar('RPM_POSTPROCESS_COMMANDS', '')
ml_package_archs = ""
ml_prefix_list = ""
@@ -224,6 +224,6 @@ python () {
ml_package_archs += " " + package_archs
ml_prefix_list += " " + eext[1]
#bb.note("ML_PACKAGE_ARCHS %s %s %s" % (eext[1], localdata.getVar("PACKAGE_ARCHS", True) or "(none)", overrides))
- bb.data.setVar('MULTILIB_PACKAGE_ARCHS', ml_package_archs, d)
- bb.data.setVar('MULTILIB_PREFIX_LIST', ml_prefix_list, d)
+ d.setVar('MULTILIB_PACKAGE_ARCHS', ml_package_archs)
+ d.setVar('MULTILIB_PREFIX_LIST', ml_prefix_list)
}
diff --git a/meta/classes/sanity.bbclass b/meta/classes/sanity.bbclass
index 838448f33..53b82d73e 100644
--- a/meta/classes/sanity.bbclass
+++ b/meta/classes/sanity.bbclass
@@ -14,7 +14,7 @@ def raise_sanity_error(msg):
def check_conf_exists(fn, data):
bbpath = []
fn = bb.data.expand(fn, data)
- vbbpath = bb.data.getVar("BBPATH", data)
+ vbbpath = data.getVar("BBPATH")
if vbbpath:
bbpath += vbbpath.split(":")
for p in bbpath:
@@ -87,12 +87,12 @@ def check_connectivity(d):
# URI's to check can be set in the CONNECTIVITY_CHECK_URIS variable
# using the same syntax as for SRC_URI. If the variable is not set
# the check is skipped
- test_uris = (bb.data.getVar('CONNECTIVITY_CHECK_URIS', d, True) or "").split()
+ test_uris = (d.getVar('CONNECTIVITY_CHECK_URIS', True) or "").split()
retval = ""
# Only check connectivity if network enabled and the
# CONNECTIVITY_CHECK_URIS are set
- network_enabled = not bb.data.getVar('BB_NO_NETWORK', d, True)
+ network_enabled = not d.getVar('BB_NO_NETWORK', True)
check_enabled = len(test_uris)
# Take a copy of the data store and unset MIRRORS and PREMIRROS
data = bb.data.createCopy(d)
@@ -105,7 +105,7 @@ def check_connectivity(d):
except Exception:
# Allow the message to be configured so that users can be
# pointed to a support mechanism.
- msg = bb.data.getVar('CONNECTIVITY_CHECK_MSG', data, True) or ""
+ msg = data.getVar('CONNECTIVITY_CHECK_MSG', True) or ""
if len(msg) == 0:
msg = "Failed to fetch test data from the network. Please ensure your network is configured correctly.\n"
retval = msg
@@ -450,7 +450,7 @@ def check_sanity(e):
addhandler check_sanity_eventhandler
python check_sanity_eventhandler() {
- if bb.event.getName(e) == "ConfigParsed" and bb.data.getVar("BB_WORKERCONTEXT", e.data, True) != "1":
+ if bb.event.getName(e) == "ConfigParsed" and e.data.getVar("BB_WORKERCONTEXT", True) != "1":
check_sanity(e)
return
diff --git a/meta/classes/siteconfig.bbclass b/meta/classes/siteconfig.bbclass
index 0813c2543..880dcad1f 100644
--- a/meta/classes/siteconfig.bbclass
+++ b/meta/classes/siteconfig.bbclass
@@ -2,7 +2,7 @@ python siteconfig_do_siteconfig () {
shared_state = sstate_state_fromvars(d)
if shared_state['name'] != 'populate-sysroot':
return
- if not os.path.isdir(os.path.join(bb.data.getVar('FILE_DIRNAME', d, 1), 'site_config')):
+ if not os.path.isdir(os.path.join(d.getVar('FILE_DIRNAME', 1), 'site_config')):
bb.debug(1, "No site_config directory, skipping do_siteconfig")
return
bb.build.exec_func('do_siteconfig_gencache', d)
diff --git a/meta/classes/siteinfo.bbclass b/meta/classes/siteinfo.bbclass
index 02294c4d2..604c6ba6b 100644
--- a/meta/classes/siteinfo.bbclass
+++ b/meta/classes/siteinfo.bbclass
@@ -130,7 +130,7 @@ def siteinfo_get_files(d, no_cache = False):
if no_cache: return sitefiles
# Now check for siteconfig cache files
- path_siteconfig = bb.data.getVar('SITECONFIG_SYSROOTCACHE', d, 1)
+ path_siteconfig = d.getVar('SITECONFIG_SYSROOTCACHE', 1)
if os.path.isdir(path_siteconfig):
for i in os.listdir(path_siteconfig):
filename = os.path.join(path_siteconfig, i)
diff --git a/meta/classes/sourcepkg.bbclass b/meta/classes/sourcepkg.bbclass
index 64e6d7c86..38edfe4e2 100644
--- a/meta/classes/sourcepkg.bbclass
+++ b/meta/classes/sourcepkg.bbclass
@@ -6,12 +6,12 @@ DISTRO ?= "openembedded"
def get_src_tree(d):
- workdir = bb.data.getVar('WORKDIR', d, 1)
+ workdir = d.getVar('WORKDIR', 1)
if not workdir:
bb.error("WORKDIR not defined, unable to find source tree.")
return
- s = bb.data.getVar('S', d, 0)
+ s = d.getVar('S', 0)
if not s:
bb.error("S not defined, unable to find source tree.")
return
@@ -55,8 +55,8 @@ sourcepkg_do_archive_bb() {
python sourcepkg_do_dumpdata() {
- workdir = bb.data.getVar('WORKDIR', d, 1)
- distro = bb.data.getVar('DISTRO', d, 1)
+ workdir = d.getVar('WORKDIR', 1)
+ distro = d.getVar('DISTRO', 1)
s_tree = get_src_tree(d)
openembeddeddir = os.path.join(workdir, s_tree, distro)
dumpfile = os.path.join(openembeddeddir, bb.data.expand("${P}-${PR}.showdata.dump",d))
@@ -73,8 +73,8 @@ python sourcepkg_do_dumpdata() {
bb.data.emit_env(f, d, True)
# emit the metadata which isnt valid shell
for e in d.keys():
- if bb.data.getVarFlag(e, 'python', d):
- f.write("\npython %s () {\n%s}\n" % (e, bb.data.getVar(e, d, 1)))
+ if d.getVarFlag(e, 'python'):
+ f.write("\npython %s () {\n%s}\n" % (e, d.getVar(e, 1)))
f.close()
}
diff --git a/meta/classes/src_distribute.bbclass b/meta/classes/src_distribute.bbclass
index fbfbdf009..2069d652a 100644
--- a/meta/classes/src_distribute.bbclass
+++ b/meta/classes/src_distribute.bbclass
@@ -3,12 +3,12 @@ python do_distribute_sources () {
l = bb.data.createCopy(d)
bb.data.update_data(l)
- sources_dir = bb.data.getVar('SRC_DISTRIBUTEDIR', d, 1)
- src_uri = bb.data.getVar('SRC_URI', d, 1).split()
+ sources_dir = d.getVar('SRC_DISTRIBUTEDIR', 1)
+ src_uri = d.getVar('SRC_URI', 1).split()
fetcher = bb.fetch2.Fetch(src_uri, d)
ud = fetcher.ud
- licenses = bb.data.getVar('LICENSE', d, 1).replace('&', '|')
+ licenses = d.getVar('LICENSE', 1).replace('&', '|')
licenses = licenses.replace('(', '').replace(')', '')
clean_licenses = ""
for x in licenses.split():
@@ -20,20 +20,20 @@ python do_distribute_sources () {
for license in clean_licenses.split('|'):
for url in ud.values():
- cmd = bb.data.getVar('SRC_DISTRIBUTECOMMAND', d, 1)
+ cmd = d.getVar('SRC_DISTRIBUTECOMMAND', 1)
if not cmd:
raise bb.build.FuncFailed("Unable to distribute sources, SRC_DISTRIBUTECOMMAND not defined")
url.setup_localpath(d)
- bb.data.setVar('SRC', url.localpath, d)
+ d.setVar('SRC', url.localpath)
if url.type == 'file':
if url.basename == '*':
import os.path
dest_dir = os.path.basename(os.path.dirname(os.path.abspath(url.localpath)))
- bb.data.setVar('DEST', "%s_%s/" % (bb.data.getVar('PF', d, 1), dest_dir), d)
+ bb.data.setVar('DEST', "%s_%s/" % (d.getVar('PF', 1), dest_dir), d)
else:
- bb.data.setVar('DEST', "%s_%s" % (bb.data.getVar('PF', d, 1), url.basename), d)
+ bb.data.setVar('DEST', "%s_%s" % (d.getVar('PF', 1), url.basename), d)
else:
- bb.data.setVar('DEST', '', d)
+ d.setVar('DEST', '')
bb.data.setVar('SRC_DISTRIBUTEDIR', "%s/%s" % (sources_dir, license), d)
bb.build.exec_func('SRC_DISTRIBUTECOMMAND', d)
diff --git a/meta/classes/sstate.bbclass b/meta/classes/sstate.bbclass
index 6abf55bb1..a777c79bb 100644
--- a/meta/classes/sstate.bbclass
+++ b/meta/classes/sstate.bbclass
@@ -20,7 +20,7 @@ SSTATEPOSTINSTFUNCS ?= ""
python () {
if bb.data.inherits_class('native', d):
- bb.data.setVar('SSTATE_PKGARCH', bb.data.getVar('BUILD_ARCH', d), d)
+ bb.data.setVar('SSTATE_PKGARCH', d.getVar('BUILD_ARCH'), d)
elif bb.data.inherits_class('cross', d):
bb.data.setVar('SSTATE_PKGARCH', bb.data.expand("${BUILD_ARCH}_${TUNE_PKGARCH}", d), d)
bb.data.setVar('SSTATE_MANMACH', bb.data.expand("${BUILD_ARCH}_${MACHINE}", d), d)
@@ -37,19 +37,19 @@ python () {
# reused if we manipulate the paths
if bb.data.inherits_class('native', d) or bb.data.inherits_class('cross', d) or bb.data.inherits_class('sdk', d) or bb.data.inherits_class('crosssdk', d):
scan_cmd = "grep -Irl ${STAGING_DIR} ${SSTATE_BUILDDIR}"
- bb.data.setVar('SSTATE_SCAN_CMD', scan_cmd, d)
+ d.setVar('SSTATE_SCAN_CMD', scan_cmd)
- unique_tasks = set((bb.data.getVar('SSTATETASKS', d, True) or "").split())
+ unique_tasks = set((d.getVar('SSTATETASKS', True) or "").split())
d.setVar('SSTATETASKS', " ".join(unique_tasks))
namemap = []
for task in unique_tasks:
- namemap.append(bb.data.getVarFlag(task, 'sstate-name', d))
- funcs = bb.data.getVarFlag(task, 'prefuncs', d) or ""
+ namemap.append(d.getVarFlag(task, 'sstate-name'))
+ funcs = d.getVarFlag(task, 'prefuncs') or ""
funcs = "sstate_task_prefunc " + funcs
- bb.data.setVarFlag(task, 'prefuncs', funcs, d)
- funcs = bb.data.getVarFlag(task, 'postfuncs', d) or ""
+ d.setVarFlag(task, 'prefuncs', funcs)
+ funcs = d.getVarFlag(task, 'postfuncs') or ""
funcs = funcs + " sstate_task_postfunc"
- bb.data.setVarFlag(task, 'postfuncs', funcs, d)
+ d.setVarFlag(task, 'postfuncs', funcs)
d.setVar('SSTATETASKNAMES', " ".join(namemap))
}
@@ -65,18 +65,18 @@ def sstate_init(name, task, d):
def sstate_state_fromvars(d, task = None):
if task is None:
- task = bb.data.getVar('BB_CURRENTTASK', d, True)
+ task = d.getVar('BB_CURRENTTASK', True)
if not task:
bb.fatal("sstate code running without task context?!")
task = task.replace("_setscene", "")
- name = bb.data.expand(bb.data.getVarFlag("do_" + task, 'sstate-name', d), d)
- inputs = (bb.data.expand(bb.data.getVarFlag("do_" + task, 'sstate-inputdirs', d) or "", d)).split()
- outputs = (bb.data.expand(bb.data.getVarFlag("do_" + task, 'sstate-outputdirs', d) or "", d)).split()
- plaindirs = (bb.data.expand(bb.data.getVarFlag("do_" + task, 'sstate-plaindirs', d) or "", d)).split()
- lockfiles = (bb.data.expand(bb.data.getVarFlag("do_" + task, 'sstate-lockfile', d) or "", d)).split()
- lockfilesshared = (bb.data.expand(bb.data.getVarFlag("do_" + task, 'sstate-lockfile-shared', d) or "", d)).split()
- interceptfuncs = (bb.data.expand(bb.data.getVarFlag("do_" + task, 'sstate-interceptfuncs', d) or "", d)).split()
+ name = bb.data.expand(d.getVarFlag("do_" + task, 'sstate-name'), d)
+ inputs = (bb.data.expand(d.getVarFlag("do_" + task, 'sstate-inputdirs') or "", d)).split()
+ outputs = (bb.data.expand(d.getVarFlag("do_" + task, 'sstate-outputdirs') or "", d)).split()
+ plaindirs = (bb.data.expand(d.getVarFlag("do_" + task, 'sstate-plaindirs') or "", d)).split()
+ lockfiles = (bb.data.expand(d.getVarFlag("do_" + task, 'sstate-lockfile') or "", d)).split()
+ lockfilesshared = (bb.data.expand(d.getVarFlag("do_" + task, 'sstate-lockfile-shared') or "", d)).split()
+ interceptfuncs = (bb.data.expand(d.getVarFlag("do_" + task, 'sstate-interceptfuncs') or "", d)).split()
if not name or len(inputs) != len(outputs):
bb.fatal("sstate variables not setup correctly?!")
@@ -139,7 +139,7 @@ def sstate_install(ss, d):
f.write(di + "\n")
f.close()
- for postinst in (bb.data.getVar('SSTATEPOSTINSTFUNCS', d, True) or '').split():
+ for postinst in (d.getVar('SSTATEPOSTINSTFUNCS', True) or '').split():
bb.build.exec_func(postinst, d)
for lock in locks:
@@ -156,7 +156,7 @@ def sstate_installpkg(ss, d):
oe.path.remove(dir)
sstateinst = bb.data.expand("${WORKDIR}/sstate-install-%s/" % ss['name'], d)
- sstatepkg = bb.data.getVar('SSTATE_PKG', d, True) + '_' + ss['name'] + ".tgz"
+ sstatepkg = d.getVar('SSTATE_PKG', True) + '_' + ss['name'] + ".tgz"
if not os.path.exists(sstatepkg):
pstaging_fetch(sstatepkg, d)
@@ -167,16 +167,16 @@ def sstate_installpkg(ss, d):
sstate_clean(ss, d)
- bb.data.setVar('SSTATE_INSTDIR', sstateinst, d)
- bb.data.setVar('SSTATE_PKG', sstatepkg, d)
+ d.setVar('SSTATE_INSTDIR', sstateinst)
+ d.setVar('SSTATE_PKG', sstatepkg)
bb.build.exec_func('sstate_unpack_package', d)
# Fixup hardcoded paths
fixmefn = sstateinst + "fixmepath"
if os.path.isfile(fixmefn):
- staging = bb.data.getVar('STAGING_DIR', d, True)
- staging_target = bb.data.getVar('STAGING_DIR_TARGET', d, True)
- staging_host = bb.data.getVar('STAGING_DIR_HOST', d, True)
+ staging = d.getVar('STAGING_DIR', True)
+ staging_target = d.getVar('STAGING_DIR_TARGET', True)
+ staging_host = d.getVar('STAGING_DIR_HOST', True)
fixmefd = open(fixmefn, "r")
fixmefiles = fixmefd.readlines()
fixmefd.close()
@@ -206,13 +206,13 @@ def sstate_installpkg(ss, d):
def sstate_clean_cachefile(ss, d):
import oe.path
- sstatepkgdir = bb.data.getVar('SSTATE_DIR', d, True)
- sstatepkgfile = sstatepkgdir + '/' + bb.data.getVar('SSTATE_PKGSPEC', d, True) + "*_" + ss['name'] + ".tgz*"
+ sstatepkgdir = d.getVar('SSTATE_DIR', True)
+ sstatepkgfile = sstatepkgdir + '/' + d.getVar('SSTATE_PKGSPEC', True) + "*_" + ss['name'] + ".tgz*"
bb.note("Removing %s" % sstatepkgfile)
oe.path.remove(sstatepkgfile)
def sstate_clean_cachefiles(d):
- for task in (bb.data.getVar('SSTATETASKS', d, True) or "").split():
+ for task in (d.getVar('SSTATETASKS', True) or "").split():
ss = sstate_state_fromvars(d, task[3:])
sstate_clean_cachefile(ss, d)
@@ -274,10 +274,10 @@ CLEANFUNCS += "sstate_cleanall"
python sstate_cleanall() {
import fnmatch
- bb.note("Removing shared state for package %s" % bb.data.getVar('PN', d, True))
+ bb.note("Removing shared state for package %s" % d.getVar('PN', True))
- manifest_dir = bb.data.getVar('SSTATE_MANIFESTS', d, True)
- manifest_prefix = bb.data.getVar("SSTATE_MANFILEPREFIX", d, True)
+ manifest_dir = d.getVar('SSTATE_MANIFESTS', True)
+ manifest_prefix = d.getVar("SSTATE_MANFILEPREFIX", True)
manifest_pattern = os.path.basename(manifest_prefix) + ".*"
if not os.path.exists(manifest_dir):
@@ -298,7 +298,7 @@ python sstate_cleanall() {
def sstate_hardcode_path(d):
# Need to remove hardcoded paths and fix these when we install the
# staging packages.
- sstate_scan_cmd = bb.data.getVar('SSTATE_SCAN_CMD', d, True)
+ sstate_scan_cmd = d.getVar('SSTATE_SCAN_CMD', True)
p = os.popen("%s" % sstate_scan_cmd)
file_list = p.read()
@@ -306,10 +306,10 @@ def sstate_hardcode_path(d):
p.close()
return
- staging = bb.data.getVar('STAGING_DIR', d, True)
- staging_target = bb.data.getVar('STAGING_DIR_TARGET', d, True)
- staging_host = bb.data.getVar('STAGING_DIR_HOST', d, True)
- sstate_builddir = bb.data.getVar('SSTATE_BUILDDIR', d, True)
+ staging = d.getVar('STAGING_DIR', True)
+ staging_target = d.getVar('STAGING_DIR_TARGET', True)
+ staging_host = d.getVar('STAGING_DIR_HOST', True)
+ sstate_builddir = d.getVar('SSTATE_BUILDDIR', True)
for i in file_list.split('\n'):
if not i:
@@ -349,10 +349,10 @@ def sstate_package(ss, d):
os.remove(path)
os.symlink(base, path)
- tmpdir = bb.data.getVar('TMPDIR', d, True)
+ tmpdir = d.getVar('TMPDIR', True)
sstatebuild = bb.data.expand("${WORKDIR}/sstate-build-%s/" % ss['name'], d)
- sstatepkg = bb.data.getVar('SSTATE_PKG', d, True) + '_'+ ss['name'] + ".tgz"
+ sstatepkg = d.getVar('SSTATE_PKG', True) + '_'+ ss['name'] + ".tgz"
bb.mkdirhier(sstatebuild)
bb.mkdirhier(os.path.dirname(sstatepkg))
for state in ss['dirs']:
@@ -369,15 +369,15 @@ def sstate_package(ss, d):
bb.debug(2, "Preparing tree %s for packaging at %s" % (state[1], sstatebuild + state[0]))
oe.path.copytree(state[1], sstatebuild + state[0])
- workdir = bb.data.getVar('WORKDIR', d, True)
+ workdir = d.getVar('WORKDIR', True)
for plain in ss['plaindirs']:
pdir = plain.replace(workdir, sstatebuild)
bb.mkdirhier(plain)
bb.mkdirhier(pdir)
oe.path.copytree(plain, pdir)
- bb.data.setVar('SSTATE_BUILDDIR', sstatebuild, d)
- bb.data.setVar('SSTATE_PKG', sstatepkg, d)
+ d.setVar('SSTATE_BUILDDIR', sstatebuild)
+ d.setVar('SSTATE_PKG', sstatepkg)
sstate_hardcode_path(d)
bb.build.exec_func('sstate_create_package', d)
@@ -389,7 +389,7 @@ def pstaging_fetch(sstatepkg, d):
import bb.fetch2
# Only try and fetch if the user has configured a mirror
- mirrors = bb.data.getVar('SSTATE_MIRRORS', d, True)
+ mirrors = d.getVar('SSTATE_MIRRORS', True)
if not mirrors:
return
@@ -402,9 +402,9 @@ def pstaging_fetch(sstatepkg, d):
bb.mkdirhier(dldir)
- bb.data.setVar('DL_DIR', dldir, localdata)
- bb.data.setVar('PREMIRRORS', mirrors, localdata)
- bb.data.setVar('SRC_URI', srcuri, localdata)
+ localdata.setVar('DL_DIR', dldir)
+ localdata.setVar('PREMIRRORS', mirrors)
+ localdata.setVar('SRC_URI', srcuri)
# Try a fetch from the sstate mirror, if it fails just return and
# we will build the package
@@ -493,15 +493,15 @@ def sstate_checkhashes(sq_fn, sq_task, sq_hash, sq_hashfn, d):
else:
bb.debug(2, "SState: Looked for but didn't find file %s" % sstatefile)
- mirrors = bb.data.getVar("SSTATE_MIRRORS", d, True)
+ mirrors = d.getVar("SSTATE_MIRRORS", True)
if mirrors:
# Copy the data object and override DL_DIR and SRC_URI
localdata = bb.data.createCopy(d)
bb.data.update_data(localdata)
dldir = bb.data.expand("${SSTATE_DIR}", localdata)
- bb.data.setVar('DL_DIR', dldir, localdata)
- bb.data.setVar('PREMIRRORS', mirrors, localdata)
+ localdata.setVar('DL_DIR', dldir)
+ localdata.setVar('PREMIRRORS', mirrors)
bb.debug(2, "SState using premirror of: %s" % mirrors)
@@ -513,7 +513,7 @@ def sstate_checkhashes(sq_fn, sq_task, sq_hash, sq_hashfn, d):
sstatefile = sstatefile.replace("${BB_TASKHASH}", sq_hash[task])
srcuri = "file://" + os.path.basename(sstatefile)
- bb.data.setVar('SRC_URI', srcuri, localdata)
+ localdata.setVar('SRC_URI', srcuri)
bb.debug(2, "SState: Attempting to fetch %s" % srcuri)
try:
diff --git a/meta/classes/staging.bbclass b/meta/classes/staging.bbclass
index 04d51ede4..831840456 100644
--- a/meta/classes/staging.bbclass
+++ b/meta/classes/staging.bbclass
@@ -84,7 +84,7 @@ python do_populate_sysroot () {
#
bb.build.exec_func("sysroot_stage_all", d)
- for f in (bb.data.getVar('SYSROOT_PREPROCESS_FUNCS', d, True) or '').split():
+ for f in (d.getVar('SYSROOT_PREPROCESS_FUNCS', True) or '').split():
bb.build.exec_func(f, d)
}
@@ -100,8 +100,8 @@ python do_populate_sysroot_setscene () {
addtask do_populate_sysroot_setscene
python () {
- if bb.data.getVar('do_stage', d, True) is not None:
- bb.fatal("Legacy staging found for %s as it has a do_stage function. This will need conversion to a do_install or often simply removal to work with Poky" % bb.data.getVar("FILE", d, True))
+ if d.getVar('do_stage', True) is not None:
+ bb.fatal("Legacy staging found for %s as it has a do_stage function. This will need conversion to a do_install or often simply removal to work with Poky" % d.getVar("FILE", True))
}
diff --git a/meta/classes/syslinux.bbclass b/meta/classes/syslinux.bbclass
index fb7597470..0cc6b851b 100644
--- a/meta/classes/syslinux.bbclass
+++ b/meta/classes/syslinux.bbclass
@@ -8,12 +8,12 @@ python build_syslinux_menu () {
import copy
import sys
- workdir = bb.data.getVar('WORKDIR', d, 1)
+ workdir = d.getVar('WORKDIR', 1)
if not workdir:
bb.error("WORKDIR is not defined")
return
- labels = bb.data.getVar('LABELS', d, 1)
+ labels = d.getVar('LABELS', 1)
if not labels:
bb.debug(1, "LABELS not defined, nothing to do")
return
@@ -22,7 +22,7 @@ python build_syslinux_menu () {
bb.debug(1, "No labels, nothing to do")
return
- cfile = bb.data.getVar('SYSLINUXMENU', d, 1)
+ cfile = d.getVar('SYSLINUXMENU', 1)
if not cfile:
raise bb.build.FuncFailed('Unable to read SYSLINUXMENU')
@@ -45,15 +45,15 @@ python build_syslinux_menu () {
from copy import deepcopy
localdata = deepcopy(d)
- overrides = bb.data.getVar('OVERRIDES', localdata)
+ overrides = localdata.getVar('OVERRIDES')
if not overrides:
raise bb.build.FuncFailed('OVERRIDES not defined')
overrides = bb.data.expand(overrides, localdata)
- bb.data.setVar('OVERRIDES', label + ':' + overrides, localdata)
+ localdata.setVar('OVERRIDES', label + ':' + overrides)
bb.data.update_data(localdata)
- usage = bb.data.getVar('USAGE', localdata, 1)
+ usage = localdata.getVar('USAGE', 1)
cfgfile.write(' \x0F\x30\x3E%16s\x0F\x30\x37: ' % (label))
cfgfile.write('%s\n' % (usage))
@@ -67,12 +67,12 @@ python build_syslinux_cfg () {
import copy
import sys
- workdir = bb.data.getVar('WORKDIR', d, 1)
+ workdir = d.getVar('WORKDIR', 1)
if not workdir:
bb.error("WORKDIR not defined, unable to package")
return
- labels = bb.data.getVar('LABELS', d, 1)
+ labels = d.getVar('LABELS', 1)
if not labels:
bb.debug(1, "LABELS not defined, nothing to do")
return
@@ -81,7 +81,7 @@ python build_syslinux_cfg () {
bb.debug(1, "No labels, nothing to do")
return
- cfile = bb.data.getVar('SYSLINUXCFG', d, 1)
+ cfile = d.getVar('SYSLINUXCFG', 1)
if not cfile:
raise bb.build.FuncFailed('Unable to read SYSLINUXCFG')
@@ -98,7 +98,7 @@ python build_syslinux_cfg () {
cfgfile.write('# Automatically created by OE\n')
- opts = bb.data.getVar('SYSLINUX_OPTS', d, 1)
+ opts = d.getVar('SYSLINUX_OPTS', 1)
if opts:
for opt in opts.split(';'):
@@ -107,7 +107,7 @@ python build_syslinux_cfg () {
cfgfile.write('ALLOWOPTIONS 1\n');
cfgfile.write('DEFAULT %s\n' % (labels.split()[0]))
- timeout = bb.data.getVar('SYSLINUX_TIMEOUT', d, 1)
+ timeout = d.getVar('SYSLINUX_TIMEOUT', 1)
if timeout:
cfgfile.write('TIMEOUT %s\n' % timeout)
@@ -116,29 +116,29 @@ python build_syslinux_cfg () {
cfgfile.write('PROMPT 1\n')
- menu = bb.data.getVar('AUTO_SYSLINUXMENU', d, 1)
+ menu = d.getVar('AUTO_SYSLINUXMENU', 1)
# This is ugly. My bad.
if menu:
bb.build.exec_func('build_syslinux_menu', d)
- mfile = bb.data.getVar('SYSLINUXMENU', d, 1)
+ mfile = d.getVar('SYSLINUXMENU', 1)
cfgfile.write('DISPLAY %s\n' % (mfile.split('/')[-1]) )
for label in labels.split():
localdata = bb.data.createCopy(d)
- overrides = bb.data.getVar('OVERRIDES', localdata, True)
+ overrides = localdata.getVar('OVERRIDES', True)
if not overrides:
raise bb.build.FuncFailed('OVERRIDES not defined')
- bb.data.setVar('OVERRIDES', label + ':' + overrides, localdata)
+ localdata.setVar('OVERRIDES', label + ':' + overrides)
bb.data.update_data(localdata)
cfgfile.write('LABEL %s\nKERNEL vmlinuz\n' % (label))
- append = bb.data.getVar('APPEND', localdata, 1)
- initrd = bb.data.getVar('INITRD', localdata, 1)
+ append = localdata.getVar('APPEND', 1)
+ initrd = localdata.getVar('INITRD', 1)
if append:
cfgfile.write('APPEND ')
diff --git a/meta/classes/task.bbclass b/meta/classes/task.bbclass
index 4edd70482..516d1a1c2 100644
--- a/meta/classes/task.bbclass
+++ b/meta/classes/task.bbclass
@@ -17,7 +17,7 @@ PACKAGE_ARCH = "all"
# to the list. Their dependencies (RRECOMMENDS) are handled as usual
# by package_depchains in a following step.
python () {
- packages = bb.data.getVar('PACKAGES', d, 1).split()
+ packages = d.getVar('PACKAGES', 1).split()
genpackages = []
for pkg in packages:
for postfix in ['-dbg', '-dev']:
diff --git a/meta/classes/toolchain-scripts.bbclass b/meta/classes/toolchain-scripts.bbclass
index 875efed02..c936a27a4 100644
--- a/meta/classes/toolchain-scripts.bbclass
+++ b/meta/classes/toolchain-scripts.bbclass
@@ -137,8 +137,8 @@ toolchain_create_sdk_version () {
}
python __anonymous () {
- deps = bb.data.getVarFlag('do_configure', 'depends', d) or ""
- for dep in (bb.data.getVar('TOOLCHAIN_NEED_CONFIGSITE_CACHE', d, True) or "").split():
+ deps = d.getVarFlag('do_configure', 'depends') or ""
+ for dep in (d.getVar('TOOLCHAIN_NEED_CONFIGSITE_CACHE', True) or "").split():
deps += " %s:do_populate_sysroot" % dep
- bb.data.setVarFlag('do_configure', 'depends', deps, d)
+ d.setVarFlag('do_configure', 'depends', deps)
}
diff --git a/meta/classes/update-alternatives.bbclass b/meta/classes/update-alternatives.bbclass
index ba812210a..e5ba6550d 100644
--- a/meta/classes/update-alternatives.bbclass
+++ b/meta/classes/update-alternatives.bbclass
@@ -78,38 +78,38 @@ fi
}
def update_alternatives_after_parse(d):
- if bb.data.getVar('ALTERNATIVE_LINKS', d) != None:
- doinstall = bb.data.getVar('do_install', d, 0)
- doinstall += bb.data.getVar('update_alternatives_batch_doinstall', d, 0)
- bb.data.setVar('do_install', doinstall, d)
+ if d.getVar('ALTERNATIVE_LINKS') != None:
+ doinstall = d.getVar('do_install', 0)
+ doinstall += d.getVar('update_alternatives_batch_doinstall', 0)
+ d.setVar('do_install', doinstall)
return
- if bb.data.getVar('ALTERNATIVE_NAME', d) == None:
- raise bb.build.FuncFailed, "%s inherits update-alternatives but doesn't set ALTERNATIVE_NAME" % bb.data.getVar('FILE', d)
- if bb.data.getVar('ALTERNATIVE_PATH', d) == None:
- raise bb.build.FuncFailed, "%s inherits update-alternatives but doesn't set ALTERNATIVE_PATH" % bb.data.getVar('FILE', d)
+ if d.getVar('ALTERNATIVE_NAME') == None:
+ raise bb.build.FuncFailed, "%s inherits update-alternatives but doesn't set ALTERNATIVE_NAME" % d.getVar('FILE')
+ if d.getVar('ALTERNATIVE_PATH') == None:
+ raise bb.build.FuncFailed, "%s inherits update-alternatives but doesn't set ALTERNATIVE_PATH" % d.getVar('FILE')
python __anonymous() {
update_alternatives_after_parse(d)
}
python populate_packages_prepend () {
- pkg = bb.data.getVar('PN', d, 1)
+ pkg = d.getVar('PN', 1)
bb.note('adding update-alternatives calls to postinst/postrm for %s' % pkg)
- postinst = bb.data.getVar('pkg_postinst_%s' % pkg, d, 1) or bb.data.getVar('pkg_postinst', d, 1)
+ postinst = d.getVar('pkg_postinst_%s' % pkg, 1) or d.getVar('pkg_postinst', 1)
if not postinst:
postinst = '#!/bin/sh\n'
- if bb.data.getVar('ALTERNATIVE_LINKS', d) != None:
- postinst += bb.data.getVar('update_alternatives_batch_postinst', d, 1)
+ if d.getVar('ALTERNATIVE_LINKS') != None:
+ postinst += d.getVar('update_alternatives_batch_postinst', 1)
else:
- postinst += bb.data.getVar('update_alternatives_postinst', d, 1)
- bb.data.setVar('pkg_postinst_%s' % pkg, postinst, d)
- postrm = bb.data.getVar('pkg_postrm_%s' % pkg, d, 1) or bb.data.getVar('pkg_postrm', d, 1)
+ postinst += d.getVar('update_alternatives_postinst', 1)
+ d.setVar('pkg_postinst_%s' % pkg, postinst)
+ postrm = d.getVar('pkg_postrm_%s' % pkg, 1) or d.getVar('pkg_postrm', 1)
if not postrm:
postrm = '#!/bin/sh\n'
- if bb.data.getVar('ALTERNATIVE_LINKS', d) != None:
- postrm += bb.data.getVar('update_alternatives_batch_postrm', d, 1)
+ if d.getVar('ALTERNATIVE_LINKS') != None:
+ postrm += d.getVar('update_alternatives_batch_postrm', 1)
else:
- postrm += bb.data.getVar('update_alternatives_postrm', d, 1)
- bb.data.setVar('pkg_postrm_%s' % pkg, postrm, d)
+ postrm += d.getVar('update_alternatives_postrm', 1)
+ d.setVar('pkg_postrm_%s' % pkg, postrm)
}
diff --git a/meta/classes/update-rc.d.bbclass b/meta/classes/update-rc.d.bbclass
index 7e4dda7cb..492c5fba2 100644
--- a/meta/classes/update-rc.d.bbclass
+++ b/meta/classes/update-rc.d.bbclass
@@ -30,11 +30,11 @@ update-rc.d $D ${INITSCRIPT_NAME} remove
def update_rc_after_parse(d):
- if bb.data.getVar('INITSCRIPT_PACKAGES', d) == None:
- if bb.data.getVar('INITSCRIPT_NAME', d) == None:
- raise bb.build.FuncFailed, "%s inherits update-rc.d but doesn't set INITSCRIPT_NAME" % bb.data.getVar('FILE', d)
- if bb.data.getVar('INITSCRIPT_PARAMS', d) == None:
- raise bb.build.FuncFailed, "%s inherits update-rc.d but doesn't set INITSCRIPT_PARAMS" % bb.data.getVar('FILE', d)
+ if d.getVar('INITSCRIPT_PACKAGES') == None:
+ if d.getVar('INITSCRIPT_NAME') == None:
+ raise bb.build.FuncFailed, "%s inherits update-rc.d but doesn't set INITSCRIPT_NAME" % d.getVar('FILE')
+ if d.getVar('INITSCRIPT_PARAMS') == None:
+ raise bb.build.FuncFailed, "%s inherits update-rc.d but doesn't set INITSCRIPT_PARAMS" % d.getVar('FILE')
python __anonymous() {
update_rc_after_parse(d)
@@ -44,7 +44,7 @@ python populate_packages_prepend () {
def update_rcd_package(pkg):
bb.debug(1, 'adding update-rc.d calls to postinst/postrm for %s' % pkg)
localdata = bb.data.createCopy(d)
- overrides = bb.data.getVar("OVERRIDES", localdata, 1)
+ overrides = localdata.getVar("OVERRIDES", 1)
bb.data.setVar("OVERRIDES", "%s:%s" % (pkg, overrides), localdata)
bb.data.update_data(localdata)
@@ -53,28 +53,28 @@ python populate_packages_prepend () {
execute on the target. Not doing so may cause update_rc.d postinst invoked
twice to cause unwanted warnings.
"""
- postinst = bb.data.getVar('pkg_postinst', localdata, 1)
+ postinst = localdata.getVar('pkg_postinst', 1)
if not postinst:
postinst = '#!/bin/sh\n'
- postinst += bb.data.getVar('updatercd_postinst', localdata, 1)
- bb.data.setVar('pkg_postinst_%s' % pkg, postinst, d)
+ postinst += localdata.getVar('updatercd_postinst', 1)
+ d.setVar('pkg_postinst_%s' % pkg, postinst)
- prerm = bb.data.getVar('pkg_prerm', localdata, 1)
+ prerm = localdata.getVar('pkg_prerm', 1)
if not prerm:
prerm = '#!/bin/sh\n'
- prerm += bb.data.getVar('updatercd_prerm', localdata, 1)
- bb.data.setVar('pkg_prerm_%s' % pkg, prerm, d)
+ prerm += localdata.getVar('updatercd_prerm', 1)
+ d.setVar('pkg_prerm_%s' % pkg, prerm)
- postrm = bb.data.getVar('pkg_postrm', localdata, 1)
+ postrm = localdata.getVar('pkg_postrm', 1)
if not postrm:
postrm = '#!/bin/sh\n'
- postrm += bb.data.getVar('updatercd_postrm', localdata, 1)
- bb.data.setVar('pkg_postrm_%s' % pkg, postrm, d)
+ postrm += localdata.getVar('updatercd_postrm', 1)
+ d.setVar('pkg_postrm_%s' % pkg, postrm)
- pkgs = bb.data.getVar('INITSCRIPT_PACKAGES', d, 1)
+ pkgs = d.getVar('INITSCRIPT_PACKAGES', 1)
if pkgs == None:
- pkgs = bb.data.getVar('UPDATERCPN', d, 1)
- packages = (bb.data.getVar('PACKAGES', d, 1) or "").split()
+ pkgs = d.getVar('UPDATERCPN', 1)
+ packages = (d.getVar('PACKAGES', 1) or "").split()
if not pkgs in packages and packages != []:
pkgs = packages[0]
for pkg in pkgs.split():
diff --git a/meta/classes/useradd.bbclass b/meta/classes/useradd.bbclass
index f3dd8236e..928f619b7 100644
--- a/meta/classes/useradd.bbclass
+++ b/meta/classes/useradd.bbclass
@@ -107,11 +107,11 @@ def update_useradd_after_parse(d):
useradd_packages = d.getVar('USERADD_PACKAGES', True)
if not useradd_packages:
- raise bb.build.FuncFailed, "%s inherits useradd but doesn't set USERADD_PACKAGES" % bb.data.getVar('FILE', d)
+ raise bb.build.FuncFailed, "%s inherits useradd but doesn't set USERADD_PACKAGES" % d.getVar('FILE')
for pkg in useradd_packages.split():
if not d.getVar('USERADD_PARAM_%s' % pkg, True) and not d.getVar('GROUPADD_PARAM_%s' % pkg, True):
- raise bb.build.FuncFailed, "%s inherits useradd but doesn't set USERADD_PARAM or GROUPADD_PARAM for package %s" % (bb.data.getVar('FILE', d), pkg)
+ raise bb.build.FuncFailed, "%s inherits useradd but doesn't set USERADD_PARAM or GROUPADD_PARAM for package %s" % (d.getVar('FILE'), pkg)
python __anonymous() {
update_useradd_after_parse(d)
@@ -147,12 +147,12 @@ fakeroot python populate_packages_prepend () {
if not preinst:
preinst = '#!/bin/sh\n'
preinst += d.getVar('useradd_preinst', True)
- bb.data.setVar('pkg_preinst_%s' % pkg, preinst, d)
+ d.setVar('pkg_preinst_%s' % pkg, preinst)
# RDEPENDS setup
rdepends = d.getVar("RDEPENDS_%s" % pkg, True) or ""
rdepends += " base-passwd shadow"
- bb.data.setVar("RDEPENDS_%s" % pkg, rdepends, d)
+ d.setVar("RDEPENDS_%s" % pkg, rdepends)
# Add the user/group preinstall scripts and RDEPENDS requirements
# to packages specified by USERADD_PACKAGES
diff --git a/meta/classes/utility-tasks.bbclass b/meta/classes/utility-tasks.bbclass
index 64bd84a5d..009ef1fd0 100644
--- a/meta/classes/utility-tasks.bbclass
+++ b/meta/classes/utility-tasks.bbclass
@@ -6,7 +6,7 @@ python do_listtasks() {
#bb.data.emit_env(sys.__stdout__, d)
# emit the metadata which isnt valid shell
for e in d.keys():
- if bb.data.getVarFlag(e, 'task', d):
+ if d.getVarFlag(e, 'task'):
bb.plain("%s" % e)
}
@@ -20,18 +20,18 @@ python do_clean() {
bb.note("Removing " + dir)
oe.path.remove(dir)
- dir = "%s.*" % bb.data.expand(bb.data.getVar('STAMP', d), d)
+ dir = "%s.*" % bb.data.expand(d.getVar('STAMP'), d)
bb.note("Removing " + dir)
oe.path.remove(dir)
- for f in (bb.data.getVar('CLEANFUNCS', d, 1) or '').split():
+ for f in (d.getVar('CLEANFUNCS', 1) or '').split():
bb.build.exec_func(f, d)
}
addtask checkuri
do_checkuri[nostamp] = "1"
python do_checkuri() {
- src_uri = (bb.data.getVar('SRC_URI', d, True) or "").split()
+ src_uri = (d.getVar('SRC_URI', True) or "").split()
if len(src_uri) == 0:
return
diff --git a/meta/classes/utils.bbclass b/meta/classes/utils.bbclass
index 3c2e342f9..103fa9a54 100644
--- a/meta/classes/utils.bbclass
+++ b/meta/classes/utils.bbclass
@@ -331,12 +331,12 @@ def explode_deps(s):
def base_set_filespath(path, d):
filespath = []
- extrapaths = (bb.data.getVar("FILESEXTRAPATHS", d, True) or "")
+ extrapaths = (d.getVar("FILESEXTRAPATHS", True) or "")
# Don't prepend empty strings to the path list
if extrapaths != "":
path = extrapaths.split(":") + path
# The ":" ensures we have an 'empty' override
- overrides = (bb.data.getVar("OVERRIDES", d, 1) or "") + ":"
+ overrides = (d.getVar("OVERRIDES", 1) or "") + ":"
for p in path:
if p != "":
for o in overrides.split(":"):