meta: Don't use deprecated bitbake API
These have been deprecated for a long time, convert the remaining references to the correct modules and prepare for removal of the compatibility support from bitbake. (From OE-Core rev: 6a39835af2b2b3c7797fe05479341d71a3f3aaf6) Signed-off-by: Richard Purdie <richard.purdie@linuxfoundation.org>
This commit is contained in:
parent
6d9e3a1938
commit
554c892ccf
|
@ -99,7 +99,7 @@ def get_bb_inc(d):
|
||||||
licenses = get_licenses(d)
|
licenses = get_licenses(d)
|
||||||
script_logs = os.path.join(work_dir, 'script-logs/'+ target_sys + '/' + licenses + '/' + pf + '/script-logs')
|
script_logs = os.path.join(work_dir, 'script-logs/'+ target_sys + '/' + licenses + '/' + pf + '/script-logs')
|
||||||
bb_inc = os.path.join(script_logs, 'bb_inc')
|
bb_inc = os.path.join(script_logs, 'bb_inc')
|
||||||
bb.mkdirhier(bb_inc)
|
bb.utils.mkdirhier(bb_inc)
|
||||||
|
|
||||||
def find_file(dir, file):
|
def find_file(dir, file):
|
||||||
for root, dirs, files in os.walk(dir):
|
for root, dirs, files in os.walk(dir):
|
||||||
|
@ -139,7 +139,7 @@ def get_logs(d):
|
||||||
script_logs = os.path.join(work_dir, 'script-logs/'+ target_sys + '/' + licenses + '/' + pf + '/script-logs')
|
script_logs = os.path.join(work_dir, 'script-logs/'+ target_sys + '/' + licenses + '/' + pf + '/script-logs')
|
||||||
|
|
||||||
try:
|
try:
|
||||||
bb.mkdirhier(os.path.join(script_logs, 'temp'))
|
bb.utils.mkdirhier(os.path.join(script_logs, 'temp'))
|
||||||
oe.path.copytree(os.path.join(work_dir, 'temp'), os.path.join(script_logs, 'temp'))
|
oe.path.copytree(os.path.join(work_dir, 'temp'), os.path.join(script_logs, 'temp'))
|
||||||
except (IOError, AttributeError):
|
except (IOError, AttributeError):
|
||||||
pass
|
pass
|
||||||
|
@ -158,7 +158,7 @@ def get_series(d):
|
||||||
s = d.getVar('S', True)
|
s = d.getVar('S', True)
|
||||||
dest = os.path.join(work_dir, pf + '-series')
|
dest = os.path.join(work_dir, pf + '-series')
|
||||||
shutil.rmtree(dest, ignore_errors=True)
|
shutil.rmtree(dest, ignore_errors=True)
|
||||||
bb.mkdirhier(dest)
|
bb.utils.mkdirhier(dest)
|
||||||
|
|
||||||
src_uri = d.getVar('SRC_URI', True).split()
|
src_uri = d.getVar('SRC_URI', True).split()
|
||||||
fetch = bb.fetch2.Fetch(src_uri, d)
|
fetch = bb.fetch2.Fetch(src_uri, d)
|
||||||
|
@ -175,7 +175,7 @@ def get_series(d):
|
||||||
shutil.copy(patch, dest)
|
shutil.copy(patch, dest)
|
||||||
except IOError:
|
except IOError:
|
||||||
if os.path.isdir(patch):
|
if os.path.isdir(patch):
|
||||||
bb.mkdirhier(os.path.join(dest, patch))
|
bb.utils.mkdirhier(os.path.join(dest, patch))
|
||||||
oe.path.copytree(patch, os.path.join(dest, patch))
|
oe.path.copytree(patch, os.path.join(dest, patch))
|
||||||
return dest
|
return dest
|
||||||
|
|
||||||
|
@ -190,11 +190,11 @@ def get_applying_patches(d):
|
||||||
work_dir = d.getVar('WORKDIR', True)
|
work_dir = d.getVar('WORKDIR', True)
|
||||||
dest = os.path.join(work_dir, pf + '-patches')
|
dest = os.path.join(work_dir, pf + '-patches')
|
||||||
shutil.rmtree(dest, ignore_errors=True)
|
shutil.rmtree(dest, ignore_errors=True)
|
||||||
bb.mkdirhier(dest)
|
bb.utils.mkdirhier(dest)
|
||||||
|
|
||||||
patches = src_patches(d)
|
patches = src_patches(d)
|
||||||
for patch in patches:
|
for patch in patches:
|
||||||
_, _, local, _, _, parm = bb.decodeurl(patch)
|
_, _, local, _, _, parm = bb.fetch.decodeurl(patch)
|
||||||
if local:
|
if local:
|
||||||
shutil.copy(local, dest)
|
shutil.copy(local, dest)
|
||||||
return dest
|
return dest
|
||||||
|
@ -357,7 +357,7 @@ def move_tarball_deploy(d, tarball_list):
|
||||||
work_dir = d.getVar('WORKDIR', True)
|
work_dir = d.getVar('WORKDIR', True)
|
||||||
tar_sources = d.getVar('DEPLOY_DIR', True) + '/sources/' + target_sys + '/' + licenses + '/' + pf
|
tar_sources = d.getVar('DEPLOY_DIR', True) + '/sources/' + target_sys + '/' + licenses + '/' + pf
|
||||||
if not os.path.exists(tar_sources):
|
if not os.path.exists(tar_sources):
|
||||||
bb.mkdirhier(tar_sources)
|
bb.utils.mkdirhier(tar_sources)
|
||||||
for source in tarball_list:
|
for source in tarball_list:
|
||||||
if source:
|
if source:
|
||||||
if os.path.exists(os.path.join(tar_sources, source)):
|
if os.path.exists(os.path.join(tar_sources, source)):
|
||||||
|
@ -459,7 +459,7 @@ def dumpdata(d):
|
||||||
licenses = get_licenses(d)
|
licenses = get_licenses(d)
|
||||||
dumpdir = os.path.join(workdir, 'diffgz-envdata/'+ target_sys + '/' + licenses + '/' + pf )
|
dumpdir = os.path.join(workdir, 'diffgz-envdata/'+ target_sys + '/' + licenses + '/' + pf )
|
||||||
if not os.path.exists(dumpdir):
|
if not os.path.exists(dumpdir):
|
||||||
bb.mkdirhier(dumpdir)
|
bb.utils.mkdirhier(dumpdir)
|
||||||
|
|
||||||
dumpfile = os.path.join(dumpdir, bb.data.expand("${P}-${PR}.showdata.dump", d))
|
dumpfile = os.path.join(dumpdir, bb.data.expand("${P}-${PR}.showdata.dump", d))
|
||||||
|
|
||||||
|
@ -499,7 +499,7 @@ def create_diff_gz(d):
|
||||||
distro = d.getVar('DISTRO',True) or ""
|
distro = d.getVar('DISTRO',True) or ""
|
||||||
dest = s + '/' + distro + '/files'
|
dest = s + '/' + distro + '/files'
|
||||||
if not os.path.exists(dest):
|
if not os.path.exists(dest):
|
||||||
bb.mkdirhier(dest)
|
bb.utils.mkdirhier(dest)
|
||||||
for i in os.listdir(os.getcwd()):
|
for i in os.listdir(os.getcwd()):
|
||||||
if os.path.isfile(i):
|
if os.path.isfile(i):
|
||||||
try:
|
try:
|
||||||
|
|
|
@ -573,7 +573,7 @@ python () {
|
||||||
d.appendVarFlag('do_fetch', 'depends', ' osc-native:do_populate_sysroot')
|
d.appendVarFlag('do_fetch', 'depends', ' osc-native:do_populate_sysroot')
|
||||||
|
|
||||||
# *.xz should depends on xz-native for unpacking
|
# *.xz should depends on xz-native for unpacking
|
||||||
# Not endswith because of "*.patch.xz;patch=1". Need bb.decodeurl in future
|
# Not endswith because of "*.patch.xz;patch=1". Need bb.fetch.decodeurl in future
|
||||||
if '.xz' in srcuri:
|
if '.xz' in srcuri:
|
||||||
d.appendVarFlag('do_unpack', 'depends', ' xz-native:do_populate_sysroot')
|
d.appendVarFlag('do_unpack', 'depends', ' xz-native:do_populate_sysroot')
|
||||||
|
|
||||||
|
|
|
@ -175,7 +175,7 @@ python run_buildstats () {
|
||||||
# set the buildname
|
# set the buildname
|
||||||
########################################################################
|
########################################################################
|
||||||
try:
|
try:
|
||||||
bb.mkdirhier(e.data.getVar('BUILDSTATS_BASE', True))
|
bb.utils.mkdirhier(e.data.getVar('BUILDSTATS_BASE', True))
|
||||||
except:
|
except:
|
||||||
pass
|
pass
|
||||||
set_bn(e)
|
set_bn(e)
|
||||||
|
@ -185,7 +185,7 @@ python run_buildstats () {
|
||||||
|
|
||||||
bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn)
|
bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn)
|
||||||
try:
|
try:
|
||||||
bb.mkdirhier(bsdir)
|
bb.utils.mkdirhier(bsdir)
|
||||||
except:
|
except:
|
||||||
pass
|
pass
|
||||||
if device != "NoLogicalDevice":
|
if device != "NoLogicalDevice":
|
||||||
|
@ -236,7 +236,7 @@ python run_buildstats () {
|
||||||
set_diskdata("__diskdata_task", device, e.data)
|
set_diskdata("__diskdata_task", device, e.data)
|
||||||
set_timedata("__timedata_task", e.data)
|
set_timedata("__timedata_task", e.data)
|
||||||
try:
|
try:
|
||||||
bb.mkdirhier(taskdir)
|
bb.utils.mkdirhier(taskdir)
|
||||||
except:
|
except:
|
||||||
pass
|
pass
|
||||||
# write into the task event file the name and start time
|
# write into the task event file the name and start time
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
CCACHE = "${@bb.which(d.getVar('PATH', True), 'ccache') and 'ccache '}"
|
CCACHE = "${@bb.utils.which(d.getVar('PATH', True), 'ccache') and 'ccache '}"
|
||||||
export CCACHE_DIR ?= "${TMPDIR}/ccache/${MULTIMACH_HOST_SYS}/${PN}"
|
export CCACHE_DIR ?= "${TMPDIR}/ccache/${MULTIMACH_HOST_SYS}/${PN}"
|
||||||
CCACHE_DISABLE[unexport] = "1"
|
CCACHE_DISABLE[unexport] = "1"
|
||||||
|
|
||||||
|
|
|
@ -33,7 +33,7 @@ python do_prepare_copyleft_sources () {
|
||||||
pf = d.getVar('PF', True)
|
pf = d.getVar('PF', True)
|
||||||
dest = os.path.join(sources_dir, pf)
|
dest = os.path.join(sources_dir, pf)
|
||||||
shutil.rmtree(dest, ignore_errors=True)
|
shutil.rmtree(dest, ignore_errors=True)
|
||||||
bb.mkdirhier(dest)
|
bb.utils.mkdirhier(dest)
|
||||||
|
|
||||||
for u in ud.values():
|
for u in ud.values():
|
||||||
local = os.path.normpath(fetch.localpath(u.url))
|
local = os.path.normpath(fetch.localpath(u.url))
|
||||||
|
@ -51,7 +51,7 @@ python do_prepare_copyleft_sources () {
|
||||||
|
|
||||||
patches = src_patches(d)
|
patches = src_patches(d)
|
||||||
for patch in patches:
|
for patch in patches:
|
||||||
_, _, local, _, _, parm = bb.decodeurl(patch)
|
_, _, local, _, _, parm = bb.fetch.decodeurl(patch)
|
||||||
patchdir = parm.get('patchdir')
|
patchdir = parm.get('patchdir')
|
||||||
if patchdir:
|
if patchdir:
|
||||||
series = os.path.join(dest, 'series.subdir.%s' % patchdir.replace('/', '_'))
|
series = os.path.join(dest, 'series.subdir.%s' % patchdir.replace('/', '_'))
|
||||||
|
|
|
@ -591,7 +591,7 @@ python do_checkpkg() {
|
||||||
pupver = "N/A"
|
pupver = "N/A"
|
||||||
pstatus = "ErrUnknown"
|
pstatus = "ErrUnknown"
|
||||||
|
|
||||||
(type, host, path, user, pswd, parm) = bb.decodeurl(uri)
|
(type, host, path, user, pswd, parm) = bb.fetch.decodeurl(uri)
|
||||||
if type in ['http', 'https', 'ftp']:
|
if type in ['http', 'https', 'ftp']:
|
||||||
if d.getVar('PRSPV', True):
|
if d.getVar('PRSPV', True):
|
||||||
pcurver = d.getVar('PRSPV', True)
|
pcurver = d.getVar('PRSPV', True)
|
||||||
|
@ -621,7 +621,7 @@ python do_checkpkg() {
|
||||||
dirver = m.group().strip("/")
|
dirver = m.group().strip("/")
|
||||||
|
|
||||||
"""use new path and remove param. for wget only param is md5sum"""
|
"""use new path and remove param. for wget only param is md5sum"""
|
||||||
alturi = bb.encodeurl([type, host, altpath, user, pswd, {}])
|
alturi = bb.fetch.encodeurl([type, host, altpath, user, pswd, {}])
|
||||||
my_uri = d.getVar('REGEX_URI', True)
|
my_uri = d.getVar('REGEX_URI', True)
|
||||||
if my_uri:
|
if my_uri:
|
||||||
if d.getVar('PRSPV', True):
|
if d.getVar('PRSPV', True):
|
||||||
|
@ -647,7 +647,7 @@ python do_checkpkg() {
|
||||||
|
|
||||||
chk_uri = d.getVar('REGEX_URI', True)
|
chk_uri = d.getVar('REGEX_URI', True)
|
||||||
if not chk_uri:
|
if not chk_uri:
|
||||||
alturi = bb.encodeurl([type, host, altpath, user, pswd, {}])
|
alturi = bb.fetch.encodeurl([type, host, altpath, user, pswd, {}])
|
||||||
else:
|
else:
|
||||||
alturi = chk_uri
|
alturi = chk_uri
|
||||||
newver = check_new_version(alturi, curname, d)
|
newver = check_new_version(alturi, curname, d)
|
||||||
|
|
|
@ -149,7 +149,7 @@ def get_devtable_list(d):
|
||||||
if devtables == None:
|
if devtables == None:
|
||||||
devtables = 'files/device_table-minimal.txt'
|
devtables = 'files/device_table-minimal.txt'
|
||||||
for devtable in devtables.split():
|
for devtable in devtables.split():
|
||||||
str += " %s" % bb.which(d.getVar('BBPATH', True), devtable)
|
str += " %s" % bb.utils.which(d.getVar('BBPATH', True), devtable)
|
||||||
return str
|
return str
|
||||||
|
|
||||||
IMAGE_CLASSES ?= "image_types"
|
IMAGE_CLASSES ?= "image_types"
|
||||||
|
|
|
@ -594,7 +594,7 @@ def package_qa_check_license(workdir, d):
|
||||||
srcdir = d.getVar('S', True)
|
srcdir = d.getVar('S', True)
|
||||||
|
|
||||||
for url in lic_files.split():
|
for url in lic_files.split():
|
||||||
(type, host, path, user, pswd, parm) = bb.decodeurl(url)
|
(type, host, path, user, pswd, parm) = bb.fetch.decodeurl(url)
|
||||||
srclicfile = os.path.join(srcdir, path)
|
srclicfile = os.path.join(srcdir, path)
|
||||||
if not os.path.isfile(srclicfile):
|
if not os.path.isfile(srclicfile):
|
||||||
raise bb.build.FuncFailed( pn + ": LIC_FILES_CHKSUM points to an invalid file: " + srclicfile)
|
raise bb.build.FuncFailed( pn + ": LIC_FILES_CHKSUM points to an invalid file: " + srclicfile)
|
||||||
|
|
|
@ -9,7 +9,7 @@ def find_patches(d):
|
||||||
patches = src_patches(d)
|
patches = src_patches(d)
|
||||||
patch_list=[]
|
patch_list=[]
|
||||||
for p in patches:
|
for p in patches:
|
||||||
_, _, local, _, _, _ = bb.decodeurl(p)
|
_, _, local, _, _, _ = bb.fetch.decodeurl(p)
|
||||||
patch_list.append(local)
|
patch_list.append(local)
|
||||||
|
|
||||||
return patch_list
|
return patch_list
|
||||||
|
|
|
@ -129,9 +129,9 @@ def add_package_and_files(d):
|
||||||
d.setVar('RRECOMMENDS_' + pn, "%s" % (pn_lic))
|
d.setVar('RRECOMMENDS_' + pn, "%s" % (pn_lic))
|
||||||
|
|
||||||
def copy_license_files(lic_files_paths, destdir):
|
def copy_license_files(lic_files_paths, destdir):
|
||||||
bb.mkdirhier(destdir)
|
bb.utils.mkdirhier(destdir)
|
||||||
for (basename, path) in lic_files_paths:
|
for (basename, path) in lic_files_paths:
|
||||||
ret = bb.copyfile(path, os.path.join(destdir, basename))
|
ret = bb.utils.copyfile(path, os.path.join(destdir, basename))
|
||||||
# If the copy didn't occur, something horrible went wrong and we fail out
|
# If the copy didn't occur, something horrible went wrong and we fail out
|
||||||
if not ret:
|
if not ret:
|
||||||
bb.warn("%s could not be copied for some reason. It may not exist. WARN for now." % path)
|
bb.warn("%s could not be copied for some reason. It may not exist. WARN for now." % path)
|
||||||
|
@ -188,7 +188,7 @@ def find_license_files(d):
|
||||||
|
|
||||||
def find_license(license_type):
|
def find_license(license_type):
|
||||||
try:
|
try:
|
||||||
bb.mkdirhier(gen_lic_dest)
|
bb.utils.mkdirhier(gen_lic_dest)
|
||||||
except:
|
except:
|
||||||
pass
|
pass
|
||||||
spdx_generic = None
|
spdx_generic = None
|
||||||
|
@ -227,7 +227,7 @@ def find_license_files(d):
|
||||||
return lic_files_paths
|
return lic_files_paths
|
||||||
|
|
||||||
for url in lic_files.split():
|
for url in lic_files.split():
|
||||||
(type, host, path, user, pswd, parm) = bb.decodeurl(url)
|
(type, host, path, user, pswd, parm) = bb.fetch.decodeurl(url)
|
||||||
# We want the license filename and path
|
# We want the license filename and path
|
||||||
srclicfile = os.path.join(srcdir, path)
|
srclicfile = os.path.join(srcdir, path)
|
||||||
lic_files_paths.append((os.path.basename(path), srclicfile))
|
lic_files_paths.append((os.path.basename(path), srclicfile))
|
||||||
|
|
|
@ -220,7 +220,7 @@ python do_package_deb () {
|
||||||
basedir = os.path.join(os.path.dirname(root))
|
basedir = os.path.join(os.path.dirname(root))
|
||||||
|
|
||||||
pkgoutdir = os.path.join(outdir, localdata.getVar('PACKAGE_ARCH', True))
|
pkgoutdir = os.path.join(outdir, localdata.getVar('PACKAGE_ARCH', True))
|
||||||
bb.mkdirhier(pkgoutdir)
|
bb.utils.mkdirhier(pkgoutdir)
|
||||||
|
|
||||||
os.chdir(root)
|
os.chdir(root)
|
||||||
from glob import glob
|
from glob import glob
|
||||||
|
@ -236,7 +236,7 @@ python do_package_deb () {
|
||||||
continue
|
continue
|
||||||
|
|
||||||
controldir = os.path.join(root, 'DEBIAN')
|
controldir = os.path.join(root, 'DEBIAN')
|
||||||
bb.mkdirhier(controldir)
|
bb.utils.mkdirhier(controldir)
|
||||||
os.chmod(controldir, 0755)
|
os.chmod(controldir, 0755)
|
||||||
try:
|
try:
|
||||||
ctrlfile = open(os.path.join(controldir, 'control'), 'w')
|
ctrlfile = open(os.path.join(controldir, 'control'), 'w')
|
||||||
|
|
|
@ -253,7 +253,7 @@ python do_package_ipk () {
|
||||||
basedir = os.path.join(os.path.dirname(root))
|
basedir = os.path.join(os.path.dirname(root))
|
||||||
arch = localdata.getVar('PACKAGE_ARCH', True)
|
arch = localdata.getVar('PACKAGE_ARCH', True)
|
||||||
pkgoutdir = "%s/%s" % (outdir, arch)
|
pkgoutdir = "%s/%s" % (outdir, arch)
|
||||||
bb.mkdirhier(pkgoutdir)
|
bb.utils.mkdirhier(pkgoutdir)
|
||||||
os.chdir(root)
|
os.chdir(root)
|
||||||
from glob import glob
|
from glob import glob
|
||||||
g = glob('*')
|
g = glob('*')
|
||||||
|
@ -268,7 +268,7 @@ python do_package_ipk () {
|
||||||
continue
|
continue
|
||||||
|
|
||||||
controldir = os.path.join(root, 'CONTROL')
|
controldir = os.path.join(root, 'CONTROL')
|
||||||
bb.mkdirhier(controldir)
|
bb.utils.mkdirhier(controldir)
|
||||||
try:
|
try:
|
||||||
ctrlfile = open(os.path.join(controldir, 'control'), 'w')
|
ctrlfile = open(os.path.join(controldir, 'control'), 'w')
|
||||||
except OSError:
|
except OSError:
|
||||||
|
|
|
@ -1069,7 +1069,7 @@ python do_package_rpm () {
|
||||||
clean_licenses = get_licenses(d)
|
clean_licenses = get_licenses(d)
|
||||||
pkgwritesrpmdir = bb.data.expand('${PKGWRITEDIRSRPM}/${PACKAGE_ARCH_EXTEND}', d)
|
pkgwritesrpmdir = bb.data.expand('${PKGWRITEDIRSRPM}/${PACKAGE_ARCH_EXTEND}', d)
|
||||||
pkgwritesrpmdir = pkgwritesrpmdir + '/' + clean_licenses
|
pkgwritesrpmdir = pkgwritesrpmdir + '/' + clean_licenses
|
||||||
bb.mkdirhier(pkgwritesrpmdir)
|
bb.utils.mkdirhier(pkgwritesrpmdir)
|
||||||
os.chmod(pkgwritesrpmdir, 0755)
|
os.chmod(pkgwritesrpmdir, 0755)
|
||||||
return pkgwritesrpmdir
|
return pkgwritesrpmdir
|
||||||
|
|
||||||
|
@ -1123,7 +1123,7 @@ python do_package_rpm () {
|
||||||
pkgwritedir = d.expand('${PKGWRITEDIRRPM}/${PACKAGE_ARCH_EXTEND}')
|
pkgwritedir = d.expand('${PKGWRITEDIRRPM}/${PACKAGE_ARCH_EXTEND}')
|
||||||
pkgarch = d.expand('${PACKAGE_ARCH_EXTEND}${TARGET_VENDOR}-${TARGET_OS}')
|
pkgarch = d.expand('${PACKAGE_ARCH_EXTEND}${TARGET_VENDOR}-${TARGET_OS}')
|
||||||
magicfile = d.expand('${STAGING_DIR_NATIVE}${datadir_native}/misc/magic.mgc')
|
magicfile = d.expand('${STAGING_DIR_NATIVE}${datadir_native}/misc/magic.mgc')
|
||||||
bb.mkdirhier(pkgwritedir)
|
bb.utils.mkdirhier(pkgwritedir)
|
||||||
os.chmod(pkgwritedir, 0755)
|
os.chmod(pkgwritedir, 0755)
|
||||||
|
|
||||||
cmd = rpmbuild
|
cmd = rpmbuild
|
||||||
|
|
|
@ -44,7 +44,7 @@ def src_patches(d, all = False ):
|
||||||
if patchdir:
|
if patchdir:
|
||||||
patchparm['patchdir'] = patchdir
|
patchparm['patchdir'] = patchdir
|
||||||
|
|
||||||
localurl = bb.encodeurl(('file', '', local, '', '', patchparm))
|
localurl = bb.fetch.encodeurl(('file', '', local, '', '', patchparm))
|
||||||
patches.append(localurl)
|
patches.append(localurl)
|
||||||
|
|
||||||
if all:
|
if all:
|
||||||
|
@ -147,7 +147,7 @@ python patch_do_patch() {
|
||||||
os.environ['TMPDIR'] = process_tmpdir
|
os.environ['TMPDIR'] = process_tmpdir
|
||||||
|
|
||||||
for patch in src_patches(d):
|
for patch in src_patches(d):
|
||||||
_, _, local, _, _, parm = bb.decodeurl(patch)
|
_, _, local, _, _, parm = bb.fetch.decodeurl(patch)
|
||||||
|
|
||||||
if "patchdir" in parm:
|
if "patchdir" in parm:
|
||||||
patchdir = parm["patchdir"]
|
patchdir = parm["patchdir"]
|
||||||
|
|
|
@ -73,7 +73,7 @@ def can_delete_FILESPATH(cfgdata, d):
|
||||||
|
|
||||||
def can_delete_FILESDIR(cfgdata, d):
|
def can_delete_FILESDIR(cfgdata, d):
|
||||||
expected = cfgdata.get("FILESDIR")
|
expected = cfgdata.get("FILESDIR")
|
||||||
#expected = "${@bb.which(d.getVar('FILESPATH', True), '.')}"
|
#expected = "${@bb.utils.which(d.getVar('FILESPATH', True), '.')}"
|
||||||
unexpanded = d.getVar("FILESDIR", 0)
|
unexpanded = d.getVar("FILESDIR", 0)
|
||||||
if unexpanded is None:
|
if unexpanded is None:
|
||||||
return False
|
return False
|
||||||
|
|
|
@ -45,9 +45,9 @@ python do_spdx () {
|
||||||
cur_ver_code = get_ver_code( info['sourcedir'] )
|
cur_ver_code = get_ver_code( info['sourcedir'] )
|
||||||
cache_cur = False
|
cache_cur = False
|
||||||
if not os.path.exists( spdx_sstate_dir ):
|
if not os.path.exists( spdx_sstate_dir ):
|
||||||
bb.mkdirhier( spdx_sstate_dir )
|
bb.utils.mkdirhier( spdx_sstate_dir )
|
||||||
if not os.path.exists( info['spdx_temp_dir'] ):
|
if not os.path.exists( info['spdx_temp_dir'] ):
|
||||||
bb.mkdirhier( info['spdx_temp_dir'] )
|
bb.utils.mkdirhier( info['spdx_temp_dir'] )
|
||||||
if os.path.exists( sstatefile ):
|
if os.path.exists( sstatefile ):
|
||||||
## cache for this package exists. read it in
|
## cache for this package exists. read it in
|
||||||
cached_spdx = get_cached_spdx( sstatefile )
|
cached_spdx = get_cached_spdx( sstatefile )
|
||||||
|
|
|
@ -129,7 +129,7 @@ def sstate_install(ss, d):
|
||||||
|
|
||||||
sharedfiles = []
|
sharedfiles = []
|
||||||
shareddirs = []
|
shareddirs = []
|
||||||
bb.mkdirhier(d.expand("${SSTATE_MANIFESTS}"))
|
bb.utils.mkdirhier(d.expand("${SSTATE_MANIFESTS}"))
|
||||||
|
|
||||||
d2 = d.createCopy()
|
d2 = d.createCopy()
|
||||||
extrainf = d.getVarFlag("do_" + ss['task'], 'stamp-extra-info', True)
|
extrainf = d.getVarFlag("do_" + ss['task'], 'stamp-extra-info', True)
|
||||||
|
@ -215,7 +215,7 @@ def sstate_installpkg(ss, d):
|
||||||
# remove dir if it exists, ensure any parent directories do exist
|
# remove dir if it exists, ensure any parent directories do exist
|
||||||
if os.path.exists(dir):
|
if os.path.exists(dir):
|
||||||
oe.path.remove(dir)
|
oe.path.remove(dir)
|
||||||
bb.mkdirhier(dir)
|
bb.utils.mkdirhier(dir)
|
||||||
oe.path.remove(dir)
|
oe.path.remove(dir)
|
||||||
|
|
||||||
sstateinst = d.expand("${WORKDIR}/sstate-install-%s/" % ss['name'])
|
sstateinst = d.expand("${WORKDIR}/sstate-install-%s/" % ss['name'])
|
||||||
|
@ -281,7 +281,7 @@ def sstate_installpkg(ss, d):
|
||||||
workdir = d.getVar('WORKDIR', True)
|
workdir = d.getVar('WORKDIR', True)
|
||||||
src = sstateinst + "/" + plain.replace(workdir, '')
|
src = sstateinst + "/" + plain.replace(workdir, '')
|
||||||
dest = plain
|
dest = plain
|
||||||
bb.mkdirhier(src)
|
bb.utils.mkdirhier(src)
|
||||||
prepdir(dest)
|
prepdir(dest)
|
||||||
os.rename(src, dest)
|
os.rename(src, dest)
|
||||||
|
|
||||||
|
@ -456,8 +456,8 @@ def sstate_package(ss, d):
|
||||||
sstatebuild = d.expand("${WORKDIR}/sstate-build-%s/" % ss['name'])
|
sstatebuild = d.expand("${WORKDIR}/sstate-build-%s/" % ss['name'])
|
||||||
sstatepkg = d.getVar('SSTATE_PKG', True) + '_'+ ss['name'] + ".tgz"
|
sstatepkg = d.getVar('SSTATE_PKG', True) + '_'+ ss['name'] + ".tgz"
|
||||||
bb.utils.remove(sstatebuild, recurse=True)
|
bb.utils.remove(sstatebuild, recurse=True)
|
||||||
bb.mkdirhier(sstatebuild)
|
bb.utils.mkdirhier(sstatebuild)
|
||||||
bb.mkdirhier(os.path.dirname(sstatepkg))
|
bb.utils.mkdirhier(os.path.dirname(sstatepkg))
|
||||||
for state in ss['dirs']:
|
for state in ss['dirs']:
|
||||||
if not os.path.exists(state[1]):
|
if not os.path.exists(state[1]):
|
||||||
continue
|
continue
|
||||||
|
@ -477,8 +477,8 @@ def sstate_package(ss, d):
|
||||||
workdir = d.getVar('WORKDIR', True)
|
workdir = d.getVar('WORKDIR', True)
|
||||||
for plain in ss['plaindirs']:
|
for plain in ss['plaindirs']:
|
||||||
pdir = plain.replace(workdir, sstatebuild)
|
pdir = plain.replace(workdir, sstatebuild)
|
||||||
bb.mkdirhier(plain)
|
bb.utils.mkdirhier(plain)
|
||||||
bb.mkdirhier(pdir)
|
bb.utils.mkdirhier(pdir)
|
||||||
oe.path.copyhardlinktree(plain, pdir)
|
oe.path.copyhardlinktree(plain, pdir)
|
||||||
|
|
||||||
d.setVar('SSTATE_BUILDDIR', sstatebuild)
|
d.setVar('SSTATE_BUILDDIR', sstatebuild)
|
||||||
|
@ -503,7 +503,7 @@ def pstaging_fetch(sstatefetch, sstatepkg, d):
|
||||||
bb.data.update_data(localdata)
|
bb.data.update_data(localdata)
|
||||||
|
|
||||||
dldir = localdata.expand("${SSTATE_DIR}")
|
dldir = localdata.expand("${SSTATE_DIR}")
|
||||||
bb.mkdirhier(dldir)
|
bb.utils.mkdirhier(dldir)
|
||||||
|
|
||||||
localdata.delVar('MIRRORS')
|
localdata.delVar('MIRRORS')
|
||||||
localdata.delVar('FILESPATH')
|
localdata.delVar('FILESPATH')
|
||||||
|
|
|
@ -20,7 +20,7 @@ def emit_terminal_func(command, envdata, d):
|
||||||
runfmt = d.getVar('BB_RUNFMT', True) or "run.{func}.{pid}"
|
runfmt = d.getVar('BB_RUNFMT', True) or "run.{func}.{pid}"
|
||||||
runfile = runfmt.format(func=cmd_func, task=cmd_func, taskfunc=cmd_func, pid=os.getpid())
|
runfile = runfmt.format(func=cmd_func, task=cmd_func, taskfunc=cmd_func, pid=os.getpid())
|
||||||
runfile = os.path.join(d.getVar('T', True), runfile)
|
runfile = os.path.join(d.getVar('T', True), runfile)
|
||||||
bb.mkdirhier(os.path.dirname(runfile))
|
bb.utils.mkdirhier(os.path.dirname(runfile))
|
||||||
|
|
||||||
with open(runfile, 'w') as script:
|
with open(runfile, 'w') as script:
|
||||||
script.write('#!/bin/sh -e\n')
|
script.write('#!/bin/sh -e\n')
|
||||||
|
|
|
@ -114,7 +114,7 @@ def populate_toolchain_links(d):
|
||||||
bb.fatal("Unable to populate toolchain binary symlinks in %s" % pattern)
|
bb.fatal("Unable to populate toolchain binary symlinks in %s" % pattern)
|
||||||
|
|
||||||
bindir = d.getVar('STAGING_BINDIR_TOOLCHAIN', True)
|
bindir = d.getVar('STAGING_BINDIR_TOOLCHAIN', True)
|
||||||
bb.mkdirhier(bindir)
|
bb.utils.mkdirhier(bindir)
|
||||||
for f in files:
|
for f in files:
|
||||||
base = os.path.basename(f)
|
base = os.path.basename(f)
|
||||||
newpath = os.path.join(bindir, base)
|
newpath = os.path.join(bindir, base)
|
||||||
|
|
|
@ -331,7 +331,7 @@ class QuiltTree(PatchSet):
|
||||||
patch = self.patches[kwargs["patch"]]
|
patch = self.patches[kwargs["patch"]]
|
||||||
if not patch:
|
if not patch:
|
||||||
raise PatchError("No patch found at index %s in patchset." % kwargs["patch"])
|
raise PatchError("No patch found at index %s in patchset." % kwargs["patch"])
|
||||||
(type, host, path, user, pswd, parm) = bb.decodeurl(patch["remote"])
|
(type, host, path, user, pswd, parm) = bb.fetch.decodeurl(patch["remote"])
|
||||||
if type == "file":
|
if type == "file":
|
||||||
import shutil
|
import shutil
|
||||||
if not patch.get("file") and patch.get("remote"):
|
if not patch.get("file") and patch.get("remote"):
|
||||||
|
|
|
@ -13,6 +13,6 @@ do_configure() {
|
||||||
oe_runconf
|
oe_runconf
|
||||||
}
|
}
|
||||||
|
|
||||||
export AUTOMAKE = "${@bb.which('automake', d.getVar('PATH', True))}"
|
export AUTOMAKE = "${@bb.utils.which('automake', d.getVar('PATH', True))}"
|
||||||
|
|
||||||
FILES_${PN} += "${datadir}/automake* ${datadir}/aclocal*"
|
FILES_${PN} += "${datadir}/automake* ${datadir}/aclocal*"
|
||||||
|
|
|
@ -34,7 +34,7 @@ python gcc_multilib_setup() {
|
||||||
rel_path = os.path.relpath(fn, src_conf_dir)
|
rel_path = os.path.relpath(fn, src_conf_dir)
|
||||||
parent_dir = os.path.dirname(rel_path)
|
parent_dir = os.path.dirname(rel_path)
|
||||||
bb.utils.mkdirhier('%s/%s' % (build_conf_dir, parent_dir))
|
bb.utils.mkdirhier('%s/%s' % (build_conf_dir, parent_dir))
|
||||||
bb.copyfile(fn, '%s/%s' % (build_conf_dir, rel_path))
|
bb.utils.copyfile(fn, '%s/%s' % (build_conf_dir, rel_path))
|
||||||
|
|
||||||
multilibs = (d.getVar('MULTILIB_VARIANTS', True) or '').split()
|
multilibs = (d.getVar('MULTILIB_VARIANTS', True) or '').split()
|
||||||
if not multilibs:
|
if not multilibs:
|
||||||
|
|
|
@ -16,7 +16,7 @@ SRC_URI[sha256sum] = "c603957a4966811c04af5f6048c71cfb4966ec93312d7b3118116ed9f3
|
||||||
|
|
||||||
S = "${WORKDIR}/pcre-${PV}"
|
S = "${WORKDIR}/pcre-${PV}"
|
||||||
|
|
||||||
FILESPATH .= ":${@base_set_filespath([bb.which(BBPATH, 'recipes-support/libpcre/files', direction=True)], d)}"
|
FILESPATH .= ":${@base_set_filespath([bb.utils.which(BBPATH, 'recipes-support/libpcre/files', direction=True)], d)}"
|
||||||
|
|
||||||
PROVIDES += "pcre"
|
PROVIDES += "pcre"
|
||||||
DEPENDS += "bzip2 zlib"
|
DEPENDS += "bzip2 zlib"
|
||||||
|
|
Loading…
Reference in New Issue