meta/classes: Various python whitespace fixes
It was pointed out we have a number of weird indentations in the python functions. This patch cleans up 3, 7 and other weird indentations for the core bbclass files. It also fixes some wierd (odd) shell function indentation which my searches picked up. (From OE-Core rev: 8385d6d74624000d68814f4e3266d47bc8885942) Signed-off-by: Richard Purdie <richard.purdie@linuxfoundation.org>
This commit is contained in:
parent
7c40daab58
commit
48619958d5
|
@ -205,8 +205,8 @@ def preferred_ml_updates(d):
|
||||||
continue
|
continue
|
||||||
virt = ""
|
virt = ""
|
||||||
if pkg.startswith("virtual/"):
|
if pkg.startswith("virtual/"):
|
||||||
pkg = pkg.replace("virtual/", "")
|
pkg = pkg.replace("virtual/", "")
|
||||||
virt = "virtual/"
|
virt = "virtual/"
|
||||||
for p in prefixes:
|
for p in prefixes:
|
||||||
newname = "PREFERRED_PROVIDER_" + virt + p + "-" + pkg
|
newname = "PREFERRED_PROVIDER_" + virt + p + "-" + pkg
|
||||||
if pkg != "kernel":
|
if pkg != "kernel":
|
||||||
|
@ -353,13 +353,13 @@ python () {
|
||||||
appends = bb.utils.explode_deps(d.expand(" ".join(appends)))
|
appends = bb.utils.explode_deps(d.expand(" ".join(appends)))
|
||||||
newappends = []
|
newappends = []
|
||||||
for a in appends:
|
for a in appends:
|
||||||
if a.endswith("-native") or a.endswith("-cross"):
|
if a.endswith("-native") or a.endswith("-cross"):
|
||||||
newappends.append(a)
|
newappends.append(a)
|
||||||
elif a.startswith("virtual/"):
|
elif a.startswith("virtual/"):
|
||||||
subs = a.split("/", 1)[1]
|
subs = a.split("/", 1)[1]
|
||||||
newappends.append("virtual/" + prefix + subs + extension)
|
newappends.append("virtual/" + prefix + subs + extension)
|
||||||
else:
|
else:
|
||||||
newappends.append(prefix + a + extension)
|
newappends.append(prefix + a + extension)
|
||||||
return newappends
|
return newappends
|
||||||
|
|
||||||
def appendVar(varname, appends):
|
def appendVar(varname, appends):
|
||||||
|
|
|
@ -105,8 +105,8 @@ python bugzilla_eventhandler() {
|
||||||
data = e.data
|
data = e.data
|
||||||
name = bb.event.getName(event)
|
name = bb.event.getName(event)
|
||||||
if name == "MsgNote":
|
if name == "MsgNote":
|
||||||
# avoid recursion
|
# avoid recursion
|
||||||
return
|
return
|
||||||
|
|
||||||
if name == "TaskFailed":
|
if name == "TaskFailed":
|
||||||
xmlrpc = data.getVar("BUGZILLA_XMLRPC", True)
|
xmlrpc = data.getVar("BUGZILLA_XMLRPC", True)
|
||||||
|
|
|
@ -65,7 +65,7 @@ def set_device(e):
|
||||||
try:
|
try:
|
||||||
for line in open("/proc/diskstats", "r"):
|
for line in open("/proc/diskstats", "r"):
|
||||||
if majordev == int(line.split()[0]) and minordev == int(line.split()[1]):
|
if majordev == int(line.split()[0]) and minordev == int(line.split()[1]):
|
||||||
rdev=line.split()[2]
|
rdev=line.split()[2]
|
||||||
except:
|
except:
|
||||||
pass
|
pass
|
||||||
file = open(e.data.getVar('DEVFILE', True), "w")
|
file = open(e.data.getVar('DEVFILE', True), "w")
|
||||||
|
@ -100,10 +100,10 @@ def get_diskdata(var, dev, data):
|
||||||
olddiskdata = data.getVar(var, False)
|
olddiskdata = data.getVar(var, False)
|
||||||
diskdata = {}
|
diskdata = {}
|
||||||
if olddiskdata is None:
|
if olddiskdata is None:
|
||||||
return
|
return
|
||||||
newdiskdata = get_diskstats(dev)
|
newdiskdata = get_diskstats(dev)
|
||||||
for key in olddiskdata.iterkeys():
|
for key in olddiskdata.iterkeys():
|
||||||
diskdata["Start"+key] = str(int(olddiskdata[key]))
|
diskdata["Start"+key] = str(int(olddiskdata[key]))
|
||||||
diskdata["End"+key] = str(int(newdiskdata[key]))
|
diskdata["End"+key] = str(int(newdiskdata[key]))
|
||||||
return diskdata
|
return diskdata
|
||||||
|
|
||||||
|
|
|
@ -54,12 +54,12 @@ python build_grub_cfg() {
|
||||||
if not workdir:
|
if not workdir:
|
||||||
bb.error("WORKDIR not defined, unable to package")
|
bb.error("WORKDIR not defined, unable to package")
|
||||||
return
|
return
|
||||||
|
|
||||||
labels = d.getVar('LABELS', True)
|
labels = d.getVar('LABELS', True)
|
||||||
if not labels:
|
if not labels:
|
||||||
bb.debug(1, "LABELS not defined, nothing to do")
|
bb.debug(1, "LABELS not defined, nothing to do")
|
||||||
return
|
return
|
||||||
|
|
||||||
if labels == []:
|
if labels == []:
|
||||||
bb.debug(1, "No labels, nothing to do")
|
bb.debug(1, "No labels, nothing to do")
|
||||||
return
|
return
|
||||||
|
@ -109,7 +109,7 @@ python build_grub_cfg() {
|
||||||
if append:
|
if append:
|
||||||
cfgfile.write('%s' % (append))
|
cfgfile.write('%s' % (append))
|
||||||
cfgfile.write('\n')
|
cfgfile.write('\n')
|
||||||
|
|
||||||
if initrd:
|
if initrd:
|
||||||
cfgfile.write('initrd /initrd')
|
cfgfile.write('initrd /initrd')
|
||||||
cfgfile.write('\n}\n')
|
cfgfile.write('\n}\n')
|
||||||
|
|
|
@ -51,13 +51,13 @@ python() {
|
||||||
# and cross packages which aren't swabber-native or one of its dependencies
|
# and cross packages which aren't swabber-native or one of its dependencies
|
||||||
# I have ignored them for now...
|
# I have ignored them for now...
|
||||||
if not bb.data.inherits_class('native', d) and not bb.data.inherits_class('nativesdk', d) and not bb.data.inherits_class('cross', d):
|
if not bb.data.inherits_class('native', d) and not bb.data.inherits_class('nativesdk', d) and not bb.data.inherits_class('cross', d):
|
||||||
deps = (d.getVarFlag('do_setscene', 'depends') or "").split()
|
deps = (d.getVarFlag('do_setscene', 'depends') or "").split()
|
||||||
deps.append('strace-native:do_populate_sysroot')
|
deps.append('strace-native:do_populate_sysroot')
|
||||||
d.setVarFlag('do_setscene', 'depends', " ".join(deps))
|
d.setVarFlag('do_setscene', 'depends', " ".join(deps))
|
||||||
logdir = d.expand("${TRACE_LOGDIR}")
|
logdir = d.expand("${TRACE_LOGDIR}")
|
||||||
bb.utils.mkdirhier(logdir)
|
bb.utils.mkdirhier(logdir)
|
||||||
else:
|
else:
|
||||||
d.setVar('STRACEFUNC', '')
|
d.setVar('STRACEFUNC', '')
|
||||||
}
|
}
|
||||||
|
|
||||||
STRACEPID = "${@os.getpid()}"
|
STRACEPID = "${@os.getpid()}"
|
||||||
|
@ -76,23 +76,23 @@ imageswab_attachstrace () {
|
||||||
|
|
||||||
do_generate_swabber_report () {
|
do_generate_swabber_report () {
|
||||||
|
|
||||||
update_distro ${HOST_DATA}
|
update_distro ${HOST_DATA}
|
||||||
|
|
||||||
# Swabber can't create the directory for us
|
# Swabber can't create the directory for us
|
||||||
mkdir -p ${SWABBER_REPORT}
|
mkdir -p ${SWABBER_REPORT}
|
||||||
|
|
||||||
REPORTSTAMP=${SWAB_ORIG_TASK}-`date +%2m%2d%2H%2M%Y`
|
REPORTSTAMP=${SWAB_ORIG_TASK}-`date +%2m%2d%2H%2M%Y`
|
||||||
|
|
||||||
if [ `which ccache` ] ; then
|
if [ `which ccache` ] ; then
|
||||||
CCACHE_DIR=`( ccache -s | grep "cache directory" | grep -o '[^ ]*$' 2> /dev/null )`
|
CCACHE_DIR=`( ccache -s | grep "cache directory" | grep -o '[^ ]*$' 2> /dev/null )`
|
||||||
fi
|
fi
|
||||||
|
|
||||||
if [ "$(ls -A ${HOST_DATA})" ]; then
|
if [ "$(ls -A ${HOST_DATA})" ]; then
|
||||||
echo "Generating swabber report"
|
echo "Generating swabber report"
|
||||||
swabber -d ${HOST_DATA} -l ${SWABBER_LOGS} -o ${SWABBER_REPORT}/report-${REPORTSTAMP}.txt -r ${SWABBER_REPORT}/extra_report-${REPORTSTAMP}.txt -c all -p ${TOPDIR} -f ${OEROOT}/meta/conf/swabber ${TOPDIR} ${OEROOT} ${CCACHE_DIR}
|
swabber -d ${HOST_DATA} -l ${SWABBER_LOGS} -o ${SWABBER_REPORT}/report-${REPORTSTAMP}.txt -r ${SWABBER_REPORT}/extra_report-${REPORTSTAMP}.txt -c all -p ${TOPDIR} -f ${OEROOT}/meta/conf/swabber ${TOPDIR} ${OEROOT} ${CCACHE_DIR}
|
||||||
else
|
else
|
||||||
echo "No host data, cannot generate swabber report."
|
echo "No host data, cannot generate swabber report."
|
||||||
fi
|
fi
|
||||||
}
|
}
|
||||||
addtask generate_swabber_report after do_${SWAB_ORIG_TASK}
|
addtask generate_swabber_report after do_${SWAB_ORIG_TASK}
|
||||||
do_generate_swabber_report[depends] = "swabber-native:do_populate_sysroot"
|
do_generate_swabber_report[depends] = "swabber-native:do_populate_sysroot"
|
||||||
|
|
|
@ -109,7 +109,7 @@ def package_qa_get_machine_dict():
|
||||||
"linux-gnux32" : {
|
"linux-gnux32" : {
|
||||||
"x86_64": (62, 0, 0, True, 32),
|
"x86_64": (62, 0, 0, True, 32),
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
# Currently not being used by default "desktop"
|
# Currently not being used by default "desktop"
|
||||||
|
@ -719,19 +719,19 @@ Rerun configure task after fixing this. The path was '%s'""" % root)
|
||||||
|
|
||||||
cnf = d.getVar('EXTRA_OECONF', True) or ""
|
cnf = d.getVar('EXTRA_OECONF', True) or ""
|
||||||
if "gettext" not in d.getVar('P', True) and "gcc-runtime" not in d.getVar('P', True) and "--disable-nls" not in cnf:
|
if "gettext" not in d.getVar('P', True) and "gcc-runtime" not in d.getVar('P', True) and "--disable-nls" not in cnf:
|
||||||
ml = d.getVar("MLPREFIX", True) or ""
|
ml = d.getVar("MLPREFIX", True) or ""
|
||||||
if bb.data.inherits_class('native', d) or bb.data.inherits_class('cross', d) or bb.data.inherits_class('crosssdk', d) or bb.data.inherits_class('nativesdk', d):
|
if bb.data.inherits_class('native', d) or bb.data.inherits_class('cross', d) or bb.data.inherits_class('crosssdk', d) or bb.data.inherits_class('nativesdk', d):
|
||||||
gt = "gettext-native"
|
gt = "gettext-native"
|
||||||
elif bb.data.inherits_class('cross-canadian', d):
|
elif bb.data.inherits_class('cross-canadian', d):
|
||||||
gt = "gettext-nativesdk"
|
gt = "gettext-nativesdk"
|
||||||
else:
|
else:
|
||||||
gt = "virtual/" + ml + "gettext"
|
gt = "virtual/" + ml + "gettext"
|
||||||
deps = bb.utils.explode_deps(d.getVar('DEPENDS', True) or "")
|
deps = bb.utils.explode_deps(d.getVar('DEPENDS', True) or "")
|
||||||
if gt not in deps:
|
if gt not in deps:
|
||||||
for config in configs:
|
for config in configs:
|
||||||
gnu = "grep \"^[[:space:]]*AM_GNU_GETTEXT\" %s >/dev/null" % config
|
gnu = "grep \"^[[:space:]]*AM_GNU_GETTEXT\" %s >/dev/null" % config
|
||||||
if subprocess.call(gnu, shell=True) == 0:
|
if subprocess.call(gnu, shell=True) == 0:
|
||||||
bb.fatal("""%s required but not in DEPENDS for file %s.
|
bb.fatal("""%s required but not in DEPENDS for file %s.
|
||||||
Missing inherit gettext?""" % (gt, config))
|
Missing inherit gettext?""" % (gt, config))
|
||||||
|
|
||||||
if not package_qa_check_license(workdir, d):
|
if not package_qa_check_license(workdir, d):
|
||||||
|
|
|
@ -230,42 +230,42 @@ def splitfile2(debugsrcdir, d):
|
||||||
|
|
||||||
sourcefile = d.expand("${WORKDIR}/debugsources.list")
|
sourcefile = d.expand("${WORKDIR}/debugsources.list")
|
||||||
if debugsrcdir and os.path.isfile(sourcefile):
|
if debugsrcdir and os.path.isfile(sourcefile):
|
||||||
dvar = d.getVar('PKGD', True)
|
dvar = d.getVar('PKGD', True)
|
||||||
pathprefix = "export PATH=%s; " % d.getVar('PATH', True)
|
pathprefix = "export PATH=%s; " % d.getVar('PATH', True)
|
||||||
strip = d.getVar("STRIP", True)
|
strip = d.getVar("STRIP", True)
|
||||||
objcopy = d.getVar("OBJCOPY", True)
|
objcopy = d.getVar("OBJCOPY", True)
|
||||||
debugedit = d.expand("${STAGING_LIBDIR_NATIVE}/rpm/bin/debugedit")
|
debugedit = d.expand("${STAGING_LIBDIR_NATIVE}/rpm/bin/debugedit")
|
||||||
workdir = d.getVar("WORKDIR", True)
|
workdir = d.getVar("WORKDIR", True)
|
||||||
workparentdir = os.path.dirname(workdir)
|
workparentdir = os.path.dirname(workdir)
|
||||||
workbasedir = os.path.basename(workdir)
|
workbasedir = os.path.basename(workdir)
|
||||||
|
|
||||||
nosuchdir = []
|
nosuchdir = []
|
||||||
basepath = dvar
|
basepath = dvar
|
||||||
for p in debugsrcdir.split("/"):
|
for p in debugsrcdir.split("/"):
|
||||||
basepath = basepath + "/" + p
|
basepath = basepath + "/" + p
|
||||||
if not os.path.exists(basepath):
|
if not os.path.exists(basepath):
|
||||||
nosuchdir.append(basepath)
|
nosuchdir.append(basepath)
|
||||||
bb.mkdirhier(basepath)
|
bb.mkdirhier(basepath)
|
||||||
|
|
||||||
processdebugsrc = "LC_ALL=C ; sort -z -u '%s' | egrep -v -z '(<internal>|<built-in>)$' | "
|
processdebugsrc = "LC_ALL=C ; sort -z -u '%s' | egrep -v -z '(<internal>|<built-in>)$' | "
|
||||||
# We need to ignore files that are not actually ours
|
# We need to ignore files that are not actually ours
|
||||||
# we do this by only paying attention to items from this package
|
# we do this by only paying attention to items from this package
|
||||||
processdebugsrc += "fgrep -z '%s' | "
|
processdebugsrc += "fgrep -z '%s' | "
|
||||||
processdebugsrc += "(cd '%s' ; cpio -pd0mL --no-preserve-owner '%s%s' 2>/dev/null)"
|
processdebugsrc += "(cd '%s' ; cpio -pd0mL --no-preserve-owner '%s%s' 2>/dev/null)"
|
||||||
|
|
||||||
subprocess.call(processdebugsrc % (sourcefile, workbasedir, workparentdir, dvar, debugsrcdir), shell=True)
|
subprocess.call(processdebugsrc % (sourcefile, workbasedir, workparentdir, dvar, debugsrcdir), shell=True)
|
||||||
|
|
||||||
# The copy by cpio may have resulted in some empty directories! Remove these
|
# The copy by cpio may have resulted in some empty directories! Remove these
|
||||||
for root, dirs, files in os.walk("%s%s" % (dvar, debugsrcdir)):
|
for root, dirs, files in os.walk("%s%s" % (dvar, debugsrcdir)):
|
||||||
for d in dirs:
|
for d in dirs:
|
||||||
dir = os.path.join(root, d)
|
dir = os.path.join(root, d)
|
||||||
#bb.note("rmdir -p %s" % dir)
|
#bb.note("rmdir -p %s" % dir)
|
||||||
subprocess.call("rmdir -p %s 2>/dev/null" % dir, shell=True)
|
subprocess.call("rmdir -p %s 2>/dev/null" % dir, shell=True)
|
||||||
|
|
||||||
# Also remove debugsrcdir if its empty
|
# Also remove debugsrcdir if its empty
|
||||||
for p in nosuchdir[::-1]:
|
for p in nosuchdir[::-1]:
|
||||||
if os.path.exists(p) and not os.listdir(p):
|
if os.path.exists(p) and not os.listdir(p):
|
||||||
os.rmdir(p)
|
os.rmdir(p)
|
||||||
|
|
||||||
def runstrip(file, elftype, d):
|
def runstrip(file, elftype, d):
|
||||||
# Function to strip a single file, called from split_and_strip_files below
|
# Function to strip a single file, called from split_and_strip_files below
|
||||||
|
@ -735,7 +735,7 @@ python split_and_strip_files () {
|
||||||
file_list = {}
|
file_list = {}
|
||||||
file_links = {}
|
file_links = {}
|
||||||
if (d.getVar('INHIBIT_PACKAGE_DEBUG_SPLIT', True) != '1') and \
|
if (d.getVar('INHIBIT_PACKAGE_DEBUG_SPLIT', True) != '1') and \
|
||||||
(d.getVar('INHIBIT_PACKAGE_STRIP', True) != '1'):
|
(d.getVar('INHIBIT_PACKAGE_STRIP', True) != '1'):
|
||||||
for root, dirs, files in os.walk(dvar):
|
for root, dirs, files in os.walk(dvar):
|
||||||
for f in files:
|
for f in files:
|
||||||
file = os.path.join(root, f)
|
file = os.path.join(root, f)
|
||||||
|
|
|
@ -172,23 +172,23 @@ package_install_internal_ipk() {
|
||||||
}
|
}
|
||||||
|
|
||||||
ipk_log_check() {
|
ipk_log_check() {
|
||||||
target="$1"
|
target="$1"
|
||||||
lf_path="$2"
|
lf_path="$2"
|
||||||
|
|
||||||
lf_txt="`cat $lf_path`"
|
lf_txt="`cat $lf_path`"
|
||||||
for keyword_die in "exit 1" "Collected errors" ERR Fail
|
for keyword_die in "exit 1" "Collected errors" ERR Fail
|
||||||
do
|
do
|
||||||
if (echo "$lf_txt" | grep -v log_check | grep "$keyword_die") >/dev/null 2>&1
|
if (echo "$lf_txt" | grep -v log_check | grep "$keyword_die") >/dev/null 2>&1
|
||||||
then
|
then
|
||||||
echo "log_check: There were error messages in the logfile"
|
echo "log_check: There were error messages in the logfile"
|
||||||
echo -e "log_check: Matched keyword: [$keyword_die]\n"
|
echo -e "log_check: Matched keyword: [$keyword_die]\n"
|
||||||
echo "$lf_txt" | grep -v log_check | grep -C 5 -i "$keyword_die"
|
echo "$lf_txt" | grep -v log_check | grep -C 5 -i "$keyword_die"
|
||||||
echo ""
|
echo ""
|
||||||
do_exit=1
|
do_exit=1
|
||||||
fi
|
fi
|
||||||
done
|
done
|
||||||
test "$do_exit" = 1 && exit 1
|
test "$do_exit" = 1 && exit 1
|
||||||
true
|
true
|
||||||
}
|
}
|
||||||
|
|
||||||
#
|
#
|
||||||
|
|
|
@ -101,23 +101,23 @@ package_generate_rpm_conf_common() {
|
||||||
}
|
}
|
||||||
|
|
||||||
rpm_log_check() {
|
rpm_log_check() {
|
||||||
target="$1"
|
target="$1"
|
||||||
lf_path="$2"
|
lf_path="$2"
|
||||||
|
|
||||||
lf_txt="`cat $lf_path`"
|
lf_txt="`cat $lf_path`"
|
||||||
for keyword_die in "Cannot find package" "exit 1" ERR Fail
|
for keyword_die in "Cannot find package" "exit 1" ERR Fail
|
||||||
do
|
do
|
||||||
if (echo "$lf_txt" | grep -v log_check | grep "$keyword_die") >/dev/null 2>&1
|
if (echo "$lf_txt" | grep -v log_check | grep "$keyword_die") >/dev/null 2>&1
|
||||||
then
|
then
|
||||||
echo "log_check: There were error messages in the logfile"
|
echo "log_check: There were error messages in the logfile"
|
||||||
echo -e "log_check: Matched keyword: [$keyword_die]\n"
|
echo -e "log_check: Matched keyword: [$keyword_die]\n"
|
||||||
echo "$lf_txt" | grep -v log_check | grep -C 5 -i "$keyword_die"
|
echo "$lf_txt" | grep -v log_check | grep -C 5 -i "$keyword_die"
|
||||||
echo ""
|
echo ""
|
||||||
do_exit=1
|
do_exit=1
|
||||||
fi
|
fi
|
||||||
done
|
done
|
||||||
test "$do_exit" = 1 && exit 1
|
test "$do_exit" = 1 && exit 1
|
||||||
true
|
true
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -4,7 +4,7 @@ def __note(msg, d):
|
||||||
__recipe_sanity_badruntimevars = "RDEPENDS RPROVIDES RRECOMMENDS RCONFLICTS"
|
__recipe_sanity_badruntimevars = "RDEPENDS RPROVIDES RRECOMMENDS RCONFLICTS"
|
||||||
def bad_runtime_vars(cfgdata, d):
|
def bad_runtime_vars(cfgdata, d):
|
||||||
if bb.data.inherits_class("native", d) or \
|
if bb.data.inherits_class("native", d) or \
|
||||||
bb.data.inherits_class("cross", d):
|
bb.data.inherits_class("cross", d):
|
||||||
return
|
return
|
||||||
|
|
||||||
for var in d.getVar("__recipe_sanity_badruntimevars", True).split():
|
for var in d.getVar("__recipe_sanity_badruntimevars", True).split():
|
||||||
|
@ -42,7 +42,7 @@ def var_renames_overwrite(cfgdata, d):
|
||||||
|
|
||||||
def incorrect_nonempty_PACKAGES(cfgdata, d):
|
def incorrect_nonempty_PACKAGES(cfgdata, d):
|
||||||
if bb.data.inherits_class("native", d) or \
|
if bb.data.inherits_class("native", d) or \
|
||||||
bb.data.inherits_class("cross", d):
|
bb.data.inherits_class("cross", d):
|
||||||
if d.getVar("PACKAGES", True):
|
if d.getVar("PACKAGES", True):
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
|
|
@ -280,10 +280,10 @@ def check_sanity_validmachine(sanity_data):
|
||||||
tunefound = True
|
tunefound = True
|
||||||
|
|
||||||
if len(dups):
|
if len(dups):
|
||||||
messages = messages + "Error, the PACKAGE_ARCHS variable contains duplicates. The following archs are listed more than once: %s" % " ".join(dups)
|
messages = messages + "Error, the PACKAGE_ARCHS variable contains duplicates. The following archs are listed more than once: %s" % " ".join(dups)
|
||||||
|
|
||||||
if tunefound == False:
|
if tunefound == False:
|
||||||
messages = messages + "Error, the PACKAGE_ARCHS variable does not contain TUNE_PKGARCH (%s)." % tunepkg
|
messages = messages + "Error, the PACKAGE_ARCHS variable does not contain TUNE_PKGARCH (%s)." % tunepkg
|
||||||
|
|
||||||
return messages
|
return messages
|
||||||
|
|
||||||
|
|
|
@ -174,7 +174,7 @@ def sstate_installpkg(ss, d):
|
||||||
sstatepkg = d.getVar('SSTATE_PKG', True) + '_' + ss['name'] + ".tgz"
|
sstatepkg = d.getVar('SSTATE_PKG', True) + '_' + ss['name'] + ".tgz"
|
||||||
|
|
||||||
if not os.path.exists(sstatepkg):
|
if not os.path.exists(sstatepkg):
|
||||||
pstaging_fetch(sstatefetch, sstatepkg, d)
|
pstaging_fetch(sstatefetch, sstatepkg, d)
|
||||||
|
|
||||||
if not os.path.isfile(sstatepkg):
|
if not os.path.isfile(sstatepkg):
|
||||||
bb.note("Staging package %s does not exist" % sstatepkg)
|
bb.note("Staging package %s does not exist" % sstatepkg)
|
||||||
|
@ -259,10 +259,10 @@ def sstate_clean_manifest(manifest, d):
|
||||||
# so we ignore errors here.
|
# so we ignore errors here.
|
||||||
try:
|
try:
|
||||||
if entry.endswith("/"):
|
if entry.endswith("/"):
|
||||||
if os.path.islink(entry[:-1]):
|
if os.path.islink(entry[:-1]):
|
||||||
os.remove(entry[:-1])
|
os.remove(entry[:-1])
|
||||||
elif os.path.exists(entry) and len(os.listdir(entry)) == 0:
|
elif os.path.exists(entry) and len(os.listdir(entry)) == 0:
|
||||||
os.rmdir(entry[:-1])
|
os.rmdir(entry[:-1])
|
||||||
else:
|
else:
|
||||||
oe.path.remove(entry)
|
oe.path.remove(entry)
|
||||||
except OSError:
|
except OSError:
|
||||||
|
@ -314,14 +314,14 @@ python sstate_cleanall() {
|
||||||
|
|
||||||
for manifest in (os.listdir(manifest_dir)):
|
for manifest in (os.listdir(manifest_dir)):
|
||||||
if fnmatch.fnmatch(manifest, manifest_pattern):
|
if fnmatch.fnmatch(manifest, manifest_pattern):
|
||||||
name = manifest.replace(manifest_pattern[:-1], "")
|
name = manifest.replace(manifest_pattern[:-1], "")
|
||||||
namemap = d.getVar('SSTATETASKNAMES', True).split()
|
namemap = d.getVar('SSTATETASKNAMES', True).split()
|
||||||
tasks = d.getVar('SSTATETASKS', True).split()
|
tasks = d.getVar('SSTATETASKS', True).split()
|
||||||
if name not in namemap:
|
if name not in namemap:
|
||||||
continue
|
continue
|
||||||
taskname = tasks[namemap.index(name)]
|
taskname = tasks[namemap.index(name)]
|
||||||
shared_state = sstate_state_fromvars(d, taskname[3:])
|
shared_state = sstate_state_fromvars(d, taskname[3:])
|
||||||
sstate_clean(shared_state, d)
|
sstate_clean(shared_state, d)
|
||||||
}
|
}
|
||||||
|
|
||||||
def sstate_hardcode_path(d):
|
def sstate_hardcode_path(d):
|
||||||
|
|
|
@ -70,8 +70,8 @@ SYSROOT_LOCK = "${STAGING_DIR}/staging.lock"
|
||||||
|
|
||||||
# We clean out any existing sstate from the sysroot if we rerun configure
|
# We clean out any existing sstate from the sysroot if we rerun configure
|
||||||
python sysroot_cleansstate () {
|
python sysroot_cleansstate () {
|
||||||
ss = sstate_state_fromvars(d, "populate_sysroot")
|
ss = sstate_state_fromvars(d, "populate_sysroot")
|
||||||
sstate_clean(ss, d)
|
sstate_clean(ss, d)
|
||||||
}
|
}
|
||||||
do_configure[prefuncs] += "sysroot_cleansstate"
|
do_configure[prefuncs] += "sysroot_cleansstate"
|
||||||
|
|
||||||
|
|
|
@ -2,20 +2,20 @@ def tinder_http_post(server, selector, content_type, body):
|
||||||
import httplib
|
import httplib
|
||||||
# now post it
|
# now post it
|
||||||
for i in range(0,5):
|
for i in range(0,5):
|
||||||
try:
|
try:
|
||||||
h = httplib.HTTP(server)
|
h = httplib.HTTP(server)
|
||||||
h.putrequest('POST', selector)
|
h.putrequest('POST', selector)
|
||||||
h.putheader('content-type', content_type)
|
h.putheader('content-type', content_type)
|
||||||
h.putheader('content-length', str(len(body)))
|
h.putheader('content-length', str(len(body)))
|
||||||
h.endheaders()
|
h.endheaders()
|
||||||
h.send(body)
|
h.send(body)
|
||||||
errcode, errmsg, headers = h.getreply()
|
errcode, errmsg, headers = h.getreply()
|
||||||
#print errcode, errmsg, headers
|
#print errcode, errmsg, headers
|
||||||
return (errcode,errmsg, headers, h.file)
|
return (errcode,errmsg, headers, h.file)
|
||||||
except:
|
except:
|
||||||
print "Error sending the report!"
|
print "Error sending the report!"
|
||||||
# try again
|
# try again
|
||||||
pass
|
pass
|
||||||
|
|
||||||
# return some garbage
|
# return some garbage
|
||||||
return (-1, "unknown", "unknown", None)
|
return (-1, "unknown", "unknown", None)
|
||||||
|
|
|
@ -150,22 +150,22 @@ def update_alternatives_after_parse(d):
|
||||||
UPDALTVARS = "ALTERNATIVE ALTERNATIVE_LINK_NAME ALTERNATIVE_TARGET ALTERNATIVE_PRIORITY"
|
UPDALTVARS = "ALTERNATIVE ALTERNATIVE_LINK_NAME ALTERNATIVE_TARGET ALTERNATIVE_PRIORITY"
|
||||||
|
|
||||||
def gen_updatealternativesvardeps(d):
|
def gen_updatealternativesvardeps(d):
|
||||||
pkgs = (d.getVar("PACKAGES", True) or "").split()
|
pkgs = (d.getVar("PACKAGES", True) or "").split()
|
||||||
vars = (d.getVar("UPDALTVARS", True) or "").split()
|
vars = (d.getVar("UPDALTVARS", True) or "").split()
|
||||||
|
|
||||||
# First compute them for non_pkg versions
|
# First compute them for non_pkg versions
|
||||||
for v in vars:
|
for v in vars:
|
||||||
for flag in (d.getVarFlags(v) or {}):
|
for flag in (d.getVarFlags(v) or {}):
|
||||||
if flag == "doc" or flag == "vardeps" or flag == "vardepsexp":
|
|
||||||
continue
|
|
||||||
d.appendVar('%s_VARDEPS' % (v), ' %s:%s' % (flag, d.getVarFlag(v, flag, False)))
|
|
||||||
|
|
||||||
for p in pkgs:
|
|
||||||
for v in vars:
|
|
||||||
for flag in (d.getVarFlags("%s_%s" % (v,p)) or {}):
|
|
||||||
if flag == "doc" or flag == "vardeps" or flag == "vardepsexp":
|
if flag == "doc" or flag == "vardeps" or flag == "vardepsexp":
|
||||||
continue
|
continue
|
||||||
d.appendVar('%s_VARDEPS_%s' % (v,p), ' %s:%s' % (flag, d.getVarFlag('%s_%s' % (v,p), flag, False)))
|
d.appendVar('%s_VARDEPS' % (v), ' %s:%s' % (flag, d.getVarFlag(v, flag, False)))
|
||||||
|
|
||||||
|
for p in pkgs:
|
||||||
|
for v in vars:
|
||||||
|
for flag in (d.getVarFlags("%s_%s" % (v,p)) or {}):
|
||||||
|
if flag == "doc" or flag == "vardeps" or flag == "vardepsexp":
|
||||||
|
continue
|
||||||
|
d.appendVar('%s_VARDEPS_%s' % (v,p), ' %s:%s' % (flag, d.getVarFlag('%s_%s' % (v,p), flag, False)))
|
||||||
|
|
||||||
python __anonymous() {
|
python __anonymous() {
|
||||||
# deprecated stuff...
|
# deprecated stuff...
|
||||||
|
@ -176,18 +176,18 @@ python __anonymous() {
|
||||||
}
|
}
|
||||||
|
|
||||||
def gen_updatealternativesvars(d):
|
def gen_updatealternativesvars(d):
|
||||||
ret = []
|
ret = []
|
||||||
pkgs = (d.getVar("PACKAGES", True) or "").split()
|
pkgs = (d.getVar("PACKAGES", True) or "").split()
|
||||||
vars = (d.getVar("UPDALTVARS", True) or "").split()
|
vars = (d.getVar("UPDALTVARS", True) or "").split()
|
||||||
|
|
||||||
for v in vars:
|
for v in vars:
|
||||||
ret.append(v + "_VARDEPS")
|
ret.append(v + "_VARDEPS")
|
||||||
|
|
||||||
for p in pkgs:
|
for p in pkgs:
|
||||||
for v in vars:
|
for v in vars:
|
||||||
ret.append(v + "_" + p)
|
ret.append(v + "_" + p)
|
||||||
ret.append(v + "_VARDEPS_" + p)
|
ret.append(v + "_VARDEPS_" + p)
|
||||||
return " ".join(ret)
|
return " ".join(ret)
|
||||||
|
|
||||||
# First the deprecated items...
|
# First the deprecated items...
|
||||||
populate_packages[vardeps] += "ALTERNATIVE_LINKS ALTERNATIVE_NAME ALTERNATIVE_PATH"
|
populate_packages[vardeps] += "ALTERNATIVE_LINKS ALTERNATIVE_NAME ALTERNATIVE_PATH"
|
||||||
|
|
|
@ -246,49 +246,49 @@ oe_machinstall() {
|
||||||
}
|
}
|
||||||
|
|
||||||
create_cmdline_wrapper () {
|
create_cmdline_wrapper () {
|
||||||
# Create a wrapper script
|
# Create a wrapper script
|
||||||
#
|
#
|
||||||
# These are useful to work around relocation issues, by setting environment
|
# These are useful to work around relocation issues, by setting environment
|
||||||
# variables which point to paths in the filesystem.
|
# variables which point to paths in the filesystem.
|
||||||
#
|
#
|
||||||
# Usage: create_wrapper FILENAME [[VAR=VALUE]..]
|
# Usage: create_wrapper FILENAME [[VAR=VALUE]..]
|
||||||
|
|
||||||
cmd=$1
|
cmd=$1
|
||||||
shift
|
shift
|
||||||
|
|
||||||
echo "Generating wrapper script for $cmd"
|
echo "Generating wrapper script for $cmd"
|
||||||
|
|
||||||
mv $cmd $cmd.real
|
mv $cmd $cmd.real
|
||||||
cmdname=`basename $cmd`.real
|
cmdname=`basename $cmd`.real
|
||||||
cat <<END >$cmd
|
cat <<END >$cmd
|
||||||
#!/bin/sh
|
#!/bin/sh
|
||||||
realpath=\`readlink -fn \$0\`
|
realpath=\`readlink -fn \$0\`
|
||||||
exec \`dirname \$realpath\`/$cmdname $@ "\$@"
|
exec \`dirname \$realpath\`/$cmdname $@ "\$@"
|
||||||
END
|
END
|
||||||
chmod +x $cmd
|
chmod +x $cmd
|
||||||
}
|
}
|
||||||
|
|
||||||
create_wrapper () {
|
create_wrapper () {
|
||||||
# Create a wrapper script
|
# Create a wrapper script
|
||||||
#
|
#
|
||||||
# These are useful to work around relocation issues, by setting environment
|
# These are useful to work around relocation issues, by setting environment
|
||||||
# variables which point to paths in the filesystem.
|
# variables which point to paths in the filesystem.
|
||||||
#
|
#
|
||||||
# Usage: create_wrapper FILENAME [[VAR=VALUE]..]
|
# Usage: create_wrapper FILENAME [[VAR=VALUE]..]
|
||||||
|
|
||||||
cmd=$1
|
cmd=$1
|
||||||
shift
|
shift
|
||||||
|
|
||||||
echo "Generating wrapper script for $cmd"
|
echo "Generating wrapper script for $cmd"
|
||||||
|
|
||||||
mv $cmd $cmd.real
|
mv $cmd $cmd.real
|
||||||
cmdname=`basename $cmd`.real
|
cmdname=`basename $cmd`.real
|
||||||
cat <<END >$cmd
|
cat <<END >$cmd
|
||||||
#!/bin/sh
|
#!/bin/sh
|
||||||
realpath=\`readlink -fn \$0\`
|
realpath=\`readlink -fn \$0\`
|
||||||
exec env $@ \`dirname \$realpath\`/$cmdname "\$@"
|
exec env $@ \`dirname \$realpath\`/$cmdname "\$@"
|
||||||
END
|
END
|
||||||
chmod +x $cmd
|
chmod +x $cmd
|
||||||
}
|
}
|
||||||
|
|
||||||
def check_app_exists(app, d):
|
def check_app_exists(app, d):
|
||||||
|
|
Loading…
Reference in New Issue