mirror of
https://git.yoctoproject.org/poky
synced 2026-02-09 18:23:02 +01:00
First, commit 81efd77987f6decf256967fa16521a40c14d3518 created a copy of __expand_var_regexp__ and __expand_python_regexp__ when creating the initial version of data_smart.py. A while later commit db1c998b31da06d7f3eb09fc6f59a1915b7b549e dropped all references to these variables from data.py. This leaves us today with two versions of these variables in the global scope. However, only those defined in data_smart.py are being used, in data_smart.py. Unfortunately there was no indication in the commit log for commit db1c998b31da indicating why the variables were left in place despite the functions referencing them were being removed. Additionally data.py imports data_smart, thus the versions of __expand_var_regexp__ and __expand_python_regexp__ defined in data_smart.py would be used by all bitbake code, beyond, potentially, some code in data.py which we know has no references to these variables. To remove any potential confusion around these variables drop the old definitions from data.py. (Bitbake rev: 60f43d0428d43c981b44b6c8d125f77440f6c8f9) Signed-off-by: Mark Asselstine <mark.asselstine@windriver.com> Signed-off-by: Richard Purdie <richard.purdie@linuxfoundation.org>
445 lines
16 KiB
Python
445 lines
16 KiB
Python
"""
|
|
BitBake 'Data' implementations
|
|
|
|
Functions for interacting with the data structure used by the
|
|
BitBake build tools.
|
|
|
|
The expandKeys and update_data are the most expensive
|
|
operations. At night the cookie monster came by and
|
|
suggested 'give me cookies on setting the variables and
|
|
things will work out'. Taking this suggestion into account
|
|
applying the skills from the not yet passed 'Entwurf und
|
|
Analyse von Algorithmen' lecture and the cookie
|
|
monster seems to be right. We will track setVar more carefully
|
|
to have faster update_data and expandKeys operations.
|
|
|
|
This is a trade-off between speed and memory again but
|
|
the speed is more critical here.
|
|
"""
|
|
|
|
# Copyright (C) 2003, 2004 Chris Larson
|
|
# Copyright (C) 2005 Holger Hans Peter Freyther
|
|
#
|
|
# SPDX-License-Identifier: GPL-2.0-only
|
|
#
|
|
# Based on functions from the base bb module, Copyright 2003 Holger Schurig
|
|
|
|
import sys, os, re
|
|
import hashlib
|
|
if sys.argv[0][-5:] == "pydoc":
|
|
path = os.path.dirname(os.path.dirname(sys.argv[1]))
|
|
else:
|
|
path = os.path.dirname(os.path.dirname(sys.argv[0]))
|
|
sys.path.insert(0, path)
|
|
from itertools import groupby
|
|
|
|
from bb import data_smart
|
|
from bb import codeparser
|
|
import bb
|
|
|
|
logger = data_smart.logger
|
|
_dict_type = data_smart.DataSmart
|
|
|
|
def init():
|
|
"""Return a new object representing the Bitbake data"""
|
|
return _dict_type()
|
|
|
|
def init_db(parent = None):
|
|
"""Return a new object representing the Bitbake data,
|
|
optionally based on an existing object"""
|
|
if parent is not None:
|
|
return parent.createCopy()
|
|
else:
|
|
return _dict_type()
|
|
|
|
def createCopy(source):
|
|
"""Link the source set to the destination
|
|
If one does not find the value in the destination set,
|
|
search will go on to the source set to get the value.
|
|
Value from source are copy-on-write. i.e. any try to
|
|
modify one of them will end up putting the modified value
|
|
in the destination set.
|
|
"""
|
|
return source.createCopy()
|
|
|
|
def initVar(var, d):
|
|
"""Non-destructive var init for data structure"""
|
|
d.initVar(var)
|
|
|
|
def keys(d):
|
|
"""Return a list of keys in d"""
|
|
return d.keys()
|
|
|
|
def expand(s, d, varname = None):
|
|
"""Variable expansion using the data store"""
|
|
return d.expand(s, varname)
|
|
|
|
def expandKeys(alterdata, readdata = None):
|
|
if readdata is None:
|
|
readdata = alterdata
|
|
|
|
todolist = {}
|
|
for key in alterdata:
|
|
if not '${' in key:
|
|
continue
|
|
|
|
ekey = expand(key, readdata)
|
|
if key == ekey:
|
|
continue
|
|
todolist[key] = ekey
|
|
|
|
# These two for loops are split for performance to maximise the
|
|
# usefulness of the expand cache
|
|
for key in sorted(todolist):
|
|
ekey = todolist[key]
|
|
newval = alterdata.getVar(ekey, False)
|
|
if newval is not None:
|
|
val = alterdata.getVar(key, False)
|
|
if val is not None:
|
|
bb.warn("Variable key %s (%s) replaces original key %s (%s)." % (key, val, ekey, newval))
|
|
alterdata.renameVar(key, ekey)
|
|
|
|
def inheritFromOS(d, savedenv, permitted):
|
|
"""Inherit variables from the initial environment."""
|
|
exportlist = bb.utils.preserved_envvars_exported()
|
|
for s in savedenv.keys():
|
|
if s in permitted:
|
|
try:
|
|
d.setVar(s, savedenv.getVar(s), op = 'from env')
|
|
if s in exportlist:
|
|
d.setVarFlag(s, "export", True, op = 'auto env export')
|
|
except TypeError:
|
|
pass
|
|
|
|
def emit_var(var, o=sys.__stdout__, d = init(), all=False):
|
|
"""Emit a variable to be sourced by a shell."""
|
|
func = d.getVarFlag(var, "func", False)
|
|
if d.getVarFlag(var, 'python', False) and func:
|
|
return False
|
|
|
|
export = d.getVarFlag(var, "export", False)
|
|
unexport = d.getVarFlag(var, "unexport", False)
|
|
if not all and not export and not unexport and not func:
|
|
return False
|
|
|
|
try:
|
|
if all:
|
|
oval = d.getVar(var, False)
|
|
val = d.getVar(var)
|
|
except (KeyboardInterrupt):
|
|
raise
|
|
except Exception as exc:
|
|
o.write('# expansion of %s threw %s: %s\n' % (var, exc.__class__.__name__, str(exc)))
|
|
return False
|
|
|
|
if all:
|
|
d.varhistory.emit(var, oval, val, o, d)
|
|
|
|
if (var.find("-") != -1 or var.find(".") != -1 or var.find('{') != -1 or var.find('}') != -1 or var.find('+') != -1) and not all:
|
|
return False
|
|
|
|
varExpanded = d.expand(var)
|
|
|
|
if unexport:
|
|
o.write('unset %s\n' % varExpanded)
|
|
return False
|
|
|
|
if val is None:
|
|
return False
|
|
|
|
val = str(val)
|
|
|
|
if varExpanded.startswith("BASH_FUNC_"):
|
|
varExpanded = varExpanded[10:-2]
|
|
val = val[3:] # Strip off "() "
|
|
o.write("%s() %s\n" % (varExpanded, val))
|
|
o.write("export -f %s\n" % (varExpanded))
|
|
return True
|
|
|
|
if func:
|
|
# Write a comment indicating where the shell function came from (line number and filename) to make it easier
|
|
# for the user to diagnose task failures. This comment is also used by build.py to determine the metadata
|
|
# location of shell functions.
|
|
o.write("# line: {0}, file: {1}\n".format(
|
|
d.getVarFlag(var, "lineno", False),
|
|
d.getVarFlag(var, "filename", False)))
|
|
# NOTE: should probably check for unbalanced {} within the var
|
|
val = val.rstrip('\n')
|
|
o.write("%s() {\n%s\n}\n" % (varExpanded, val))
|
|
return 1
|
|
|
|
if export:
|
|
o.write('export ')
|
|
|
|
# if we're going to output this within doublequotes,
|
|
# to a shell, we need to escape the quotes in the var
|
|
alter = re.sub('"', '\\"', val)
|
|
alter = re.sub('\n', ' \\\n', alter)
|
|
alter = re.sub('\\$', '\\\\$', alter)
|
|
o.write('%s="%s"\n' % (varExpanded, alter))
|
|
return False
|
|
|
|
def emit_env(o=sys.__stdout__, d = init(), all=False):
|
|
"""Emits all items in the data store in a format such that it can be sourced by a shell."""
|
|
|
|
isfunc = lambda key: bool(d.getVarFlag(key, "func", False))
|
|
keys = sorted((key for key in d.keys() if not key.startswith("__")), key=isfunc)
|
|
grouped = groupby(keys, isfunc)
|
|
for isfunc, keys in grouped:
|
|
for key in sorted(keys):
|
|
emit_var(key, o, d, all and not isfunc) and o.write('\n')
|
|
|
|
def exported_keys(d):
|
|
return (key for key in d.keys() if not key.startswith('__') and
|
|
d.getVarFlag(key, 'export', False) and
|
|
not d.getVarFlag(key, 'unexport', False))
|
|
|
|
def exported_vars(d):
|
|
k = list(exported_keys(d))
|
|
for key in k:
|
|
try:
|
|
value = d.getVar(key)
|
|
except Exception as err:
|
|
bb.warn("%s: Unable to export ${%s}: %s" % (d.getVar("FILE"), key, err))
|
|
continue
|
|
|
|
if value is not None:
|
|
yield key, str(value)
|
|
|
|
def emit_func(func, o=sys.__stdout__, d = init()):
|
|
"""Emits all items in the data store in a format such that it can be sourced by a shell."""
|
|
|
|
keys = (key for key in d.keys() if not key.startswith("__") and not d.getVarFlag(key, "func", False))
|
|
for key in sorted(keys):
|
|
emit_var(key, o, d, False)
|
|
|
|
o.write('\n')
|
|
emit_var(func, o, d, False) and o.write('\n')
|
|
newdeps = bb.codeparser.ShellParser(func, logger).parse_shell(d.getVar(func))
|
|
newdeps |= set((d.getVarFlag(func, "vardeps") or "").split())
|
|
seen = set()
|
|
while newdeps:
|
|
deps = newdeps
|
|
seen |= deps
|
|
newdeps = set()
|
|
for dep in sorted(deps):
|
|
if d.getVarFlag(dep, "func", False) and not d.getVarFlag(dep, "python", False):
|
|
emit_var(dep, o, d, False) and o.write('\n')
|
|
newdeps |= bb.codeparser.ShellParser(dep, logger).parse_shell(d.getVar(dep))
|
|
newdeps |= set((d.getVarFlag(dep, "vardeps") or "").split())
|
|
newdeps -= seen
|
|
|
|
_functionfmt = """
|
|
def {function}(d):
|
|
{body}"""
|
|
|
|
def emit_func_python(func, o=sys.__stdout__, d = init()):
|
|
"""Emits all items in the data store in a format such that it can be sourced by a shell."""
|
|
|
|
def write_func(func, o, call = False):
|
|
body = d.getVar(func, False)
|
|
if not body.startswith("def"):
|
|
body = _functionfmt.format(function=func, body=body)
|
|
|
|
o.write(body.strip() + "\n\n")
|
|
if call:
|
|
o.write(func + "(d)" + "\n\n")
|
|
|
|
write_func(func, o, True)
|
|
pp = bb.codeparser.PythonParser(func, logger)
|
|
pp.parse_python(d.getVar(func, False))
|
|
newdeps = pp.execs
|
|
newdeps |= set((d.getVarFlag(func, "vardeps") or "").split())
|
|
seen = set()
|
|
while newdeps:
|
|
deps = newdeps
|
|
seen |= deps
|
|
newdeps = set()
|
|
for dep in deps:
|
|
if d.getVarFlag(dep, "func", False) and d.getVarFlag(dep, "python", False):
|
|
write_func(dep, o)
|
|
pp = bb.codeparser.PythonParser(dep, logger)
|
|
pp.parse_python(d.getVar(dep, False))
|
|
newdeps |= pp.execs
|
|
newdeps |= set((d.getVarFlag(dep, "vardeps") or "").split())
|
|
newdeps -= seen
|
|
|
|
def update_data(d):
|
|
"""Performs final steps upon the datastore, including application of overrides"""
|
|
d.finalize(parent = True)
|
|
|
|
def build_dependencies(key, keys, shelldeps, varflagsexcl, ignored_vars, d):
|
|
deps = set()
|
|
try:
|
|
if key[-1] == ']':
|
|
vf = key[:-1].split('[')
|
|
if vf[1] == "vardepvalueexclude":
|
|
return deps, ""
|
|
value, parser = d.getVarFlag(vf[0], vf[1], False, retparser=True)
|
|
deps |= parser.references
|
|
deps = deps | (keys & parser.execs)
|
|
return deps, value
|
|
varflags = d.getVarFlags(key, ["vardeps", "vardepvalue", "vardepsexclude", "exports", "postfuncs", "prefuncs", "lineno", "filename"]) or {}
|
|
vardeps = varflags.get("vardeps")
|
|
exclusions = varflags.get("vardepsexclude", "").split()
|
|
|
|
def handle_contains(value, contains, exclusions, d):
|
|
newvalue = []
|
|
if value:
|
|
newvalue.append(str(value))
|
|
for k in sorted(contains):
|
|
if k in exclusions or k in ignored_vars:
|
|
continue
|
|
l = (d.getVar(k) or "").split()
|
|
for item in sorted(contains[k]):
|
|
for word in item.split():
|
|
if not word in l:
|
|
newvalue.append("\n%s{%s} = Unset" % (k, item))
|
|
break
|
|
else:
|
|
newvalue.append("\n%s{%s} = Set" % (k, item))
|
|
return "".join(newvalue)
|
|
|
|
def handle_remove(value, deps, removes, d):
|
|
for r in sorted(removes):
|
|
r2 = d.expandWithRefs(r, None)
|
|
value += "\n_remove of %s" % r
|
|
deps |= r2.references
|
|
deps = deps | (keys & r2.execs)
|
|
return value
|
|
|
|
if "vardepvalue" in varflags:
|
|
value = varflags.get("vardepvalue")
|
|
elif varflags.get("func"):
|
|
if varflags.get("python"):
|
|
value = d.getVarFlag(key, "_content", False)
|
|
parser = bb.codeparser.PythonParser(key, logger)
|
|
parser.parse_python(value, filename=varflags.get("filename"), lineno=varflags.get("lineno"))
|
|
deps = deps | parser.references
|
|
deps = deps | (keys & parser.execs)
|
|
value = handle_contains(value, parser.contains, exclusions, d)
|
|
else:
|
|
value, parsedvar = d.getVarFlag(key, "_content", False, retparser=True)
|
|
parser = bb.codeparser.ShellParser(key, logger)
|
|
parser.parse_shell(parsedvar.value)
|
|
deps = deps | shelldeps
|
|
deps = deps | parsedvar.references
|
|
deps = deps | (keys & parser.execs) | (keys & parsedvar.execs)
|
|
value = handle_contains(value, parsedvar.contains, exclusions, d)
|
|
if hasattr(parsedvar, "removes"):
|
|
value = handle_remove(value, deps, parsedvar.removes, d)
|
|
if vardeps is None:
|
|
parser.log.flush()
|
|
if "prefuncs" in varflags:
|
|
deps = deps | set(varflags["prefuncs"].split())
|
|
if "postfuncs" in varflags:
|
|
deps = deps | set(varflags["postfuncs"].split())
|
|
if "exports" in varflags:
|
|
deps = deps | set(varflags["exports"].split())
|
|
else:
|
|
value, parser = d.getVarFlag(key, "_content", False, retparser=True)
|
|
deps |= parser.references
|
|
deps = deps | (keys & parser.execs)
|
|
value = handle_contains(value, parser.contains, exclusions, d)
|
|
if hasattr(parser, "removes"):
|
|
value = handle_remove(value, deps, parser.removes, d)
|
|
|
|
if "vardepvalueexclude" in varflags:
|
|
exclude = varflags.get("vardepvalueexclude")
|
|
for excl in exclude.split('|'):
|
|
if excl:
|
|
value = value.replace(excl, '')
|
|
|
|
# Add varflags, assuming an exclusion list is set
|
|
if varflagsexcl:
|
|
varfdeps = []
|
|
for f in varflags:
|
|
if f not in varflagsexcl:
|
|
varfdeps.append('%s[%s]' % (key, f))
|
|
if varfdeps:
|
|
deps |= set(varfdeps)
|
|
|
|
deps |= set((vardeps or "").split())
|
|
deps -= set(exclusions)
|
|
except bb.parse.SkipRecipe:
|
|
raise
|
|
except Exception as e:
|
|
bb.warn("Exception during build_dependencies for %s" % key)
|
|
raise
|
|
return deps, value
|
|
#bb.note("Variable %s references %s and calls %s" % (key, str(deps), str(execs)))
|
|
#d.setVarFlag(key, "vardeps", deps)
|
|
|
|
def generate_dependencies(d, ignored_vars):
|
|
|
|
keys = set(key for key in d if not key.startswith("__"))
|
|
shelldeps = set(key for key in d.getVar("__exportlist", False) if d.getVarFlag(key, "export", False) and not d.getVarFlag(key, "unexport", False))
|
|
varflagsexcl = d.getVar('BB_SIGNATURE_EXCLUDE_FLAGS')
|
|
|
|
deps = {}
|
|
values = {}
|
|
|
|
tasklist = d.getVar('__BBTASKS', False) or []
|
|
for task in tasklist:
|
|
deps[task], values[task] = build_dependencies(task, keys, shelldeps, varflagsexcl, ignored_vars, d)
|
|
newdeps = deps[task]
|
|
seen = set()
|
|
while newdeps:
|
|
nextdeps = newdeps - ignored_vars
|
|
seen |= nextdeps
|
|
newdeps = set()
|
|
for dep in nextdeps:
|
|
if dep not in deps:
|
|
deps[dep], values[dep] = build_dependencies(dep, keys, shelldeps, varflagsexcl, ignored_vars, d)
|
|
newdeps |= deps[dep]
|
|
newdeps -= seen
|
|
#print "For %s: %s" % (task, str(deps[task]))
|
|
return tasklist, deps, values
|
|
|
|
def generate_dependency_hash(tasklist, gendeps, lookupcache, ignored_vars, fn):
|
|
taskdeps = {}
|
|
basehash = {}
|
|
|
|
for task in tasklist:
|
|
data = lookupcache[task]
|
|
|
|
if data is None:
|
|
bb.error("Task %s from %s seems to be empty?!" % (task, fn))
|
|
data = []
|
|
else:
|
|
data = [data]
|
|
|
|
gendeps[task] -= ignored_vars
|
|
newdeps = gendeps[task]
|
|
seen = set()
|
|
while newdeps:
|
|
nextdeps = newdeps
|
|
seen |= nextdeps
|
|
newdeps = set()
|
|
for dep in nextdeps:
|
|
if dep in ignored_vars:
|
|
continue
|
|
gendeps[dep] -= ignored_vars
|
|
newdeps |= gendeps[dep]
|
|
newdeps -= seen
|
|
|
|
alldeps = sorted(seen)
|
|
for dep in alldeps:
|
|
data.append(dep)
|
|
var = lookupcache[dep]
|
|
if var is not None:
|
|
data.append(str(var))
|
|
k = fn + ":" + task
|
|
basehash[k] = hashlib.sha256("".join(data).encode("utf-8")).hexdigest()
|
|
taskdeps[task] = alldeps
|
|
|
|
return taskdeps, basehash
|
|
|
|
def inherits_class(klass, d):
|
|
val = d.getVar('__inherit_cache', False) or []
|
|
needle = '/%s.bbclass' % klass
|
|
for v in val:
|
|
if v.endswith(needle):
|
|
return True
|
|
return False
|