mirror of
https://git.yoctoproject.org/poky
synced 2026-04-19 06:32:13 +02:00
After the recent locale changes there were warnings about many unpackaged files. Fix this by directing libc-package.bbclass to operate directly on the files in the sysroot and adding packaging for .debug files in this package. Also sync up the eglibc and glibc versions of this code more closely. (From OE-Core rev: 1a3a76e4652b95b640c287cb3672ccc146b023ca) Signed-off-by: Richard Purdie <richard.purdie@linuxfoundation.org>
378 lines
14 KiB
Plaintext
378 lines
14 KiB
Plaintext
#
|
|
# This class knows how to package up [e]glibc. Its shared since prebuild binary toolchains
|
|
# may need packaging and its pointless to duplicate this code.
|
|
#
|
|
# Caller should set GLIBC_INTERNAL_USE_BINARY_LOCALE to one of:
|
|
# "compile" - Use QEMU to generate the binary locale files
|
|
# "precompiled" - The binary locale files are pregenerated and already present
|
|
# "ondevice" - The device will build the locale files upon first boot through the postinst
|
|
|
|
GLIBC_INTERNAL_USE_BINARY_LOCALE ?= "ondevice"
|
|
|
|
python __anonymous () {
|
|
enabled = bb.data.getVar("ENABLE_BINARY_LOCALE_GENERATION", d, 1)
|
|
|
|
pn = d.getVar("PN", True)
|
|
if pn.endswith("-initial"):
|
|
enabled = False
|
|
|
|
if enabled and int(enabled):
|
|
import re
|
|
|
|
target_arch = bb.data.getVar("TARGET_ARCH", d, 1)
|
|
binary_arches = bb.data.getVar("BINARY_LOCALE_ARCHES", d, 1) or ""
|
|
use_cross_localedef = bb.data.getVar("LOCALE_GENERATION_WITH_CROSS-LOCALEDEF", d, 1) or ""
|
|
|
|
for regexp in binary_arches.split(" "):
|
|
r = re.compile(regexp)
|
|
|
|
if r.match(target_arch):
|
|
depends = bb.data.getVar("DEPENDS", d, 1)
|
|
if use_cross_localedef == "1" :
|
|
depends = "%s cross-localedef-native" % depends
|
|
else:
|
|
depends = "%s qemu-native" % depends
|
|
bb.data.setVar("DEPENDS", depends, d)
|
|
bb.data.setVar("GLIBC_INTERNAL_USE_BINARY_LOCALE", "compile", d)
|
|
break
|
|
}
|
|
|
|
OVERRIDES_append = ":${TARGET_ARCH}-${TARGET_OS}"
|
|
|
|
do_configure_prepend() {
|
|
sed -e "s#@BASH@#/bin/sh#" -i ${S}/elf/ldd.bash.in
|
|
}
|
|
|
|
|
|
|
|
# indentation removed on purpose
|
|
locale_base_postinst() {
|
|
#!/bin/sh
|
|
|
|
if [ "x$D" != "x" ]; then
|
|
exit 1
|
|
fi
|
|
|
|
rm -rf ${TMP_LOCALE}
|
|
mkdir -p ${TMP_LOCALE}
|
|
if [ -f ${libdir}/locale/locale-archive ]; then
|
|
cp ${libdir}/locale/locale-archive ${TMP_LOCALE}/
|
|
fi
|
|
localedef --inputfile=${datadir}/i18n/locales/%s --charmap=%s --prefix=/tmp/locale %s
|
|
mkdir -p ${libdir}/locale/
|
|
mv ${TMP_LOCALE}/locale-archive ${libdir}/locale/
|
|
rm -rf ${TMP_LOCALE}
|
|
}
|
|
|
|
# indentation removed on purpose
|
|
locale_base_postrm() {
|
|
#!/bin/sh
|
|
|
|
rm -rf ${TMP_LOCALE}
|
|
mkdir -p ${TMP_LOCALE}
|
|
if [ -f ${libdir}/locale/locale-archive ]; then
|
|
cp ${libdir}/locale/locale-archive ${TMP_LOCALE}/
|
|
fi
|
|
localedef --delete-from-archive --inputfile=${datadir}/locales/%s --charmap=%s --prefix=/tmp/locale %s
|
|
mv ${TMP_LOCALE}/locale-archive ${libdir}/locale/
|
|
rm -rf ${TMP_LOCALE}
|
|
}
|
|
|
|
|
|
TMP_LOCALE="/tmp/locale${libdir}/locale"
|
|
LOCALETREESRC ?= "{PKGD}"
|
|
|
|
do_prep_locale_tree() {
|
|
treedir=${WORKDIR}/locale-tree
|
|
rm -rf $treedir
|
|
mkdir -p $treedir/${base_bindir} $treedir/${base_libdir} $treedir/${datadir} $treedir/${libdir}/locale
|
|
tar -cf - -C ${LOCALETREESRC}${datadir} -ps i18n | tar -xf - -C $treedir/${datadir}
|
|
# unzip to avoid parsing errors
|
|
for i in $treedir/${datadir}/i18n/charmaps/*gz; do
|
|
gunzip $i
|
|
done
|
|
tar -cf - -C ${LOCALETREESRC}${base_libdir} -ps . | tar -xf - -C $treedir/${base_libdir}
|
|
if [ -f ${STAGING_DIR_NATIVE}${prefix_native}/lib/libgcc_s.* ]; then
|
|
tar -cf - -C ${STAGING_DIR_NATIVE}/${prefix_native}/${base_libdir} -ps libgcc_s.* | tar -xf - -C $treedir/${base_libdir}
|
|
fi
|
|
install -m 0755 ${LOCALETREESRC}${bindir}/localedef $treedir/${base_bindir}
|
|
}
|
|
|
|
do_collect_bins_from_locale_tree() {
|
|
treedir=${WORKDIR}/locale-tree
|
|
|
|
mkdir -p ${PKGD}${libdir}
|
|
tar -cf - -C $treedir/${libdir} -ps locale | tar -xf - -C ${PKGD}${libdir}
|
|
}
|
|
|
|
inherit qemu
|
|
|
|
python package_do_split_gconvs () {
|
|
import os, re
|
|
if (bb.data.getVar('PACKAGE_NO_GCONV', d, 1) == '1'):
|
|
bb.note("package requested not splitting gconvs")
|
|
return
|
|
|
|
if not bb.data.getVar('PACKAGES', d, 1):
|
|
return
|
|
|
|
bpn = bb.data.getVar('BPN', d, 1)
|
|
libdir = bb.data.getVar('libdir', d, 1)
|
|
if not libdir:
|
|
bb.error("libdir not defined")
|
|
return
|
|
datadir = bb.data.getVar('datadir', d, 1)
|
|
if not datadir:
|
|
bb.error("datadir not defined")
|
|
return
|
|
|
|
gconv_libdir = base_path_join(libdir, "gconv")
|
|
charmap_dir = base_path_join(datadir, "i18n", "charmaps")
|
|
locales_dir = base_path_join(datadir, "i18n", "locales")
|
|
binary_locales_dir = base_path_join(libdir, "locale")
|
|
|
|
def calc_gconv_deps(fn, pkg, file_regex, output_pattern, group):
|
|
deps = []
|
|
f = open(fn, "r")
|
|
c_re = re.compile('^copy "(.*)"')
|
|
i_re = re.compile('^include "(\w+)".*')
|
|
for l in f.readlines():
|
|
m = c_re.match(l) or i_re.match(l)
|
|
if m:
|
|
dp = legitimize_package_name('%s-gconv-%s' % (bpn, m.group(1)))
|
|
if not dp in deps:
|
|
deps.append(dp)
|
|
f.close()
|
|
if deps != []:
|
|
bb.data.setVar('RDEPENDS_%s' % pkg, " ".join(deps), d)
|
|
if bpn != 'glibc':
|
|
bb.data.setVar('RPROVIDES_%s' % pkg, pkg.replace(bpn, 'glibc'), d)
|
|
|
|
do_split_packages(d, gconv_libdir, file_regex='^(.*)\.so$', output_pattern=bpn+'-gconv-%s', \
|
|
description='gconv module for character set %s', hook=calc_gconv_deps, \
|
|
extra_depends=bpn+'-gconv')
|
|
|
|
def calc_charmap_deps(fn, pkg, file_regex, output_pattern, group):
|
|
deps = []
|
|
f = open(fn, "r")
|
|
c_re = re.compile('^copy "(.*)"')
|
|
i_re = re.compile('^include "(\w+)".*')
|
|
for l in f.readlines():
|
|
m = c_re.match(l) or i_re.match(l)
|
|
if m:
|
|
dp = legitimize_package_name('%s-charmap-%s' % (bpn, m.group(1)))
|
|
if not dp in deps:
|
|
deps.append(dp)
|
|
f.close()
|
|
if deps != []:
|
|
bb.data.setVar('RDEPENDS_%s' % pkg, " ".join(deps), d)
|
|
if bpn != 'glibc':
|
|
bb.data.setVar('RPROVIDES_%s' % pkg, pkg.replace(bpn, 'glibc'), d)
|
|
|
|
do_split_packages(d, charmap_dir, file_regex='^(.*)\.gz$', output_pattern=bpn+'-charmap-%s', \
|
|
description='character map for %s encoding', hook=calc_charmap_deps, extra_depends='')
|
|
|
|
def calc_locale_deps(fn, pkg, file_regex, output_pattern, group):
|
|
deps = []
|
|
f = open(fn, "r")
|
|
c_re = re.compile('^copy "(.*)"')
|
|
i_re = re.compile('^include "(\w+)".*')
|
|
for l in f.readlines():
|
|
m = c_re.match(l) or i_re.match(l)
|
|
if m:
|
|
dp = legitimize_package_name(bpn+'-localedata-%s' % m.group(1))
|
|
if not dp in deps:
|
|
deps.append(dp)
|
|
f.close()
|
|
if deps != []:
|
|
bb.data.setVar('RDEPENDS_%s' % pkg, " ".join(deps), d)
|
|
if bpn != 'glibc':
|
|
bb.data.setVar('RPROVIDES_%s' % pkg, pkg.replace(bpn, 'glibc'), d)
|
|
|
|
do_split_packages(d, locales_dir, file_regex='(.*)', output_pattern=bpn+'-localedata-%s', \
|
|
description='locale definition for %s', hook=calc_locale_deps, extra_depends='')
|
|
bb.data.setVar('PACKAGES', bb.data.getVar('PACKAGES', d) + ' ' + bpn + '-gconv', d)
|
|
|
|
use_bin = bb.data.getVar("GLIBC_INTERNAL_USE_BINARY_LOCALE", d, 1)
|
|
|
|
dot_re = re.compile("(.*)\.(.*)")
|
|
|
|
#GLIBC_GENERATE_LOCALES var specifies which locales to be supported, empty or "all" means all locales
|
|
if use_bin != "precompiled":
|
|
supported = bb.data.getVar('GLIBC_GENERATE_LOCALES', d, 1)
|
|
if not supported or supported == "all":
|
|
f = open(base_path_join(bb.data.getVar('WORKDIR', d, 1), "SUPPORTED"), "r")
|
|
supported = f.readlines()
|
|
f.close()
|
|
else:
|
|
supported = supported.split()
|
|
supported = map(lambda s:s.replace(".", " ") + "\n", supported)
|
|
else:
|
|
supported = []
|
|
full_bin_path = bb.data.getVar('PKGD', d, True) + binary_locales_dir
|
|
for dir in os.listdir(full_bin_path):
|
|
dbase = dir.split(".")
|
|
d2 = " "
|
|
if len(dbase) > 1:
|
|
d2 = "." + dbase[1].upper() + " "
|
|
supported.append(dbase[0] + d2)
|
|
|
|
# Collate the locales by base and encoding
|
|
utf8_only = int(bb.data.getVar('LOCALE_UTF8_ONLY', d, 1) or 0)
|
|
encodings = {}
|
|
for l in supported:
|
|
l = l[:-1]
|
|
(locale, charset) = l.split(" ")
|
|
if utf8_only and charset != 'UTF-8':
|
|
continue
|
|
m = dot_re.match(locale)
|
|
if m:
|
|
locale = m.group(1)
|
|
if not encodings.has_key(locale):
|
|
encodings[locale] = []
|
|
encodings[locale].append(charset)
|
|
|
|
def output_locale_source(name, pkgname, locale, encoding):
|
|
bb.data.setVar('RDEPENDS_%s' % pkgname, 'localedef %s-localedata-%s %s-charmap-%s' % \
|
|
(bpn, legitimize_package_name(locale), bpn, legitimize_package_name(encoding)), d)
|
|
bb.data.setVar('pkg_postinst_%s' % pkgname, bb.data.getVar('locale_base_postinst', d, 1) \
|
|
% (locale, encoding, locale), d)
|
|
bb.data.setVar('pkg_postrm_%s' % pkgname, bb.data.getVar('locale_base_postrm', d, 1) % \
|
|
(locale, encoding, locale), d)
|
|
|
|
def output_locale_binary_rdepends(name, pkgname, locale, encoding):
|
|
m = re.match("(.*)_(.*)", name)
|
|
if m:
|
|
libc_name = "%s.%s" % (m.group(1), m.group(2).lower().replace("-",""))
|
|
else:
|
|
libc_name = name
|
|
bb.data.setVar('RDEPENDS_%s' % pkgname, legitimize_package_name('%s-binary-localedata-%s' \
|
|
% (bpn, libc_name)), d)
|
|
rprovides = (bb.data.getVar('RPROVIDES_%s' % pkgname, d, True) or "").split()
|
|
rprovides.append(legitimize_package_name('%s-binary-localedata-%s' % (bpn, libc_name)))
|
|
bb.data.setVar('RPROVIDES_%s' % pkgname, " ".join(rprovides), d)
|
|
|
|
def output_locale_binary(name, pkgname, locale, encoding):
|
|
treedir = base_path_join(bb.data.getVar("WORKDIR", d, 1), "locale-tree")
|
|
ldlibdir = base_path_join(treedir, bb.data.getVar("base_libdir", d, 1))
|
|
path = bb.data.getVar("PATH", d, 1)
|
|
i18npath = base_path_join(treedir, datadir, "i18n")
|
|
gconvpath = base_path_join(treedir, "iconvdata")
|
|
outputpath = base_path_join(treedir, libdir, "locale")
|
|
|
|
use_cross_localedef = bb.data.getVar("LOCALE_GENERATION_WITH_CROSS-LOCALEDEF", d, 1) or "0"
|
|
if use_cross_localedef == "1":
|
|
target_arch = bb.data.getVar('TARGET_ARCH', d, True)
|
|
locale_arch_options = { \
|
|
"arm": " --uint32-align=4 --little-endian ", \
|
|
"powerpc": " --uint32-align=4 --big-endian ", \
|
|
"mips": " --uint32-align=4 --big-endian ", \
|
|
"mipsel": " --uint32-align=4 --little-endian ", \
|
|
"i586": " --uint32-align=4 --little-endian ", \
|
|
"x86_64": " --uint32-align=4 --little-endian " }
|
|
|
|
if target_arch in locale_arch_options:
|
|
localedef_opts = locale_arch_options[target_arch]
|
|
else:
|
|
bb.error("locale_arch_options not found for target_arch=" + target_arch)
|
|
raise bb.build.FuncFailed("unknown arch:" + target_arch + " for locale_arch_options")
|
|
|
|
localedef_opts += " --force --old-style --no-archive --prefix=%s \
|
|
--inputfile=%s/%s/i18n/locales/%s --charmap=%s %s/%s" \
|
|
% (treedir, treedir, datadir, locale, encoding, outputpath, name)
|
|
|
|
cmd = "PATH=\"%s\" I18NPATH=\"%s\" GCONV_PATH=\"%s\" cross-localedef %s" % \
|
|
(path, i18npath, gconvpath, localedef_opts)
|
|
else: # earlier slower qemu way
|
|
qemu = qemu_target_binary(d)
|
|
localedef_opts = "--force --old-style --no-archive --prefix=%s \
|
|
--inputfile=%s/i18n/locales/%s --charmap=%s %s" \
|
|
% (treedir, datadir, locale, encoding, name)
|
|
|
|
qemu_options = bb.data.getVar("QEMU_OPTIONS_%s" % bb.data.getVar('PACKAGE_ARCH', d, 1), d, 1)
|
|
if not qemu_options:
|
|
qemu_options = bb.data.getVar('QEMU_OPTIONS', d, 1)
|
|
|
|
cmd = "PSEUDO_RELOADED=YES PATH=\"%s\" I18NPATH=\"%s\" %s -L %s \
|
|
-E LD_LIBRARY_PATH=%s %s %s/bin/localedef %s" % \
|
|
(path, i18npath, qemu, treedir, ldlibdir, qemu_options, treedir, localedef_opts)
|
|
|
|
bb.note("generating locale %s (%s)" % (locale, encoding))
|
|
import subprocess
|
|
process = subprocess.Popen(cmd, shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
|
|
if process.wait() != 0:
|
|
bb.note("cmd:")
|
|
bb.note(cmd)
|
|
bb.note("stdout:")
|
|
bb.note(process.stdout.read())
|
|
bb.note("stderr:")
|
|
bb.note(process.stderr.read())
|
|
raise bb.build.FuncFailed("localedef returned an error")
|
|
|
|
def output_locale(name, locale, encoding):
|
|
pkgname = 'locale-base-' + legitimize_package_name(name)
|
|
bb.data.setVar('ALLOW_EMPTY_%s' % pkgname, '1', d)
|
|
bb.data.setVar('PACKAGES', '%s %s' % (pkgname, bb.data.getVar('PACKAGES', d, 1)), d)
|
|
rprovides = ' virtual-locale-%s' % legitimize_package_name(name)
|
|
m = re.match("(.*)_(.*)", name)
|
|
if m:
|
|
rprovides += ' virtual-locale-%s' % m.group(1)
|
|
bb.data.setVar('RPROVIDES_%s' % pkgname, rprovides, d)
|
|
|
|
if use_bin == "compile":
|
|
output_locale_binary_rdepends(name, pkgname, locale, encoding)
|
|
output_locale_binary(name, pkgname, locale, encoding)
|
|
elif use_bin == "precompiled":
|
|
output_locale_binary_rdepends(name, pkgname, locale, encoding)
|
|
else:
|
|
output_locale_source(name, pkgname, locale, encoding)
|
|
|
|
if use_bin == "compile":
|
|
bb.note("preparing tree for binary locale generation")
|
|
bb.build.exec_func("do_prep_locale_tree", d)
|
|
|
|
# Reshuffle names so that UTF-8 is preferred over other encodings
|
|
non_utf8 = []
|
|
for l in encodings.keys():
|
|
if len(encodings[l]) == 1:
|
|
output_locale(l, l, encodings[l][0])
|
|
if encodings[l][0] != "UTF-8":
|
|
non_utf8.append(l)
|
|
else:
|
|
if "UTF-8" in encodings[l]:
|
|
output_locale(l, l, "UTF-8")
|
|
encodings[l].remove("UTF-8")
|
|
else:
|
|
non_utf8.append(l)
|
|
for e in encodings[l]:
|
|
output_locale('%s.%s' % (l, e), l, e)
|
|
|
|
if non_utf8 != [] and use_bin != "precompiled":
|
|
bb.note("the following locales are supported only in legacy encodings:")
|
|
bb.note(" " + " ".join(non_utf8))
|
|
|
|
if use_bin == "compile":
|
|
bb.note("collecting binary locales from locale tree")
|
|
bb.build.exec_func("do_collect_bins_from_locale_tree", d)
|
|
do_split_packages(d, binary_locales_dir, file_regex='(.*)', \
|
|
output_pattern=bpn+'-binary-localedata-%s', \
|
|
description='binary locale definition for %s', extra_depends='', allow_dirs=True)
|
|
elif use_bin == "precompiled":
|
|
do_split_packages(d, binary_locales_dir, file_regex='(.*)', \
|
|
output_pattern=bpn+'-binary-localedata-%s', \
|
|
description='binary locale definition for %s', extra_depends='', allow_dirs=True)
|
|
else:
|
|
bb.note("generation of binary locales disabled. this may break i18n!")
|
|
|
|
}
|
|
|
|
# We want to do this indirection so that we can safely 'return'
|
|
# from the called function even though we're prepending
|
|
python populate_packages_prepend () {
|
|
if bb.data.getVar('DEBIAN_NAMES', d, 1):
|
|
bpn = bb.data.getVar('BPN', d, 1)
|
|
bb.data.setVar('PKG_'+bpn, 'libc6', d)
|
|
bb.data.setVar('PKG_'+bpn+'-dev', 'libc6-dev', d)
|
|
bb.build.exec_func('package_do_split_gconvs', d)
|
|
}
|