blob: 9bccd76f0cf6a7b3bcbdec25eede1c43243e793e [file]
# Development tool - ide-sdk command plugin
#
# Copyright (C) 2023-2024 Siemens AG
#
# SPDX-License-Identifier: GPL-2.0-only
#
"""Devtool ide-sdk plugin"""
import json
import logging
import os
import re
import shutil
import stat
import subprocess
import sys
import shlex
import glob
from argparse import RawTextHelpFormatter
from enum import Enum
import scriptutils
import bb
from devtool import exec_build_env_command, setup_tinfoil, check_workspace_recipe, DevtoolError, parse_recipe
from devtool.standard import get_real_srctree
from devtool.ide_plugins import BuildTool
from oe.kernel_module import kernel_module_os_env
logger = logging.getLogger('devtool')
# dict of classes derived from IdeBase
ide_plugins = {}
class DevtoolIdeMode(Enum):
"""Different modes are supported by the ide-sdk plugin.
The enum might be extended by more advanced modes in the future. Some ideas:
- auto: modified if all recipes are modified, shared if none of the recipes is modified.
- mixed: modified mode for modified recipes, shared mode for all other recipes.
"""
modified = 'modified'
shared = 'shared'
class TargetDevice:
"""SSH remote login parameters"""
def __init__(self, args):
self.extraoptions = []
if args.no_host_check:
self.extraoptions += ['-o', 'UserKnownHostsFile=/dev/null', '-o', 'StrictHostKeyChecking=no']
self.ssh_sshexec = 'ssh'
if args.ssh_exec:
self.ssh_sshexec = args.ssh_exec
self.ssh_port = ''
if args.port:
self.ssh_port = ['-p', args.port]
if args.key:
self.extraoptions += ['-i', args.key]
self.target = args.target
target_sp = args.target.split('@')
if len(target_sp) == 1:
self.login = ""
self.host = target_sp[0]
elif len(target_sp) == 2:
self.login = target_sp[0]
self.host = target_sp[1]
else:
logger.error("Invalid target argument: %s" % args.target)
class RecipeNative:
"""Base class for calling bitbake to provide a -native recipe"""
def __init__(self, name, target_arch=None):
self.name = name
self.target_arch = target_arch
self.bootstrap_tasks = [self.name + ':do_addto_recipe_sysroot']
self.staging_bindir_native = None
self.target_sys = None
self.__native_bin = None
def _initialize(self, config, workspace, tinfoil):
"""Get the parsed recipe"""
recipe_d = parse_recipe(
config, tinfoil, self.name, appends=True, filter_workspace=False)
if not recipe_d:
raise DevtoolError("Parsing %s recipe failed" % self.name)
self.staging_bindir_native = os.path.realpath(
recipe_d.getVar('STAGING_BINDIR_NATIVE'))
self.target_sys = recipe_d.getVar('TARGET_SYS')
return recipe_d
def initialize(self, config, workspace, tinfoil):
"""Basic initialization that can be overridden by a derived class"""
self._initialize(config, workspace, tinfoil)
@property
def native_bin(self):
if not self.__native_bin:
raise DevtoolError("native binary name is not defined.")
return self.__native_bin
class RecipeGdbCross(RecipeNative):
"""Handle gdb-cross on the host and the gdbserver on the target device"""
def __init__(self, args, target_arch, target_device):
super().__init__('gdb-cross-' + target_arch, target_arch)
self.target_device = target_device
self.gdb = None
self.gdbserver_port_next = int(args.gdbserver_port_start)
self.config_db = {}
def __find_gdbserver(self, config, tinfoil):
"""Absolute path of the gdbserver"""
recipe_d_gdb = parse_recipe(
config, tinfoil, 'gdb', appends=True, filter_workspace=False)
if not recipe_d_gdb:
raise DevtoolError("Parsing gdb recipe failed")
return os.path.join(recipe_d_gdb.getVar('bindir'), 'gdbserver')
def initialize(self, config, workspace, tinfoil):
super()._initialize(config, workspace, tinfoil)
gdb_bin = self.target_sys + '-gdb'
gdb_path = os.path.join(
self.staging_bindir_native, self.target_sys, gdb_bin)
self.gdb = gdb_path
self.gdbserver_path = self.__find_gdbserver(config, tinfoil)
@property
def host(self):
return self.target_device.host
class RecipeImage:
"""Handle some image recipe related properties
Most workflows require firmware that runs on the target device.
This firmware must be consistent with the setup of the host system.
In particular, the debug symbols must be compatible. For this, the
rootfs must be created as part of the SDK.
"""
def __init__(self, name):
self.combine_dbg_image = False
self.gdbserver_missing = False
self.name = name
self.rootfs = None
self.__rootfs_dbg = None
self.bootstrap_tasks = [self.name + ':do_build']
def initialize(self, config, tinfoil):
image_d = parse_recipe(
config, tinfoil, self.name, appends=True, filter_workspace=False)
if not image_d:
raise DevtoolError(
"Parsing image recipe %s failed" % self.name)
self.combine_dbg_image = bb.data.inherits_class(
'image-combined-dbg', image_d)
workdir = image_d.getVar('WORKDIR')
self.rootfs = os.path.join(workdir, 'rootfs')
if image_d.getVar('IMAGE_GEN_DEBUGFS') == "1":
self.__rootfs_dbg = os.path.join(workdir, 'rootfs-dbg')
self.gdbserver_missing = 'gdbserver' not in image_d.getVar(
'IMAGE_INSTALL') and 'tools-debug' not in image_d.getVar('IMAGE_FEATURES')
@property
def debug_support(self):
return bool(self.rootfs_dbg)
@property
def rootfs_dbg(self):
if self.__rootfs_dbg and os.path.isdir(self.__rootfs_dbg):
return self.__rootfs_dbg
return None
class RecipeMetaIdeSupport:
"""For the shared sysroots mode meta-ide-support is needed
For use cases where just a cross tool-chain is required but
no recipe is used, devtool ide-sdk abstracts calling bitbake meta-ide-support
and bitbake build-sysroots. This also allows to expose the cross-toolchains
to IDEs. For example VSCode support different tool-chains with e.g. cmake-kits.
"""
def __init__(self):
self.bootstrap_tasks = ['meta-ide-support:do_build']
self.topdir = None
self.datadir = None
self.deploy_dir_image = None
self.build_sys = None
# From toolchain-scripts
self.real_multimach_target_sys = None
def initialize(self, config, tinfoil):
meta_ide_support_d = parse_recipe(
config, tinfoil, 'meta-ide-support', appends=True, filter_workspace=False)
if not meta_ide_support_d:
raise DevtoolError("Parsing meta-ide-support recipe failed")
self.topdir = meta_ide_support_d.getVar('TOPDIR')
self.datadir = meta_ide_support_d.getVar('datadir')
self.deploy_dir_image = meta_ide_support_d.getVar(
'DEPLOY_DIR_IMAGE')
self.build_sys = meta_ide_support_d.getVar('BUILD_SYS')
self.real_multimach_target_sys = meta_ide_support_d.getVar(
'REAL_MULTIMACH_TARGET_SYS')
class RecipeBuildSysroots:
"""For the shared sysroots mode build-sysroots is needed"""
def __init__(self):
self.standalone_sysroot = None
self.standalone_sysroot_native = None
self.bootstrap_tasks = [
'build-sysroots:do_build_target_sysroot',
'build-sysroots:do_build_native_sysroot'
]
def initialize(self, config, tinfoil):
build_sysroots_d = parse_recipe(
config, tinfoil, 'build-sysroots', appends=True, filter_workspace=False)
if not build_sysroots_d:
raise DevtoolError("Parsing build-sysroots recipe failed")
self.standalone_sysroot = build_sysroots_d.getVar(
'STANDALONE_SYSROOT')
self.standalone_sysroot_native = build_sysroots_d.getVar(
'STANDALONE_SYSROOT_NATIVE')
class SharedSysrootsEnv:
"""Handle the shared sysroots based workflow
Support the workflow with just a tool-chain without a recipe.
It's basically like:
bitbake some-dependencies
bitbake meta-ide-support
bitbake build-sysroots
Use the environment-* file found in the deploy folder
"""
def __init__(self):
self.ide_support = None
self.build_sysroots = None
def initialize(self, ide_support, build_sysroots):
self.ide_support = ide_support
self.build_sysroots = build_sysroots
def setup_ide(self, ide):
ide.setup(self)
class RecipeNotModified:
"""Handling of recipes added to the Direct DSK shared sysroots."""
def __init__(self, name):
self.name = name
self.bootstrap_tasks = [name + ':do_populate_sysroot']
class ExecutableBinary:
"""Represent an installed executable binary of a modified recipe"""
def __init__(self, image_dir_d, binary_path,
systemd_services, init_scripts):
self.image_dir_d = image_dir_d
self.binary_path = binary_path
self.init_script = None
self.systemd_service = None
self._init_service_for_binary(systemd_services)
self._init_init_script_for_binary(init_scripts)
def _init_service_for_binary(self, systemd_services):
"""Find systemd service file that handles this binary"""
service_dirs = [
'etc/systemd/system',
'lib/systemd/system',
'usr/lib/systemd/system'
]
for _, services in systemd_services.items():
for service in services:
for service_dir in service_dirs:
service_path = os.path.join(self.image_dir_d, service_dir, service)
if os.path.exists(service_path):
try:
with open(service_path, 'r') as f:
for line in f:
if line.strip().startswith('ExecStart='):
exec_start = line.strip()[10:].strip() # Remove 'ExecStart='
# Remove any leading modifiers like '-' or '@'
exec_start = exec_start.lstrip('-@')
# Get the first word (the executable path)
exec_binary = exec_start.split()[0] if exec_start.split() else ''
if exec_binary == self.binary_path or exec_binary.endswith('/' + self.binary_path.lstrip('/')):
logger.debug("Found systemd service for binary %s: %s" % (self.binary_path, service))
self.systemd_service = service
except (IOError, OSError):
continue
def _init_init_script_for_binary(self, init_scripts):
"""Find SysV init script that handles this binary"""
init_dirs = [
'etc/init.d',
'etc/rc.d/init.d'
]
for _, init_scripts in init_scripts.items():
for init_script in init_scripts:
for init_dir in init_dirs:
init_path = os.path.join(self.image_dir_d, init_dir, init_script)
if os.path.exists(init_path):
init_script_path = os.path.join("/", init_dir, init_script)
binary_name = os.path.basename(self.binary_path)
# if the init script file name is equal to the binary file name, return it directly
if os.path.basename(init_script) == binary_name:
logger.debug("Found SysV init script for binary %s: %s" % (self.binary_path, init_script_path))
self.init_script = init_script_path
# Otherwise check if the script containes a reference to the binary
try:
with open(init_path, 'r') as f:
content = f.read()
pattern = r'\b' + re.escape(binary_name) + r'\b'
if re.search(pattern, content):
logger.debug("Found SysV init script for binary %s: %s" % (self.binary_path, init_script_path))
self.init_script = init_script_path
return
except (IOError, OSError):
continue
@property
def binary_host_path(self):
"""Get the absolute path of this binary on the host"""
return os.path.join(self.image_dir_d, self.binary_path.lstrip('/'))
@property
def runs_as_service(self):
"""Check if this binary is run by a service or init script"""
return self.systemd_service is not None or self.init_script is not None
@property
def start_command(self):
"""Get the command to start this binary"""
if self.systemd_service:
return "systemctl start %s" % self.systemd_service
if self.init_script:
return "%s start" % self.init_script
return None
@property
def stop_command(self):
"""Get the command to stop this binary"""
if self.systemd_service:
return "systemctl stop %s" % self.systemd_service
if self.init_script:
return "%s stop" % self.init_script
return None
@property
def pid_command(self):
"""Get the command to get the PID of this binary"""
if self.systemd_service:
return "systemctl show --property MainPID --value %s" % self.systemd_service
if self.init_script:
return "pidof %s" % os.path.basename(self.binary_path)
return None
class RecipeModified:
"""Handling of recipes in the workspace created by devtool modify"""
OE_INIT_BUILD_ENV = 'oe-init-build-env'
INIT_BUILD_ENV = 'init-build-env'
VALID_BASH_ENV_NAME_CHARS = re.compile(r"^[a-zA-Z0-9_]*$")
def __init__(self, name):
self.name = name
self.bootstrap_tasks = [name + ':do_install']
self.gdb_cross = None
# workspace
self.real_srctree = None
self.srctree = None
self.ide_sdk_dir = None
self.ide_sdk_scripts_dir = None
self.bbappend = None
# recipe variables from d.getVar
self.b = None
self.base_libdir = None
self.bblayers = None
self.bitbakepath = None
self.bpn = None
self.d = None
self.debug_build = None
self.reverse_debug_prefix_map = {}
self.fakerootcmd = None
self.fakerootenv = None
self.libdir = None
self.max_process = None
self.package_arch = None
self.package_debug_split_style = None
self.path = None
self.pn = None
self.recipe_id = None
self.recipe_sysroot = None
self.recipe_sysroot_native = None
self.s = None
self.staging_incdir = None
self.strip_cmd = None
self.target_arch = None
self.tcoverride = None
self.topdir = None
self.workdir = None
# Service management
self.systemd_services = {}
self.init_scripts = {}
# replicate bitbake build environment
self.exported_vars = None
self.cmd_compile = None
self.__oe_init_dir = None
# main build tool used by this recipe
self.build_tool = BuildTool.UNDEFINED
# Whether this recipe benefits from gdbserver and rootfs-dbg in the image.
self.wants_gdbserver = True
# Whether to warn when DEBUG_BUILD is not set. Kernel modules are built
# by the kernel's build system and DEBUG_BUILD does not influence them.
self.wants_debug_build = True
# build_tool = cmake
self.oecmake_generator = None
self.cmake_cache_vars = None
# build_tool = meson
self.meson_buildtype = None
self.meson_wrapper = None
self.mesonopts = None
self.extra_oemeson = None
self.meson_cross_file = None
# kernel module
self.make_targets = None
self.extra_oemake = None
self.kernel_cc = None
self.staging_kernel_dir = None
# Populated after bitbake built all the recipes
self._installed_binaries = None
self._gdb_pretty_print_scripts = None
def initialize(self, config, workspace, tinfoil):
recipe_d = parse_recipe(
config, tinfoil, self.name, appends=True, filter_workspace=False)
if not recipe_d:
raise DevtoolError("Parsing %s recipe failed" % self.name)
# Verify this recipe is built as externalsrc setup by devtool modify
workspacepn = check_workspace_recipe(
workspace, self.name, bbclassextend=True)
self.srctree = workspace[workspacepn]['srctree']
# Need to grab this here in case the source is within a subdirectory
self.real_srctree = get_real_srctree(
self.srctree, recipe_d.getVar('S'), recipe_d.getVar('UNPACKDIR'))
self.bbappend = workspace[workspacepn]['bbappend']
self.ide_sdk_dir = os.path.join(
config.workspace_path, 'ide-sdk', self.name)
if os.path.exists(self.ide_sdk_dir):
shutil.rmtree(self.ide_sdk_dir)
self.ide_sdk_scripts_dir = os.path.join(self.ide_sdk_dir, 'scripts')
self.b = recipe_d.getVar('B')
self.base_libdir = recipe_d.getVar('base_libdir')
self.bblayers = recipe_d.getVar('BBLAYERS').split()
self.bitbakepath = recipe_d.getVar('BITBAKEPATH')
self.bpn = recipe_d.getVar('BPN')
self.cxx = recipe_d.getVar('CXX')
self.d = recipe_d.getVar('D')
self.debug_build = recipe_d.getVar('DEBUG_BUILD')
self.fakerootcmd = recipe_d.getVar('FAKEROOTCMD')
self.fakerootenv = recipe_d.getVar('FAKEROOTENV')
self.libdir = recipe_d.getVar('libdir')
self.max_process = int(recipe_d.getVar(
"BB_NUMBER_THREADS") or os.cpu_count() or 1)
self.package_arch = recipe_d.getVar('PACKAGE_ARCH')
self.package_debug_split_style = recipe_d.getVar(
'PACKAGE_DEBUG_SPLIT_STYLE')
self.path = recipe_d.getVar('PATH')
self.pn = recipe_d.getVar('PN')
self.recipe_sysroot = os.path.realpath(
recipe_d.getVar('RECIPE_SYSROOT'))
self.recipe_sysroot_native = os.path.realpath(
recipe_d.getVar('RECIPE_SYSROOT_NATIVE'))
self.s = recipe_d.getVar('S')
self.staging_bindir_toolchain = os.path.realpath(
recipe_d.getVar('STAGING_BINDIR_TOOLCHAIN'))
self.staging_incdir = os.path.realpath(
recipe_d.getVar('STAGING_INCDIR'))
self.strip_cmd = recipe_d.getVar('STRIP')
self.target_arch = recipe_d.getVar('TARGET_ARCH')
self.tcoverride = recipe_d.getVar('TCOVERRIDE')
self.topdir = recipe_d.getVar('TOPDIR')
self.workdir = os.path.realpath(recipe_d.getVar('WORKDIR'))
self.__init_exported_variables(recipe_d)
self.__init_systemd_services(recipe_d)
self.__init_init_scripts(recipe_d)
if bb.data.inherits_class('cmake', recipe_d):
self.oecmake_generator = recipe_d.getVar('OECMAKE_GENERATOR')
self.__init_cmake_preset_cache(recipe_d)
self.build_tool = BuildTool.CMAKE
elif bb.data.inherits_class('meson', recipe_d):
self.meson_buildtype = recipe_d.getVar('MESON_BUILDTYPE')
self.mesonopts = recipe_d.getVar('MESONOPTS')
self.extra_oemeson = recipe_d.getVar('EXTRA_OEMESON')
self.meson_cross_file = recipe_d.getVar('MESON_CROSS_FILE')
self.build_tool = BuildTool.MESON
elif bb.data.inherits_class('module', recipe_d):
self.build_tool = BuildTool.KERNEL_MODULE
self.wants_gdbserver = False
self.wants_debug_build = False
make_targets = recipe_d.getVar('MAKE_TARGETS')
if make_targets:
self.make_targets = shlex.split(make_targets)
else:
self.make_targets = ["all"]
extra_oemake = recipe_d.getVar('EXTRA_OEMAKE')
if extra_oemake:
self.extra_oemake = shlex.split(extra_oemake)
else:
self.extra_oemake = []
self.kernel_cc = recipe_d.getVar('KERNEL_CC')
self.staging_kernel_dir = recipe_d.getVar('STAGING_KERNEL_DIR')
# Export up the environment for building kernel modules
kernel_module_os_env(recipe_d, self.exported_vars)
# For the kernel the KERNEL_CC variable contains the prefix-map arguments
if self.build_tool is BuildTool.KERNEL_MODULE:
self.reverse_debug_prefix_map = self._init_reverse_debug_prefix_map(
self.kernel_cc)
else:
self.reverse_debug_prefix_map = self._init_reverse_debug_prefix_map(
recipe_d.getVar('DEBUG_PREFIX_MAP'))
# Recipe ID is the identifier for IDE config sections
self.recipe_id = self.bpn + "-" + self.package_arch
self.recipe_id_pretty = self.bpn + ": " + self.package_arch
@staticmethod
def is_valid_shell_variable(var):
"""Skip strange shell variables like systemd
prevent from strange bugs because of strange variables which
are not used in this context but break various tools.
"""
if RecipeModified.VALID_BASH_ENV_NAME_CHARS.match(var):
bb.debug(1, "ignoring variable: %s" % var)
return True
return False
def solib_search_path(self, image):
"""Search for debug symbols in the rootfs and rootfs-dbg
The debug symbols of shared libraries which are provided by other packages
are grabbed from the -dbg packages in the rootfs-dbg.
But most cross debugging tools like gdb, perf, and systemtap need to find
executable/library first and through it debuglink note find corresponding
symbols file. Therefore the library paths from the rootfs are added as well.
Note: For the devtool modified recipe compiled from the IDE, the debug
symbols are taken from the unstripped binaries in the image folder.
Also, devtool deploy-target takes the files from the image folder.
debug symbols in the image folder refer to the corresponding source files
with absolute paths of the build machine. Debug symbols found in the
rootfs-dbg are relocated and contain paths which refer to the source files
installed on the target device e.g. /usr/src/...
"""
base_libdir = self.base_libdir.lstrip('/')
libdir = self.libdir.lstrip('/')
so_paths = [
# debug symbols for package_debug_split_style: debug-with-srcpkg or .debug
os.path.join(image.rootfs_dbg, base_libdir, ".debug"),
os.path.join(image.rootfs_dbg, libdir, ".debug"),
# debug symbols for package_debug_split_style: debug-file-directory
os.path.join(image.rootfs_dbg, "usr", "lib", "debug"),
# The binaries are required as well, the debug packages are not enough
# With image-combined-dbg.bbclass the binaries are copied into rootfs-dbg
os.path.join(image.rootfs_dbg, base_libdir),
os.path.join(image.rootfs_dbg, libdir),
# Without image-combined-dbg.bbclass the binaries are only in rootfs.
# Note: Stepping into source files located in rootfs-dbg does not
# work without image-combined-dbg.bbclass yet.
os.path.join(image.rootfs, base_libdir),
os.path.join(image.rootfs, libdir)
]
return so_paths
def solib_search_path_str(self, image):
"""Return a : separated list of paths usable by GDB's set solib-search-path"""
return ':'.join(self.solib_search_path(image))
def _init_reverse_debug_prefix_map(self, debug_prefix_map):
"""Parses GCC map options and returns a mapping of target to host paths.
This function scans a string containing GCC options such as -fdebug-prefix-map,
-fmacro-prefix-map, and -ffile-prefix-map (in both '--option value' and '--option=value'
forms), extracting all mappings from target paths (used in debug info) to host source
paths. If multiple mappings for the same target path are found, the most suitable
host path is selected (preferring 'sources' over 'build' directories).
"""
prefixes = ("-fdebug-prefix-map", "-fmacro-prefix-map", "-ffile-prefix-map")
all_mappings = {}
args = shlex.split(debug_prefix_map)
i = 0
# Collect all mappings, storing potentially multiple host paths per target path
while i < len(args):
arg = args[i]
mapping = None
for prefix in prefixes:
if arg == prefix:
i += 1
mapping = args[i]
break
elif arg.startswith(prefix + '='):
mapping = arg[len(prefix)+1:]
break
if mapping:
host_path, target_path = mapping.split('=', 1)
if target_path:
if target_path not in all_mappings:
all_mappings[target_path] = []
all_mappings[target_path].append(os.path.realpath(host_path))
i += 1
# Select the best host path for each target path (only 1:1 mappings are supported by GDB)
mappings = {}
unused_host_paths = []
for target_path, host_paths in all_mappings.items():
if len(host_paths) == 1:
mappings[target_path] = host_paths[0]
else:
# First priority path for sources is the source directory S
# Second priority path is any other directory
# Least priority is the build directory B, which probably contains only generated source files
sources_paths = [path for path in host_paths if os.path.realpath(path).startswith(os.path.realpath(self.s))]
if sources_paths:
mappings[target_path] = sources_paths[0]
unused_host_paths.extend([path for path in host_paths if path != sources_paths[0]])
else:
# If no 'sources' path, prefer non-'build' paths
non_build_paths = [path for path in host_paths if not os.path.realpath(path).startswith(os.path.realpath(self.b))]
if non_build_paths:
mappings[target_path] = non_build_paths[0]
unused_host_paths.extend([path for path in host_paths if path != non_build_paths[0]])
else:
# Fall back to first path if all are build paths
mappings[target_path] = host_paths[0]
unused_host_paths.extend(host_paths[1:])
if unused_host_paths:
logger.info("Some source directories mapped by -fdebug-prefix-map are not included in the debugger search paths. Ignored host paths: %s", unused_host_paths)
return mappings
@property
def gdb_pretty_print_scripts(self):
if self._gdb_pretty_print_scripts is None:
if self.tcoverride == "toolchain-gcc":
gcc_python_helpers_pattern = os.path.join(self.recipe_sysroot, "usr", "share", "gcc-*", "python")
gcc_python_helpers_dirs = glob.glob(gcc_python_helpers_pattern)
if gcc_python_helpers_dirs:
gcc_python_helpers = gcc_python_helpers_dirs[0]
else:
logger.warning("Could not find gcc python helpers directory matching: %s", gcc_python_helpers_pattern)
gcc_python_helpers = ""
pretty_print_scripts = [
"import sys",
"sys.path.insert(0, '" + gcc_python_helpers + "')",
"from libstdcxx.v6.printers import register_libstdcxx_printers",
"register_libstdcxx_printers(None)"
]
self._gdb_pretty_print_scripts = pretty_print_scripts
else:
self._gdb_pretty_print_scripts = ""
return self._gdb_pretty_print_scripts
def __init_exported_variables(self, d):
"""Find all variables with export flag set.
This allows to generate IDE configurations which compile with the same
environment as bitbake does. That's at least a reasonable default behavior.
"""
exported_vars = {}
vars = (key for key in d.keys() if not key.startswith(
"__") and not d.getVarFlag(key, "func", False))
for var in sorted(vars):
func = d.getVarFlag(var, "func", False)
if d.getVarFlag(var, 'python', False) and func:
continue
export = d.getVarFlag(var, "export", False)
unexport = d.getVarFlag(var, "unexport", False)
if not export and not unexport and not func:
continue
if unexport:
continue
val = d.getVar(var)
if val is None:
continue
if set(var) & set("-.{}+"):
logger.warn(
"Warning: Found invalid character in variable name %s", str(var))
continue
varExpanded = d.expand(var)
val = str(val)
if not RecipeModified.is_valid_shell_variable(varExpanded):
continue
if func:
code_line = "line: {0}, file: {1}\n".format(
d.getVarFlag(var, "lineno", False),
d.getVarFlag(var, "filename", False))
val = val.rstrip('\n')
logger.warn("Warning: exported shell function %s() is not exported (%s)" %
(varExpanded, code_line))
continue
if export:
exported_vars[varExpanded] = val.strip()
continue
self.exported_vars = exported_vars
def __init_systemd_services(self, d):
"""Find all systemd service files for the recipe."""
services = {}
if bb.data.inherits_class('systemd', d):
systemd_packages = d.getVar('SYSTEMD_PACKAGES')
if systemd_packages:
for package in systemd_packages.split():
services[package] = d.getVar('SYSTEMD_SERVICE:' + package).split()
self.systemd_services = services
def __init_init_scripts(self, d):
"""Find all SysV init scripts for the recipe."""
init_scripts = {}
if bb.data.inherits_class('update-rc.d', d):
script_packages = d.getVar('INITSCRIPT_PACKAGES')
if script_packages:
for package in script_packages.split():
initscript_name = d.getVar('INITSCRIPT_NAME:' + package)
if initscript_name:
# Handle both single script and multiple scripts
scripts = initscript_name.split()
if scripts:
init_scripts[package] = scripts
else:
# If INITSCRIPT_PACKAGES is not set, check for default INITSCRIPT_NAME
initscript_name = d.getVar('INITSCRIPT_NAME')
if initscript_name:
scripts = initscript_name.split()
if scripts:
# Use PN as the default package name when INITSCRIPT_PACKAGES is not set
pn = d.getVar('PN')
if pn:
init_scripts[pn] = scripts
self.init_scripts = init_scripts
def __init_cmake_preset_cache(self, d):
"""Get the arguments passed to cmake
Replicate the cmake configure arguments with all details to
share on build folder between bitbake and SDK.
"""
site_file = os.path.join(self.workdir, 'site-file.cmake')
if os.path.exists(site_file):
print("Warning: site-file.cmake is not supported")
cache_vars = {}
oecmake_args = d.getVar('OECMAKE_ARGS').split()
extra_oecmake = d.getVar('EXTRA_OECMAKE').split()
for param in sorted(oecmake_args + extra_oecmake):
d_pref = "-D"
if param.startswith(d_pref):
param = param[len(d_pref):]
else:
print("Error: expected a -D")
param_s = param.split('=', 1)
param_nt = param_s[0].split(':', 1)
def handle_undefined_variable(var):
if var.startswith('${') and var.endswith('}'):
return ''
else:
return var
# Example: FOO=ON
if len(param_nt) == 1:
cache_vars[param_s[0]] = handle_undefined_variable(param_s[1])
# Example: FOO:PATH=/tmp
elif len(param_nt) == 2:
cache_vars[param_nt[0]] = {
"type": param_nt[1],
"value": handle_undefined_variable(param_s[1]),
}
else:
print("Error: cannot parse %s" % param)
self.cmake_cache_vars = cache_vars
def cmake_preset(self):
"""Create a preset for cmake that mimics how bitbake calls cmake"""
toolchain_file = os.path.join(self.workdir, 'toolchain.cmake')
cmake_executable = os.path.join(
self.recipe_sysroot_native, 'usr', 'bin', 'cmake')
self.cmd_compile = cmake_executable + " --build --preset " + self.recipe_id
preset_dict_configure = {
"name": self.recipe_id,
"displayName": self.recipe_id_pretty,
"description": "Bitbake build environment for the recipe %s compiled for %s" % (self.bpn, self.package_arch),
"binaryDir": self.b,
"generator": self.oecmake_generator,
"toolchainFile": toolchain_file,
"cacheVariables": self.cmake_cache_vars,
"environment": self.exported_vars,
"cmakeExecutable": cmake_executable
}
preset_dict_build = {
"name": self.recipe_id,
"displayName": self.recipe_id_pretty,
"description": "Bitbake build environment for the recipe %s compiled for %s" % (self.bpn, self.package_arch),
"configurePreset": self.recipe_id,
"inheritConfigureEnvironment": True
}
preset_dict_test = {
"name": self.recipe_id,
"displayName": self.recipe_id_pretty,
"description": "Bitbake build environment for the recipe %s compiled for %s" % (self.bpn, self.package_arch),
"configurePreset": self.recipe_id,
"inheritConfigureEnvironment": True
}
preset_dict = {
"version": 3, # cmake 3.21, backward compatible with kirkstone
"configurePresets": [preset_dict_configure],
"buildPresets": [preset_dict_build],
"testPresets": [preset_dict_test]
}
# Finally write the json file
json_file = 'CMakeUserPresets.json'
json_path = os.path.join(self.real_srctree, json_file)
logger.info("Updating CMake preset: %s (%s)" % (json_file, json_path))
if not os.path.exists(self.real_srctree):
os.makedirs(self.real_srctree)
try:
with open(json_path) as f:
orig_dict = json.load(f)
except json.decoder.JSONDecodeError:
logger.info(
"Decoding %s failed. Probably because of comments in the json file" % json_path)
orig_dict = {}
except FileNotFoundError:
orig_dict = {}
# Add or update the presets for the recipe and keep other presets
for k, v in preset_dict.items():
if isinstance(v, list):
update_preset = v[0]
preset_added = False
if k in orig_dict:
for index, orig_preset in enumerate(orig_dict[k]):
if 'name' in orig_preset:
if orig_preset['name'] == update_preset['name']:
logger.debug("Updating preset: %s" %
orig_preset['name'])
orig_dict[k][index] = update_preset
preset_added = True
break
else:
logger.debug("keeping preset: %s" %
orig_preset['name'])
else:
logger.warn("preset without a name found")
if not preset_added:
if not k in orig_dict:
orig_dict[k] = []
orig_dict[k].append(update_preset)
logger.debug("Added preset: %s" %
update_preset['name'])
else:
orig_dict[k] = v
with open(json_path, 'w') as f:
json.dump(orig_dict, f, indent=4)
def gen_meson_wrapper(self):
"""Generate a wrapper script to call meson with the cross environment"""
bb.utils.mkdirhier(self.ide_sdk_scripts_dir)
meson_wrapper = os.path.join(self.ide_sdk_scripts_dir, 'meson')
meson_real = os.path.join(
self.recipe_sysroot_native, 'usr', 'bin', 'meson.real')
with open(meson_wrapper, 'w') as mwrap:
mwrap.write("#!/bin/sh" + os.linesep)
for var, val in self.exported_vars.items():
mwrap.write('export %s="%s"' % (var, val) + os.linesep)
mwrap.write("unset CC CXX CPP LD AR NM STRIP" + os.linesep)
private_temp = os.path.join(self.b, "meson-private", "tmp")
mwrap.write('mkdir -p "%s"' % private_temp + os.linesep)
mwrap.write('export TMPDIR="%s"' % private_temp + os.linesep)
mwrap.write('exec "%s" "$@"' % meson_real + os.linesep)
st = os.stat(meson_wrapper)
os.chmod(meson_wrapper, st.st_mode | stat.S_IEXEC)
self.meson_wrapper = meson_wrapper
self.cmd_compile = meson_wrapper + " compile -C " + self.b
def which(self, executable):
bin_path = shutil.which(executable, path=self.path)
if not bin_path:
raise DevtoolError(
'Cannot find %s. Probably the recipe %s is not built yet.' % (executable, self.bpn))
return bin_path
@staticmethod
def is_elf_file(file_path):
with open(file_path, "rb") as f:
data = f.read(4)
if data == b'\x7fELF':
return True
return False
@property
def installed_binaries(self):
"""find all executable elf files in the image directory"""
if self._installed_binaries:
return self._installed_binaries
binaries = {}
d_len = len(self.d)
re_so = re.compile(r'.*\.so[.0-9]*$')
for root, _, files in os.walk(self.d, followlinks=False):
for file in files:
if os.path.islink(file):
continue
if re_so.match(file):
continue
abs_name = os.path.join(root, file)
if os.access(abs_name, os.X_OK) and RecipeModified.is_elf_file(abs_name):
binary_path = abs_name[d_len:]
binary = ExecutableBinary(self.d, binary_path,
self.systemd_services, self.init_scripts)
binaries[binary_path] = binary
self._installed_binaries = dict(sorted(binaries.items()))
return self._installed_binaries
def gen_deploy_target_script(self, args):
"""Generate a script which does what devtool deploy-target does
This script is much quicker than devtool target-deploy. Because it
does not need to start a bitbake server. All information from tinfoil
is hard-coded in the generated script.
"""
cmd_lines = ['#!%s' % str(sys.executable)]
cmd_lines.append('import sys')
cmd_lines.append('devtool_sys_path = %s' % str(sys.path))
cmd_lines.append('devtool_sys_path.reverse()')
cmd_lines.append('for p in devtool_sys_path:')
cmd_lines.append(' if p not in sys.path:')
cmd_lines.append(' sys.path.insert(0, p)')
cmd_lines.append('from devtool.deploy import deploy_no_d')
args_filter = ['debug', 'dry_run', 'key', 'no_check_space', 'no_host_check',
'no_preserve', 'port', 'show_status', 'ssh_exec', 'strip', 'target']
filtered_args_dict = {key: value for key, value in vars(
args).items() if key in args_filter}
cmd_lines.append('filtered_args_dict = %s' % str(filtered_args_dict))
cmd_lines.append('class Dict2Class(object):')
cmd_lines.append(' def __init__(self, my_dict):')
cmd_lines.append(' for key in my_dict:')
cmd_lines.append(' setattr(self, key, my_dict[key])')
cmd_lines.append('filtered_args = Dict2Class(filtered_args_dict)')
cmd_lines.append('if len(sys.argv) > 2:')
cmd_lines.append(' if sys.argv[1] == "-t" or sys.argv[1] == "--target":')
cmd_lines.append(' setattr(filtered_args, "target", sys.argv[2])')
cmd_lines.append(
'setattr(filtered_args, "recipename", "%s")' % self.bpn)
cmd_lines.append('deploy_no_d("%s", "%s", "%s", "%s", "%s", "%s", %d, "%s", "%s", filtered_args)' %
(self.d, self.workdir, self.path, self.strip_cmd,
self.libdir, self.base_libdir, self.max_process,
self.fakerootcmd, self.fakerootenv))
return self.write_script(cmd_lines, 'deploy_target')
def gen_install_deploy_script(self, args):
"""Generate a script which does install and deploy"""
cmd_lines = ['#!/bin/sh']
# . oe-init-build-env $BUILDDIR $BITBAKEDIR
# Using 'set' to pass the build directory to oe-init-build-env in sh syntax
cmd_lines.append('cd "%s" || { echo "cd %s failed"; exit 1; }' % (
self.oe_init_dir, self.oe_init_dir))
cmd_lines.append('set %s %s' % (self.topdir, self.bitbakepath.rstrip('/bin')))
cmd_lines.append('. "%s" || { echo ". %s %s failed"; exit 1; }' % (
self.oe_init_build_env, self.oe_init_build_env, self.topdir))
# bitbake -c install
cmd_lines.append(
'bitbake %s -c install --force || { echo "bitbake %s -c install --force failed"; exit 1; }' % (self.bpn, self.bpn))
# Self contained devtool deploy-target
cmd_lines.append(self.gen_deploy_target_script(args) + ' "$@"')
return self.write_script(cmd_lines, 'install_and_deploy')
def write_script(self, cmd_lines, script_name):
bb.utils.mkdirhier(self.ide_sdk_scripts_dir)
script_name_arch = script_name + '_' + self.recipe_id
script_file = os.path.join(self.ide_sdk_scripts_dir, script_name_arch)
with open(script_file, 'w') as script_f:
script_f.write(os.linesep.join(cmd_lines))
st = os.stat(script_file)
os.chmod(script_file, st.st_mode | stat.S_IEXEC)
return script_file
@property
def oe_init_build_env(self):
"""Find the init-build-env used for this setup"""
# bitbake-setup mode
bb_setup_init = os.path.join(self.topdir, RecipeModified.INIT_BUILD_ENV)
if os.path.exists(bb_setup_init):
return os.path.abspath(bb_setup_init)
# poky mode
oe_init_dir = self.oe_init_dir
if oe_init_dir:
return os.path.join(oe_init_dir, RecipeModified.OE_INIT_BUILD_ENV)
return None
@property
def oe_init_dir(self):
"""Find the directory where the oe-init-build-env is located
Assumption: There might be a layer with higher priority than poky
which provides to oe-init-build-env in the layer's toplevel folder.
"""
if not self.__oe_init_dir:
for layer in reversed(self.bblayers):
result = subprocess.run(
['git', 'rev-parse', '--show-toplevel'], cwd=layer, capture_output=True)
if result.returncode == 0:
oe_init_dir = result.stdout.decode('utf-8').strip()
oe_init_path = os.path.join(
oe_init_dir, RecipeModified.OE_INIT_BUILD_ENV)
if os.path.exists(oe_init_path):
logger.debug("Using %s from: %s" % (
RecipeModified.OE_INIT_BUILD_ENV, oe_init_path))
self.__oe_init_dir = oe_init_dir
break
if not self.__oe_init_dir:
logger.error("Cannot find the bitbake top level folder")
return self.__oe_init_dir
def ide_setup(args, config, basepath, workspace):
"""Generate the IDE configuration for the workspace"""
# Explicitely passing some special recipes does not make sense
for recipe in args.recipenames:
if recipe in ['meta-ide-support', 'build-sysroots']:
raise DevtoolError("Invalid recipe: %s." % recipe)
# Collect information about tasks which need to be bitbaked
bootstrap_tasks = []
bootstrap_tasks_late = []
tinfoil = setup_tinfoil(config_only=False, basepath=basepath)
try:
# define mode depending on recipes which need to be processed
recipes_image_names = []
recipes_modified_names = []
recipes_other_names = []
for recipe in args.recipenames:
try:
check_workspace_recipe(
workspace, recipe, bbclassextend=True)
recipes_modified_names.append(recipe)
except DevtoolError:
recipe_d = parse_recipe(
config, tinfoil, recipe, appends=True, filter_workspace=False)
if not recipe_d:
raise DevtoolError("Parsing recipe %s failed" % recipe)
if bb.data.inherits_class('image', recipe_d):
recipes_image_names.append(recipe)
else:
recipes_other_names.append(recipe)
invalid_params = False
if args.mode == DevtoolIdeMode.shared:
if len(recipes_modified_names):
logger.error("In shared sysroots mode modified recipes %s cannot be handled." % str(
recipes_modified_names))
invalid_params = True
if args.mode == DevtoolIdeMode.modified:
if len(recipes_other_names):
logger.error("Only in shared sysroots mode not modified recipes %s can be handled." % str(
recipes_other_names))
invalid_params = True
if len(recipes_image_names) != 1:
logger.error(
"One image recipe is required as the rootfs for the remote development.")
invalid_params = True
for modified_recipe_name in recipes_modified_names:
if modified_recipe_name.startswith('nativesdk-') or modified_recipe_name.endswith('-native'):
logger.error(
"Only cross compiled recipes are support. %s is not cross." % modified_recipe_name)
invalid_params = True
if invalid_params:
raise DevtoolError("Invalid parameters are passed.")
# For the shared sysroots mode, add all dependencies of all the images to the sysroots
# For the modified mode provide one rootfs and the corresponding debug symbols via rootfs-dbg
recipes_images = []
for recipes_image_name in recipes_image_names:
logger.info("Using image: %s" % recipes_image_name)
recipe_image = RecipeImage(recipes_image_name)
recipe_image.initialize(config, tinfoil)
bootstrap_tasks += recipe_image.bootstrap_tasks
recipes_images.append(recipe_image)
# Provide a Direct SDK with shared sysroots
recipes_not_modified = []
if args.mode == DevtoolIdeMode.shared:
ide_support = RecipeMetaIdeSupport()
ide_support.initialize(config, tinfoil)
bootstrap_tasks += ide_support.bootstrap_tasks
logger.info("Adding %s to the Direct SDK sysroots." %
str(recipes_other_names))
for recipe_name in recipes_other_names:
recipe_not_modified = RecipeNotModified(recipe_name)
bootstrap_tasks += recipe_not_modified.bootstrap_tasks
recipes_not_modified.append(recipe_not_modified)
build_sysroots = RecipeBuildSysroots()
build_sysroots.initialize(config, tinfoil)
bootstrap_tasks_late += build_sysroots.bootstrap_tasks
shared_env = SharedSysrootsEnv()
shared_env.initialize(ide_support, build_sysroots)
recipes_modified = []
if args.mode == DevtoolIdeMode.modified:
logger.info("Setting up workspaces for modified recipe: %s" %
str(recipes_modified_names))
gdbs_cross = {}
for recipe_name in recipes_modified_names:
recipe_modified = RecipeModified(recipe_name)
recipe_modified.initialize(config, workspace, tinfoil)
bootstrap_tasks += recipe_modified.bootstrap_tasks
recipes_modified.append(recipe_modified)
if recipe_modified.target_arch not in gdbs_cross:
target_device = TargetDevice(args)
gdb_cross = RecipeGdbCross(
args, recipe_modified.target_arch, target_device)
gdb_cross.initialize(config, workspace, tinfoil)
bootstrap_tasks += gdb_cross.bootstrap_tasks
gdbs_cross[recipe_modified.target_arch] = gdb_cross
recipe_modified.gdb_cross = gdbs_cross[recipe_modified.target_arch]
finally:
tinfoil.shutdown()
if not args.skip_bitbake:
bb_cmd = 'bitbake '
if args.bitbake_k:
bb_cmd += "-k "
bb_cmd_early = bb_cmd + ' '.join(bootstrap_tasks)
exec_build_env_command(
config.init_path, basepath, bb_cmd_early, watch=True)
if bootstrap_tasks_late:
bb_cmd_late = bb_cmd + ' '.join(bootstrap_tasks_late)
exec_build_env_command(
config.init_path, basepath, bb_cmd_late, watch=True)
wants_gdbserver = any(
r.wants_gdbserver for r in recipes_modified)
for recipe_image in recipes_images:
if wants_gdbserver and recipe_image.gdbserver_missing:
logger.warning(
"gdbserver not installed in image %s. Remote debugging will not be available" % recipe_image)
if wants_gdbserver and recipe_image.combine_dbg_image is False:
logger.warning(
'IMAGE_CLASSES += "image-combined-dbg" is missing for image %s. Remote debugging will not find debug symbols from rootfs-dbg.' % recipe_image)
# Instantiate the active IDE plugin
ide = ide_plugins[args.ide]()
if args.mode == DevtoolIdeMode.shared:
ide.setup_shared_sysroots(shared_env)
elif args.mode == DevtoolIdeMode.modified:
for recipe_modified in recipes_modified:
if recipe_modified.build_tool is BuildTool.CMAKE:
recipe_modified.cmake_preset()
if recipe_modified.build_tool is BuildTool.MESON:
recipe_modified.gen_meson_wrapper()
ide.setup_modified_recipe(
args, recipe_image, recipe_modified)
if recipe_modified.wants_debug_build and recipe_modified.debug_build != '1':
logger.warn(
'Recipe %s is compiled with release build configuration. '
'You might want to add DEBUG_BUILD = "1" to %s. '
'Note that devtool modify --debug-build can do this automatically.',
recipe_modified.name, recipe_modified.bbappend)
else:
raise DevtoolError("Must not end up here.")
def register_commands(subparsers, context):
"""Register devtool subcommands from this plugin"""
# The ide-sdk command bootstraps the SDK from the bitbake environment before the IDE
# configuration is generated. In the case of the eSDK, the bootstrapping is performed
# during the installation of the eSDK installer. Running the ide-sdk plugin from an
# eSDK installer-based setup would require skipping the bootstrapping and probably
# taking some other differences into account when generating the IDE configurations.
# This would be possible. But it is not implemented.
if context.fixed_setup:
return
global ide_plugins
# Search for IDE plugins in all sub-folders named ide_plugins where devtool seraches for plugins.
pluginpaths = [os.path.join(path, 'ide_plugins')
for path in context.pluginpaths]
ide_plugin_modules = []
for pluginpath in pluginpaths:
scriptutils.load_plugins(logger, ide_plugin_modules, pluginpath)
for ide_plugin_module in ide_plugin_modules:
if hasattr(ide_plugin_module, 'register_ide_plugin'):
ide_plugin_module.register_ide_plugin(ide_plugins)
# Sort plugins according to their priority. The first entry is the default IDE plugin.
ide_plugins = dict(sorted(ide_plugins.items(),
key=lambda p: p[1].ide_plugin_priority(), reverse=True))
parser_ide_sdk = subparsers.add_parser('ide-sdk', group='working', order=50, formatter_class=RawTextHelpFormatter,
help='Setup the SDK and configure the IDE')
parser_ide_sdk.add_argument(
'recipenames', nargs='+', help='Generate an IDE configuration suitable to work on the given recipes.\n'
'Depending on the --mode parameter different types of SDKs and IDE configurations are generated.')
parser_ide_sdk.add_argument(
'-m', '--mode', type=DevtoolIdeMode, default=DevtoolIdeMode.modified,
help='Different SDK types are supported:\n'
'- "' + DevtoolIdeMode.modified.name + '" (default):\n'
' devtool modify creates a workspace to work on the source code of a recipe.\n'
' devtool ide-sdk builds the SDK and generates the IDE configuration(s) in the workspace directorie(s)\n'
' Usage example:\n'
' devtool modify cmake-example\n'
' devtool ide-sdk cmake-example core-image-minimal\n'
' Start the IDE in the workspace folder\n'
' At least one devtool modified recipe plus one image recipe are required:\n'
' The image recipe is used to generate the target image and the remote debug configuration.\n'
'- "' + DevtoolIdeMode.shared.name + '":\n'
' Usage example:\n'
' devtool ide-sdk -m ' + DevtoolIdeMode.shared.name + ' recipe(s)\n'
' This command generates a cross-toolchain as well as the corresponding shared sysroot directories.\n'
' To use this tool-chain the environment-* file found in the deploy..image folder needs to be sourced into a shell.\n'
' In case of VSCode and cmake the tool-chain is also exposed as a cmake-kit')
default_ide = list(ide_plugins.keys())[0]
parser_ide_sdk.add_argument(
'-i', '--ide', choices=ide_plugins.keys(), default=default_ide,
help='Setup the configuration for this IDE (default: %s)' % default_ide)
parser_ide_sdk.add_argument(
'-t', '--target', default='root@192.168.7.2',
help='Live target machine running an ssh server: user@hostname.')
parser_ide_sdk.add_argument(
'-G', '--gdbserver-port-start', default="1234", help='port where gdbserver is listening.')
parser_ide_sdk.add_argument(
'-c', '--no-host-check', help='Disable ssh host key checking', action='store_true')
parser_ide_sdk.add_argument(
'-e', '--ssh-exec', help='Executable to use in place of ssh')
parser_ide_sdk.add_argument(
'-P', '--port', help='Specify ssh port to use for connection to the target')
parser_ide_sdk.add_argument(
'-I', '--key', help='Specify ssh private key for connection to the target')
parser_ide_sdk.add_argument(
'--skip-bitbake', help='Generate IDE configuration but skip calling bitbake to update the SDK', action='store_true')
parser_ide_sdk.add_argument(
'-k', '--bitbake-k', help='Pass -k parameter to bitbake', action='store_true')
parser_ide_sdk.add_argument(
'--no-strip', help='Do not strip executables prior to deploy', dest='strip', action='store_false')
parser_ide_sdk.add_argument(
'-n', '--dry-run', help='List files to be undeployed only', action='store_true')
parser_ide_sdk.add_argument(
'-s', '--show-status', help='Show progress/status output', action='store_true')
parser_ide_sdk.add_argument(
'-p', '--no-preserve', help='Do not preserve existing files', action='store_true')
parser_ide_sdk.add_argument(
'--no-check-space', help='Do not check for available space before deploying', action='store_true')
parser_ide_sdk.set_defaults(func=ide_setup)