""" Contains definitions for creation of external C/C++ build rules (for building external libraries with CMake, configure/make, autotools) """ load("@bazel_skylib//lib:collections.bzl", "collections") load("@bazel_tools//tools/cpp:toolchain_utils.bzl", "find_cpp_toolchain") load("//foreign_cc:providers.bzl", "ForeignCcArtifact", "ForeignCcDeps") load("//foreign_cc/private:detect_root.bzl", "detect_root", "filter_containing_dirs_from_inputs") load( "//foreign_cc/private/framework:helpers.bzl", "convert_shell_script", "create_function", "script_extension", "shebang", ) load("//foreign_cc/private/framework:platform.bzl", "os_name") load( "//toolchains/native_tools:tool_access.bzl", "get_make_data", "get_ninja_data", ) load( ":cc_toolchain_util.bzl", "LibrariesToLinkInfo", "create_linking_info", "get_env_vars", "targets_windows", ) load( ":run_shell_file_utils.bzl", "copy_directory", "fictive_file_in_genroot", ) # Dict with definitions of the context attributes, that customize cc_external_rule_impl function. # Many of the attributes have default values. # # Typically, the concrete external library rule will use this structure to create the attributes # description dict. See cmake.bzl as an example. # CC_EXTERNAL_RULE_ATTRIBUTES = { "additional_inputs": attr.label_list( doc = ( "Optional additional inputs to be declared as needed for the shell script action." + "Not used by the shell script part in cc_external_rule_impl." ), mandatory = False, allow_files = True, default = [], ), "additional_tools": attr.label_list( doc = ( "Optional additional tools needed for the building. " + "Not used by the shell script part in cc_external_rule_impl." ), mandatory = False, allow_files = True, cfg = "exec", default = [], ), "alwayslink": attr.bool( doc = ( "Optional. if true, link all the object files from the static library, " + "even if they are not used." ), mandatory = False, default = False, ), "data": attr.label_list( doc = "Files needed by this rule at runtime. May list file or rule targets. Generally allows any target.", mandatory = False, allow_files = True, default = [], ), "defines": attr.string_list( doc = ( "Optional compilation definitions to be passed to the dependencies of this library. " + "They are NOT passed to the compiler, you should duplicate them in the configuration options." ), mandatory = False, default = [], ), "deps": attr.label_list( doc = ( "Optional dependencies to be copied into the directory structure. " + "Typically those directly required for the external building of the library/binaries. " + "(i.e. those that the external buidl system will be looking for and paths to which are " + "provided by the calling rule)" ), mandatory = False, allow_files = True, default = [], ), "env": attr.string_dict( doc = ( "Environment variables to set during the build. " + "`$(execpath)` macros may be used to point at files which are listed as data deps, tools_deps, or additional_tools, " + "but unlike with other rules, these will be replaced with absolute paths to those files, " + "because the build does not run in the exec root. " + "No other macros are supported." ), ), "lib_name": attr.string( doc = ( "Library name. Defines the name of the install directory and the name of the static library, " + "if no output files parameters are defined (any of static_libraries, shared_libraries, " + "interface_libraries, binaries_names) " + "Optional. If not defined, defaults to the target's name." ), mandatory = False, ), "lib_source": attr.label( doc = ( "Label with source code to build. Typically a filegroup for the source of remote repository. " + "Mandatory." ), mandatory = True, allow_files = True, ), "linkopts": attr.string_list( doc = "Optional link options to be passed up to the dependencies of this library", mandatory = False, default = [], ), "make_commands": attr.string_list( doc = "Optional make commands.", mandatory = False, default = ["make", "make install"], ), "out_bin_dir": attr.string( doc = "Optional name of the output subdirectory with the binary files, defaults to 'bin'.", mandatory = False, default = "bin", ), "out_binaries": attr.string_list( doc = "Optional names of the resulting binaries.", mandatory = False, ), "out_headers_only": attr.bool( doc = "Flag variable to indicate that the library produces only headers", mandatory = False, default = False, ), "out_include_dir": attr.string( doc = "Optional name of the output subdirectory with the header files, defaults to 'include'.", mandatory = False, default = "include", ), "out_interface_libs": attr.string_list( doc = "Optional names of the resulting interface libraries.", mandatory = False, ), "out_lib_dir": attr.string( doc = "Optional name of the output subdirectory with the library files, defaults to 'lib'.", mandatory = False, default = "lib", ), "out_shared_libs": attr.string_list( doc = "Optional names of the resulting shared libraries.", mandatory = False, ), "out_static_libs": attr.string_list( doc = ( "Optional names of the resulting static libraries. Note that if `out_headers_only`, `out_static_libs`, " + "`out_shared_libs`, and `out_binaries` are not set, default `lib_name.a`/`lib_name.lib` static " + "library is assumed" ), mandatory = False, ), "postfix_script": attr.string( doc = "Optional part of the shell script to be added after the make commands", mandatory = False, ), "targets": attr.string_list( doc = ( "A list of targets with in the foreign build system to produce. An empty string (`\"\"`) will result in " + "a call to the underlying build system with no explicit target set" ), mandatory = False, ), "tools_deps": attr.label_list( doc = ( "Optional tools to be copied into the directory structure. " + "Similar to deps, those directly required for the external building of the library/binaries." ), mandatory = False, allow_files = True, cfg = "exec", default = [], ), # we need to declare this attribute to access cc_toolchain "_cc_toolchain": attr.label( default = Label("@bazel_tools//tools/cpp:current_cc_toolchain"), ), "_foreign_cc_framework_platform": attr.label( doc = "Information about the execution platform", cfg = "exec", default = Label("@rules_foreign_cc//foreign_cc/private/framework:platform_info"), ), } # A list of common fragments required by rules using this framework CC_EXTERNAL_RULE_FRAGMENTS = [ "cpp", ] # buildifier: disable=function-docstring-header # buildifier: disable=function-docstring-args # buildifier: disable=function-docstring-return def create_attrs(attr_struct, configure_name, create_configure_script, **kwargs): """Function for adding/modifying context attributes struct (originally from ctx.attr), provided by user, to be passed to the cc_external_rule_impl function as a struct. Copies a struct 'attr_struct' values (with attributes from CC_EXTERNAL_RULE_ATTRIBUTES) to the resulting struct, adding or replacing attributes passed in 'configure_name', 'configure_script', and '**kwargs' parameters. """ attrs = {} for key in CC_EXTERNAL_RULE_ATTRIBUTES: if not key.startswith("_") and hasattr(attr_struct, key): attrs[key] = getattr(attr_struct, key) attrs["configure_name"] = configure_name attrs["create_configure_script"] = create_configure_script for arg in kwargs: attrs[arg] = kwargs[arg] return struct(**attrs) # buildifier: disable=name-conventions ConfigureParameters = provider( doc = """Parameters of create_configure_script callback function, called by cc_external_rule_impl function. create_configure_script creates the configuration part of the script, and allows to reuse the inputs structure, created by the framework.""", fields = dict( ctx = "Rule context", attrs = """Attributes struct, created by create_attrs function above""", inputs = """InputFiles provider: summarized information on rule inputs, created by framework function, to be reused in script creator. Contains in particular merged compilation and linking dependencies.""", ), ) def cc_external_rule_impl(ctx, attrs): """Framework function for performing external C/C++ building. To be used to build external libraries or/and binaries with CMake, configure/make, autotools etc., and use results in Bazel. It is possible to use it to build a group of external libraries, that depend on each other or on Bazel library, and pass nessesary tools. Accepts the actual commands for build configuration/execution in attrs. Creates and runs a shell script, which: 1. prepares directory structure with sources, dependencies, and tools symlinked into subdirectories of the execroot directory. Adds tools into PATH. 2. defines the correct absolute paths in tools with the script paths, see 7 3. defines the following environment variables: EXT_BUILD_ROOT: execroot directory EXT_BUILD_DEPS: subdirectory of execroot, which contains the following subdirectories: For cmake_external built dependencies: symlinked install directories of the dependencies for Bazel built/imported dependencies: include - here the include directories are symlinked lib - here the library files are symlinked lib/pkgconfig - here the pkgconfig files are symlinked bin - here the tools are copied INSTALLDIR: subdirectory of the execroot (named by the lib_name), where the library/binary will be installed These variables should be used by the calling rule to refer to the created directory structure. 4. calls 'attrs.create_configure_script' 5. calls 'attrs.make_commands' 6. calls 'attrs.postfix_script' 7. replaces absolute paths in possibly created scripts with a placeholder value Please see cmake.bzl for example usage. Args: ctx: calling rule context attrs: attributes struct, created by create_attrs function above. Contains fields from CC_EXTERNAL_RULE_ATTRIBUTES (see descriptions there), two mandatory fields: - configure_name: name of the configuration tool, to be used in action mnemonic, - create_configure_script(ConfigureParameters): function that creates configuration script, accepts ConfigureParameters and some other fields provided by the rule, which have been passed to create_attrs. Returns: A list of providers """ lib_name = attrs.lib_name or ctx.attr.name inputs = _define_inputs(attrs) outputs = _define_outputs(ctx, attrs, lib_name) out_cc_info = _define_out_cc_info(ctx, attrs, inputs, outputs) cc_env = _correct_path_variable(get_env_vars(ctx)) set_cc_envs = [] execution_os_name = os_name(ctx) if execution_os_name != "macos": set_cc_envs = ["export {}=\"{}\"".format(key, cc_env[key]) for key in cc_env] lib_header = "Bazel external C/C++ Rules. Building library '{}'".format(lib_name) # We can not declare outputs of the action, which are in parent-child relashion, # so we need to have a (symlinked) copy of the output directory to provide # both the C/C++ artifacts - libraries, headers, and binaries, # and the install directory as a whole (which is mostly nessesary for chained external builds). # # We want the install directory output of this rule to have the same name as the library, # so symlink it under the same name but in a subdirectory installdir_copy = copy_directory(ctx.actions, "$$INSTALLDIR$$", "copy_{}/{}".format(lib_name, lib_name)) # we need this fictive file in the root to get the path of the root in the script empty = fictive_file_in_genroot(ctx.actions, ctx.label.name) data_dependencies = ctx.attr.data + ctx.attr.tools_deps + ctx.attr.additional_tools define_variables = set_cc_envs + [ "export EXT_BUILD_ROOT=##pwd##", "export INSTALLDIR=$$EXT_BUILD_ROOT$$/" + empty.file.dirname + "/" + lib_name, "export BUILD_TMPDIR=$$INSTALLDIR$$.build_tmpdir", "export EXT_BUILD_DEPS=$$INSTALLDIR$$.ext_build_deps", ] + [ "export {key}={value}".format( key = key, # Prepend the exec root to each $(execpath ) lookup because the working directory will not be the exec root. value = ctx.expand_location(value.replace("$(execpath ", "$$EXT_BUILD_ROOT$$/$(execpath "), data_dependencies), ) for key, value in dict( getattr(ctx.attr, "env", {}).items() + getattr(attrs, "env", {}).items(), ).items() ] make_commands, build_tools = _generate_make_commands(ctx) postfix_script = [attrs.postfix_script] if not attrs.postfix_script: postfix_script = [] script_lines = [ "##echo## \"\"", "##echo## \"{}\"".format(lib_header), "##echo## \"\"", "##script_prelude##", ] + define_variables + [ "##path## $$EXT_BUILD_ROOT$$", "##mkdirs## $$INSTALLDIR$$", "##mkdirs## $$BUILD_TMPDIR$$", "##mkdirs## $$EXT_BUILD_DEPS$$", ] + _print_env() + _copy_deps_and_tools(inputs) + [ "cd $$BUILD_TMPDIR$$", ] + attrs.create_configure_script(ConfigureParameters(ctx = ctx, attrs = attrs, inputs = inputs)) + make_commands + postfix_script + [ # replace references to the root directory when building ($BUILD_TMPDIR) # and the root where the dependencies were installed ($EXT_BUILD_DEPS) # for the results which are in $INSTALLDIR (with placeholder) "##replace_absolute_paths## $$INSTALLDIR$$ $$BUILD_TMPDIR$$", "##replace_absolute_paths## $$INSTALLDIR$$ $$EXT_BUILD_DEPS$$", installdir_copy.script, empty.script, "cd $$EXT_BUILD_ROOT$$", ] script_text = "\n".join([ shebang(ctx), convert_shell_script(ctx, script_lines), ]) wrapped_outputs = wrap_outputs(ctx, lib_name, attrs.configure_name, script_text) rule_outputs = outputs.declared_outputs + [installdir_copy.file] cc_toolchain = find_cpp_toolchain(ctx) execution_requirements = {"block-network": ""} if "requires-network" in ctx.attr.tags: execution_requirements = {"requires-network": ""} # The use of `run_shell` here is intended to ensure bash is correctly setup on windows # environments. This should not be replaced with `run` until a cross platform implementation # is found that guarantees bash exists or appropriately errors out. ctx.actions.run_shell( mnemonic = "Cc" + attrs.configure_name.capitalize() + "MakeRule", inputs = depset(inputs.declared_inputs), outputs = rule_outputs + [ empty.file, wrapped_outputs.log_file, ], tools = depset( [wrapped_outputs.script_file, wrapped_outputs.wrapper_script_file] + ctx.files.data + ctx.files.tools_deps + ctx.files.additional_tools, transitive = [cc_toolchain.all_files] + [data[DefaultInfo].default_runfiles.files for data in data_dependencies] + build_tools, ), # TODO: Default to never using the default shell environment to make builds more hermetic. For now, every platform # but MacOS will take the default PATH passed by Bazel, not that from cc_toolchain. use_default_shell_env = execution_os_name != "macos", command = wrapped_outputs.wrapper_script_file.path, execution_requirements = execution_requirements, # this is ignored if use_default_shell_env = True env = cc_env, ) # Gather runfiles transitively as per the documentation in: # https://docs.bazel.build/versions/master/skylark/rules.html#runfiles runfiles = ctx.runfiles(files = ctx.files.data) for target in [ctx.attr.lib_source] + ctx.attr.additional_inputs + ctx.attr.deps + ctx.attr.data: runfiles = runfiles.merge(target[DefaultInfo].default_runfiles) externally_built = ForeignCcArtifact( gen_dir = installdir_copy.file, bin_dir_name = attrs.out_bin_dir, lib_dir_name = attrs.out_lib_dir, include_dir_name = attrs.out_include_dir, ) output_groups = _declare_output_groups(installdir_copy.file, outputs.out_binary_files) wrapped_files = [ wrapped_outputs.script_file, wrapped_outputs.log_file, wrapped_outputs.wrapper_script_file, ] output_groups[attrs.configure_name + "_logs"] = wrapped_files return [ DefaultInfo( files = depset(direct = rule_outputs), runfiles = runfiles, ), OutputGroupInfo(**output_groups), ForeignCcDeps(artifacts = depset( [externally_built], transitive = _get_transitive_artifacts(attrs.deps), )), CcInfo( compilation_context = out_cc_info.compilation_context, linking_context = out_cc_info.linking_context, ), ] # buildifier: disable=name-conventions WrappedOutputs = provider( doc = "Structure for passing the log and scripts file information, and wrapper script text.", fields = { "log_file": "Execution log file", "script_file": "Main script file", "wrapper_script": "Wrapper script text to execute", "wrapper_script_file": "Wrapper script file (output for debugging purposes)", }, ) # buildifier: disable=function-docstring def wrap_outputs(ctx, lib_name, configure_name, script_text, build_script_file = None): extension = script_extension(ctx) build_log_file = ctx.actions.declare_file("{}_foreign_cc/{}.log".format(lib_name, configure_name)) build_script_file = ctx.actions.declare_file("{}_foreign_cc/build_script{}".format(lib_name, extension)) wrapper_script_file = ctx.actions.declare_file("{}_foreign_cc/wrapper_build_script{}".format(lib_name, extension)) ctx.actions.write( output = build_script_file, content = script_text, is_executable = True, ) cleanup_on_success_function = create_function( ctx, "cleanup_on_success", "rm -rf $BUILD_TMPDIR $EXT_BUILD_DEPS", ) cleanup_on_failure_function = create_function( ctx, "cleanup_on_failure", "\n".join([ "##echo## \"rules_foreign_cc: Build failed!\"", "##echo## \"rules_foreign_cc: Keeping temp build directory $$BUILD_TMPDIR$$ and dependencies directory $$EXT_BUILD_DEPS$$ for debug.\"", "##echo## \"rules_foreign_cc: Please note that the directories inside a sandbox are still cleaned unless you specify '--sandbox_debug' Bazel command line flag.\"", "##echo## \"rules_foreign_cc: Printing build logs:\"", "##echo## \"_____ BEGIN BUILD LOGS _____\"", "##cat## $$BUILD_LOG$$", "##echo## \"_____ END BUILD LOGS _____\"", "##echo## \"rules_foreign_cc: Build wrapper script location: $$BUILD_WRAPPER_SCRIPT$$\"", "##echo## \"rules_foreign_cc: Build script location: $$BUILD_SCRIPT$$\"", "##echo## \"rules_foreign_cc: Build log location: $$BUILD_LOG$$\"", "##echo## \"\"", ]), ) trap_function = "##cleanup_function## cleanup_on_success cleanup_on_failure" build_command_lines = [ "##assert_script_errors##", cleanup_on_success_function, cleanup_on_failure_function, # the call trap is defined inside, in a way how the shell function should be called # see, for instance, linux_commands.bzl trap_function, "export BUILD_WRAPPER_SCRIPT=\"{}\"".format(wrapper_script_file.path), "export BUILD_SCRIPT=\"{}\"".format(build_script_file.path), "export BUILD_LOG=\"{}\"".format(build_log_file.path), # sometimes the log file is not created, we do not want our script to fail because of this "##touch## $$BUILD_LOG$$", "##redirect_out_err## $$BUILD_SCRIPT$$ $$BUILD_LOG$$", ] build_command = "\n".join([ shebang(ctx), convert_shell_script(ctx, build_command_lines), "", ]) ctx.actions.write( output = wrapper_script_file, content = build_command, is_executable = True, ) return WrappedOutputs( script_file = build_script_file, log_file = build_log_file, wrapper_script_file = wrapper_script_file, wrapper_script = build_command, ) def _declare_output_groups(installdir, outputs): dict_ = {} dict_["gen_dir"] = depset([installdir]) for output in outputs: dict_[output.basename] = [output] return dict_ def _get_transitive_artifacts(deps): artifacts = [] for dep in deps: foreign_dep = get_foreign_cc_dep(dep) if foreign_dep: artifacts.append(foreign_dep.artifacts) return artifacts def _print_env(): return [ "##echo## \"Environment:______________\"", "##env##", "##echo## \"__________________________\"", ] def _correct_path_variable(env): value = env.get("PATH", "") if not value: return env value = env.get("PATH", "").replace("C:\\", "/c/") value = value.replace("\\", "/") value = value.replace(";", ":") env["PATH"] = "$PATH:" + value return env def _depset(item): if item == None: return depset() return depset([item]) def _list(item): if item: return [item] return [] def _copy_deps_and_tools(files): lines = [] lines += _symlink_contents_to_dir("lib", files.libs) lines += _symlink_contents_to_dir("include", files.headers + files.include_dirs) if files.tools_files: lines.append("##mkdirs## $$EXT_BUILD_DEPS$$/bin") for tool in files.tools_files: lines.append("##symlink_to_dir## $$EXT_BUILD_ROOT$$/{} $$EXT_BUILD_DEPS$$/bin/".format(tool)) for ext_dir in files.ext_build_dirs: lines.append("##symlink_to_dir## $$EXT_BUILD_ROOT$$/{} $$EXT_BUILD_DEPS$$".format(_file_path(ext_dir))) lines.append("##children_to_path## $$EXT_BUILD_DEPS$$/bin") lines.append("##path## $$EXT_BUILD_DEPS$$/bin") return lines def _symlink_contents_to_dir(dir_name, files_list): # It is possible that some duplicate libraries will be passed as inputs # to cmake_external or configure_make. Filter duplicates out here. files_list = collections.uniq(files_list) if len(files_list) == 0: return [] lines = ["##mkdirs## $$EXT_BUILD_DEPS$$/" + dir_name] for file in files_list: path = _file_path(file).strip() if path: lines.append("##symlink_contents_to_dir## \ $$EXT_BUILD_ROOT$$/{} $$EXT_BUILD_DEPS$$/{}".format(path, dir_name)) return lines def _file_path(file): return file if type(file) == "string" else file.path _FORBIDDEN_FOR_FILENAME = ["\\", "/", ":", "*", "\"", "<", ">", "|"] def _check_file_name(var): if (len(var) == 0): fail("Library name cannot be an empty string.") for index in range(0, len(var) - 1): letter = var[index] if letter in _FORBIDDEN_FOR_FILENAME: fail("Symbol '%s' is forbidden in library name '%s'." % (letter, var)) # buildifier: disable=name-conventions _Outputs = provider( doc = "Provider to keep different kinds of the external build output files and directories", fields = dict( out_include_dir = "Directory with header files (relative to install directory)", out_binary_files = "Binary files, which will be created by the action", libraries = "Library files, which will be created by the action", declared_outputs = "All output files and directories of the action", ), ) def _define_outputs(ctx, attrs, lib_name): attr_binaries_libs = [] attr_headers_only = attrs.out_headers_only attr_interface_libs = [] attr_shared_libs = [] attr_static_libs = [] # TODO: Until the the deprecated attributes are removed, we must # create a mutatable list so we can ensure they're being included attr_binaries_libs.extend(getattr(attrs, "out_binaries", [])) attr_interface_libs.extend(getattr(attrs, "out_interface_libs", [])) attr_shared_libs.extend(getattr(attrs, "out_shared_libs", [])) attr_static_libs.extend(getattr(attrs, "out_static_libs", [])) # TODO: These names are deprecated, remove if getattr(attrs, "binaries", []): # buildifier: disable=print print("The `binaries` attr is deprecated in favor of `out_binaries`. Please update the target `{}`".format(ctx.label)) attr_binaries_libs.extend(getattr(attrs, "binaries", [])) if getattr(attrs, "headers_only", False): # buildifier: disable=print print("The `headers_only` attr is deprecated in favor of `out_headers_only`. Please update the target `{}`".format(ctx.label)) attr_headers_only = attrs.headers_only if getattr(attrs, "interface_libraries", []): # buildifier: disable=print print("The `interface_libraries` attr is deprecated in favor of `out_interface_libs`. Please update the target `{}`".format(ctx.label)) attr_interface_libs.extend(getattr(attrs, "interface_libraries", [])) if getattr(attrs, "shared_libraries", []): # buildifier: disable=print print("The `shared_libraries` attr is deprecated in favor of `out_shared_libs`. Please update the target `{}`".format(ctx.label)) attr_shared_libs.extend(getattr(attrs, "shared_libraries", [])) if getattr(attrs, "static_libraries", []): # buildifier: disable=print print("The `static_libraries` attr is deprecated in favor of `out_static_libs`. Please update the target `{}`".format(ctx.label)) attr_static_libs.extend(getattr(attrs, "static_libraries", [])) static_libraries = [] if not attr_headers_only: if not attr_static_libs and not attr_shared_libs and not attr_binaries_libs and not attr_interface_libs: static_libraries = [lib_name + (".lib" if targets_windows(ctx, None) else ".a")] else: static_libraries = attr_static_libs _check_file_name(lib_name) out_include_dir = ctx.actions.declare_directory(lib_name + "/" + attrs.out_include_dir) out_binary_files = _declare_out(ctx, lib_name, attrs.out_bin_dir, attr_binaries_libs) libraries = LibrariesToLinkInfo( static_libraries = _declare_out(ctx, lib_name, attrs.out_lib_dir, static_libraries), shared_libraries = _declare_out(ctx, lib_name, attrs.out_lib_dir, attr_shared_libs), interface_libraries = _declare_out(ctx, lib_name, attrs.out_lib_dir, attr_interface_libs), ) declared_outputs = [out_include_dir] + out_binary_files declared_outputs += libraries.static_libraries declared_outputs += libraries.shared_libraries + libraries.interface_libraries return _Outputs( out_include_dir = out_include_dir, out_binary_files = out_binary_files, libraries = libraries, declared_outputs = declared_outputs, ) def _declare_out(ctx, lib_name, dir_, files): if files and len(files) > 0: return [ctx.actions.declare_file("/".join([lib_name, dir_, file])) for file in files] return [] # buildifier: disable=name-conventions InputFiles = provider( doc = ( "Provider to keep different kinds of input files, directories, " + "and C/C++ compilation and linking info from dependencies" ), fields = dict( headers = "Include files built by Bazel. Will be copied into $EXT_BUILD_DEPS/include.", include_dirs = ( "Include directories built by Bazel. Will be copied " + "into $EXT_BUILD_DEPS/include." ), libs = "Library files built by Bazel. Will be copied into $EXT_BUILD_DEPS/lib.", tools_files = ( "Files and directories with tools needed for configuration/building " + "to be copied into the bin folder, which is added to the PATH" ), ext_build_dirs = ( "Directories with libraries, built by framework function. " + "This directories should be copied into $EXT_BUILD_DEPS/lib-name as is, with all contents." ), deps_compilation_info = "Merged CcCompilationInfo from deps attribute", deps_linking_info = "Merged CcLinkingInfo from deps attribute", declared_inputs = "All files and directories that must be declared as action inputs", ), ) def _define_inputs(attrs): cc_infos = [] bazel_headers = [] bazel_system_includes = [] bazel_libs = [] # This framework function-built libraries: copy result directories under # $EXT_BUILD_DEPS/lib-name ext_build_dirs = [] for dep in attrs.deps: external_deps = get_foreign_cc_dep(dep) cc_infos.append(dep[CcInfo]) if external_deps: ext_build_dirs += [artifact.gen_dir for artifact in external_deps.artifacts.to_list()] else: headers_info = _get_headers(dep[CcInfo].compilation_context) bazel_headers += headers_info.headers bazel_system_includes += headers_info.include_dirs bazel_libs += _collect_libs(dep[CcInfo].linking_context) # Keep the order of the transitive foreign dependencies # (the order is important for the correct linking), # but filter out repeating directories ext_build_dirs = uniq_list_keep_order(ext_build_dirs) tools_roots = [] tools_files = [] input_files = [] for tool in attrs.tools_deps: tool_root = detect_root(tool) tools_roots.append(tool_root) for file_list in tool.files.to_list(): tools_files += _list(file_list) for tool in attrs.additional_tools: for file_list in tool.files.to_list(): tools_files += _list(file_list) for input in attrs.additional_inputs: for file_list in input.files.to_list(): input_files += _list(file_list) # These variables are needed for correct C/C++ providers constraction, # they should contain all libraries and include directories. cc_info_merged = cc_common.merge_cc_infos(cc_infos = cc_infos) return InputFiles( headers = bazel_headers, include_dirs = bazel_system_includes, libs = bazel_libs, tools_files = tools_roots, deps_compilation_info = cc_info_merged.compilation_context, deps_linking_info = cc_info_merged.linking_context, ext_build_dirs = ext_build_dirs, declared_inputs = filter_containing_dirs_from_inputs(attrs.lib_source.files.to_list()) + bazel_libs + tools_files + input_files + cc_info_merged.compilation_context.headers.to_list() + ext_build_dirs, ) # buildifier: disable=function-docstring def uniq_list_keep_order(list): result = [] contains_map = {} for item in list: if contains_map.get(item): continue contains_map[item] = 1 result.append(item) return result def get_foreign_cc_dep(dep): return dep[ForeignCcDeps] if ForeignCcDeps in dep else None # consider optimization here to do not iterate both collections def _get_headers(compilation_info): include_dirs = compilation_info.system_includes.to_list() + \ compilation_info.includes.to_list() # do not use quote includes, currently they do not contain # library-specific information include_dirs = collections.uniq(include_dirs) headers = [] for header in compilation_info.headers.to_list(): path = header.path included = False for dir_ in include_dirs: if path.startswith(dir_): included = True break if not included: headers.append(header) return struct( headers = headers, include_dirs = include_dirs, ) def _define_out_cc_info(ctx, attrs, inputs, outputs): compilation_info = cc_common.create_compilation_context( headers = depset([outputs.out_include_dir]), system_includes = depset([outputs.out_include_dir.path]), includes = depset([]), quote_includes = depset([]), defines = depset(attrs.defines), ) linking_info = create_linking_info(ctx, attrs.linkopts, outputs.libraries) cc_info = CcInfo( compilation_context = compilation_info, linking_context = linking_info, ) inputs_info = CcInfo( compilation_context = inputs.deps_compilation_info, linking_context = inputs.deps_linking_info, ) return cc_common.merge_cc_infos(cc_infos = [cc_info, inputs_info]) def _extract_libraries(library_to_link): return [ library_to_link.static_library, library_to_link.pic_static_library, library_to_link.dynamic_library, library_to_link.interface_library, ] def _collect_libs(cc_linking): libs = [] for li in cc_linking.linker_inputs.to_list(): for library_to_link in li.libraries: for library in _extract_libraries(library_to_link): if library: libs.append(library) return collections.uniq(libs) def _generate_make_commands(ctx): make_commands = getattr(ctx.attr, "make_commands", []) tools_deps = [] # Early out if there are no commands set if not make_commands: return make_commands, tools_deps if _uses_tool(ctx.attr.make_commands, "make"): make_data = get_make_data(ctx) tools_deps += make_data.deps make_commands = [command.replace("make", make_data.path) for command in make_commands] if _uses_tool(ctx.attr.make_commands, "ninja"): ninja_data = get_ninja_data(ctx) tools_deps += ninja_data.deps make_commands = [command.replace("ninja", ninja_data.path) for command in make_commands] return make_commands, [tool.files for tool in tools_deps] def _uses_tool(make_commands, tool): for command in make_commands: (before, separator, after) = command.partition(" ") if before == tool: return True return False