mirror of
https://github.com/bazelbuild/bazel-skylib
synced 2024-11-28 08:43:51 +00:00
505 lines
17 KiB
Python
505 lines
17 KiB
Python
# Copyright 2017 The Bazel Authors. All rights reserved.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
|
|
"""Unit testing support.
|
|
|
|
Unlike most Skylib files, this exports two modules: `unittest` which contains
|
|
functions to declare and define unit tests, and `asserts` which contains the
|
|
assertions used to within tests.
|
|
"""
|
|
|
|
load(":new_sets.bzl", new_sets = "sets")
|
|
load(":old_sets.bzl", "sets")
|
|
load(":types.bzl", "types")
|
|
|
|
# The following function should only be called from WORKSPACE files and workspace macros.
|
|
def register_unittest_toolchains():
|
|
"""Registers the toolchains for unittest users."""
|
|
native.register_toolchains(
|
|
"@bazel_skylib//toolchains/unittest:cmd_toolchain",
|
|
"@bazel_skylib//toolchains/unittest:bash_toolchain",
|
|
)
|
|
|
|
TOOLCHAIN_TYPE = "@bazel_skylib//toolchains/unittest:toolchain_type"
|
|
|
|
_UnittestToolchainInfo = provider(
|
|
doc = "Execution platform information for rules in the bazel_skylib repository.",
|
|
fields = ["file_ext", "success_templ", "failure_templ", "join_on"],
|
|
)
|
|
|
|
def _unittest_toolchain_impl(ctx):
|
|
return [
|
|
platform_common.ToolchainInfo(
|
|
unittest_toolchain_info = _UnittestToolchainInfo(
|
|
file_ext = ctx.attr.file_ext,
|
|
success_templ = ctx.attr.success_templ,
|
|
failure_templ = ctx.attr.failure_templ,
|
|
join_on = ctx.attr.join_on,
|
|
),
|
|
),
|
|
]
|
|
|
|
unittest_toolchain = rule(
|
|
implementation = _unittest_toolchain_impl,
|
|
attrs = {
|
|
"failure_templ": attr.string(mandatory = True),
|
|
"file_ext": attr.string(mandatory = True),
|
|
"join_on": attr.string(mandatory = True),
|
|
"success_templ": attr.string(mandatory = True),
|
|
},
|
|
)
|
|
|
|
def _impl_function_name(impl):
|
|
"""Derives the name of the given rule implementation function.
|
|
|
|
This can be used for better test feedback.
|
|
|
|
Args:
|
|
impl: the rule implementation function
|
|
|
|
Returns:
|
|
The name of the given function
|
|
"""
|
|
|
|
# Starlark currently stringifies a function as "<function NAME>", so we use
|
|
# that knowledge to parse the "NAME" portion out. If this behavior ever
|
|
# changes, we'll need to update this.
|
|
# TODO(bazel-team): Expose a ._name field on functions to avoid this.
|
|
impl_name = str(impl)
|
|
impl_name = impl_name.partition("<function ")[-1]
|
|
return impl_name.rpartition(">")[0]
|
|
|
|
def _make(impl, attrs = {}):
|
|
"""Creates a unit test rule from its implementation function.
|
|
|
|
Each unit test is defined in an implementation function that must then be
|
|
associated with a rule so that a target can be built. This function handles
|
|
the boilerplate to create and return a test rule and captures the
|
|
implementation function's name so that it can be printed in test feedback.
|
|
|
|
The optional `attrs` argument can be used to define dependencies for this
|
|
test, in order to form unit tests of rules.
|
|
|
|
An example of a unit test:
|
|
|
|
```
|
|
def _your_test(ctx):
|
|
env = unittest.begin(ctx)
|
|
|
|
# Assert statements go here
|
|
|
|
return unittest.end(env)
|
|
|
|
your_test = unittest.make(_your_test)
|
|
```
|
|
|
|
Recall that names of test rules must end in `_test`.
|
|
|
|
Args:
|
|
impl: The implementation function of the unit test.
|
|
attrs: An optional dictionary to supplement the attrs passed to the
|
|
unit test's `rule()` constructor.
|
|
|
|
Returns:
|
|
A rule definition that should be stored in a global whose name ends in
|
|
`_test`.
|
|
"""
|
|
attrs = dict(attrs)
|
|
attrs["_impl_name"] = attr.string(default = _impl_function_name(impl))
|
|
|
|
return rule(
|
|
impl,
|
|
attrs = attrs,
|
|
_skylark_testable = True,
|
|
test = True,
|
|
toolchains = [TOOLCHAIN_TYPE],
|
|
)
|
|
|
|
_ActionInfo = provider(fields = ["actions"])
|
|
|
|
def _action_retrieving_aspect_impl(target, ctx):
|
|
_ignore = [ctx]
|
|
return [_ActionInfo(actions = target.actions)]
|
|
|
|
_action_retrieving_aspect = aspect(
|
|
attr_aspects = [],
|
|
implementation = _action_retrieving_aspect_impl,
|
|
)
|
|
|
|
# TODO(cparsons): Provide more full documentation on analysis testing in README.
|
|
def _make_analysis_test(impl, expect_failure = False, attrs = {}, config_settings = {}):
|
|
"""Creates an analysis test rule from its implementation function.
|
|
|
|
An analysis test verifies the behavior of a "real" rule target by examining
|
|
and asserting on the providers given by the real target.
|
|
|
|
Each analysis test is defined in an implementation function that must then be
|
|
associated with a rule so that a target can be built. This function handles
|
|
the boilerplate to create and return a test rule and captures the
|
|
implementation function's name so that it can be printed in test feedback.
|
|
|
|
An example of an analysis test:
|
|
|
|
```
|
|
def _your_test(ctx):
|
|
env = analysistest.begin(ctx)
|
|
|
|
# Assert statements go here
|
|
|
|
return analysistest.end(env)
|
|
|
|
your_test = analysistest.make(_your_test)
|
|
```
|
|
|
|
Recall that names of test rules must end in `_test`.
|
|
|
|
Args:
|
|
impl: The implementation function of the unit test.
|
|
expect_failure: If true, the analysis test will expect the target_under_test
|
|
to fail. Assertions can be made on the underlying failure using asserts.expect_failure
|
|
attrs: An optional dictionary to supplement the attrs passed to the
|
|
unit test's `rule()` constructor.
|
|
config_settings: A dictionary of configuration settings to change for the target under
|
|
test and its dependencies. This may be used to essentially change 'build flags' for
|
|
the target under test, and may thus be utilized to test multiple targets with different
|
|
flags in a single build
|
|
|
|
Returns:
|
|
A rule definition that should be stored in a global whose name ends in
|
|
`_test`.
|
|
"""
|
|
attrs = dict(attrs)
|
|
attrs["_impl_name"] = attr.string(default = _impl_function_name(impl))
|
|
|
|
changed_settings = dict(config_settings)
|
|
if expect_failure:
|
|
changed_settings["//command_line_option:allow_analysis_failures"] = "True"
|
|
|
|
if changed_settings:
|
|
test_transition = analysis_test_transition(
|
|
settings = changed_settings,
|
|
)
|
|
attrs["target_under_test"] = attr.label(
|
|
aspects = [_action_retrieving_aspect],
|
|
cfg = test_transition,
|
|
mandatory = True,
|
|
)
|
|
else:
|
|
attrs["target_under_test"] = attr.label(
|
|
aspects = [_action_retrieving_aspect],
|
|
mandatory = True,
|
|
)
|
|
|
|
return rule(
|
|
impl,
|
|
attrs = attrs,
|
|
test = True,
|
|
toolchains = [TOOLCHAIN_TYPE],
|
|
analysis_test = True,
|
|
)
|
|
|
|
def _suite(name, *test_rules):
|
|
"""Defines a `test_suite` target that contains multiple tests.
|
|
|
|
After defining your test rules in a `.bzl` file, you need to create targets
|
|
from those rules so that `blaze test` can execute them. Doing this manually
|
|
in a BUILD file would consist of listing each test in your `load` statement
|
|
and then creating each target one by one. To reduce duplication, we recommend
|
|
writing a macro in your `.bzl` file to instantiate all targets, and calling
|
|
that macro from your BUILD file so you only have to load one symbol.
|
|
|
|
For the case where your unit tests do not take any (non-default) attributes --
|
|
i.e., if your unit tests do not test rules -- you can use this function to
|
|
create the targets and wrap them in a single test_suite target. In your
|
|
`.bzl` file, write:
|
|
|
|
```
|
|
def your_test_suite():
|
|
unittest.suite(
|
|
"your_test_suite",
|
|
your_test,
|
|
your_other_test,
|
|
yet_another_test,
|
|
)
|
|
```
|
|
|
|
Then, in your `BUILD` file, simply load the macro and invoke it to have all
|
|
of the targets created:
|
|
|
|
```
|
|
load("//path/to/your/package:tests.bzl", "your_test_suite")
|
|
your_test_suite()
|
|
```
|
|
|
|
If you pass _N_ unit test rules to `unittest.suite`, _N_ + 1 targets will be
|
|
created: a `test_suite` target named `${name}` (where `${name}` is the name
|
|
argument passed in here) and targets named `${name}_test_${i}`, where `${i}`
|
|
is the index of the test in the `test_rules` list, which is used to uniquely
|
|
name each target.
|
|
|
|
Args:
|
|
name: The name of the `test_suite` target, and the prefix of all the test
|
|
target names.
|
|
*test_rules: A list of test rules defines by `unittest.test`.
|
|
"""
|
|
test_names = []
|
|
for index, test_rule in enumerate(test_rules):
|
|
test_name = "%s_test_%d" % (name, index)
|
|
test_rule(name = test_name)
|
|
test_names.append(test_name)
|
|
|
|
native.test_suite(
|
|
name = name,
|
|
tests = [":%s" % t for t in test_names],
|
|
)
|
|
|
|
def _begin(ctx):
|
|
"""Begins a unit test.
|
|
|
|
This should be the first function called in a unit test implementation
|
|
function. It initializes a "test environment" that is used to collect
|
|
assertion failures so that they can be reported and logged at the end of the
|
|
test.
|
|
|
|
Args:
|
|
ctx: The Skylark context. Pass the implementation function's `ctx` argument
|
|
in verbatim.
|
|
|
|
Returns:
|
|
A test environment struct that must be passed to assertions and finally to
|
|
`unittest.end`. Do not rely on internal details about the fields in this
|
|
struct as it may change.
|
|
"""
|
|
return struct(ctx = ctx, failures = [])
|
|
|
|
def _end_analysis_test(env):
|
|
"""Ends an analysis test and logs the results.
|
|
|
|
This must be called and returned at the end of an analysis test implementation function so
|
|
that the results are reported.
|
|
|
|
Args:
|
|
env: The test environment returned by `analysistest.begin`.
|
|
|
|
Returns:
|
|
A list of providers needed to automatically register the analysis test result.
|
|
"""
|
|
return [AnalysisTestResultInfo(
|
|
success = (len(env.failures) == 0),
|
|
message = "\n".join(env.failures),
|
|
)]
|
|
|
|
def _end(env):
|
|
"""Ends a unit test and logs the results.
|
|
|
|
This must be called and returned at the end of a unit test implementation function so
|
|
that the results are reported.
|
|
|
|
Args:
|
|
env: The test environment returned by `unittest.begin`.
|
|
|
|
Returns:
|
|
A list of providers needed to automatically register the test result.
|
|
"""
|
|
|
|
tc = env.ctx.toolchains[TOOLCHAIN_TYPE].unittest_toolchain_info
|
|
testbin = env.ctx.actions.declare_file(env.ctx.label.name + tc.file_ext)
|
|
if env.failures:
|
|
cmd = tc.failure_templ % tc.join_on.join(env.failures)
|
|
else:
|
|
cmd = tc.success_templ
|
|
|
|
env.ctx.actions.write(
|
|
output = testbin,
|
|
content = cmd,
|
|
is_executable = True,
|
|
)
|
|
return [DefaultInfo(executable = testbin)]
|
|
|
|
def _fail(env, msg):
|
|
"""Unconditionally causes the current test to fail.
|
|
|
|
Args:
|
|
env: The test environment returned by `unittest.begin`.
|
|
msg: The message to log describing the failure.
|
|
"""
|
|
full_msg = "In test %s: %s" % (env.ctx.attr._impl_name, msg)
|
|
# There isn't a better way to output the message in Starlark, so use print.
|
|
# buildifier: disable=print
|
|
print(full_msg)
|
|
env.failures.append(full_msg)
|
|
|
|
def _assert_true(
|
|
env,
|
|
condition,
|
|
msg = "Expected condition to be true, but was false."):
|
|
"""Asserts that the given `condition` is true.
|
|
|
|
Args:
|
|
env: The test environment returned by `unittest.begin`.
|
|
condition: A value that will be evaluated in a Boolean context.
|
|
msg: An optional message that will be printed that describes the failure.
|
|
If omitted, a default will be used.
|
|
"""
|
|
if not condition:
|
|
_fail(env, msg)
|
|
|
|
def _assert_false(
|
|
env,
|
|
condition,
|
|
msg = "Expected condition to be false, but was true."):
|
|
"""Asserts that the given `condition` is false.
|
|
|
|
Args:
|
|
env: The test environment returned by `unittest.begin`.
|
|
condition: A value that will be evaluated in a Boolean context.
|
|
msg: An optional message that will be printed that describes the failure.
|
|
If omitted, a default will be used.
|
|
"""
|
|
if condition:
|
|
_fail(env, msg)
|
|
|
|
def _assert_equals(env, expected, actual, msg = None):
|
|
"""Asserts that the given `expected` and `actual` values are equal.
|
|
|
|
Args:
|
|
env: The test environment returned by `unittest.begin`.
|
|
expected: The expected value of some computation.
|
|
actual: The actual value returned by some computation.
|
|
msg: An optional message that will be printed that describes the failure.
|
|
If omitted, a default will be used.
|
|
"""
|
|
if expected != actual:
|
|
expectation_msg = 'Expected "%s", but got "%s"' % (expected, actual)
|
|
if msg:
|
|
full_msg = "%s (%s)" % (msg, expectation_msg)
|
|
else:
|
|
full_msg = expectation_msg
|
|
_fail(env, full_msg)
|
|
|
|
def _assert_set_equals(env, expected, actual, msg = None):
|
|
"""Asserts that the given `expected` and `actual` sets are equal.
|
|
|
|
Args:
|
|
env: The test environment returned by `unittest.begin`.
|
|
expected: The expected set resulting from some computation.
|
|
actual: The actual set returned by some computation.
|
|
msg: An optional message that will be printed that describes the failure.
|
|
If omitted, a default will be used.
|
|
"""
|
|
if type(actual) != type(depset()) or not sets.is_equal(expected, actual):
|
|
expectation_msg = "Expected %r, but got %r" % (expected, actual)
|
|
if msg:
|
|
full_msg = "%s (%s)" % (msg, expectation_msg)
|
|
else:
|
|
full_msg = expectation_msg
|
|
_fail(env, full_msg)
|
|
|
|
def _assert_new_set_equals(env, expected, actual, msg = None):
|
|
"""Asserts that the given `expected` and `actual` sets are equal.
|
|
|
|
Args:
|
|
env: The test environment returned by `unittest.begin`.
|
|
expected: The expected set resulting from some computation.
|
|
actual: The actual set returned by some computation.
|
|
msg: An optional message that will be printed that describes the failure.
|
|
If omitted, a default will be used.
|
|
"""
|
|
if not new_sets.is_equal(expected, actual):
|
|
expectation_msg = "Expected %r, but got %r" % (expected, actual)
|
|
if msg:
|
|
full_msg = "%s (%s)" % (msg, expectation_msg)
|
|
else:
|
|
full_msg = expectation_msg
|
|
_fail(env, full_msg)
|
|
|
|
def _expect_failure(env, expected_failure_msg = ""):
|
|
"""Asserts that the target under test has failed with a given error message.
|
|
|
|
This requires that the analysis test is created with `analysistest.make()` and
|
|
`expect_failures = True` is specified.
|
|
|
|
Args:
|
|
env: The test environment returned by `analysistest.begin`.
|
|
expected_failure_msg: The error message to expect as a result of analysis failures.
|
|
"""
|
|
dep = _target_under_test(env)
|
|
if AnalysisFailureInfo in dep:
|
|
actual_errors = ""
|
|
for cause in dep[AnalysisFailureInfo].causes.to_list():
|
|
actual_errors += cause.message + "\n"
|
|
if actual_errors.find(expected_failure_msg) < 0:
|
|
expectation_msg = "Expected errors to contain '%s' but did not. " % expected_failure_msg
|
|
expectation_msg += "Actual errors:%s" % actual_errors
|
|
_fail(env, expectation_msg)
|
|
else:
|
|
_fail(env, "Expected failure of target_under_test, but found success")
|
|
|
|
def _target_actions(env):
|
|
"""Returns a list of actions registered by the target under test.
|
|
|
|
Args:
|
|
env: The test environment returned by `analysistest.begin`.
|
|
|
|
Returns:
|
|
A list of actions registered by the target under test
|
|
"""
|
|
|
|
# Validate?
|
|
dep = _target_under_test(env)
|
|
return dep[_ActionInfo].actions
|
|
|
|
def _target_under_test(env):
|
|
"""Returns the target under test.
|
|
|
|
Args:
|
|
env: The test environment returned by `analysistest.begin`.
|
|
|
|
Returns:
|
|
The target under test.
|
|
"""
|
|
result = getattr(env.ctx.attr, "target_under_test")
|
|
if types.is_list(result):
|
|
if result:
|
|
return result[0]
|
|
else:
|
|
fail("test rule does not have a target_under_test")
|
|
return result
|
|
|
|
asserts = struct(
|
|
expect_failure = _expect_failure,
|
|
equals = _assert_equals,
|
|
false = _assert_false,
|
|
set_equals = _assert_set_equals,
|
|
new_set_equals = _assert_new_set_equals,
|
|
true = _assert_true,
|
|
)
|
|
|
|
unittest = struct(
|
|
make = _make,
|
|
suite = _suite,
|
|
begin = _begin,
|
|
end = _end,
|
|
fail = _fail,
|
|
)
|
|
|
|
analysistest = struct(
|
|
make = _make_analysis_test,
|
|
begin = _begin,
|
|
end = _end_analysis_test,
|
|
fail = _fail,
|
|
target_actions = _target_actions,
|
|
target_under_test = _target_under_test,
|
|
)
|