|  | """Utilities for testing bazel.""" | 
|  | # | 
|  | # Copyright 2015 The Bazel Authors. All rights reserved. | 
|  | # | 
|  | # Licensed under the Apache License, Version 2.0 (the "License"); | 
|  | # you may not use this file except in compliance with the License. | 
|  | # You may obtain a copy of the License at | 
|  | # | 
|  | #    http://www.apache.org/licenses/LICENSE-2.0 | 
|  | # | 
|  | # Unless required by applicable law or agreed to in writing, software | 
|  | # distributed under the License is distributed on an "AS IS" BASIS, | 
|  | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | 
|  | # See the License for the specific language governing permissions and | 
|  | # limitations under the License. | 
|  |  | 
|  | ### First, trivial tests that either always pass, always fail, | 
|  | ### or sometimes pass depending on a trivial computation. | 
|  |  | 
|  | def success_target(ctx, msg): | 
|  | """Return a success for an analysis test. | 
|  |  | 
|  | The test rule must have an executable output. | 
|  |  | 
|  | Args: | 
|  | ctx: the Bazel rule context | 
|  | msg: an informative message to display | 
|  |  | 
|  | Returns: | 
|  | a suitable rule implementation struct(), | 
|  | with actions that always succeed at execution time. | 
|  | """ | 
|  | exe = ctx.outputs.executable | 
|  | dat = ctx.new_file(ctx.genfiles_dir, exe, ".dat") | 
|  | ctx.actions.write( | 
|  | output = dat, | 
|  | content = msg, | 
|  | ) | 
|  | ctx.actions.write( | 
|  | output = exe, | 
|  | content = "cat " + dat.path + " ; echo", | 
|  | is_executable = True, | 
|  | ) | 
|  | return struct(runfiles = ctx.runfiles([exe, dat])) | 
|  |  | 
|  | def _successful_test_impl(ctx): | 
|  | return success_target(ctx, ctx.attr.msg) | 
|  |  | 
|  | successful_test = rule( | 
|  | attrs = {"msg": attr.string(mandatory = True)}, | 
|  | executable = True, | 
|  | test = True, | 
|  | implementation = _successful_test_impl, | 
|  | ) | 
|  |  | 
|  | def failure_target(ctx, msg): | 
|  | """Return a failure for an analysis test. | 
|  |  | 
|  | The test rule must have an executable output. | 
|  |  | 
|  | Args: | 
|  | ctx: the Bazel rule context | 
|  | msg: an informative message to display | 
|  |  | 
|  | Returns: | 
|  | a suitable rule implementation struct(), | 
|  | with actions that always fail at execution time. | 
|  | """ | 
|  |  | 
|  | ### fail(msg) ### <--- This would fail at analysis time. | 
|  | exe = ctx.outputs.executable | 
|  | dat = ctx.new_file(ctx.genfiles_dir, exe, ".dat") | 
|  | ctx.file_action( | 
|  | output = dat, | 
|  | content = msg, | 
|  | ) | 
|  | ctx.file_action( | 
|  | output = exe, | 
|  | content = "(cat " + dat.short_path + " ; echo ) >&2 ; exit 1", | 
|  | executable = True, | 
|  | ) | 
|  | return struct(runfiles = ctx.runfiles([exe, dat])) | 
|  |  | 
|  | def _failed_test_impl(ctx): | 
|  | return failure_target(ctx, ctx.attr.msg) | 
|  |  | 
|  | failed_test = rule( | 
|  | attrs = {"msg": attr.string(mandatory = True)}, | 
|  | executable = True, | 
|  | test = True, | 
|  | implementation = _failed_test_impl, | 
|  | ) | 
|  |  | 
|  | ### Second, general purpose utilities | 
|  |  | 
|  | def assert_(condition, string = "assertion failed", *args): | 
|  | """Trivial assertion mechanism. | 
|  |  | 
|  | Args: | 
|  | condition: a generalized boolean expected to be true | 
|  | string: a format string for the error message should the assertion fail | 
|  | *args: format arguments for the error message should the assertion fail | 
|  |  | 
|  | Returns: | 
|  | None. | 
|  |  | 
|  | Raises: | 
|  | an error if the condition isn't true. | 
|  | """ | 
|  |  | 
|  | if not condition: | 
|  | fail(string % args) | 
|  |  | 
|  | def strip_prefix(prefix, string): | 
|  | assert_( | 
|  | string.startswith(prefix), | 
|  | "%s does not start with %s", | 
|  | string, | 
|  | prefix, | 
|  | ) | 
|  | return string[len(prefix):len(string)] | 
|  |  | 
|  | def expectation_description(expect = None, expect_failure = None): | 
|  | """Turn expectation of result or error into a string.""" | 
|  | if expect_failure: | 
|  | return "failure " + str(expect_failure) | 
|  | else: | 
|  | return "result " + repr(expect) | 
|  |  | 
|  | def check_results(result, failure, expect, expect_failure): | 
|  | """See if actual computation results match expectations. | 
|  |  | 
|  | Args: | 
|  | result: the result returned by the test if it ran to completion | 
|  | failure: the failure message caught while testing, if any | 
|  | expect: the expected result for a successful test, if no failure expected | 
|  | expect_failure: the expected failure message for the test, if any | 
|  |  | 
|  | Returns: | 
|  | a pair (tuple) of a boolean (true if success) and a message (string). | 
|  | """ | 
|  | wanted = expectation_description(expect, expect_failure) | 
|  | found = expectation_description(result, failure) | 
|  | if wanted == found: | 
|  | return (True, "successfully computed " + wanted) | 
|  | else: | 
|  | return (False, "expect " + wanted + " but found " + found) | 
|  |  | 
|  | def load_results( | 
|  | name, | 
|  | result = None, | 
|  | failure = None, | 
|  | expect = None, | 
|  | expect_failure = None): | 
|  | """issue load-time results of a test. | 
|  |  | 
|  | Args: | 
|  | name: the name of the Bazel rule at load time. | 
|  | result: the result returned by the test if it ran to completion | 
|  | failure: the failure message caught while testing, if any | 
|  | expect: the expected result for a successful test, if no failure expected | 
|  | expect_failure: the expected failure message for the test, if any | 
|  |  | 
|  | Returns: | 
|  | None, after issuing a rule that will succeed at execution time if | 
|  | expectations were met. | 
|  | """ | 
|  | (is_success, msg) = check_results(result, failure, expect, expect_failure) | 
|  | this_test = successful_test if is_success else failed_test | 
|  | return this_test(name = name, msg = msg) | 
|  |  | 
|  | def analysis_results( | 
|  | ctx, | 
|  | result = None, | 
|  | failure = None, | 
|  | expect = None, | 
|  | expect_failure = None): | 
|  | """issue analysis-time results of a test. | 
|  |  | 
|  | Args: | 
|  | ctx: the Bazel rule context | 
|  | result: the result returned by the test if it ran to completion | 
|  | failure: the failure message caught while testing, if any | 
|  | expect: the expected result for a successful test, if no failure expected | 
|  | expect_failure: the expected failure message for the test, if any | 
|  |  | 
|  | Returns: | 
|  | a suitable rule implementation struct(), | 
|  | with actions that succeed at execution time if expectation were met, | 
|  | or fail at execution time if they didn't. | 
|  | """ | 
|  | (is_success, msg) = check_results(result, failure, expect, expect_failure) | 
|  | this_test = success_target if is_success else failure_target | 
|  | return this_test(ctx, msg) | 
|  |  | 
|  | ### Simple tests | 
|  |  | 
|  | def _rule_test_impl(ctx): | 
|  | """check that a rule generates the desired outputs and providers.""" | 
|  | rule_ = ctx.attr.rule | 
|  | rule_name = str(rule_.label) | 
|  | exe = ctx.outputs.executable | 
|  | if ctx.attr.generates: | 
|  | # Generate the proper prefix to remove from generated files. | 
|  | prefix_parts = [] | 
|  |  | 
|  | if rule_.label.workspace_root: | 
|  | # Create a prefix that is correctly relative to the output of this rule. | 
|  | prefix_parts = ["..", strip_prefix("external/", rule_.label.workspace_root)] | 
|  |  | 
|  | if rule_.label.package: | 
|  | prefix_parts.append(rule_.label.package) | 
|  |  | 
|  | prefix = "/".join(prefix_parts) | 
|  |  | 
|  | if prefix: | 
|  | # If the prefix isn't empty, it needs a trailing slash. | 
|  | prefix = prefix + "/" | 
|  |  | 
|  | # TODO(bazel-team): Use set() instead of sorted() once | 
|  | # set comparison is implemented. | 
|  | # TODO(bazel-team): Use a better way to determine if two paths refer to | 
|  | # the same file. | 
|  | generates = sorted(ctx.attr.generates) | 
|  | generated = sorted([ | 
|  | strip_prefix(prefix, f.short_path) | 
|  | for f in rule_.files.to_list() | 
|  | ]) | 
|  | if generates != generated: | 
|  | fail("rule %s generates %s not %s" % | 
|  | (rule_name, repr(generated), repr(generates))) | 
|  | provides = ctx.attr.provides | 
|  | if provides: | 
|  | files = [] | 
|  | commands = [] | 
|  | for k in provides.keys(): | 
|  | if hasattr(rule_, k): | 
|  | v = repr(getattr(rule_, k)) | 
|  | else: | 
|  | fail(("rule %s doesn't provide attribute %s. " + | 
|  | "Its list of attributes is: %s") % | 
|  | (rule_name, k, dir(rule_))) | 
|  | file_ = ctx.new_file(ctx.genfiles_dir, exe, "." + k) | 
|  | files += [file_] | 
|  | regexp = provides[k] | 
|  | commands += [ | 
|  | "if ! grep %s %s ; then echo 'bad %s:' ; cat %s ; echo ; exit 1 ; fi" % | 
|  | (repr(regexp), file_.short_path, k, file_.short_path), | 
|  | ] | 
|  | ctx.file_action(output = file_, content = v) | 
|  | script = "\n".join(commands + ["true"]) | 
|  | ctx.file_action(output = exe, content = script, executable = True) | 
|  | return struct(runfiles = ctx.runfiles([exe] + files)) | 
|  | else: | 
|  | return success_target(ctx, "success") | 
|  |  | 
|  | rule_test = rule( | 
|  | attrs = { | 
|  | "rule": attr.label(mandatory = True), | 
|  | "generates": attr.string_list(), | 
|  | "provides": attr.string_dict(), | 
|  | }, | 
|  | executable = True, | 
|  | test = True, | 
|  | implementation = _rule_test_impl, | 
|  | ) | 
|  |  | 
|  | def _file_test_impl(ctx): | 
|  | """check that a file has a given content.""" | 
|  | exe = ctx.outputs.executable | 
|  | file_ = ctx.file.file | 
|  | content = ctx.attr.content | 
|  | regexp = ctx.attr.regexp | 
|  | matches = ctx.attr.matches | 
|  | if bool(content) == bool(regexp): | 
|  | fail("Must specify one and only one of content or regexp") | 
|  | if content and matches != -1: | 
|  | fail("matches only makes sense with regexp") | 
|  | if content: | 
|  | dat = ctx.new_file(ctx.genfiles_dir, exe, ".dat") | 
|  | ctx.file_action( | 
|  | output = dat, | 
|  | content = content, | 
|  | ) | 
|  | ctx.file_action( | 
|  | output = exe, | 
|  | content = "diff -u %s %s" % (dat.short_path, file_.short_path), | 
|  | executable = True, | 
|  | ) | 
|  | return struct(runfiles = ctx.runfiles([exe, dat, file_])) | 
|  | if matches != -1: | 
|  | script = "[ %s == $(grep -c %s %s) ]" % ( | 
|  | matches, | 
|  | repr(regexp), | 
|  | file_.short_path, | 
|  | ) | 
|  | else: | 
|  | script = "grep %s %s" % (repr(regexp), file_.short_path) | 
|  | ctx.file_action( | 
|  | output = exe, | 
|  | content = script, | 
|  | executable = True, | 
|  | ) | 
|  | return struct(runfiles = ctx.runfiles([exe, file_])) | 
|  |  | 
|  | file_test = rule( | 
|  | attrs = { | 
|  | "file": attr.label( | 
|  | mandatory = True, | 
|  | allow_files = True, | 
|  | single_file = True, | 
|  | ), | 
|  | "content": attr.string(default = ""), | 
|  | "regexp": attr.string(default = ""), | 
|  | "matches": attr.int(default = -1), | 
|  | }, | 
|  | executable = True, | 
|  | test = True, | 
|  | implementation = _file_test_impl, | 
|  | ) |