|  | #!/usr/bin/env python3 | 
|  | # | 
|  | # Copyright 2018 The Bazel Authors. All rights reserved. | 
|  | # | 
|  | # Licensed under the Apache License, Version 2.0 (the "License"); | 
|  | # you may not use this file except in compliance with the License. | 
|  | # You may obtain a copy of the License at | 
|  | # | 
|  | #    http://www.apache.org/licenses/LICENSE-2.0 | 
|  | # | 
|  | # Unless required by applicable law or agreed to in writing, software | 
|  | # distributed under the License is distributed on an "AS IS" BASIS, | 
|  | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | 
|  | # See the License for the specific language governing permissions and | 
|  | # limitations under the License. | 
|  |  | 
|  | import argparse | 
|  | import base64 | 
|  | import codecs | 
|  | import collections | 
|  | import concurrent.futures | 
|  | import copy | 
|  | import datetime | 
|  | from glob import glob | 
|  | import hashlib | 
|  | import itertools | 
|  | import json | 
|  | import multiprocessing | 
|  | import os | 
|  | import os.path | 
|  | import platform as platform_module | 
|  | import random | 
|  | import re | 
|  | import requests | 
|  | import shutil | 
|  | import stat | 
|  | import subprocess | 
|  | import sys | 
|  | import tarfile | 
|  | import tempfile | 
|  | import threading | 
|  | import time | 
|  | from typing import Sequence | 
|  | import urllib.error | 
|  | import urllib.request | 
|  | import yaml | 
|  |  | 
|  | # Initialize the random number generator. | 
|  | random.seed() | 
|  |  | 
|  | BUILDKITE_ORG = os.environ["BUILDKITE_ORGANIZATION_SLUG"] | 
|  | THIS_IS_PRODUCTION = BUILDKITE_ORG == "bazel" | 
|  | THIS_IS_TESTING = BUILDKITE_ORG == "bazel-testing" | 
|  | THIS_IS_TRUSTED = BUILDKITE_ORG == "bazel-trusted" | 
|  | THIS_IS_SPARTA = True | 
|  |  | 
|  | CLOUD_PROJECT = "bazel-public" if THIS_IS_TRUSTED else "bazel-untrusted" | 
|  |  | 
|  | GITHUB_BRANCH = {"bazel": "master", "bazel-trusted": "master", "bazel-testing": "testing"}[ | 
|  | BUILDKITE_ORG | 
|  | ] | 
|  |  | 
|  | SCRIPT_URL = "https://raw.githubusercontent.com/bazelbuild/continuous-integration/{}/buildkite/bazelci.py".format( | 
|  | GITHUB_BRANCH | 
|  | ) | 
|  |  | 
|  | AGGREGATE_INCOMPATIBLE_TEST_RESULT_URL = "https://raw.githubusercontent.com/bazelbuild/continuous-integration/{}/buildkite/aggregate_incompatible_flags_test_result.py?{}".format( | 
|  | GITHUB_BRANCH, int(time.time()) | 
|  | ) | 
|  |  | 
|  | EMERGENCY_FILE_URL = "https://raw.githubusercontent.com/bazelbuild/continuous-integration/{}/buildkite/emergency.yml?{}".format( | 
|  | GITHUB_BRANCH, int(time.time()) | 
|  | ) | 
|  |  | 
|  | FLAKY_TESTS_BUCKET = { | 
|  | "bazel-testing": "gs://bazel-testing-buildkite-stats/flaky-tests-bep/", | 
|  | "bazel-trusted": "gs://bazel-buildkite-stats/flaky-tests-bep/", | 
|  | "bazel": "gs://bazel-buildkite-stats/flaky-tests-bep/", | 
|  | }[BUILDKITE_ORG] | 
|  |  | 
|  | KZIPS_BUCKET = { | 
|  | "bazel-testing": "gs://bazel-kzips-testing/", | 
|  | "bazel-trusted": "gs://bazel-kzips/", | 
|  | "bazel": "gs://bazel-kzips/", | 
|  | }[BUILDKITE_ORG] | 
|  |  | 
|  | # We don't collect logs in the trusted org | 
|  | LOG_BUCKET = { | 
|  | "bazel-testing": "https://storage.googleapis.com/bazel-testing-buildkite-artifacts", | 
|  | "bazel-trusted": "", | 
|  | "bazel": "https://storage.googleapis.com/bazel-untrusted-buildkite-artifacts", | 
|  | }[BUILDKITE_ORG] | 
|  |  | 
|  | # Projects can opt out of receiving GitHub issues from --notify by adding `"do_not_notify": True` to their respective downstream entry. | 
|  | DOWNSTREAM_PROJECTS_PRODUCTION = { | 
|  | "Android Studio Plugin": { | 
|  | "git_repository": "https://github.com/bazelbuild/intellij.git", | 
|  | "file_config": ".bazelci/android-studio.yml", | 
|  | "pipeline_slug": "android-studio-plugin", | 
|  | }, | 
|  | "Android Studio Plugin Google": { | 
|  | "git_repository": "https://github.com/bazelbuild/intellij.git", | 
|  | "file_config": ".bazelci/android-studio.yml", | 
|  | "pipeline_slug": "android-studio-plugin-google", | 
|  | }, | 
|  | "Android Testing": { | 
|  | "git_repository": "https://github.com/googlesamples/android-testing.git", | 
|  | "file_config": "bazelci/buildkite-pipeline.yml", | 
|  | "pipeline_slug": "android-testing", | 
|  | "disabled_reason": "https://github.com/android/testing-samples/issues/417", | 
|  | }, | 
|  | "Bazel": { | 
|  | "git_repository": "https://github.com/bazelbuild/bazel.git", | 
|  | "file_config": ".bazelci/postsubmit.yml", | 
|  | "pipeline_slug": "bazel-bazel", | 
|  | }, | 
|  | "Bazel Bench": { | 
|  | "git_repository": "https://github.com/bazelbuild/bazel-bench.git", | 
|  | "file_config": ".bazelci/postsubmit.yml", | 
|  | "pipeline_slug": "bazel-bench", | 
|  | }, | 
|  | "Bazel Codelabs": { | 
|  | "git_repository": "https://github.com/bazelbuild/codelabs.git", | 
|  | "pipeline_slug": "bazel-codelabs", | 
|  | "disabled_reason": "https://github.com/bazelbuild/codelabs/issues/38", | 
|  | }, | 
|  | "Bazel Examples": { | 
|  | "git_repository": "https://github.com/bazelbuild/examples.git", | 
|  | "pipeline_slug": "bazel-bazel-examples", | 
|  | }, | 
|  | "Bazel Remote Cache": { | 
|  | "git_repository": "https://github.com/buchgr/bazel-remote.git", | 
|  | "pipeline_slug": "bazel-remote-cache", | 
|  | }, | 
|  | "Bazel skylib": { | 
|  | "git_repository": "https://github.com/bazelbuild/bazel-skylib.git", | 
|  | "pipeline_slug": "bazel-skylib", | 
|  | "owned_by_bazel": True, | 
|  | }, | 
|  | "Bazel toolchains": { | 
|  | "git_repository": "https://github.com/bazelbuild/bazel-toolchains.git", | 
|  | "pipeline_slug": "bazel-toolchains", | 
|  | }, | 
|  | "Bazel watcher": { | 
|  | "git_repository": "https://github.com/bazelbuild/bazel-watcher.git", | 
|  | "pipeline_slug": "bazel-watcher", | 
|  | "disabled_reason": "https://github.com/bazelbuild/bazel-watcher/issues/590", | 
|  | }, | 
|  | "Bazelisk": { | 
|  | "git_repository": "https://github.com/bazelbuild/bazelisk.git", | 
|  | "file_config": ".bazelci/config.yml", | 
|  | "pipeline_slug": "bazelisk", | 
|  | }, | 
|  | "Buildfarm": { | 
|  | "git_repository": "https://github.com/bazelbuild/bazel-buildfarm.git", | 
|  | "pipeline_slug": "buildfarm-farmer", | 
|  | }, | 
|  | "Buildtools": { | 
|  | "git_repository": "https://github.com/bazelbuild/buildtools.git", | 
|  | "pipeline_slug": "buildtools", | 
|  | }, | 
|  | "CLion Plugin": { | 
|  | "git_repository": "https://github.com/bazelbuild/intellij.git", | 
|  | "file_config": ".bazelci/clion.yml", | 
|  | "pipeline_slug": "clion-plugin", | 
|  | }, | 
|  | "CLion Plugin Google": { | 
|  | "git_repository": "https://github.com/bazelbuild/intellij.git", | 
|  | "file_config": ".bazelci/clion.yml", | 
|  | "pipeline_slug": "clion-plugin-google", | 
|  | }, | 
|  | "Cartographer": { | 
|  | "git_repository": "https://github.com/googlecartographer/cartographer.git", | 
|  | "pipeline_slug": "cartographer", | 
|  | "disabled_reason": "https://github.com/cartographer-project/cartographer/issues/1938#issuecomment-1804795113", | 
|  | }, | 
|  | "Cloud Robotics Core": { | 
|  | "git_repository": "https://github.com/googlecloudrobotics/core.git", | 
|  | "http_config": "https://raw.githubusercontent.com/bazelbuild/continuous-integration/master/pipelines/cloud-robotics.yml", | 
|  | "pipeline_slug": "cloud-robotics-core", | 
|  | }, | 
|  | "Envoy": { | 
|  | "git_repository": "https://github.com/envoyproxy/envoy.git", | 
|  | "pipeline_slug": "envoy", | 
|  | "disabled_reason": "https://github.com/envoyproxy/envoy/issues/29723", | 
|  | }, | 
|  | "FlatBuffers": { | 
|  | "git_repository": "https://github.com/google/flatbuffers.git", | 
|  | "pipeline_slug": "flatbuffers", | 
|  | "disabled_reason": "https://github.com/google/flatbuffers/issues/7992", | 
|  | }, | 
|  | "Flogger": { | 
|  | "git_repository": "https://github.com/google/flogger.git", | 
|  | "http_config": "https://raw.githubusercontent.com/bazelbuild/continuous-integration/master/pipelines/flogger.yml", | 
|  | "pipeline_slug": "flogger", | 
|  | }, | 
|  | "Gerrit": { | 
|  | "git_repository": "https://gerrit.googlesource.com/gerrit.git", | 
|  | "http_config": "https://raw.githubusercontent.com/bazelbuild/continuous-integration/master/pipelines/gerrit.yml", | 
|  | "pipeline_slug": "gerrit", | 
|  | }, | 
|  | "Google Logging": { | 
|  | "git_repository": "https://github.com/google/glog.git", | 
|  | "pipeline_slug": "google-logging", | 
|  | }, | 
|  | "IntelliJ Plugin": { | 
|  | "git_repository": "https://github.com/bazelbuild/intellij.git", | 
|  | "file_config": ".bazelci/intellij.yml", | 
|  | "pipeline_slug": "intellij-plugin", | 
|  | }, | 
|  | "IntelliJ Plugin Google": { | 
|  | "git_repository": "https://github.com/bazelbuild/intellij.git", | 
|  | "file_config": ".bazelci/intellij.yml", | 
|  | "pipeline_slug": "intellij-plugin-google", | 
|  | }, | 
|  | "IntelliJ UE Plugin": { | 
|  | "git_repository": "https://github.com/bazelbuild/intellij.git", | 
|  | "file_config": ".bazelci/intellij-ue.yml", | 
|  | "pipeline_slug": "intellij-ue-plugin", | 
|  | }, | 
|  | "IntelliJ UE Plugin Google": { | 
|  | "git_repository": "https://github.com/bazelbuild/intellij.git", | 
|  | "file_config": ".bazelci/intellij-ue.yml", | 
|  | "pipeline_slug": "intellij-ue-plugin-google", | 
|  | }, | 
|  | "IntelliJ Plugin Aspect": { | 
|  | "git_repository": "https://github.com/bazelbuild/intellij.git", | 
|  | "file_config": ".bazelci/aspect.yml", | 
|  | "pipeline_slug": "intellij-plugin-aspect", | 
|  | }, | 
|  | "IntelliJ Plugin Aspect Google": { | 
|  | "git_repository": "https://github.com/bazelbuild/intellij.git", | 
|  | "file_config": ".bazelci/aspect.yml", | 
|  | "pipeline_slug": "intellij-plugin-aspect-google", | 
|  | }, | 
|  | "Kythe": { | 
|  | "git_repository": "https://github.com/kythe/kythe.git", | 
|  | "pipeline_slug": "kythe", | 
|  | }, | 
|  | "Protobuf": { | 
|  | "git_repository": "https://github.com/google/protobuf.git", | 
|  | "http_config": "https://raw.githubusercontent.com/bazelbuild/continuous-integration/master/pipelines/protobuf.yml", | 
|  | "pipeline_slug": "protobuf", | 
|  | "owned_by_bazel": True, | 
|  | }, | 
|  | "Stardoc": { | 
|  | "git_repository": "https://github.com/bazelbuild/stardoc.git", | 
|  | "pipeline_slug": "stardoc", | 
|  | "owned_by_bazel": True, | 
|  | }, | 
|  | "TensorFlow": { | 
|  | "git_repository": "https://github.com/tensorflow/tensorflow.git", | 
|  | "http_config": "https://raw.githubusercontent.com/bazelbuild/continuous-integration/master/pipelines/tensorflow.yml", | 
|  | "pipeline_slug": "tensorflow", | 
|  | "disabled_reason": "https://github.com/tensorflow/tensorflow/issues/60508", | 
|  | }, | 
|  | "re2": { | 
|  | "git_repository": "https://github.com/google/re2.git", | 
|  | "http_config": "https://raw.githubusercontent.com/bazelbuild/continuous-integration/master/pipelines/re2.yml", | 
|  | "pipeline_slug": "re2", | 
|  | }, | 
|  | "rules_android": { | 
|  | "git_repository": "https://github.com/bazelbuild/rules_android.git", | 
|  | "pipeline_slug": "rules-android", | 
|  | }, | 
|  | "rules_android_ndk": { | 
|  | "git_repository": "https://github.com/bazelbuild/rules_android_ndk.git", | 
|  | "pipeline_slug": "rules-android-ndk", | 
|  | }, | 
|  | "rules_apple": { | 
|  | "git_repository": "https://github.com/bazelbuild/rules_apple.git", | 
|  | "pipeline_slug": "rules-apple-darwin", | 
|  | }, | 
|  | "rules_cc": { | 
|  | "git_repository": "https://github.com/bazelbuild/rules_cc.git", | 
|  | "pipeline_slug": "rules-cc", | 
|  | "owned_by_bazel": True, | 
|  | "disabled_reason": "https://github.com/bazelbuild/rules_cc/issues/190", | 
|  | }, | 
|  | "rules_closure": { | 
|  | "git_repository": "https://github.com/bazelbuild/rules_closure.git", | 
|  | "pipeline_slug": "rules-closure-closure-compiler", | 
|  | "owned_by_bazel": True, | 
|  | }, | 
|  | "rules_dotnet": { | 
|  | "git_repository": "https://github.com/bazelbuild/rules_dotnet.git", | 
|  | "pipeline_slug": "rules-dotnet-edge", | 
|  | }, | 
|  | "rules_foreign_cc": { | 
|  | "git_repository": "https://github.com/bazelbuild/rules_foreign_cc.git", | 
|  | "file_config": ".bazelci/config.yaml", | 
|  | "pipeline_slug": "rules-foreign-cc", | 
|  | "owned_by_bazel": True, | 
|  | }, | 
|  | "rules_go": { | 
|  | "git_repository": "https://github.com/bazelbuild/rules_go.git", | 
|  | "pipeline_slug": "rules-go-golang", | 
|  | }, | 
|  | "rules_graalvm": { | 
|  | "git_repository": "https://github.com/sgammon/rules_graalvm.git", | 
|  | "pipeline_slug": "rules-graalvm", | 
|  | "disabled_reason": "https://github.com/sgammon/rules_graalvm/issues/409", | 
|  | }, | 
|  | "rules_haskell": { | 
|  | "git_repository": "https://github.com/tweag/rules_haskell.git", | 
|  | "pipeline_slug": "rules-haskell-haskell", | 
|  | }, | 
|  | "rules_jsonnet": { | 
|  | "git_repository": "https://github.com/bazelbuild/rules_jsonnet.git", | 
|  | "pipeline_slug": "rules-jsonnet", | 
|  | }, | 
|  | "rules_jvm_external": { | 
|  | "git_repository": "https://github.com/bazelbuild/rules_jvm_external.git", | 
|  | "pipeline_slug": "rules-jvm-external", | 
|  | "owned_by_bazel": True, | 
|  | }, | 
|  | "rules_jvm_external - examples": { | 
|  | "git_repository": "https://github.com/bazelbuild/rules_jvm_external.git", | 
|  | "file_config": ".bazelci/examples.yml", | 
|  | "pipeline_slug": "rules-jvm-external-examples", | 
|  | "owned_by_bazel": True, | 
|  | }, | 
|  | "rules_kotlin": { | 
|  | "git_repository": "https://github.com/bazelbuild/rules_kotlin.git", | 
|  | "pipeline_slug": "rules-kotlin-kotlin", | 
|  | }, | 
|  | "rules_nodejs": { | 
|  | "git_repository": "https://github.com/bazelbuild/rules_nodejs.git", | 
|  | "pipeline_slug": "rules-nodejs-nodejs", | 
|  | }, | 
|  | "rules_perl": { | 
|  | "git_repository": "https://github.com/bazelbuild/rules_perl.git", | 
|  | "pipeline_slug": "rules-perl", | 
|  | }, | 
|  | "rules_platform": { | 
|  | "git_repository": "https://github.com/bazelbuild/rules_platform.git", | 
|  | "pipeline_slug": "rules-platform", | 
|  | }, | 
|  | "rules_proto": { | 
|  | "git_repository": "https://github.com/bazelbuild/rules_proto.git", | 
|  | "pipeline_slug": "rules-proto", | 
|  | "owned_by_bazel": True, | 
|  | }, | 
|  | "rules_python": { | 
|  | "git_repository": "https://github.com/bazelbuild/rules_python.git", | 
|  | "pipeline_slug": "rules-python-python", | 
|  | "owned_by_bazel": True, | 
|  | }, | 
|  | "rules_rust": { | 
|  | "git_repository": "https://github.com/bazelbuild/rules_rust.git", | 
|  | "pipeline_slug": "rules-rust-rustlang", | 
|  | "disabled_reason": "https://github.com/bazelbuild/rules_rust/issues/2519, https://github.com/bazelbuild/rules_rust/issues/2464", | 
|  | }, | 
|  | "rules_scala": { | 
|  | "git_repository": "https://github.com/bazelbuild/rules_scala.git", | 
|  | "pipeline_slug": "rules-scala-scala", | 
|  | "disabled_reason": "waiting on https://github.com/bazelbuild/rules_scala/pull/1422", | 
|  | }, | 
|  | "rules_swift": { | 
|  | "git_repository": "https://github.com/bazelbuild/rules_swift.git", | 
|  | "pipeline_slug": "rules-swift-swift", | 
|  | }, | 
|  | "rules_testing": { | 
|  | "git_repository": "https://github.com/bazelbuild/rules_testing.git", | 
|  | "pipeline_slug": "rules-testing", | 
|  | "owned_by_bazel": True, | 
|  | }, | 
|  | "rules_webtesting": { | 
|  | "git_repository": "https://github.com/bazelbuild/rules_webtesting.git", | 
|  | "pipeline_slug": "rules-webtesting-saucelabs", | 
|  | }, | 
|  | } | 
|  |  | 
|  | DOWNSTREAM_PROJECTS_TESTING = { | 
|  | "Bazel": DOWNSTREAM_PROJECTS_PRODUCTION["Bazel"], | 
|  | "Bazelisk": DOWNSTREAM_PROJECTS_PRODUCTION["Bazelisk"], | 
|  | "rules_go": DOWNSTREAM_PROJECTS_PRODUCTION["rules_go"], | 
|  | "rules_kotlin": DOWNSTREAM_PROJECTS_PRODUCTION["rules_kotlin"], | 
|  | "rules_nodejs": DOWNSTREAM_PROJECTS_PRODUCTION["rules_nodejs"], | 
|  | "rules_rust": DOWNSTREAM_PROJECTS_PRODUCTION["rules_rust"], | 
|  | "rules_scala": DOWNSTREAM_PROJECTS_PRODUCTION["rules_scala"], | 
|  | } | 
|  |  | 
|  | DOWNSTREAM_PROJECTS = { | 
|  | "bazel-testing": DOWNSTREAM_PROJECTS_TESTING, | 
|  | "bazel-trusted": {}, | 
|  | "bazel": DOWNSTREAM_PROJECTS_PRODUCTION, | 
|  | }[BUILDKITE_ORG] | 
|  |  | 
|  | DOCKER_REGISTRY_PREFIX = { | 
|  | "bazel-testing": "bazel-public/testing", | 
|  | "bazel-trusted": "bazel-public", | 
|  | "bazel": "bazel-public", | 
|  | }[BUILDKITE_ORG] | 
|  |  | 
|  | # A map containing all supported platform names as keys, with the values being | 
|  | # the platform name in a human readable format, and a the buildkite-agent's | 
|  | # working directory. | 
|  | PLATFORMS = { | 
|  | "centos7": { | 
|  | "name": "CentOS 7", | 
|  | "emoji-name": ":centos: CentOS 7", | 
|  | "publish_binary": ["ubuntu1404", "centos7", "linux"], | 
|  | "docker-image": f"gcr.io/{DOCKER_REGISTRY_PREFIX}/centos7", | 
|  | "python": "python3.6", | 
|  | }, | 
|  | "centos7_java11": { | 
|  | "name": "CentOS 7 (OpenJDK 11, gcc 4.8.5)", | 
|  | "emoji-name": ":centos: CentOS 7 (OpenJDK 11, gcc 4.8.5)", | 
|  | "publish_binary": [], | 
|  | "docker-image": f"gcr.io/{DOCKER_REGISTRY_PREFIX}/centos7-java11", | 
|  | "python": "python3.6", | 
|  | }, | 
|  | "centos7_java11_devtoolset10": { | 
|  | "name": "CentOS 7 (OpenJDK 11, gcc 10.2.1)", | 
|  | "emoji-name": ":centos: CentOS 7 (OpenJDK 11, gcc 10.2.1)", | 
|  | "publish_binary": [], | 
|  | "docker-image": f"gcr.io/{DOCKER_REGISTRY_PREFIX}/centos7-java11-devtoolset10", | 
|  | "python": "python3.6", | 
|  | }, | 
|  | "debian10": { | 
|  | "name": "Debian 10 Buster (OpenJDK 11, gcc 8.3.0)", | 
|  | "emoji-name": ":debian: Debian 10 Buster (OpenJDK 11, gcc 8.3.0)", | 
|  | "publish_binary": [], | 
|  | "docker-image": f"gcr.io/{DOCKER_REGISTRY_PREFIX}/debian10-java11", | 
|  | "python": "python3.7", | 
|  | }, | 
|  | "debian11": { | 
|  | "name": "Debian 11 Bullseye (OpenJDK 17, gcc 10.2.1)", | 
|  | "emoji-name": ":debian: Debian 11 Bullseye (OpenJDK 17, gcc 10.2.1)", | 
|  | "publish_binary": [], | 
|  | "docker-image": f"gcr.io/{DOCKER_REGISTRY_PREFIX}/debian11-java17", | 
|  | "python": "python3.9", | 
|  | }, | 
|  | "ubuntu1604": { | 
|  | "name": "Ubuntu 16.04 LTS (OpenJDK 8, gcc 5.4.0)", | 
|  | "emoji-name": ":ubuntu: Ubuntu 16.04 LTS (OpenJDK 8, gcc 5.4.0)", | 
|  | "publish_binary": [], | 
|  | "docker-image": f"gcr.io/{DOCKER_REGISTRY_PREFIX}/ubuntu1604-java8", | 
|  | "python": "python3.6", | 
|  | }, | 
|  | "ubuntu1804": { | 
|  | "name": "Ubuntu 18.04 LTS (OpenJDK 11, gcc 7.5.0)", | 
|  | "emoji-name": ":ubuntu: Ubuntu 18.04 LTS (OpenJDK 11, gcc 7.5.0)", | 
|  | "publish_binary": [], | 
|  | "docker-image": f"gcr.io/{DOCKER_REGISTRY_PREFIX}/ubuntu1804-java11", | 
|  | "python": "python3.6", | 
|  | }, | 
|  | "ubuntu2004_java11": { | 
|  | "name": "Ubuntu 20.04 LTS (OpenJDK 11, gcc 9.4.0)", | 
|  | "emoji-name": ":ubuntu: Ubuntu 20.04 LTS (OpenJDK 11, gcc 9.4.0)", | 
|  | "publish_binary": [], | 
|  | "docker-image": f"gcr.io/{DOCKER_REGISTRY_PREFIX}/ubuntu2004-java11", | 
|  | "python": "python3.8", | 
|  | }, | 
|  | "ubuntu2004": { | 
|  | "name": "Ubuntu 20.04 LTS", | 
|  | "emoji-name": ":ubuntu: Ubuntu 20.04 LTS", | 
|  | "publish_binary": [], | 
|  | "docker-image": f"gcr.io/{DOCKER_REGISTRY_PREFIX}/ubuntu2004", | 
|  | "python": "python3.8", | 
|  | }, | 
|  | "ubuntu2004_arm64": { | 
|  | "name": "Ubuntu 20.04 LTS ARM64", | 
|  | "emoji-name": ":ubuntu: Ubuntu 20.04 LTS ARM64", | 
|  | "publish_binary": [], | 
|  | "docker-image": f"gcr.io/{DOCKER_REGISTRY_PREFIX}/ubuntu2004", | 
|  | "python": "python3.8", | 
|  | "queue": "arm64", | 
|  | # TODO: Re-enable always-pull if we also publish docker containers for Linux ARM64 | 
|  | "always-pull": False, | 
|  | }, | 
|  | "kythe_ubuntu2004": { | 
|  | "name": "Kythe (Ubuntu 20.04 LTS)", | 
|  | "emoji-name": "Kythe (:ubuntu: Ubuntu 20.04 LTS)", | 
|  | "publish_binary": [], | 
|  | "docker-image": f"gcr.io/{DOCKER_REGISTRY_PREFIX}/ubuntu2004-kythe", | 
|  | "python": "python3.8", | 
|  | }, | 
|  | "kythe_ubuntu2204": { | 
|  | "name": "Kythe (Ubuntu 22.04 LTS)", | 
|  | "emoji-name": "Kythe (:ubuntu: Ubuntu 22.04 LTS)", | 
|  | "publish_binary": [], | 
|  | "docker-image": f"gcr.io/{DOCKER_REGISTRY_PREFIX}/ubuntu2204-kythe", | 
|  | "python": "python3", | 
|  | }, | 
|  | "ubuntu2204_java17": { | 
|  | "name": "Ubuntu 22.04 (OpenJDK 17, gcc 11.2.0)", | 
|  | "emoji-name": ":ubuntu: Ubuntu 22.04 (OpenJDK 17, gcc 11.2.0)", | 
|  | "publish_binary": [], | 
|  | "docker-image": f"gcr.io/{DOCKER_REGISTRY_PREFIX}/ubuntu2204-java17", | 
|  | "python": "python3", | 
|  | }, | 
|  | "ubuntu2204": { | 
|  | "name": "Ubuntu 22.04", | 
|  | "emoji-name": ":ubuntu: Ubuntu 22.04 LTS", | 
|  | "publish_binary": [], | 
|  | "docker-image": f"gcr.io/{DOCKER_REGISTRY_PREFIX}/ubuntu2204", | 
|  | "python": "python3", | 
|  | }, | 
|  | "fedora39": { | 
|  | "name": "Fedora 39 (OpenJDK 17, gcc 13.1.1)", | 
|  | "emoji-name": ":fedora: Fedora 39 (OpenJDK 17, gcc 13.1.1)", | 
|  | "publish_binary": [], | 
|  | "docker-image": f"gcr.io/{DOCKER_REGISTRY_PREFIX}/fedora39-java17", | 
|  | "python": "python3", | 
|  | }, | 
|  | "fedora40": { | 
|  | "name": "Fedora 40 (OpenJDK 21, gcc 14.1.1)", | 
|  | "emoji-name": ":fedora: Fedora 40 (OpenJDK 21, gcc 14.1.1)", | 
|  | "publish_binary": [], | 
|  | "docker-image": f"gcr.io/{DOCKER_REGISTRY_PREFIX}/fedora40-java21", | 
|  | "python": "python3", | 
|  | }, | 
|  | "macos": { | 
|  | "name": "macOS", | 
|  | "emoji-name": ":darwin: macOS", | 
|  | "publish_binary": ["macos"], | 
|  | "queue": "macos", | 
|  | "python": "python3", | 
|  | }, | 
|  | "macos_legacy": { | 
|  | "name": "macOS", | 
|  | "emoji-name": ":darwin: macOS", | 
|  | "publish_binary": [], | 
|  | "queue": "macos_legacy", | 
|  | "python": "python3", | 
|  | }, | 
|  | "macos_v2": { | 
|  | "name": "macOS", | 
|  | "emoji-name": ":darwin: macOS", | 
|  | "publish_binary": [], | 
|  | "queue": "macos_v2", | 
|  | "python": "python3", | 
|  | }, | 
|  | "macos_qa": { | 
|  | "name": "macOS QA", | 
|  | "emoji-name": ":darwin: :fire_extinguisher:", | 
|  | "publish_binary": [], | 
|  | "queue": "macos_qa", | 
|  | "python": "python3", | 
|  | }, | 
|  | "macos_arm64_qa": { | 
|  | "name": "macOS arm64 QA", | 
|  | "emoji-name": ":darwin: arm64 :fire_extinguisher:", | 
|  | "publish_binary": [], | 
|  | "queue": "macos_arm64_qa", | 
|  | "python": "python3", | 
|  | }, | 
|  | "macos_arm64": { | 
|  | "name": "macOS arm64", | 
|  | "emoji-name": ":darwin: macOS arm64", | 
|  | "publish_binary": ["macos_arm64"], | 
|  | "queue": "macos_arm64", | 
|  | "python": "python3", | 
|  | }, | 
|  | "macos_arm64_legacy": { | 
|  | "name": "macOS arm64", | 
|  | "emoji-name": ":darwin: macOS arm64", | 
|  | "publish_binary": [], | 
|  | "queue": "macos_arm64_legacy", | 
|  | "python": "python3", | 
|  | }, | 
|  | "macos_arm64_v2": { | 
|  | "name": "macOS arm64", | 
|  | "emoji-name": ":darwin: macOS arm64", | 
|  | "publish_binary": [], | 
|  | "queue": "macos_arm64_v2", | 
|  | "python": "python3", | 
|  | }, | 
|  | "windows": { | 
|  | "name": "Windows", | 
|  | "emoji-name": ":windows: Windows", | 
|  | "publish_binary": ["windows"], | 
|  | "queue": "windows", | 
|  | "python": "python.exe", | 
|  | }, | 
|  | "windows_arm64": { | 
|  | "name": "Windows ARM64", | 
|  | "emoji-name": ":windows: Windows arm64", | 
|  | "publish_binary": ["windows_arm64"], | 
|  | # TODO(pcloudy): Switch to windows_arm64 queue when Windows ARM64 machines are available, | 
|  | # current we just use x86_64 machines to do cross compile. | 
|  | "queue": "windows", | 
|  | "python": "python.exe", | 
|  | }, | 
|  | } | 
|  |  | 
|  | # Generate rbe_ubuntu* platforms based on ubuntu* platforms. | 
|  | for platform, platform_dict in PLATFORMS.copy().items(): | 
|  | if platform.startswith("ubuntu"): | 
|  | rbe_platform_dict = copy.deepcopy(platform_dict) | 
|  | rbe_platform_dict["name"] = "RBE {}".format(platform_dict["name"]) | 
|  | rbe_platform_dict["emoji-name"] = "RBE {}".format(platform_dict["emoji-name"]) | 
|  | PLATFORMS["rbe_{}".format(platform)] = rbe_platform_dict | 
|  |  | 
|  | BUILDIFIER_DOCKER_IMAGE = "gcr.io/bazel-public/buildifier" | 
|  |  | 
|  | # The platform used for various steps (e.g. stuff that formerly ran on the "pipeline" workers). | 
|  | DEFAULT_PLATFORM = "ubuntu1804" | 
|  |  | 
|  | # In order to test that "the one Linux binary" that we build for our official releases actually | 
|  | # works on all Linux distributions that we test on, we use the Linux binary built on our official | 
|  | # release platform for all Linux downstream tests. | 
|  | LINUX_BINARY_PLATFORM = "centos7" | 
|  |  | 
|  | XCODE_VERSION_REGEX = re.compile(r"^\d+\.\d+(\.\d+)?$") | 
|  | XCODE_VERSION_OVERRIDES = {"10.2.1": "10.3", "11.2": "11.2.1", "11.3": "11.3.1"} | 
|  |  | 
|  | BUILD_LABEL_PATTERN = re.compile(r"^Build label: (\S+)$", re.MULTILINE) | 
|  |  | 
|  | BUILDIFIER_STEP_NAME = "Buildifier" | 
|  | SHARD_SUMMARY_STEP_NAME = "Print Test Summary for Shards" | 
|  |  | 
|  | SKIP_TASKS_ENV_VAR = "CI_SKIP_TASKS" | 
|  |  | 
|  | # TODO: change to USE_BAZEL_DIFF once the feature has been tested in QA | 
|  | USE_BAZEL_DIFF_ENV_VAR = "USE_BAZEL_DIFF" | 
|  |  | 
|  | BAZEL_DIFF_ANNOTATION_CTX = "'diff'" | 
|  |  | 
|  | # TODO(fweikert): Install bazel-diff on the Docker images and on the Mac machines | 
|  | BAZEL_DIFF_URL = ( | 
|  | "https://github.com/Tinder/bazel-diff/releases/download/4.5.0/bazel-diff_deploy.jar" | 
|  | ) | 
|  |  | 
|  | AUTO_DIFFBASE_VALUES = frozenset(["1", "true", "auto"]) | 
|  |  | 
|  | # Always run all test targets if any of the paths here are modified by the current commit. | 
|  | # Values can be directory paths (with a trailing slash) or file paths. | 
|  | DISABLE_BAZEL_DIFF_IF_MODIFIED = (".bazelci/", ".bazelversion", "MODULE.bazel") | 
|  |  | 
|  | COMMIT_RE = re.compile(r"[0-9a-z]{40}") | 
|  |  | 
|  | CONFIG_FILE_EXTENSIONS = {".yml", ".yaml"} | 
|  |  | 
|  | KYTHE_DIR = "/usr/local/kythe" | 
|  |  | 
|  | INDEX_UPLOAD_POLICY_ALWAYS = "Always" | 
|  |  | 
|  | INDEX_UPLOAD_POLICY_IF_BUILD_SUCCESS = "IfBuildSuccess" | 
|  |  | 
|  | INDEX_UPLOAD_POLICY_NEVER = "Never" | 
|  |  | 
|  | ESCAPED_BACKSLASH = "%5C" | 
|  |  | 
|  | # The maximum number of tasks allowed in one pipeline yaml config file. | 
|  | # This is to prevent accidentally creating too many tasks with the martix testing feature. | 
|  | MAX_TASK_NUMBER = 80 | 
|  |  | 
|  | LAB_AGENT_PATTERNS = [ | 
|  | re.compile(r"^bk-imacpro-\d+$"), | 
|  | re.compile(r"^bk-(trusted|testing)-macpro-\d+$"), | 
|  | re.compile(r"^bk-(trusted-)?macstudio-\d+$"), | 
|  | ] | 
|  |  | 
|  | _TEST_BEP_FILE = "test_bep.json" | 
|  | _SHARD_RE = re.compile(r"(.+) \(shard (\d+)\)") | 
|  | _SLOWEST_N_TARGETS = 20 | 
|  |  | 
|  |  | 
|  | class BuildkiteException(Exception): | 
|  | """ | 
|  | Raised whenever something goes wrong and we should exit with an error. | 
|  | """ | 
|  |  | 
|  | pass | 
|  |  | 
|  |  | 
|  | class BuildkiteInfraException(Exception): | 
|  | """ | 
|  | Raised whenever something goes wrong with the CI infra and we should immediately exit with an error. | 
|  | """ | 
|  |  | 
|  | pass | 
|  |  | 
|  |  | 
|  | class BinaryUploadRaceException(Exception): | 
|  | """ | 
|  | Raised when try_publish_binaries wasn't able to publish a set of binaries, | 
|  | because the generation of the current file didn't match the expected value. | 
|  | """ | 
|  |  | 
|  | pass | 
|  |  | 
|  |  | 
|  | class BuildkiteClient(object): | 
|  | _ENCRYPTED_BUILDKITE_API_TOKEN = """ | 
|  | CiQA4DEB9ldzC+E39KomywtqXfaQ86hhulgeDsicds2BuvbCYzsSUAAqwcvXZPh9IMWlwWh94J2F | 
|  | exosKKaWB0tSRJiPKnv2NPDfEqGul0ZwVjtWeASpugwxxKeLhFhPMcgHMPfndH6j2GEIY6nkKRbP | 
|  | uwoRMCwe | 
|  | """.strip() | 
|  |  | 
|  | _ENCRYPTED_BUILDKITE_API_TESTING_TOKEN = """ | 
|  | CiQAMTBkWjL1C+F5oon3+cC1vmum5+c1y5+96WQY44p0Lxd0PeASUQAy7iU0c6E3W5EOSFYfD5fA | 
|  | MWy/SHaMno1NQSUa4xDOl5yc2kizrtxPPVkX4x9pLNuGUY/xwAn2n1DdiUdWZNWlY1bX2C4ex65e | 
|  | P9w8kNhEbw== | 
|  | """.strip() | 
|  |  | 
|  | _BUILD_STATUS_URL_TEMPLATE = ( | 
|  | "https://api.buildkite.com/v2/organizations/{}/pipelines/{}/builds/{}" | 
|  | ) | 
|  |  | 
|  | _NEW_BUILD_URL_TEMPLATE = "https://api.buildkite.com/v2/organizations/{}/pipelines/{}/builds" | 
|  |  | 
|  | _RETRY_JOB_URL_TEMPLATE = ( | 
|  | "https://api.buildkite.com/v2/organizations/{}/pipelines/{}/builds/{}/jobs/{}/retry" | 
|  | ) | 
|  |  | 
|  | _PIPELINE_INFO_URL_TEMPLATE = "https://api.buildkite.com/v2/organizations/{}/pipelines/{}" | 
|  |  | 
|  | def __init__(self, org, pipeline): | 
|  | self._org = org | 
|  | self._pipeline = pipeline | 
|  | self._token = self._get_buildkite_token() | 
|  |  | 
|  | def _get_buildkite_token(self): | 
|  | return decrypt_token( | 
|  | encrypted_token=( | 
|  | self._ENCRYPTED_BUILDKITE_API_TESTING_TOKEN | 
|  | if THIS_IS_TESTING | 
|  | else self._ENCRYPTED_BUILDKITE_API_TOKEN | 
|  | ), | 
|  | kms_key=( | 
|  | "buildkite-testing-api-token" | 
|  | if THIS_IS_TESTING | 
|  | else "buildkite-untrusted-api-token" | 
|  | ), | 
|  | ) | 
|  |  | 
|  | def _open_url(self, url, params=[]): | 
|  | try: | 
|  | params_str = "".join("&{}={}".format(k, v) for k, v in params) | 
|  | return ( | 
|  | urllib.request.urlopen("{}?access_token={}{}".format(url, self._token, params_str)) | 
|  | .read() | 
|  | .decode("utf-8", "ignore") | 
|  | ) | 
|  | except urllib.error.HTTPError as ex: | 
|  | raise BuildkiteException("Failed to open {}: {} - {}".format(url, ex.code, ex.reason)) | 
|  |  | 
|  | def get_pipeline_info(self): | 
|  | """Get details for a pipeline given its organization slug | 
|  | and pipeline slug. | 
|  | See https://buildkite.com/docs/apis/rest-api/pipelines#get-a-pipeline | 
|  |  | 
|  | Returns | 
|  | ------- | 
|  | dict | 
|  | the metadata for the pipeline | 
|  | """ | 
|  | url = self._PIPELINE_INFO_URL_TEMPLATE.format(self._org, self._pipeline) | 
|  | output = self._open_url(url) | 
|  | return json.loads(output) | 
|  |  | 
|  | def get_build_info(self, build_number): | 
|  | """Get build info for a pipeline with a given build number | 
|  | See https://buildkite.com/docs/apis/rest-api/builds#get-a-build | 
|  |  | 
|  | Parameters | 
|  | ---------- | 
|  | build_number : the build number | 
|  |  | 
|  | Returns | 
|  | ------- | 
|  | dict | 
|  | the metadata for the build | 
|  | """ | 
|  | url = self._BUILD_STATUS_URL_TEMPLATE.format(self._org, self._pipeline, build_number) | 
|  | output = self._open_url(url) | 
|  | return json.loads(output) | 
|  |  | 
|  | def get_build_info_list(self, params): | 
|  | """Get a list of build infos for this pipeline | 
|  | See https://buildkite.com/docs/apis/rest-api/builds#list-builds-for-a-pipeline | 
|  |  | 
|  | Parameters | 
|  | ---------- | 
|  | params : the parameters to filter the result | 
|  |  | 
|  | Returns | 
|  | ------- | 
|  | list of dict | 
|  | the metadata for a list of builds | 
|  | """ | 
|  | url = self._BUILD_STATUS_URL_TEMPLATE.format(self._org, self._pipeline, "") | 
|  | output = self._open_url(url, params) | 
|  | return json.loads(output) | 
|  |  | 
|  | def get_build_log(self, job): | 
|  | return self._open_url(job["raw_log_url"]) | 
|  |  | 
|  | @staticmethod | 
|  | def _check_response(response, expected_status_code): | 
|  | if response.status_code != expected_status_code: | 
|  | eprint("Exit code:", response.status_code) | 
|  | eprint("Response:\n", response.text) | 
|  | response.raise_for_status() | 
|  |  | 
|  | def trigger_new_build(self, commit, message=None, env={}): | 
|  | """Trigger a new build at a given commit and return the build metadata. | 
|  | See https://buildkite.com/docs/apis/rest-api/builds#create-a-build | 
|  |  | 
|  | Parameters | 
|  | ---------- | 
|  | commit : the commit we want to build at | 
|  | message : the message we should as the build titile | 
|  | env : (optional) the environment variables to set | 
|  |  | 
|  | Returns | 
|  | ------- | 
|  | dict | 
|  | the metadata for the build | 
|  | """ | 
|  | pipeline_info = self.get_pipeline_info() | 
|  | if not pipeline_info: | 
|  | raise BuildkiteException(f"Cannot find pipeline info for pipeline {self._pipeline}.") | 
|  |  | 
|  | url = self._NEW_BUILD_URL_TEMPLATE.format(self._org, self._pipeline) | 
|  | data = { | 
|  | "commit": commit, | 
|  | "branch": pipeline_info.get("default_branch") or "master", | 
|  | "message": message if message else f"Trigger build at {commit}", | 
|  | "env": env, | 
|  | "ignore_pipeline_branch_filters": "true", | 
|  | } | 
|  | response = requests.post(url + "?access_token=" + self._token, json=data) | 
|  | BuildkiteClient._check_response(response, requests.codes.created) | 
|  | return json.loads(response.text) | 
|  |  | 
|  | def trigger_job_retry(self, build_number, job_id): | 
|  | """Trigger a job retry and return the job metadata. | 
|  | See https://buildkite.com/docs/apis/rest-api/jobs#retry-a-job | 
|  |  | 
|  | Parameters | 
|  | ---------- | 
|  | build_number : the number of the build we want to retry | 
|  | job_id : the id of the job we want to retry | 
|  |  | 
|  | Returns | 
|  | ------- | 
|  | dict | 
|  | the metadata for the job | 
|  | """ | 
|  | url = self._RETRY_JOB_URL_TEMPLATE.format(self._org, self._pipeline, build_number, job_id) | 
|  | response = requests.put(url + "?access_token=" + self._token) | 
|  | BuildkiteClient._check_response(response, requests.codes.ok) | 
|  | return json.loads(response.text) | 
|  |  | 
|  | def wait_job_to_finish(self, build_number, job_id, interval_time=30, logger=None): | 
|  | """Wait a job to finish and return the job metadata | 
|  |  | 
|  | Parameters | 
|  | ---------- | 
|  | build_number : the number of the build we want to wait | 
|  | job_id : the id of the job we want to wait | 
|  | interval_time : (optional) the interval time to check the build status, default to 30s | 
|  | logger : (optional) a logger to report progress | 
|  |  | 
|  | Returns | 
|  | ------- | 
|  | dict | 
|  | the latest metadata for the job | 
|  | """ | 
|  | t = 0 | 
|  | build_info = self.get_build_info(build_number) | 
|  | while True: | 
|  | for job in build_info["jobs"]: | 
|  | if job["id"] == job_id: | 
|  | state = job["state"] | 
|  | if state != "scheduled" and state != "running" and state != "assigned": | 
|  | return job | 
|  | break | 
|  | else: | 
|  | raise BuildkiteException( | 
|  | f"job id {job_id} doesn't exist in build " + build_info["web_url"] | 
|  | ) | 
|  | url = build_info["web_url"] | 
|  | if logger: | 
|  | logger.log(f"Waiting for {url}, waited {t} seconds...") | 
|  | time.sleep(interval_time) | 
|  | t += interval_time | 
|  | build_info = self.get_build_info(build_number) | 
|  |  | 
|  | def wait_build_to_finish(self, build_number, interval_time=30, logger=None): | 
|  | """Wait a build to finish and return the build metadata | 
|  |  | 
|  | Parameters | 
|  | ---------- | 
|  | build_number : the number of the build we want to wait | 
|  | interval_time : (optional) the interval time to check the build status, default to 30s | 
|  | logger : (optional) a logger to report progress | 
|  |  | 
|  | Returns | 
|  | ------- | 
|  | dict | 
|  | the latest metadata for the build | 
|  | """ | 
|  | t = 0 | 
|  | build_info = self.get_build_info(build_number) | 
|  | while build_info["state"] == "scheduled" or build_info["state"] == "running": | 
|  | url = build_info["web_url"] | 
|  | if logger: | 
|  | logger.log(f"Waiting for {url}, waited {t} seconds...") | 
|  | time.sleep(interval_time) | 
|  | t += interval_time | 
|  | build_info = self.get_build_info(build_number) | 
|  | return build_info | 
|  |  | 
|  |  | 
|  | def decrypt_token(encrypted_token, kms_key, project="bazel-untrusted"): | 
|  | try: | 
|  | result = subprocess.run( | 
|  | [ | 
|  | gcloud_command(), | 
|  | "kms", | 
|  | "decrypt", | 
|  | "--project", | 
|  | project, | 
|  | "--location", | 
|  | "global", | 
|  | "--keyring", | 
|  | "buildkite", | 
|  | "--key", | 
|  | kms_key, | 
|  | "--ciphertext-file", | 
|  | "-", | 
|  | "--plaintext-file", | 
|  | "-", | 
|  | ], | 
|  | input=base64.b64decode(encrypted_token), | 
|  | env=os.environ, | 
|  | check=True, | 
|  | stdout=subprocess.PIPE,  # We cannot use capture_output since some workers run Python <3.7 | 
|  | stderr=subprocess.PIPE,  # We cannot use capture_output since some workers run Python <3.7 | 
|  | ) | 
|  | return result.stdout.decode("utf-8").strip() | 
|  | except subprocess.CalledProcessError as ex: | 
|  | cause = ex.stderr.decode("utf-8") | 
|  | raise BuildkiteException(f"Failed to decrypt token:\n{cause}") | 
|  |  | 
|  |  | 
|  | def eprint(*args, **kwargs): | 
|  | """ | 
|  | Print to stderr and flush (just in case). | 
|  | """ | 
|  | print(*args, flush=True, file=sys.stderr, **kwargs) | 
|  |  | 
|  |  | 
|  | def is_windows(): | 
|  | return os.name == "nt" | 
|  |  | 
|  |  | 
|  | def is_mac(): | 
|  | return platform_module.system() == "Darwin" | 
|  |  | 
|  |  | 
|  | def is_lab_machine(): | 
|  | agent = os.getenv("BUILDKITE_AGENT_NAME") | 
|  | return any(p.match(agent) for p in LAB_AGENT_PATTERNS) | 
|  |  | 
|  |  | 
|  | def is_ipv6_mac(): | 
|  | return is_mac() and not is_lab_machine() | 
|  |  | 
|  |  | 
|  | def gsutil_command(): | 
|  | return "gsutil.cmd" if is_windows() else "gsutil" | 
|  |  | 
|  |  | 
|  | def gcloud_command(): | 
|  | return "gcloud.cmd" if is_windows() else "gcloud" | 
|  |  | 
|  |  | 
|  | def match_matrix_attr_pattern(s): | 
|  | return re.match("^\${{\s*(\w+)\s*}}$", s) | 
|  |  | 
|  |  | 
|  | def get_matrix_attributes(task): | 
|  | """Get unexpanded matrix attributes from the given task. | 
|  |  | 
|  | If a value of field matches "${{<name>}}", then <name> is a wanted matrix attribute. | 
|  | eg. platform: ${{ platform }} | 
|  | """ | 
|  | attributes = [] | 
|  | for key, value in task.items(): | 
|  | if type(value) is str: | 
|  | res = match_matrix_attr_pattern(value) | 
|  | if res: | 
|  | attributes.append(res.groups()[0]) | 
|  | return list(set(attributes)) | 
|  |  | 
|  |  | 
|  | def get_combinations(matrix, attributes): | 
|  | """Given a matrix and the wanted attributes, return all possible combinations. | 
|  |  | 
|  | eg. | 
|  | With matrix = {'a': [1, 2], 'b': [1], 'c': [1]}, | 
|  | if attributes = ['a', 'b'], then returns [[('a', 1), ('b', 1)], [('a', 2), ('b', 1)]] | 
|  | if attributes = ['b', 'c'], then returns [[('b', 1), ('c', 1)]] | 
|  | if attributes = ['c'], then returns [[('c', 1)]] | 
|  | """ | 
|  | # Sort the attributes to make the output deterministic. | 
|  | attributes.sort() | 
|  | for attr in attributes: | 
|  | if attr not in matrix: | 
|  | raise BuildkiteException("${{ %s }} is not defined in `matrix` section." % attr) | 
|  | pairs = [[(attr, value) for value in matrix[attr]] for attr in attributes] | 
|  | return sorted(itertools.product(*pairs)) | 
|  |  | 
|  |  | 
|  | def get_expanded_task(task, combination): | 
|  | """Expand a task with the given combination of values of attributes.""" | 
|  | combination = dict(combination) | 
|  | expanded_task = copy.deepcopy(task) | 
|  | for key, value in task.items(): | 
|  | if type(value) is str: | 
|  | res = match_matrix_attr_pattern(value) | 
|  | if res: | 
|  | attr = res.groups()[0] | 
|  | expanded_task[key] = combination[attr] | 
|  | return expanded_task | 
|  |  | 
|  |  | 
|  | def fetch_configs(http_url, file_config): | 
|  | """ | 
|  | If specified fetches the build configuration from file_config or http_url, else tries to | 
|  | read it from .bazelci/presubmit.yml. | 
|  | Returns the json configuration as a python data structure. | 
|  | """ | 
|  | if file_config is not None and http_url is not None: | 
|  | raise BuildkiteException("file_config and http_url cannot be set at the same time") | 
|  |  | 
|  | return load_config(http_url, file_config) | 
|  |  | 
|  |  | 
|  | def expand_task_config(config): | 
|  | # Expand tasks that uses attributes defined in the matrix section. | 
|  | # The original task definition expands to multiple tasks for each possible combination. | 
|  | tasks_to_expand = [] | 
|  | expanded_tasks = {} | 
|  | matrix = config.pop("matrix", {}) | 
|  | for key, value in matrix.items(): | 
|  | if type(key) is not str or type(value) is not list: | 
|  | raise BuildkiteException("Expect `matrix` is a map of str -> list") | 
|  |  | 
|  | for task in config["tasks"]: | 
|  | attributes = get_matrix_attributes(config["tasks"][task]) | 
|  | if attributes: | 
|  | tasks_to_expand.append(task) | 
|  | count = 1 | 
|  | for combination in get_combinations(matrix, attributes): | 
|  | expanded_task_name = "%s_config_%.2d" % (task, count) | 
|  | count += 1 | 
|  | expanded_tasks[expanded_task_name] = get_expanded_task( | 
|  | config["tasks"][task], combination | 
|  | ) | 
|  |  | 
|  | for task in tasks_to_expand: | 
|  | config["tasks"].pop(task) | 
|  | config["tasks"].update(expanded_tasks) | 
|  |  | 
|  |  | 
|  | def load_config(http_url, file_config, allow_imports=True): | 
|  | if http_url: | 
|  | config = load_remote_yaml_file(http_url) | 
|  | else: | 
|  | file_config = file_config or ".bazelci/presubmit.yml" | 
|  | with open(file_config, "r") as fd: | 
|  | config = yaml.safe_load(fd) | 
|  |  | 
|  | # Legacy mode means that there is exactly one task per platform (e.g. ubuntu1604_nojdk), | 
|  | # which means that we can get away with using the platform name as task ID. | 
|  | # No other updates are needed since get_platform_for_task() falls back to using the | 
|  | # task ID as platform if there is no explicit "platforms" field. | 
|  | if "platforms" in config: | 
|  | config["tasks"] = config.pop("platforms") | 
|  |  | 
|  | if "tasks" not in config: | 
|  | config["tasks"] = {} | 
|  |  | 
|  | expand_task_config(config) | 
|  |  | 
|  | imports = config.pop("imports", None) | 
|  | if imports: | 
|  | if not allow_imports: | 
|  | raise BuildkiteException("Nested imports are not allowed") | 
|  |  | 
|  | for i in imports: | 
|  | imported_tasks = load_imported_tasks(i, http_url, file_config) | 
|  | config["tasks"].update(imported_tasks) | 
|  |  | 
|  | if len(config["tasks"]) > MAX_TASK_NUMBER: | 
|  | raise BuildkiteException( | 
|  | "The number of tasks in one config file is limited to %s!" % MAX_TASK_NUMBER | 
|  | ) | 
|  |  | 
|  | return config | 
|  |  | 
|  |  | 
|  | def load_remote_yaml_file(http_url): | 
|  | with urllib.request.urlopen(http_url) as resp: | 
|  | reader = codecs.getreader("utf-8") | 
|  | return yaml.safe_load(reader(resp)) | 
|  |  | 
|  |  | 
|  | def load_imported_tasks(import_name, http_url, file_config): | 
|  | if "/" in import_name: | 
|  | raise BuildkiteException("Invalid import '%s'" % import_name) | 
|  |  | 
|  | old_path = http_url or file_config | 
|  | new_path = "%s%s" % (old_path[: old_path.rfind("/") + 1], import_name) | 
|  | if http_url: | 
|  | http_url = new_path | 
|  | else: | 
|  | file_config = new_path | 
|  |  | 
|  | imported_config = load_config(http_url=http_url, file_config=file_config, allow_imports=False) | 
|  |  | 
|  | namespace = import_name.partition(".")[0] | 
|  | tasks = {} | 
|  | for task_name, task_config in imported_config["tasks"].items(): | 
|  | fix_imported_task_platform(task_name, task_config) | 
|  | fix_imported_task_name(namespace, task_config) | 
|  | fix_imported_task_working_directory(namespace, task_config) | 
|  | tasks["%s_%s" % (namespace, task_name)] = task_config | 
|  |  | 
|  | return tasks | 
|  |  | 
|  |  | 
|  | def fix_imported_task_platform(task_name, task_config): | 
|  | if "platform" not in task_config: | 
|  | task_config["platform"] = task_name | 
|  |  | 
|  |  | 
|  | def fix_imported_task_name(namespace, task_config): | 
|  | old_name = task_config.get("name") | 
|  | task_config["name"] = "%s (%s)" % (namespace, old_name) if old_name else namespace | 
|  |  | 
|  |  | 
|  | def fix_imported_task_working_directory(namespace, task_config): | 
|  | old_dir = task_config.get("working_directory") | 
|  | task_config["working_directory"] = os.path.join(namespace, old_dir) if old_dir else namespace | 
|  |  | 
|  |  | 
|  | def print_collapsed_group(name): | 
|  | eprint("\n\n--- {0}\n\n".format(name)) | 
|  |  | 
|  |  | 
|  | def print_expanded_group(name): | 
|  | eprint("\n\n+++ {0}\n\n".format(name)) | 
|  |  | 
|  |  | 
|  | def is_trueish(s): | 
|  | return str(s).lower() in ["true", "1", "t", "y", "yes"] | 
|  |  | 
|  |  | 
|  | def use_bazelisk_migrate(): | 
|  | """ | 
|  | If USE_BAZELISK_MIGRATE is set, we use `bazelisk --migrate` to test incompatible flags. | 
|  | """ | 
|  | return is_trueish(os.environ.get("USE_BAZELISK_MIGRATE")) | 
|  |  | 
|  |  | 
|  | def is_downstream_pipeline(): | 
|  | """ | 
|  | Return true if BAZELCI_DOWNSTREAM_PIPELINE is set | 
|  | """ | 
|  | return is_trueish(os.environ.get("BAZELCI_DOWNSTREAM_PIPELINE")) | 
|  |  | 
|  |  | 
|  | def local_run_only(): | 
|  | """ | 
|  | If BAZELCI_LOCAL_RUN is set, run bazelci in local-only mode, with no attempt | 
|  | to use remote cache/execution, and no attempt to upload test results. Only | 
|  | compatible with `bazelci.py runner`, not other subcommands. | 
|  | """ | 
|  | return is_trueish(os.environ.get("BAZELCI_LOCAL_RUN", "false")) | 
|  |  | 
|  |  | 
|  | def bazelisk_flags(): | 
|  | return ["--migrate"] if use_bazelisk_migrate() else [] | 
|  |  | 
|  |  | 
|  | def calculate_flags(task_config, task_config_key, action_key, tmpdir, test_env_vars): | 
|  | include_json_profile = task_config.get("include_json_profile", []) | 
|  | capture_corrupted_outputs = task_config.get("capture_corrupted_outputs", []) | 
|  |  | 
|  | json_profile_flags = [] | 
|  | json_profile_out = None | 
|  | if action_key in include_json_profile: | 
|  | json_profile_out = os.path.join(tmpdir, "{}.profile.gz".format(action_key)) | 
|  | json_profile_flags = ["--profile={}".format(json_profile_out)] | 
|  |  | 
|  | capture_corrupted_outputs_flags = [] | 
|  | capture_corrupted_outputs_dir = None | 
|  | if action_key in capture_corrupted_outputs: | 
|  | capture_corrupted_outputs_dir = os.path.join( | 
|  | tmpdir, "{}_corrupted_outputs".format(action_key) | 
|  | ) | 
|  | capture_corrupted_outputs_flags = [ | 
|  | "--experimental_remote_capture_corrupted_outputs={}".format( | 
|  | capture_corrupted_outputs_dir | 
|  | ) | 
|  | ] | 
|  |  | 
|  | flags = list(task_config.get(task_config_key, [])) | 
|  | flags += json_profile_flags | 
|  | flags += capture_corrupted_outputs_flags | 
|  | # We have to add --test_env flags to `build`, too, otherwise Bazel | 
|  | # discards its analysis cache between `build` and `test`. | 
|  | if test_env_vars: | 
|  | flags += ["--test_env={}".format(v) for v in test_env_vars] | 
|  |  | 
|  | return flags, json_profile_out, capture_corrupted_outputs_dir | 
|  |  | 
|  |  | 
|  | def execute_commands( | 
|  | task_config, | 
|  | platform, | 
|  | use_bazel_at_commit, | 
|  | use_but, | 
|  | save_but, | 
|  | needs_clean, | 
|  | build_only, | 
|  | test_only, | 
|  | monitor_flaky_tests, | 
|  | bazel_version=None, | 
|  | ): | 
|  | if use_bazelisk_migrate(): | 
|  | # If we are testing incompatible flags with Bazelisk, | 
|  | # use Bazel@last_green if USE_BAZEL_VERSION env var is not set explicitly. | 
|  | if "USE_BAZEL_VERSION" not in os.environ: | 
|  | bazel_version = "last_green" | 
|  |  | 
|  | # Override use_but in case we are in the downstream pipeline so that it doesn't try to | 
|  | # download Bazel built from previous jobs. | 
|  | use_but = False | 
|  |  | 
|  | # Set BAZELISK_INCOMPATIBLE_FLAGS to tell Bazelisk which flags to test. | 
|  | os.environ["BAZELISK_INCOMPATIBLE_FLAGS"] = ",".join(fetch_incompatible_flags().keys()) | 
|  |  | 
|  | if not bazel_version: | 
|  | # The last good version of Bazel can be specified in an emergency file. | 
|  | # However, we only use last_good_bazel for pipelines that do not | 
|  | # explicitly specify a version of Bazel. | 
|  | try: | 
|  | emergency_settings = load_remote_yaml_file(EMERGENCY_FILE_URL) | 
|  | bazel_version = emergency_settings.get("last_good_bazel") | 
|  | except urllib.error.HTTPError: | 
|  | # Ignore this error. The Setup step will have already complained about | 
|  | # it by showing an error message. | 
|  | pass | 
|  |  | 
|  | if build_only and test_only: | 
|  | raise BuildkiteException("build_only and test_only cannot be true at the same time") | 
|  |  | 
|  | if use_but: | 
|  | if use_bazel_at_commit: | 
|  | raise BuildkiteException("use_bazel_at_commit cannot be set when use_but is true") | 
|  |  | 
|  | print_collapsed_group(":printer: Printing task config for downstream job...") | 
|  | eprint(json.dumps(task_config, indent=2)) | 
|  |  | 
|  | tmpdir = tempfile.mkdtemp() | 
|  | sc_process = None | 
|  | try: | 
|  | if is_mac(): | 
|  | activate_xcode(task_config) | 
|  |  | 
|  | # If the CI worker runs Bazelisk, we need to forward all required env variables to the test. | 
|  | # Otherwise any integration test that invokes Bazel (=Bazelisk in this case) will fail. | 
|  | test_env_vars = ["LocalAppData"] if platform == "windows" else ["HOME"] | 
|  |  | 
|  | # CI should have its own user agent so that we can remove it from Bazel download statistics. | 
|  | os.environ["BAZELISK_USER_AGENT"] = "Bazelisk/BazelCI" | 
|  | test_env_vars.append("BAZELISK_USER_AGENT") | 
|  |  | 
|  | # Avoid "Network is unreachable" errors in IPv6-only environments | 
|  | for e in ("COURSIER_OPTS", "JAVA_TOOL_OPTIONS", "SSL_CERT_FILE"): | 
|  | if os.getenv(e): | 
|  | test_env_vars.append(e) | 
|  |  | 
|  | # We use one binary for all Linux platforms (because we also just release one binary for all | 
|  | # Linux versions and we have to ensure that it works on all of them). | 
|  | binary_platform = platform if is_mac() or is_windows() else LINUX_BINARY_PLATFORM | 
|  |  | 
|  | bazel_binary = "bazel" | 
|  | if use_bazel_at_commit: | 
|  | print_collapsed_group(":gcloud: Downloading Bazel built at " + use_bazel_at_commit) | 
|  | # Linux binaries are published under platform name "centos7" | 
|  | if binary_platform == LINUX_BINARY_PLATFORM: | 
|  | binary_platform = "centos7" | 
|  | os.environ["USE_BAZEL_VERSION"] = download_bazel_binary_at_commit( | 
|  | tmpdir, binary_platform, use_bazel_at_commit | 
|  | ) | 
|  | print_collapsed_group(":bazel: Using Bazel at " + os.environ["USE_BAZEL_VERSION"]) | 
|  | elif use_but: | 
|  | print_collapsed_group(":gcloud: Downloading Bazel Under Test") | 
|  | os.environ["USE_BAZEL_VERSION"] = download_bazel_binary(tmpdir, binary_platform) | 
|  | print_collapsed_group(":bazel: Using Bazel at " + os.environ["USE_BAZEL_VERSION"]) | 
|  | else: | 
|  | print_collapsed_group(":bazel: Using Bazel version " + bazel_version) | 
|  | if bazel_version: | 
|  | os.environ["USE_BAZEL_VERSION"] = bazel_version | 
|  | if "USE_BAZEL_VERSION" in os.environ and not task_config.get( | 
|  | "skip_use_bazel_version_for_test", False | 
|  | ): | 
|  | # This will only work if the bazel binary in $PATH is actually a bazelisk binary | 
|  | # (https://github.com/bazelbuild/bazelisk). | 
|  | test_env_vars.append("USE_BAZEL_VERSION") | 
|  |  | 
|  | for key, value in task_config.get("environment", {}).items(): | 
|  | # We have to explicitly convert the value to a string, because sometimes YAML tries to | 
|  | # be smart and converts strings like "true" and "false" to booleans. | 
|  | os.environ[key] = os.path.expandvars(str(value)) | 
|  |  | 
|  | # Set BAZELISK_SHUTDOWN to 1 when we use bazelisk --migrate on Windows. | 
|  | # This is a workaround for https://github.com/bazelbuild/continuous-integration/issues/1012 | 
|  | if use_bazelisk_migrate() and platform == "windows": | 
|  | os.environ["BAZELISK_SHUTDOWN"] = "1" | 
|  |  | 
|  | def PrepareRepoInCwd(print_cmd_groups, initial_setup=False): | 
|  | # Allow the config to override the current working directory. | 
|  | requested_working_dir = task_config.get("working_directory") | 
|  | if requested_working_dir: | 
|  | if os.path.isabs(requested_working_dir): | 
|  | raise BuildkiteException( | 
|  | f"working_directory must be relative to the repository root, " | 
|  | "but was {requested_working_dir}" | 
|  | ) | 
|  |  | 
|  | full_requested_working_dir = os.path.abspath(requested_working_dir) | 
|  | if not os.path.isdir(full_requested_working_dir): | 
|  | raise BuildkiteException( | 
|  | f"{full_requested_working_dir} does not exist or is not a directory" | 
|  | ) | 
|  |  | 
|  | os.chdir(full_requested_working_dir) | 
|  |  | 
|  | # Dirty workaround for #1660 | 
|  | if initial_setup: | 
|  | # Set OUTPUT_BASE environment variable | 
|  | os.environ["OUTPUT_BASE"] = get_output_base(bazel_binary) | 
|  |  | 
|  | cmd_exec_func = ( | 
|  | execute_batch_commands if platform == "windows" else execute_shell_commands | 
|  | ) | 
|  | cmd_exec_func(task_config.get("setup", None)) | 
|  |  | 
|  | if platform == "windows": | 
|  | execute_batch_commands(task_config.get("batch_commands", None), print_cmd_groups) | 
|  | else: | 
|  | execute_shell_commands(task_config.get("shell_commands", None), print_cmd_groups) | 
|  |  | 
|  | PrepareRepoInCwd(True, initial_setup=True) | 
|  |  | 
|  | bazel_version = print_bazel_version_info(bazel_binary, platform) | 
|  |  | 
|  | print_environment_variables_info() | 
|  |  | 
|  | execute_bazel_run(bazel_binary, platform, task_config.get("run_targets", None)) | 
|  |  | 
|  | if needs_clean: | 
|  | execute_bazel_clean(bazel_binary, platform) | 
|  |  | 
|  | # we need the commit here in order to calculate the correct targets. | 
|  | git_commit = os.getenv("BUILDKITE_COMMIT") | 
|  | if not git_commit: | 
|  | raise BuildkiteInfraException("Unable to determine Git commit for this build") | 
|  |  | 
|  | test_flags, json_profile_out_test, capture_corrupted_outputs_dir_test = calculate_flags( | 
|  | task_config, "test_flags", "test", tmpdir, test_env_vars | 
|  | ) | 
|  |  | 
|  | build_targets, test_targets, coverage_targets, index_targets = calculate_targets( | 
|  | task_config, | 
|  | bazel_binary, | 
|  | build_only, | 
|  | test_only, | 
|  | os.getcwd(), | 
|  | PrepareRepoInCwd, | 
|  | git_commit, | 
|  | test_flags, | 
|  | ) | 
|  |  | 
|  | if build_targets: | 
|  | ( | 
|  | build_flags, | 
|  | json_profile_out_build, | 
|  | capture_corrupted_outputs_dir_build, | 
|  | ) = calculate_flags(task_config, "build_flags", "build", tmpdir, test_env_vars) | 
|  | try: | 
|  | release_name = get_release_name_from_branch_name() | 
|  | execute_bazel_build( | 
|  | bazel_version, | 
|  | bazel_binary, | 
|  | platform, | 
|  | build_flags | 
|  | + ( | 
|  | ["--stamp", "--embed_label=%s" % release_name] | 
|  | if save_but and release_name | 
|  | else [] | 
|  | ), | 
|  | build_targets, | 
|  | None, | 
|  | ) | 
|  | if save_but: | 
|  | upload_bazel_binary(platform) | 
|  | finally: | 
|  | if json_profile_out_build: | 
|  | upload_json_profile(json_profile_out_build, tmpdir) | 
|  | if capture_corrupted_outputs_dir_build: | 
|  | upload_corrupted_outputs(capture_corrupted_outputs_dir_build, tmpdir) | 
|  |  | 
|  | if test_targets: | 
|  | if not is_windows(): | 
|  | # On platforms that support sandboxing (Linux, MacOS) we have | 
|  | # to allow access to Bazelisk's cache directory. | 
|  | # However, the flag requires the directory to exist, | 
|  | # so we create it here in order to not crash when a test | 
|  | # does not invoke Bazelisk. | 
|  | bazelisk_cache_dir = get_bazelisk_cache_directory() | 
|  | os.makedirs(bazelisk_cache_dir, mode=0o755, exist_ok=True) | 
|  | test_flags.append("--sandbox_writable_path={}".format(bazelisk_cache_dir)) | 
|  |  | 
|  | # Set BUILDKITE_ANALYTICS_TOKEN so that bazelci-agent can upload test results to Test Analytics | 
|  | if "ENCRYPTED_BUILDKITE_ANALYTICS_TOKEN" in os.environ: | 
|  | if THIS_IS_TESTING: | 
|  | kms_key = "buildkite-testing-api-token" | 
|  | project = "bazel-untrusted" | 
|  | elif THIS_IS_TRUSTED: | 
|  | kms_key = "buildkite-trusted-api-token" | 
|  | project = "bazel-public" | 
|  | else: | 
|  | kms_key = "buildkite-untrusted-api-token" | 
|  | project = "bazel-untrusted" | 
|  |  | 
|  | try: | 
|  | os.environ["BUILDKITE_ANALYTICS_TOKEN"] = decrypt_token( | 
|  | encrypted_token=os.environ["ENCRYPTED_BUILDKITE_ANALYTICS_TOKEN"], | 
|  | kms_key=kms_key, | 
|  | project=project, | 
|  | ) | 
|  | except Exception as ex: | 
|  | print_collapsed_group( | 
|  | ":rotating_light: Test analytics disabled due to an error :warning:" | 
|  | ) | 
|  | eprint(ex) | 
|  |  | 
|  | test_bep_file = os.path.join(tmpdir, _TEST_BEP_FILE) | 
|  | # Create an empty test_bep_file so that the bazelci-agent can start to follow the file right away. Otherwise, | 
|  | # there is a race between when bazelci-agent starts to read the file and when Bazel creates the file. | 
|  | open(test_bep_file, "w").close() | 
|  | with concurrent.futures.ThreadPoolExecutor() as executor: | 
|  | future = executor.submit( | 
|  | upload_test_logs_from_bep, test_bep_file, tmpdir, monitor_flaky_tests | 
|  | ) | 
|  | try: | 
|  | execute_bazel_test( | 
|  | bazel_version, | 
|  | bazel_binary, | 
|  | platform, | 
|  | test_flags, | 
|  | test_targets, | 
|  | test_bep_file, | 
|  | monitor_flaky_tests, | 
|  | ) | 
|  | finally: | 
|  | if json_profile_out_test: | 
|  | upload_json_profile(json_profile_out_test, tmpdir) | 
|  | if capture_corrupted_outputs_dir_test: | 
|  | upload_corrupted_outputs(capture_corrupted_outputs_dir_test, tmpdir) | 
|  |  | 
|  | _ = future.result() | 
|  | # TODO: print results | 
|  |  | 
|  | if coverage_targets: | 
|  | ( | 
|  | coverage_flags, | 
|  | json_profile_out_coverage, | 
|  | capture_corrupted_outputs_dir_coverage, | 
|  | ) = calculate_flags(task_config, "coverage_flags", "coverage", tmpdir, test_env_vars) | 
|  | try: | 
|  | execute_bazel_coverage( | 
|  | bazel_version, | 
|  | bazel_binary, | 
|  | platform, | 
|  | coverage_flags, | 
|  | coverage_targets, | 
|  | ) | 
|  | finally: | 
|  | if json_profile_out_coverage: | 
|  | upload_json_profile(json_profile_out_coverage, tmpdir) | 
|  | if capture_corrupted_outputs_dir_coverage: | 
|  | upload_corrupted_outputs(capture_corrupted_outputs_dir_coverage, tmpdir) | 
|  |  | 
|  | if index_targets: | 
|  | ( | 
|  | index_flags, | 
|  | json_profile_out_index, | 
|  | capture_corrupted_outputs_dir_index, | 
|  | ) = calculate_flags(task_config, "index_flags", "index", tmpdir, test_env_vars) | 
|  | index_upload_policy = task_config.get("index_upload_policy", "IfBuildSuccess") | 
|  | index_upload_gcs = task_config.get("index_upload_gcs", False) | 
|  |  | 
|  | try: | 
|  | should_upload_kzip = ( | 
|  | True if index_upload_policy == INDEX_UPLOAD_POLICY_ALWAYS else False | 
|  | ) | 
|  | try: | 
|  | execute_bazel_build_with_kythe( | 
|  | bazel_version, | 
|  | bazel_binary, | 
|  | platform, | 
|  | index_flags, | 
|  | index_targets, | 
|  | None, | 
|  | ) | 
|  |  | 
|  | if index_upload_policy == INDEX_UPLOAD_POLICY_IF_BUILD_SUCCESS: | 
|  | should_upload_kzip = True | 
|  | except subprocess.CalledProcessError as e: | 
|  | # If not running with Always policy, raise the build error. | 
|  | if index_upload_policy != INDEX_UPLOAD_POLICY_ALWAYS: | 
|  | handle_bazel_failure(e, "build") | 
|  |  | 
|  | if should_upload_kzip and not is_pull_request(): | 
|  | try: | 
|  | merge_and_upload_kythe_kzip(platform, index_upload_gcs) | 
|  | except subprocess.CalledProcessError: | 
|  | raise BuildkiteException("Failed to upload kythe kzip") | 
|  | finally: | 
|  | if json_profile_out_index: | 
|  | upload_json_profile(json_profile_out_index, tmpdir) | 
|  | if capture_corrupted_outputs_dir_index: | 
|  | upload_corrupted_outputs(capture_corrupted_outputs_dir_index, tmpdir) | 
|  |  | 
|  | if platform == "windows": | 
|  | execute_batch_commands( | 
|  | task_config.get("post_batch_commands", None), | 
|  | True, | 
|  | ":batch: Post Processing (Batch Commands)", | 
|  | ) | 
|  | else: | 
|  | execute_shell_commands( | 
|  | task_config.get("post_shell_commands", None), | 
|  | True, | 
|  | ":bash: Post Processing (Shell Commands)", | 
|  | ) | 
|  |  | 
|  | finally: | 
|  | terminate_background_process(sc_process) | 
|  | if tmpdir: | 
|  | shutil.rmtree(tmpdir) | 
|  |  | 
|  |  | 
|  | def activate_xcode(task_config): | 
|  | supported_versions = sorted( | 
|  | # Stripping "Xcode" prefix and ".app" suffix from e.g. "Xcode12.0.1.app" leaves just the version number. | 
|  | [os.path.basename(x)[5:-4] for x in glob("/Applications/Xcode*.app")], | 
|  | ) | 
|  | if not supported_versions: | 
|  | raise BuildkiteInfraException("Could not find a valid Xcode installation.") | 
|  |  | 
|  | # Get the Xcode version from the config. Can be None. | 
|  | wanted_xcode_version = get_requested_xcode_version(task_config) | 
|  |  | 
|  | # This is used to replace e.g. 11.2 with 11.2.1 without having to update all configs. | 
|  | xcode_version = XCODE_VERSION_OVERRIDES.get(wanted_xcode_version, wanted_xcode_version) | 
|  |  | 
|  | # Default to the latest installed version if no version was requested | 
|  | #  or the requested version is not installed. | 
|  | if xcode_version not in supported_versions: | 
|  | xcode_version = supported_versions[-1] | 
|  |  | 
|  | if not wanted_xcode_version or wanted_xcode_version == xcode_version: | 
|  | print_collapsed_group(":xcode: Activating Xcode {}...".format(xcode_version)) | 
|  | else: | 
|  | print_collapsed_group( | 
|  | ":xcode: Fixed Xcode version: {} -> {}...".format(wanted_xcode_version, xcode_version) | 
|  | ) | 
|  | lines = [ | 
|  | "Your selected Xcode version {} was not available on the machine.".format( | 
|  | wanted_xcode_version | 
|  | ), | 
|  | "Bazel CI automatically picked a fallback version: {}.".format(xcode_version), | 
|  | "Available versions are: {}.".format(", ".join(supported_versions)), | 
|  | ] | 
|  | execute_command( | 
|  | [ | 
|  | "buildkite-agent", | 
|  | "annotate", | 
|  | "--style=warning", | 
|  | "\n".join(lines), | 
|  | "--context", | 
|  | "ctx-xcode_version_fixed", | 
|  | ] | 
|  | ) | 
|  |  | 
|  | # Now activate the specified Xcode version and let it install its required components. | 
|  | # The CI machines have a sudoers config that allows the 'buildkite' user to run exactly | 
|  | # these two commands, so don't change them without also modifying the file there. | 
|  | xcode_path = "/Applications/Xcode{}.app".format(xcode_version) | 
|  | execute_command(["/usr/bin/sudo", "/usr/bin/xcode-select", "--switch", xcode_path]) | 
|  | execute_command(["/usr/bin/sudo", "/usr/bin/xcodebuild", "-runFirstLaunch"]) | 
|  |  | 
|  |  | 
|  | def get_requested_xcode_version(task_config): | 
|  | wanted_xcode_version = task_config.get("xcode_version") | 
|  | if not wanted_xcode_version: | 
|  | return None | 
|  |  | 
|  | # Ensure it's a valid version number. | 
|  | if not isinstance(wanted_xcode_version, str): | 
|  | raise BuildkiteException( | 
|  | "Version number '{}' is not a string. Did you forget to put it in quotes?".format( | 
|  | wanted_xcode_version | 
|  | ) | 
|  | ) | 
|  | if not XCODE_VERSION_REGEX.match(wanted_xcode_version): | 
|  | raise BuildkiteException( | 
|  | "Invalid Xcode version format '{}', must match the format X.Y[.Z].".format( | 
|  | wanted_xcode_version | 
|  | ) | 
|  | ) | 
|  |  | 
|  | return wanted_xcode_version | 
|  |  | 
|  |  | 
|  | def get_bazelisk_cache_directory(): | 
|  | # The path relies on the behavior of Go's os.UserCacheDir() | 
|  | # and of the Go version of Bazelisk. | 
|  | cache_dir = "Library/Caches" if is_mac() else ".cache" | 
|  | return os.path.join(os.environ.get("HOME"), cache_dir, "bazelisk") | 
|  |  | 
|  |  | 
|  | def current_branch_is_main_branch(): | 
|  | return os.getenv("BUILDKITE_BRANCH") in ("master", "stable", "main", "google") | 
|  |  | 
|  |  | 
|  | def get_release_name_from_branch_name(): | 
|  | # TODO(pcloudy): Find a better way to do this | 
|  | if os.getenv("BUILDKITE_PIPELINE_SLUG") == "publish-bazel-binaries": | 
|  | return None | 
|  | res = re.match(r"release-(\d+\.\d+\.\d+(rc\d+)?).*", os.getenv("BUILDKITE_BRANCH")) | 
|  | return res.group(1) if res else None | 
|  |  | 
|  |  | 
|  | def is_pull_request(): | 
|  | third_party_repo = os.getenv("BUILDKITE_PULL_REQUEST_REPO", "") | 
|  | return len(third_party_repo) > 0 | 
|  |  | 
|  |  | 
|  | def print_bazel_version_info(bazel_binary, platform): | 
|  | print_collapsed_group(":information_source: Bazel Info") | 
|  | version_output = execute_command_and_get_output( | 
|  | [bazel_binary] + common_startup_flags() + ["--nosystem_rc", "--nohome_rc", "version"] | 
|  | ) | 
|  | execute_command( | 
|  | [bazel_binary] + common_startup_flags() + ["--nosystem_rc", "--nohome_rc", "info"] | 
|  | ) | 
|  |  | 
|  | match = BUILD_LABEL_PATTERN.search(version_output) | 
|  | return match.group(1) if match else "unreleased binary" | 
|  |  | 
|  |  | 
|  | def print_environment_variables_info(): | 
|  | print_collapsed_group(":information_source: Environment Variables") | 
|  | for key, value in os.environ.items(): | 
|  | eprint("%s=(%s)" % (key, value)) | 
|  |  | 
|  |  | 
|  | def upload_bazel_binary(platform): | 
|  | if local_run_only(): | 
|  | return | 
|  | print_collapsed_group(":gcloud: Uploading Bazel Under Test") | 
|  | if platform == "windows": | 
|  | binary_dir = r"bazel-bin\src" | 
|  | binary_name = r"bazel.exe" | 
|  | binary_nojdk_name = r"bazel_nojdk.exe" | 
|  | else: | 
|  | binary_dir = "bazel-bin/src" | 
|  | binary_name = "bazel" | 
|  | binary_nojdk_name = "bazel_nojdk" | 
|  | execute_command(["buildkite-agent", "artifact", "upload", binary_name], cwd=binary_dir) | 
|  | execute_command(["buildkite-agent", "artifact", "upload", binary_nojdk_name], cwd=binary_dir) | 
|  |  | 
|  |  | 
|  | def merge_and_upload_kythe_kzip(platform, index_upload_gcs): | 
|  | if local_run_only(): | 
|  | return | 
|  | print_collapsed_group(":gcloud: Uploading kythe kzip") | 
|  |  | 
|  | kzips = glob("bazel-out/*/extra_actions/**/*.kzip", recursive=True) | 
|  |  | 
|  | build_number = os.getenv("BUILDKITE_BUILD_NUMBER") | 
|  | git_commit = os.getenv("BUILDKITE_COMMIT") | 
|  | final_kzip_name = "{}-{}-{}.kzip".format(build_number, platform, git_commit) | 
|  |  | 
|  | execute_command([f"{KYTHE_DIR}/tools/kzip", "merge", "--output", final_kzip_name] + kzips) | 
|  | execute_command(["buildkite-agent", "artifact", "upload", final_kzip_name]) | 
|  |  | 
|  | if index_upload_gcs: | 
|  | pipeline = os.getenv("BUILDKITE_PIPELINE_SLUG") | 
|  | branch = os.getenv("BUILDKITE_BRANCH") | 
|  | destination = KZIPS_BUCKET + pipeline + "/" + branch + "/" + final_kzip_name | 
|  | print("Uploading to GCS {}".format(destination)) | 
|  | execute_command([gsutil_command(), "cp", final_kzip_name, destination]) | 
|  |  | 
|  |  | 
|  | def download_binary(dest_dir, platform, binary_name): | 
|  | source_step = create_label(platform, "Bazel", build_only=True) | 
|  | execute_command( | 
|  | ["buildkite-agent", "artifact", "download", binary_name, dest_dir, "--step", source_step] | 
|  | ) | 
|  | bazel_binary_path = os.path.join(dest_dir, binary_name) | 
|  | st = os.stat(bazel_binary_path) | 
|  | os.chmod(bazel_binary_path, st.st_mode | stat.S_IEXEC) | 
|  | return bazel_binary_path | 
|  |  | 
|  |  | 
|  | def download_bazel_binary(dest_dir, platform): | 
|  | binary_name = "bazel.exe" if platform == "windows" else "bazel" | 
|  | return download_binary(dest_dir, platform, binary_name) | 
|  |  | 
|  |  | 
|  | def download_bazel_nojdk_binary(dest_dir, platform): | 
|  | binary_name = "bazel_nojdk.exe" if platform == "windows" else "bazel_nojdk" | 
|  | return download_binary(dest_dir, platform, binary_name) | 
|  |  | 
|  |  | 
|  | def download_binary_at_commit(bazel_git_commit, bazel_binary_url, bazel_binary_path): | 
|  | try: | 
|  | execute_command([gsutil_command(), "cp", bazel_binary_url, bazel_binary_path]) | 
|  | except subprocess.CalledProcessError as e: | 
|  | raise BuildkiteInfraException( | 
|  | "Failed to download Bazel binary at %s, error message:\n%s" % (bazel_git_commit, str(e)) | 
|  | ) | 
|  | st = os.stat(bazel_binary_path) | 
|  | os.chmod(bazel_binary_path, st.st_mode | stat.S_IEXEC) | 
|  | return bazel_binary_path | 
|  |  | 
|  |  | 
|  | def download_bazel_binary_at_commit(dest_dir, platform, bazel_git_commit): | 
|  | url = bazelci_builds_gs_url(platform, bazel_git_commit) | 
|  | path = os.path.join(dest_dir, "bazel.exe" if platform == "windows" else "bazel") | 
|  | return download_binary_at_commit(bazel_git_commit, url, path) | 
|  |  | 
|  |  | 
|  | def download_bazel_nojdk_binary_at_commit(dest_dir, platform, bazel_git_commit): | 
|  | url = bazelci_builds_nojdk_gs_url(platform, bazel_git_commit) | 
|  | path = os.path.join(dest_dir, "bazel_nojdk.exe" if platform == "windows" else "bazel_nojdk") | 
|  | return download_binary_at_commit(bazel_git_commit, url, path) | 
|  |  | 
|  |  | 
|  | def download_bazelci_agent(dest_dir): | 
|  | repo = "bazelbuild/continuous-integration" | 
|  | if THIS_IS_TESTING and "BAZELCI_AGENT_REPO" in os.environ: | 
|  | repo = os.environ["BAZELCI_AGENT_REPO"] | 
|  | version = "0.2.2" | 
|  | if THIS_IS_TESTING and "BAZELCI_AGENT_VERSION" in os.environ: | 
|  | version = os.environ["BAZELCI_AGENT_VERSION"] | 
|  | postfix = "" | 
|  | if is_windows(): | 
|  | postfix = "x86_64-pc-windows-msvc.exe" | 
|  | elif is_mac(): | 
|  | if platform_module.machine() == "arm64": | 
|  | postfix = "aarch64-apple-darwin" | 
|  | else: | 
|  | postfix = "x86_64-apple-darwin" | 
|  | else: | 
|  | if platform_module.machine() == "aarch64": | 
|  | postfix = "aarch64-unknown-linux-musl" | 
|  | else: | 
|  | postfix = "x86_64-unknown-linux-musl" | 
|  |  | 
|  | name = "bazelci-agent-{}-{}".format(version, postfix) | 
|  | url = "https://github.com/{}/releases/download/agent-{}/{}".format(repo, version, name) | 
|  | path = os.path.join(dest_dir, "bazelci-agent.exe" if is_windows() else "bazelci-agent") | 
|  | execute_command(["curl", "-sSL", url, "-o", path]) | 
|  | st = os.stat(path) | 
|  | os.chmod(path, st.st_mode | stat.S_IEXEC) | 
|  | return path | 
|  |  | 
|  |  | 
|  | def get_mirror_root(): | 
|  | if is_mac(): | 
|  | return "/usr/local/var/bazelbuild/" | 
|  | elif is_windows(): | 
|  | return "c:\\buildkite\\bazelbuild\\" | 
|  |  | 
|  | return "/var/lib/bazelbuild/" | 
|  |  | 
|  |  | 
|  | def get_repositories_root(): | 
|  | """A root directory for preparing the reposioty to be tested.""" | 
|  | repo_root = "/var/lib/buildkite-agent/builds/${BUILDKITE_AGENT_NAME}/${BUILDKITE_ORGANIZATION_SLUG}-org-repo-root" | 
|  | if is_mac(): | 
|  | repo_root = "/Users/buildkite/builds/${BUILDKITE_AGENT_NAME}/${BUILDKITE_ORGANIZATION_SLUG}-org-repo-root" | 
|  | elif is_windows(): | 
|  | repo_root = "c:/b/${BUILDKITE_AGENT_NAME}/${BUILDKITE_ORGANIZATION_SLUG}-org-repo-root" | 
|  | # If this is a Windows machine with a local SSD, the build directory is on drive D. | 
|  | if os.path.exists("d:/b"): | 
|  | repo_root = repo_root.replace("c:/b/", "d:/b/") | 
|  |  | 
|  | repo_root = os.path.expandvars(repo_root) | 
|  |  | 
|  | if not os.path.exists(repo_root): | 
|  | os.makedirs(repo_root) | 
|  | return repo_root | 
|  |  | 
|  |  | 
|  | def clone_git_repository(git_repository, git_commit=None, suppress_stdout=False): | 
|  | def execute_git_command(args): | 
|  | execute_command(args, print_output=not suppress_stdout, suppress_stdout=suppress_stdout) | 
|  |  | 
|  | root = get_repositories_root() | 
|  | project_name = re.search(r"/([^/]+)\.git$", git_repository).group(1) | 
|  | clone_path = os.path.join(root, project_name) | 
|  | if not suppress_stdout: | 
|  | print_collapsed_group( | 
|  | "Fetching %s sources at %s" % (project_name, git_commit if git_commit else "HEAD") | 
|  | ) | 
|  |  | 
|  | mirror_path = get_mirror_root() + re.sub(r"[^0-9A-Za-z]", "-", git_repository) | 
|  |  | 
|  | if not os.path.exists(clone_path): | 
|  | if os.path.exists(mirror_path): | 
|  | execute_git_command( | 
|  | ["git", "clone", "-v", "--reference", mirror_path, git_repository, clone_path] | 
|  | ) | 
|  | else: | 
|  | execute_git_command(["git", "clone", "-v", git_repository, clone_path]) | 
|  |  | 
|  | os.chdir(clone_path) | 
|  | execute_git_command(["git", "remote", "set-url", "origin", git_repository]) | 
|  | execute_git_command(["git", "clean", "-fdqx"]) | 
|  | execute_git_command(["git", "submodule", "foreach", "--recursive", "git clean -fdqx"]) | 
|  | execute_git_command(["git", "fetch", "origin"]) | 
|  | if git_commit: | 
|  | # sync to a specific commit of this repository | 
|  | execute_git_command(["git", "reset", git_commit, "--hard"]) | 
|  | else: | 
|  | # sync to the latest commit of HEAD. Unlikely git pull this also works after a force push. | 
|  | remote_head = ( | 
|  | subprocess.check_output(["git", "symbolic-ref", "refs/remotes/origin/HEAD"]) | 
|  | .decode("utf-8") | 
|  | .rstrip() | 
|  | ) | 
|  | execute_git_command(["git", "reset", remote_head, "--hard"]) | 
|  | execute_git_command(["git", "submodule", "sync", "--recursive"]) | 
|  | execute_git_command(["git", "submodule", "update", "--init", "--recursive", "--force"]) | 
|  | execute_git_command(["git", "submodule", "foreach", "--recursive", "git reset --hard"]) | 
|  | execute_git_command(["git", "clean", "-fdqx"]) | 
|  | execute_git_command(["git", "submodule", "foreach", "--recursive", "git clean -fdqx"]) | 
|  | return clone_path | 
|  |  | 
|  |  | 
|  | def execute_batch_commands( | 
|  | commands, print_group=True, group_message=":batch: Setup (Batch Commands)" | 
|  | ): | 
|  | if not commands: | 
|  | return | 
|  |  | 
|  | if print_group: | 
|  | print_collapsed_group(group_message) | 
|  |  | 
|  | batch_commands = "&".join(commands) | 
|  | return subprocess.run(batch_commands, shell=True, check=True, env=os.environ).returncode | 
|  |  | 
|  |  | 
|  | def execute_shell_commands( | 
|  | commands, print_group=True, group_message=":bash: Setup (Shell Commands)" | 
|  | ): | 
|  | if not commands: | 
|  | return | 
|  |  | 
|  | if print_group: | 
|  | print_collapsed_group(group_message) | 
|  |  | 
|  | shell_command = "\n".join(["set -e"] + commands) | 
|  | execute_command([shell_command], shell=True) | 
|  |  | 
|  |  | 
|  | def handle_bazel_failure(exception, action): | 
|  | msg = "bazel {0} failed with exit code {1}".format(action, exception.returncode) | 
|  | if use_bazelisk_migrate(): | 
|  | print_collapsed_group(msg) | 
|  | else: | 
|  | raise BuildkiteException(msg) | 
|  |  | 
|  |  | 
|  | def execute_bazel_run(bazel_binary, platform, targets): | 
|  | if not targets: | 
|  | return | 
|  | print_collapsed_group("Setup (Run Targets)") | 
|  | for target in targets: | 
|  | try: | 
|  | execute_command( | 
|  | [bazel_binary] | 
|  | + bazelisk_flags() | 
|  | + common_startup_flags() | 
|  | + ["run"] | 
|  | + common_build_flags(None, platform) | 
|  | + [target] | 
|  | ) | 
|  | except subprocess.CalledProcessError as e: | 
|  | handle_bazel_failure(e, "run") | 
|  |  | 
|  |  | 
|  | def remote_caching_flags(platform, accept_cached=True): | 
|  | # Only enable caching for untrusted and testing builds. | 
|  | if CLOUD_PROJECT != "bazel-untrusted": | 
|  | return [] | 
|  | # We don't enable remote caching on the Linux ARM64 machine since it doesn't have access to GCS. | 
|  | if platform == "ubuntu2004_arm64": | 
|  | return [] | 
|  |  | 
|  | platform_cache_key = [ | 
|  | BUILDKITE_ORG.encode("utf-8"), | 
|  | # Whenever the remote cache was known to have been poisoned increase the number below | 
|  | "cache-poisoning-20230803".encode("utf-8"), | 
|  | platform.encode("utf-8"), | 
|  | ] | 
|  |  | 
|  | if is_mac(): | 
|  | platform_cache_key += [ | 
|  | # macOS version: | 
|  | subprocess.check_output(["/usr/bin/sw_vers", "-productVersion"]), | 
|  | # Path to Xcode: | 
|  | subprocess.check_output(["/usr/bin/xcode-select", "-p"]), | 
|  | # Xcode version: | 
|  | subprocess.check_output(["/usr/bin/xcodebuild", "-version"]), | 
|  | ] | 
|  |  | 
|  | if is_mac() and is_lab_machine(): | 
|  | # Use a local cache server for our physical macOS machines in the lab. | 
|  | flags = ["--remote_cache=grpc://[2a00:79e1:abc:8602:a28c:fdff:fed0:ec39]:9092"] | 
|  | else: | 
|  | # Use GCS for caching builds running on MacService. | 
|  | # Use RBE for caching builds running on GCE. | 
|  | remote_cache_flags = [] | 
|  | if is_mac(): | 
|  | remote_cache_flags = [ | 
|  | "--remote_cache=https://storage.googleapis.com/bazel-unstrusted-build-cache" | 
|  | ] | 
|  | else: | 
|  | remote_cache_flags = [ | 
|  | "--remote_cache=remotebuildexecution.googleapis.com", | 
|  | "--remote_instance_name=projects/{}/instances/default_instance".format( | 
|  | CLOUD_PROJECT | 
|  | ), | 
|  | ] | 
|  | flags = ( | 
|  | remote_cache_flags | 
|  | + [ | 
|  | "--google_default_credentials", | 
|  | ] | 
|  | + ( | 
|  | [] | 
|  | if is_mac() | 
|  | else [  # Re-enable on macOS once b/346751326 is resolved. | 
|  | # Enable BES / Build Results reporting. | 
|  | "--bes_backend=buildeventservice.googleapis.com", | 
|  | "--bes_timeout=360s", | 
|  | "--project_id=bazel-untrusted", | 
|  | ] | 
|  | ) | 
|  | ) | 
|  |  | 
|  | platform_cache_digest = hashlib.sha256() | 
|  | for key in platform_cache_key: | 
|  | eprint("Adding to platform cache key: {}".format(key)) | 
|  | platform_cache_digest.update(key) | 
|  | platform_cache_digest.update(b":") | 
|  |  | 
|  | remote_timeout = 3600 if is_ipv6_mac() else 60 | 
|  | flags += [ | 
|  | f"--remote_timeout={remote_timeout}", | 
|  | "--remote_max_connections=200", | 
|  | '--remote_default_platform_properties=properties:{name:"cache-silo-key" value:"%s"}' | 
|  | % platform_cache_digest.hexdigest(), | 
|  | "--remote_download_toplevel", | 
|  | ] | 
|  |  | 
|  | if not accept_cached: | 
|  | flags += ["--noremote_accept_cached"] | 
|  |  | 
|  | return flags | 
|  |  | 
|  |  | 
|  | def remote_enabled(flags): | 
|  | # Detect if the project configuration enabled its own remote caching / execution. | 
|  | remote_flags = ["--remote_executor", "--remote_cache", "--remote_http_cache"] | 
|  | for flag in flags: | 
|  | for remote_flag in remote_flags: | 
|  | if flag.startswith(remote_flag): | 
|  | return True | 
|  | return False | 
|  |  | 
|  |  | 
|  | def concurrent_jobs(platform): | 
|  | return "75" if platform.startswith("rbe_") else str(multiprocessing.cpu_count()) | 
|  |  | 
|  |  | 
|  | def concurrent_test_jobs(platform): | 
|  | if platform.startswith("rbe_"): | 
|  | return "75" | 
|  | elif is_windows(): | 
|  | return "8" | 
|  | elif is_mac() and THIS_IS_TESTING: | 
|  | return "4" | 
|  | elif is_mac(): | 
|  | # TODO(twerth): This is an experiment, remove. | 
|  | return str(int(multiprocessing.cpu_count() / 2)) | 
|  | return "12" | 
|  |  | 
|  |  | 
|  | def common_startup_flags(): | 
|  | if is_windows(): | 
|  | if os.path.exists("D:/b"): | 
|  | # This machine has a local SSD mounted as drive D. | 
|  | return ["--output_user_root=D:/b"] | 
|  | else: | 
|  | # This machine uses its PD-SSD as the build directory. | 
|  | return ["--output_user_root=C:/b"] | 
|  | elif is_ipv6_mac(): | 
|  | return ["--host_jvm_args=-Djava.net.preferIPv6Addresses=true"] | 
|  | return [] | 
|  |  | 
|  |  | 
|  | def common_build_flags(bep_file, platform): | 
|  | flags = [ | 
|  | "--show_progress_rate_limit=5", | 
|  | "--curses=yes", | 
|  | "--color=yes", | 
|  | "--terminal_columns=143", | 
|  | "--show_timestamps", | 
|  | "--verbose_failures", | 
|  | "--jobs=" + concurrent_jobs(platform), | 
|  | "--announce_rc", | 
|  | "--experimental_repository_cache_hardlinks", | 
|  | # Some projects set --disk_cache in their project-specific bazelrc, which we never want on | 
|  | # CI, so let's just disable it explicitly. | 
|  | "--disk_cache=", | 
|  | ] | 
|  |  | 
|  | if is_windows(): | 
|  | pass | 
|  | elif is_mac(): | 
|  | flags += [ | 
|  | "--sandbox_writable_path=/var/tmp/_bazel_buildkite/cache/repos/v1", | 
|  | "--test_env=REPOSITORY_CACHE=/var/tmp/_bazel_buildkite/cache/repos/v1", | 
|  | ] | 
|  | else: | 
|  | flags += ["--sandbox_tmpfs_path=/tmp"] | 
|  |  | 
|  | if bep_file: | 
|  | flags += [ | 
|  | "--experimental_build_event_json_file_path_conversion=false", | 
|  | "--build_event_json_file=" + bep_file, | 
|  | ] | 
|  |  | 
|  | if is_ipv6_mac(): | 
|  | flags += ["--jvmopt=-Djava.net.preferIPv6Addresses"] | 
|  |  | 
|  | return flags | 
|  |  | 
|  |  | 
|  | def rbe_flags(original_flags, accept_cached): | 
|  | # Enable remote execution via RBE. | 
|  | flags = [ | 
|  | "--remote_executor=remotebuildexecution.googleapis.com", | 
|  | "--remote_instance_name=projects/bazel-untrusted/instances/default_instance", | 
|  | "--remote_timeout=3600", | 
|  | "--incompatible_strict_action_env", | 
|  | "--google_default_credentials", | 
|  | "--toolchain_resolution_debug", | 
|  | "--remote_download_toplevel", | 
|  | ] | 
|  |  | 
|  | # Enable BES / Build Results reporting. | 
|  | flags += [ | 
|  | "--bes_backend=buildeventservice.googleapis.com", | 
|  | "--bes_timeout=360s", | 
|  | "--project_id=bazel-untrusted", | 
|  | ] | 
|  |  | 
|  | if not accept_cached: | 
|  | flags += ["--noremote_accept_cached"] | 
|  |  | 
|  | # Adapted from https://github.com/bazelbuild/bazel-toolchains/blob/master/bazelrc/.bazelrc | 
|  | flags += [ | 
|  | # These should NOT longer need to be modified. | 
|  | # All that is needed is updating the @bazel_toolchains repo pin | 
|  | # in projects' WORKSPACE files. | 
|  | # | 
|  | # Toolchain related flags to append at the end of your .bazelrc file. | 
|  | "--host_javabase=@buildkite_config//java:jdk", | 
|  | "--javabase=@buildkite_config//java:jdk", | 
|  | "--host_java_toolchain=@bazel_tools//tools/jdk:toolchain_hostjdk8", | 
|  | "--java_toolchain=@bazel_tools//tools/jdk:toolchain_hostjdk8", | 
|  | "--crosstool_top=@buildkite_config//cc:toolchain", | 
|  | "--action_env=BAZEL_DO_NOT_DETECT_CPP_TOOLCHAIN=1", | 
|  | ] | 
|  |  | 
|  | # Platform flags: | 
|  | # The toolchain container used for execution is defined in the target indicated | 
|  | # by "extra_execution_platforms", "host_platform" and "platforms". | 
|  | # If you are using your own toolchain container, you need to create a platform | 
|  | # target with "constraint_values" that allow for the toolchain specified with | 
|  | # "extra_toolchains" to be selected (given constraints defined in | 
|  | # "exec_compatible_with"). | 
|  | # More about platforms: https://docs.bazel.build/versions/master/platforms.html | 
|  | # Don't add platform flags if they are specified already. | 
|  | platform_flags = { | 
|  | "--extra_toolchains": "@buildkite_config//config:cc-toolchain", | 
|  | "--extra_execution_platforms": "@buildkite_config//config:platform", | 
|  | "--host_platform": "@buildkite_config//config:platform", | 
|  | "--platforms": "@buildkite_config//config:platform", | 
|  | } | 
|  | for platform_flag, value in list(platform_flags.items()): | 
|  | found = False | 
|  | for original_flag in original_flags: | 
|  | if original_flag.startswith(platform_flag): | 
|  | found = True | 
|  | break | 
|  | if not found: | 
|  | flags += [platform_flag + "=" + value] | 
|  |  | 
|  | return flags | 
|  |  | 
|  |  | 
|  | def get_output_base(bazel_binary): | 
|  | return execute_command_and_get_output( | 
|  | [bazel_binary] + common_startup_flags() + ["info", "output_base"], | 
|  | print_output=False, | 
|  | ).strip() | 
|  |  | 
|  |  | 
|  | def compute_flags(platform, flags, bep_file, bazel_binary, enable_remote_cache=False): | 
|  | aggregated_flags = common_build_flags(bep_file, platform) | 
|  |  | 
|  | if not local_run_only(): | 
|  | if not remote_enabled(flags): | 
|  | if platform.startswith("rbe_"): | 
|  | aggregated_flags += rbe_flags(flags, accept_cached=enable_remote_cache) | 
|  | else: | 
|  | aggregated_flags += remote_caching_flags( | 
|  | platform, accept_cached=enable_remote_cache | 
|  | ) | 
|  | aggregated_flags += flags | 
|  |  | 
|  | for i, flag in enumerate(aggregated_flags): | 
|  | if "$HOME" in flag: | 
|  | if is_windows(): | 
|  | if os.path.exists("D:/"): | 
|  | home = "D:" | 
|  | else: | 
|  | home = "C:/b" | 
|  | elif is_mac(): | 
|  | home = "/Users/buildkite" | 
|  | else: | 
|  | home = "/var/lib/buildkite-agent" | 
|  | aggregated_flags[i] = flag.replace("$HOME", home) | 
|  | if "$OUTPUT_BASE" in flag: | 
|  | output_base = get_output_base(bazel_binary) | 
|  | aggregated_flags[i] = flag.replace("$OUTPUT_BASE", output_base) | 
|  |  | 
|  | if is_downstream_pipeline(): | 
|  | # If we are in a downstream pipeline, turn off the lockfile update since changing Bazel version could affect the lockfile. | 
|  | aggregated_flags.append("--lockfile_mode=off") | 
|  |  | 
|  | return aggregated_flags | 
|  |  | 
|  |  | 
|  | def execute_bazel_clean(bazel_binary, platform): | 
|  | print_expanded_group(":bazel: Clean") | 
|  |  | 
|  | try: | 
|  | execute_command([bazel_binary] + common_startup_flags() + ["clean", "--expunge"]) | 
|  | except subprocess.CalledProcessError as e: | 
|  | raise BuildkiteException("bazel clean failed with exit code {}".format(e.returncode)) | 
|  |  | 
|  |  | 
|  | def kythe_startup_flags(): | 
|  | return [f"--bazelrc={KYTHE_DIR}/extractors.bazelrc"] | 
|  |  | 
|  |  | 
|  | def kythe_build_flags(): | 
|  | return [ | 
|  | "--experimental_convenience_symlinks=normal", | 
|  | f"--override_repository=kythe_release={KYTHE_DIR}", | 
|  | ] | 
|  |  | 
|  |  | 
|  | def execute_bazel_build(bazel_version, bazel_binary, platform, flags, targets, bep_file): | 
|  | print_collapsed_group(":bazel: Computing flags for build step") | 
|  | aggregated_flags = compute_flags( | 
|  | platform, | 
|  | flags, | 
|  | bep_file, | 
|  | bazel_binary, | 
|  | enable_remote_cache=True, | 
|  | ) | 
|  |  | 
|  | print_expanded_group(":bazel: Build ({})".format(bazel_version)) | 
|  | try: | 
|  | execute_command( | 
|  | [bazel_binary] | 
|  | + bazelisk_flags() | 
|  | + common_startup_flags() | 
|  | + ["build"] | 
|  | + aggregated_flags | 
|  | + ["--"] | 
|  | + targets | 
|  | ) | 
|  | except subprocess.CalledProcessError as e: | 
|  | handle_bazel_failure(e, "build") | 
|  |  | 
|  |  | 
|  | def execute_bazel_build_with_kythe(bazel_version, bazel_binary, platform, flags, targets, bep_file): | 
|  | print_collapsed_group(":bazel: Computing flags for build step") | 
|  | aggregated_flags = compute_flags( | 
|  | platform, | 
|  | flags, | 
|  | bep_file, | 
|  | bazel_binary, | 
|  | enable_remote_cache=False, | 
|  | ) | 
|  |  | 
|  | print_expanded_group(":bazel: Build ({})".format(bazel_version)) | 
|  |  | 
|  | execute_command( | 
|  | [bazel_binary] | 
|  | + bazelisk_flags() | 
|  | + common_startup_flags() | 
|  | + kythe_startup_flags() | 
|  | + ["build"] | 
|  | + kythe_build_flags() | 
|  | + aggregated_flags | 
|  | + ["--"] | 
|  | + targets | 
|  | ) | 
|  |  | 
|  |  | 
|  | def calculate_targets( | 
|  | task_config, | 
|  | bazel_binary, | 
|  | build_only, | 
|  | test_only, | 
|  | workspace_dir, | 
|  | ws_setup_func, | 
|  | git_commit, | 
|  | test_flags, | 
|  | ): | 
|  | print_collapsed_group(":dart: Calculating targets") | 
|  |  | 
|  | build_targets = [] if test_only else list(task_config.get("build_targets", [])) | 
|  | test_targets = [] if build_only else list(task_config.get("test_targets", [])) | 
|  | coverage_targets = ( | 
|  | [] if (build_only or test_only) else list(task_config.get("coverage_targets", [])) | 
|  | ) | 
|  | index_targets = [] if (build_only or test_only) else list(task_config.get("index_targets", [])) | 
|  |  | 
|  | index_targets_query = ( | 
|  | None if (build_only or test_only) else task_config.get("index_targets_query", None) | 
|  | ) | 
|  | if index_targets_query: | 
|  | output = execute_command_and_get_output( | 
|  | [bazel_binary] | 
|  | + common_startup_flags() | 
|  | + ["--nosystem_rc", "--nohome_rc", "query", "--lockfile_mode=off", index_targets_query], | 
|  | print_output=False, | 
|  | ) | 
|  | index_targets += output.strip().split("\n") | 
|  |  | 
|  | # Remove the "--" argument splitter from the list that some configs explicitly | 
|  | # include. We'll add it back again later where needed. | 
|  | build_targets = [x.strip() for x in build_targets if x.strip() != "--"] | 
|  | test_targets = [x.strip() for x in test_targets if x.strip() != "--"] | 
|  | coverage_targets = [x.strip() for x in coverage_targets if x.strip() != "--"] | 
|  | index_targets = [x.strip() for x in index_targets if x.strip() != "--"] | 
|  |  | 
|  | diffbase = os.getenv(USE_BAZEL_DIFF_ENV_VAR, "").lower() | 
|  | shard_id = int(os.getenv("BUILDKITE_PARALLEL_JOB", "-1")) | 
|  | shard_count = int(os.getenv("BUILDKITE_PARALLEL_JOB_COUNT", "-1")) | 
|  | sharding_enabled = shard_id > -1 and shard_count > -1 | 
|  |  | 
|  | use_bazel_diff = diffbase and can_use_bazel_diff(git_commit) | 
|  |  | 
|  | # Skip target expansion if we don't need to calculate test targets | 
|  | if not use_bazel_diff and not sharding_enabled: | 
|  | return build_targets, test_targets, coverage_targets, index_targets | 
|  |  | 
|  | # TODO(#1614): Fix target expansion | 
|  | expanded_test_targets = expand_test_target_patterns(bazel_binary, test_targets, test_flags) | 
|  |  | 
|  | actual_test_targets = ( | 
|  | filter_unchanged_targets( | 
|  | expanded_test_targets, | 
|  | workspace_dir, | 
|  | ws_setup_func, | 
|  | bazel_binary, | 
|  | diffbase, | 
|  | git_commit, | 
|  | ) | 
|  | if use_bazel_diff | 
|  | else expanded_test_targets | 
|  | ) | 
|  |  | 
|  | if sharding_enabled: | 
|  | print_collapsed_group( | 
|  | ":female-detective: Calculating targets for shard {}/{}".format( | 
|  | shard_id + 1, shard_count | 
|  | ) | 
|  | ) | 
|  | actual_test_targets = get_targets_for_shard(actual_test_targets, shard_id, shard_count) | 
|  |  | 
|  | return build_targets, actual_test_targets, coverage_targets, index_targets | 
|  |  | 
|  |  | 
|  | def can_use_bazel_diff(git_commit): | 
|  | matched_files = [] | 
|  | for f in get_modified_files(git_commit): | 
|  | for d in DISABLE_BAZEL_DIFF_IF_MODIFIED: | 
|  | if d.endswith("/") and f.startswith(d) or f == d: | 
|  | matched_files.append(f) | 
|  |  | 
|  | if matched_files: | 
|  | eprint( | 
|  | "Cannot enable bazel-diff since the following files were modified:\n\t{}".format( | 
|  | "\n\t".join(sorted(matched_files)) | 
|  | ) | 
|  | ) | 
|  |  | 
|  | return not matched_files | 
|  |  | 
|  |  | 
|  | def expand_test_target_patterns(bazel_binary, test_targets, test_flags): | 
|  | if not test_targets: | 
|  | return [] | 
|  |  | 
|  | print_collapsed_group(":ninja: Resolving test targets via bazel query") | 
|  |  | 
|  | output = execute_command_and_get_output( | 
|  | [bazel_binary] | 
|  | + common_startup_flags() | 
|  | + [ | 
|  | "--nosystem_rc", | 
|  | "--nohome_rc", | 
|  | "cquery" if os.getenv("EXP_USE_CQUERY") else "query", | 
|  | "--lockfile_mode=off", | 
|  | get_test_query(test_targets, test_flags), | 
|  | ], | 
|  | print_output=False, | 
|  | ).strip() | 
|  | return output.split("\n") if output else [] | 
|  |  | 
|  |  | 
|  | def get_test_query(test_targets, test_flags): | 
|  | included_targets, excluded_targets = partition_list(test_targets) | 
|  |  | 
|  | def FormatTargetList(targets): | 
|  | return " ".join("'{}'".format(t) for t in targets) | 
|  |  | 
|  | query = "let t = tests(set({})) in $t".format(FormatTargetList(included_targets)) | 
|  |  | 
|  | if excluded_targets: | 
|  | query += " except tests(set({}))".format(FormatTargetList(excluded_targets)) | 
|  |  | 
|  | included_tags, excluded_tags = get_test_tags(test_flags) | 
|  |  | 
|  | for t in excluded_tags: | 
|  | query += " except attr('tags', '\\b{}\\b', $t)".format(t) | 
|  |  | 
|  | if included_tags: | 
|  | parts = ["attr('tags', '\\b{}\\b', $tt)".format(t) for t in included_tags] | 
|  | query = "let tt = {} in {}".format(query, " union ".join(parts)) | 
|  |  | 
|  | return query | 
|  |  | 
|  |  | 
|  | def get_test_tags(test_flags): | 
|  | wanted_prefix = "--test_tag_filters=" | 
|  |  | 
|  | for f in test_flags: | 
|  | if not f.startswith(wanted_prefix): | 
|  | continue | 
|  |  | 
|  | tags = removeprefix(f, wanted_prefix).split(",") | 
|  | include, exclude = partition_list(tags) | 
|  |  | 
|  | # Skip tests tagged as "manual" by default, unless explicitly requested | 
|  | manual_tag = "manual" | 
|  | if manual_tag not in include and manual_tag not in exclude: | 
|  | exclude.append(manual_tag) | 
|  |  | 
|  | return include, exclude | 
|  |  | 
|  | return [], ["manual"] | 
|  |  | 
|  |  | 
|  | def removeprefix(s, prefix): | 
|  | def rp(p): | 
|  | if s.startswith(p): | 
|  | return s[len(p) :] | 
|  | return s | 
|  |  | 
|  | func = getattr(s, "removeprefix", rp) | 
|  | return func(prefix) | 
|  |  | 
|  |  | 
|  | def filter_unchanged_targets( | 
|  | expanded_test_targets, workspace_dir, ws_setup_func, bazel_binary, diffbase, git_commit | 
|  | ): | 
|  | print_collapsed_group( | 
|  | f":scissors: Filtering targets that haven't been affected since {diffbase}" | 
|  | ) | 
|  |  | 
|  | tmpdir = tempfile.mkdtemp() | 
|  | try: | 
|  | resolved_diffbase = resolve_diffbase(diffbase) | 
|  | eprint(f"Resolved diffbase to {resolved_diffbase}") | 
|  |  | 
|  | eprint("Cloning comparison repository...") | 
|  | diffbase_archive_url = get_commit_archive_url(resolved_diffbase) | 
|  | local_archive_path = download_file(diffbase_archive_url, tmpdir, "repo.tar.gz") | 
|  | diffbase_repo_dir = os.path.join(tmpdir, resolved_diffbase) | 
|  | extract_archive( | 
|  | local_archive_path, | 
|  | diffbase_repo_dir, | 
|  | strip_top_level_dir=not is_googlesource_repo(diffbase_archive_url), | 
|  | ) | 
|  |  | 
|  | eprint("Setting up comparison repository...") | 
|  | os.chdir(diffbase_repo_dir) | 
|  | ws_setup_func(False) | 
|  |  | 
|  | eprint(f"Downloading bazel-diff to {tmpdir}") | 
|  | bazel_diff_path = download_file(BAZEL_DIFF_URL, tmpdir, "bazel-diff.jar") | 
|  | eprint(f"Running bazel-diff for {resolved_diffbase} and {git_commit}") | 
|  |  | 
|  | affected_targets = run_bazel_diff( | 
|  | bazel_diff_path, diffbase_repo_dir, workspace_dir, bazel_binary, tmpdir | 
|  | ) | 
|  | except (BuildkiteException, BuildkiteInfraException) as ex: | 
|  | try: | 
|  | execute_command( | 
|  | [ | 
|  | "buildkite-agent", | 
|  | "annotate", | 
|  | "--style=warning", | 
|  | "--context", | 
|  | "'diff_failed'", | 
|  | "This build runs all test targets even though `{}` is set " | 
|  | "since bazel-diff failed with an error:\n```\n{}\n```".format( | 
|  | USE_BAZEL_DIFF_ENV_VAR, ex | 
|  | ), | 
|  | ] | 
|  | ) | 
|  | execute_command( | 
|  | ["buildkite-agent", "annotation", "remove", "--context", BAZEL_DIFF_ANNOTATION_CTX] | 
|  | ) | 
|  | finally: | 
|  | return expanded_test_targets | 
|  | finally: | 
|  | try: | 
|  | shutil.rmtree(tmpdir) | 
|  | except: | 
|  | pass | 
|  |  | 
|  | os.chdir(workspace_dir) | 
|  |  | 
|  | config_target_set = set(expanded_test_targets) | 
|  | remaining_targets = list(config_target_set.intersection(affected_targets)) | 
|  | if len(remaining_targets) < len(expanded_test_targets): | 
|  | print_collapsed_group( | 
|  | ":scissors: Successfully reduced test targets from {} to {}".format( | 
|  | len(expanded_test_targets), len(remaining_targets) | 
|  | ) | 
|  | ) | 
|  |  | 
|  | skipped_targets = sorted(config_target_set.difference(remaining_targets)) | 
|  | eprint("Skipped targets:\n\t{}".format("\n\t".join(skipped_targets))) | 
|  |  | 
|  | execute_command( | 
|  | [ | 
|  | "buildkite-agent", | 
|  | "annotate", | 
|  | "--style=info", | 
|  | "--context", | 
|  | BAZEL_DIFF_ANNOTATION_CTX, | 
|  | "This run only contains test targets that have been changed since " | 
|  | "{} due to the `{}` env variable".format(resolved_diffbase, USE_BAZEL_DIFF_ENV_VAR), | 
|  | ] | 
|  | ) | 
|  |  | 
|  | return remaining_targets | 
|  |  | 
|  |  | 
|  | def fetch_base_branch(): | 
|  | """Fetch the base branch for the current build, set FETCH_HEAD for git.""" | 
|  | base_branch = os.getenv("BUILDKITE_PULL_REQUEST_BASE_BRANCH", "") | 
|  | # Fallback to the default branch for this repository if BUILDKITE_PULL_REQUEST_BASE_BRANCH is not set. | 
|  | if not base_branch: | 
|  | base_branch = os.getenv("BUILDKITE_PIPELINE_DEFAULT_BRANCH", "") | 
|  | execute_command(["git", "fetch", "origin", base_branch]) | 
|  |  | 
|  |  | 
|  | def resolve_diffbase(diffbase): | 
|  | if diffbase in AUTO_DIFFBASE_VALUES: | 
|  | fetch_base_branch() | 
|  | return execute_command_and_get_output(["git", "merge-base", "HEAD", "FETCH_HEAD"]).strip() | 
|  | elif COMMIT_RE.fullmatch(diffbase): | 
|  | return diffbase | 
|  |  | 
|  | raise BuildkiteException( | 
|  | "Invalid value '{}' for `{}` env variable. Must be a Git commit hash or one of {}".format( | 
|  | diffbase, ", ".join(AUTO_DIFFBASE_VALUES) | 
|  | ) | 
|  | ) | 
|  |  | 
|  |  | 
|  | def is_googlesource_repo(repo_url): | 
|  | return "googlesource" in repo_url | 
|  |  | 
|  |  | 
|  | def get_commit_archive_url(resolved_diffbase): | 
|  | repo_url = os.getenv("BUILDKITE_REPO", "") | 
|  | prefix = "+" if is_googlesource_repo(repo_url) else "" | 
|  | return repo_url.replace(".git", "/{}archive/{}.tar.gz".format(prefix, resolved_diffbase)) | 
|  |  | 
|  |  | 
|  | def extract_archive(archive_path, dest_dir, strip_top_level_dir): | 
|  | if not os.path.isdir(dest_dir): | 
|  | os.mkdir(dest_dir) | 
|  |  | 
|  | try: | 
|  | with tarfile.open(archive_path, mode="r:gz") as archive: | 
|  | if strip_top_level_dir: | 
|  | for member in archive.getmembers(): | 
|  | member.name = "/".join(member.name.split("/")[1:]) | 
|  | archive.extract(member, dest_dir) | 
|  | else: | 
|  | archive.extractall(dest_dir) | 
|  | except tarfile.TarError as ex: | 
|  | raise BuildkiteInfraException("Failed to extract repository archive: {}".format(ex)) from ex | 
|  |  | 
|  |  | 
|  | def download_file(url, dest_dir, dest_filename): | 
|  | local_path = os.path.join(dest_dir, dest_filename) | 
|  | try: | 
|  | execute_command(["curl", "-sSL", url, "-o", local_path], capture_stderr=True) | 
|  | except subprocess.CalledProcessError as ex: | 
|  | raise BuildkiteInfraException("Failed to download {}: {}\n{}".format(url, ex, ex.stderr)) | 
|  | return local_path | 
|  |  | 
|  |  | 
|  | def run_bazel_diff(bazel_diff_path, old_workspace_dir, new_workspace_dir, bazel_binary, data_dir): | 
|  | before_json = os.path.join(data_dir, "before.json") | 
|  | after_json = os.path.join(data_dir, "after.json") | 
|  | targets_file = os.path.join(data_dir, "targets.txt") | 
|  |  | 
|  | try: | 
|  | for repo_dir, json_path in ( | 
|  | (old_workspace_dir, before_json), | 
|  | (new_workspace_dir, after_json), | 
|  | ): | 
|  | execute_command( | 
|  | [ | 
|  | "java", | 
|  | "-jar", | 
|  | bazel_diff_path, | 
|  | "generate-hashes", | 
|  | "-w", | 
|  | repo_dir, | 
|  | "-b", | 
|  | bazel_binary, | 
|  | "--bazelCommandOptions=--lockfile_mode=off", | 
|  | json_path, | 
|  | ], | 
|  | capture_stderr=True, | 
|  | ) | 
|  |  | 
|  | execute_command( | 
|  | [ | 
|  | "java", | 
|  | "-jar", | 
|  | bazel_diff_path, | 
|  | "get-impacted-targets", | 
|  | "-sh", | 
|  | before_json, | 
|  | "-fh", | 
|  | after_json, | 
|  | "-o", | 
|  | targets_file, | 
|  | ], | 
|  | capture_stderr=True, | 
|  | ) | 
|  | except subprocess.CalledProcessError as ex: | 
|  | raise BuildkiteInfraException("Failed to run bazel-diff: {}\n{}".format(ex, ex.stderr)) | 
|  |  | 
|  | with open(targets_file, "rt") as f: | 
|  | contents = f.read() | 
|  |  | 
|  | return contents.split("\n") | 
|  |  | 
|  |  | 
|  | def partition_list(items): | 
|  | included, excluded = [], [] | 
|  | for i in items: | 
|  | if i.startswith("-"): | 
|  | excluded.append(i[1:]) | 
|  | else: | 
|  | included.append(i) | 
|  |  | 
|  | return included, excluded | 
|  |  | 
|  |  | 
|  | def get_targets_for_shard(test_targets, shard_id, shard_count): | 
|  | # TODO(fweikert): implement a more sophisticated algorithm | 
|  | return sorted(test_targets)[shard_id::shard_count] | 
|  |  | 
|  |  | 
|  | def execute_bazel_test( | 
|  | bazel_version, | 
|  | bazel_binary, | 
|  | platform, | 
|  | flags, | 
|  | targets, | 
|  | bep_file, | 
|  | monitor_flaky_tests, | 
|  | ): | 
|  | aggregated_flags = [ | 
|  | "--flaky_test_attempts=3", | 
|  | "--build_tests_only", | 
|  | "--local_test_jobs=" + concurrent_test_jobs(platform), | 
|  | ] | 
|  |  | 
|  | # Don't enable remote caching if the user enabled remote execution / caching themselves | 
|  | # or flaky test monitoring is enabled, as remote caching makes tests look less flaky than | 
|  | # they are. | 
|  | print_collapsed_group(":bazel: Computing flags for test step") | 
|  | aggregated_flags += compute_flags( | 
|  | platform, | 
|  | flags, | 
|  | bep_file, | 
|  | bazel_binary, | 
|  | enable_remote_cache=not monitor_flaky_tests, | 
|  | ) | 
|  |  | 
|  | print_expanded_group(":bazel: Test ({})".format(bazel_version)) | 
|  | try: | 
|  | execute_command( | 
|  | [bazel_binary] | 
|  | + bazelisk_flags() | 
|  | + common_startup_flags() | 
|  | + ["test"] | 
|  | + aggregated_flags | 
|  | + ["--"] | 
|  | + targets | 
|  | ) | 
|  | except subprocess.CalledProcessError as e: | 
|  | handle_bazel_failure(e, "test") | 
|  |  | 
|  |  | 
|  | def execute_bazel_coverage(bazel_version, bazel_binary, platform, flags, targets): | 
|  | aggregated_flags = [ | 
|  | "--build_tests_only", | 
|  | "--local_test_jobs=" + concurrent_test_jobs(platform), | 
|  | ] | 
|  | print_collapsed_group(":bazel: Computing flags for coverage step") | 
|  | aggregated_flags += compute_flags( | 
|  | platform, | 
|  | flags, | 
|  | None, | 
|  | bazel_binary, | 
|  | enable_remote_cache=True, | 
|  | ) | 
|  |  | 
|  | print_expanded_group(":bazel: Coverage ({})".format(bazel_version)) | 
|  | try: | 
|  | execute_command( | 
|  | [bazel_binary] | 
|  | + bazelisk_flags() | 
|  | + common_startup_flags() | 
|  | + ["coverage"] | 
|  | + aggregated_flags | 
|  | + ["--"] | 
|  | + targets | 
|  | ) | 
|  | except subprocess.CalledProcessError as e: | 
|  | handle_bazel_failure(e, "coverage") | 
|  |  | 
|  |  | 
|  | def upload_test_logs_from_bep(bep_file, tmpdir, monitor_flaky_tests): | 
|  | if local_run_only(): | 
|  | return | 
|  |  | 
|  | bazelci_agent_binary = download_bazelci_agent(tmpdir) | 
|  | execute_command( | 
|  | [ | 
|  | bazelci_agent_binary, | 
|  | "artifact", | 
|  | "upload", | 
|  | "--debug",  # Force BEP upload for non-flaky failures | 
|  | "--mode=buildkite", | 
|  | "--build_event_json_file={}".format(bep_file), | 
|  | ] | 
|  | + (["--monitor_flaky_tests"] if monitor_flaky_tests else []) | 
|  | ) | 
|  |  | 
|  |  | 
|  | def upload_json_profile(json_profile_path, tmpdir): | 
|  | if local_run_only(): | 
|  | return | 
|  | if not os.path.exists(json_profile_path): | 
|  | return | 
|  | print_collapsed_group(":gcloud: Uploading JSON Profile") | 
|  | execute_command(["buildkite-agent", "artifact", "upload", json_profile_path], cwd=tmpdir) | 
|  |  | 
|  |  | 
|  | def upload_corrupted_outputs(capture_corrupted_outputs_dir, tmpdir): | 
|  | if local_run_only(): | 
|  | return | 
|  | if not os.path.exists(capture_corrupted_outputs_dir): | 
|  | return | 
|  | print_collapsed_group(":gcloud: Uploading corrupted outputs") | 
|  | execute_command( | 
|  | ["buildkite-agent", "artifact", "upload", "{}/**/*".format(capture_corrupted_outputs_dir)], | 
|  | cwd=tmpdir, | 
|  | ) | 
|  |  | 
|  |  | 
|  | def execute_command_and_get_output(args, shell=False, fail_if_nonzero=True, print_output=True): | 
|  | eprint(" ".join(args)) | 
|  | process = subprocess.run( | 
|  | args, | 
|  | shell=shell, | 
|  | check=fail_if_nonzero, | 
|  | env=os.environ, | 
|  | stdout=subprocess.PIPE,  # We cannot use capture_output since some workers run Python <3.7 | 
|  | errors="replace", | 
|  | universal_newlines=True, | 
|  | ) | 
|  | if print_output: | 
|  | eprint(process.stdout) | 
|  |  | 
|  | return process.stdout | 
|  |  | 
|  |  | 
|  | def execute_command( | 
|  | args, | 
|  | shell=False, | 
|  | fail_if_nonzero=True, | 
|  | cwd=None, | 
|  | print_output=True, | 
|  | capture_stderr=False, | 
|  | suppress_stdout=False, | 
|  | ): | 
|  | if print_output: | 
|  | eprint(" ".join(args)) | 
|  | return subprocess.run( | 
|  | args, | 
|  | shell=shell, | 
|  | check=fail_if_nonzero, | 
|  | env=os.environ, | 
|  | cwd=cwd, | 
|  | errors="replace", | 
|  | stdout=( | 
|  | subprocess.DEVNULL if suppress_stdout else None | 
|  | ),  # suppress_stdout=True when we don't want the output to be printed | 
|  | stderr=( | 
|  | subprocess.PIPE if capture_stderr else None | 
|  | ),  # capture_stderr=True when we want exceptions to contain stderr | 
|  | ).returncode | 
|  |  | 
|  |  | 
|  | def execute_command_background(args): | 
|  | eprint(" ".join(args)) | 
|  | return subprocess.Popen(args, env=os.environ) | 
|  |  | 
|  |  | 
|  | def terminate_background_process(process): | 
|  | if process: | 
|  | process.terminate() | 
|  | try: | 
|  | process.wait(timeout=10) | 
|  | except subprocess.TimeoutExpired: | 
|  | process.kill() | 
|  |  | 
|  |  | 
|  | def create_step(label, commands, platform, shards=1, soft_fail=None): | 
|  | if "docker-image" in PLATFORMS[platform]: | 
|  | step = create_docker_step( | 
|  | label, | 
|  | image=PLATFORMS[platform]["docker-image"], | 
|  | commands=commands, | 
|  | queue=PLATFORMS[platform].get("queue", "default"), | 
|  | always_pull=PLATFORMS[platform].get("always-pull", True), | 
|  | ) | 
|  | else: | 
|  | step = { | 
|  | "label": label, | 
|  | "command": commands, | 
|  | "agents": {"queue": PLATFORMS[platform]["queue"]}, | 
|  | } | 
|  |  | 
|  | if shards > 1: | 
|  | step["label"] += " (shard %n)" | 
|  | step["parallelism"] = shards | 
|  |  | 
|  | if soft_fail is not None: | 
|  | step["soft_fail"] = soft_fail | 
|  |  | 
|  | # Enforce a global 8 hour job timeout. | 
|  | step["timeout_in_minutes"] = 8 * 60 | 
|  |  | 
|  | # Automatically retry when an agent got lost (usually due to an infra flake). | 
|  | step["retry"] = { | 
|  | "automatic": [ | 
|  | {"exit_status": -1, "limit": 3},  # Buildkite internal "agent lost" exit code | 
|  | {"exit_status": 137, "limit": 3},  # SIGKILL | 
|  | {"exit_status": 143, "limit": 3},  # SIGTERM | 
|  | ] | 
|  | } | 
|  |  | 
|  | return step | 
|  |  | 
|  |  | 
|  | def create_docker_step( | 
|  | label, image, commands=None, additional_env_vars=None, queue="default", always_pull=True | 
|  | ): | 
|  | env = ["ANDROID_HOME", "ANDROID_NDK_HOME", "BUILDKITE_ARTIFACT_UPLOAD_DESTINATION"] | 
|  | if additional_env_vars: | 
|  | env += ["{}={}".format(k, v) for k, v in additional_env_vars.items()] | 
|  |  | 
|  | step = { | 
|  | "label": label, | 
|  | "command": commands, | 
|  | "agents": {"queue": queue}, | 
|  | "plugins": { | 
|  | "docker#v3.8.0": { | 
|  | "always-pull": always_pull, | 
|  | "environment": env, | 
|  | "image": image, | 
|  | "network": "host", | 
|  | "privileged": True, | 
|  | "propagate-environment": True, | 
|  | "propagate-uid-gid": True, | 
|  | "volumes": [ | 
|  | "/etc/group:/etc/group:ro", | 
|  | "/etc/passwd:/etc/passwd:ro", | 
|  | "/etc/shadow:/etc/shadow:ro", | 
|  | "/opt/android-ndk-r15c:/opt/android-ndk-r15c:ro", | 
|  | "/opt/android-ndk-r25b:/opt/android-ndk-r25b:ro", | 
|  | "/opt/android-sdk-linux:/opt/android-sdk-linux:ro", | 
|  | "/var/lib/buildkite-agent:/var/lib/buildkite-agent", | 
|  | "/var/lib/gitmirrors:/var/lib/gitmirrors:ro", | 
|  | "/var/run/docker.sock:/var/run/docker.sock", | 
|  | ], | 
|  | } | 
|  | }, | 
|  | } | 
|  | if not step["command"]: | 
|  | del step["command"] | 
|  | return step | 
|  |  | 
|  |  | 
|  | def print_project_pipeline( | 
|  | configs, | 
|  | project_name, | 
|  | http_config, | 
|  | file_config, | 
|  | git_repository, | 
|  | monitor_flaky_tests, | 
|  | use_but, | 
|  | notify, | 
|  | print_shard_summary, | 
|  | ): | 
|  | task_configs = configs.get("tasks", None) | 
|  | if not task_configs: | 
|  | raise BuildkiteException("{0} pipeline configuration is empty.".format(project_name)) | 
|  |  | 
|  | pipeline_steps = [] | 
|  | # If the repository is hosted on Git-on-borg, we show the link to the commit Gerrit review | 
|  | buildkite_repo = os.getenv("BUILDKITE_REPO") | 
|  | if is_git_on_borg_repo(buildkite_repo): | 
|  | show_gerrit_review_link(buildkite_repo, pipeline_steps) | 
|  |  | 
|  | task_configs = filter_tasks_that_should_be_skipped(task_configs, pipeline_steps) | 
|  |  | 
|  | buildifier_config = configs.get("buildifier") | 
|  | # Skip Buildifier when we test downstream projects. | 
|  | if buildifier_config and not is_downstream_pipeline(): | 
|  | buildifier_env_vars = {} | 
|  | if isinstance(buildifier_config, str): | 
|  | # Simple format: | 
|  | # --- | 
|  | # buildifier: latest | 
|  | buildifier_env_vars["BUILDIFIER_VERSION"] = buildifier_config | 
|  | else: | 
|  | # Advanced format: | 
|  | # --- | 
|  | # buildifier: | 
|  | #   version: latest | 
|  | #   warnings: all | 
|  | if "version" in buildifier_config: | 
|  | buildifier_env_vars["BUILDIFIER_VERSION"] = buildifier_config["version"] | 
|  | if "warnings" in buildifier_config: | 
|  | buildifier_env_vars["BUILDIFIER_WARNINGS"] = buildifier_config["warnings"] | 
|  |  | 
|  | if not buildifier_env_vars: | 
|  | raise BuildkiteException( | 
|  | 'Invalid buildifier configuration entry "{}"'.format(buildifier_config) | 
|  | ) | 
|  |  | 
|  | pipeline_steps.append( | 
|  | create_docker_step( | 
|  | BUILDIFIER_STEP_NAME, | 
|  | image=BUILDIFIER_DOCKER_IMAGE, | 
|  | additional_env_vars=buildifier_env_vars, | 
|  | ) | 
|  | ) | 
|  |  | 
|  | # In Bazel Downstream Project pipelines, we should test the project at the last green commit. | 
|  | git_commit = get_last_green_commit(project_name) if is_downstream_pipeline() else None | 
|  |  | 
|  | config_hashes = set() | 
|  | skipped_downstream_tasks = [] | 
|  | has_sharded_task = False | 
|  | for task, task_config in task_configs.items(): | 
|  | platform = get_platform_for_task(task, task_config) | 
|  | task_name = task_config.get("name") | 
|  | soft_fail = task_config.get("soft_fail") | 
|  |  | 
|  | # We override the Bazel version in downstream pipelines. This means that two tasks that | 
|  | # only differ in the value of their explicit "bazel" field will be identical in the | 
|  | # downstream pipeline, thus leading to duplicate work. | 
|  | # Consequently, we filter those duplicate tasks here. | 
|  | if is_downstream_pipeline(): | 
|  | h = hash_task_config(task, task_config) | 
|  | if h in config_hashes: | 
|  | skipped_downstream_tasks.append( | 
|  | "{}: {}".format( | 
|  | create_label(platform, project_name, task_name=task_name), | 
|  | "The same task already exists after ignoring bazel version.", | 
|  | ) | 
|  | ) | 
|  | continue | 
|  |  | 
|  | config_hashes.add(h) | 
|  |  | 
|  | # Skip tasks with `skip_in_bazel_downstream_pipeline` specified. | 
|  | skipped_reason = task_config.get("skip_in_bazel_downstream_pipeline", "") | 
|  | if skipped_reason: | 
|  | skipped_downstream_tasks.append( | 
|  | "{}: {}".format( | 
|  | create_label(platform, project_name, task_name=task_name), | 
|  | skipped_reason, | 
|  | ) | 
|  | ) | 
|  | continue | 
|  |  | 
|  | shards = task_config.get("shards", "1") | 
|  | try: | 
|  | shards = int(shards) | 
|  | except ValueError: | 
|  | raise BuildkiteException("Task {} has invalid shard value '{}'".format(task, shards)) | 
|  |  | 
|  | if shards > 1: | 
|  | has_sharded_task = True | 
|  |  | 
|  | step = runner_step( | 
|  | platform=platform, | 
|  | task=task, | 
|  | task_name=task_name, | 
|  | project_name=project_name, | 
|  | http_config=http_config, | 
|  | file_config=file_config, | 
|  | git_repository=git_repository, | 
|  | git_commit=git_commit, | 
|  | monitor_flaky_tests=monitor_flaky_tests, | 
|  | use_but=use_but, | 
|  | shards=shards, | 
|  | soft_fail=soft_fail, | 
|  | ) | 
|  | pipeline_steps.append(step) | 
|  |  | 
|  | if skipped_downstream_tasks: | 
|  | lines = ["\n- {}".format(s) for s in skipped_downstream_tasks] | 
|  | commands = [ | 
|  | "buildkite-agent meta-data exists 'has-skipped-annotation' || buildkite-agent annotate --style=info 'The following tasks were skipped:\n' --context 'ctx-skipped_downstream_tasks'", | 
|  | "buildkite-agent meta-data set 'has-skipped-annotation' 'true'", | 
|  | "buildkite-agent annotate --style=info '{}' --append --context 'ctx-skipped_downstream_tasks'".format( | 
|  | "".join(lines) | 
|  | ), | 
|  | ] | 
|  | pipeline_steps.append( | 
|  | create_step( | 
|  | label=":pipeline: Print information about skipped tasks", | 
|  | commands=commands, | 
|  | platform=DEFAULT_PLATFORM, | 
|  | ) | 
|  | ) | 
|  |  | 
|  | pipeline_slug = os.getenv("BUILDKITE_PIPELINE_SLUG") | 
|  | all_downstream_pipeline_slugs = [] | 
|  | for _, config in DOWNSTREAM_PROJECTS.items(): | 
|  | all_downstream_pipeline_slugs.append(config["pipeline_slug"]) | 
|  |  | 
|  | # We update last green commit in the following cases: | 
|  | #   1. This job runs on master, stable or main branch (could be a custom build launched manually) | 
|  | #   2. We intend to run the same job in downstream with Bazel@HEAD (eg. google-bazel-presubmit) | 
|  | #   3. This job is not: | 
|  | #      - a GitHub pull request | 
|  | #      - uses a custom built Bazel binary (in Bazel Downstream Projects pipeline) | 
|  | #      - testing incompatible flags | 
|  | #      - running `bazelisk --migrate` in a non-downstream pipeline | 
|  | should_update_last_green = ( | 
|  | current_branch_is_main_branch() | 
|  | and pipeline_slug in all_downstream_pipeline_slugs | 
|  | and not (is_pull_request() or use_but or use_bazelisk_migrate()) | 
|  | ) | 
|  |  | 
|  | actually_print_shard_summary = has_sharded_task and print_shard_summary | 
|  |  | 
|  | if should_update_last_green or actually_print_shard_summary: | 
|  | pipeline_steps.append({"wait": None, "continue_on_failure": True}) | 
|  |  | 
|  | if should_update_last_green: | 
|  | # We need to call "Try Update Last Green Commit" even if there are failures, | 
|  | # since we don't want a failing Buildifier step to block the update of | 
|  | # the last green commit for this project. | 
|  | # try_update_last_green_commit() ensures that we don't update the commit | 
|  | # if any build or test steps fail. | 
|  | pipeline_steps.append( | 
|  | create_step( | 
|  | label="Try Update Last Green Commit", | 
|  | commands=[ | 
|  | fetch_bazelcipy_command(), | 
|  | PLATFORMS[DEFAULT_PLATFORM]["python"] | 
|  | + " bazelci.py try_update_last_green_commit", | 
|  | ], | 
|  | platform=DEFAULT_PLATFORM, | 
|  | ) | 
|  | ) | 
|  |  | 
|  | if "validate_config" in configs and not is_downstream_pipeline(): | 
|  | pipeline_steps += create_config_validation_steps( | 
|  | git_commit or os.getenv("BUILDKITE_COMMIT") | 
|  | ) | 
|  |  | 
|  | if use_bazelisk_migrate() and not is_downstream_pipeline(): | 
|  | # Print results of bazelisk --migrate in project pipelines that explicitly set | 
|  | # the USE_BAZELISK_MIGRATE env var, but that are not being run as part of a | 
|  | # downstream pipeline. | 
|  | number = os.getenv("BUILDKITE_BUILD_NUMBER") | 
|  | pipeline_steps += get_steps_for_aggregating_migration_results(number, notify) | 
|  |  | 
|  | if actually_print_shard_summary: | 
|  | pipeline_steps.append( | 
|  | create_step( | 
|  | label=SHARD_SUMMARY_STEP_NAME, | 
|  | commands=[ | 
|  | fetch_bazelcipy_command(), | 
|  | PLATFORMS[DEFAULT_PLATFORM]["python"] + " bazelci.py print_shard_summary", | 
|  | ], | 
|  | platform=DEFAULT_PLATFORM, | 
|  | ) | 
|  | ) | 
|  |  | 
|  | print_pipeline_steps(pipeline_steps, handle_emergencies=not is_downstream_pipeline()) | 
|  |  | 
|  |  | 
|  | def show_gerrit_review_link(git_repository, pipeline_steps): | 
|  | host = re.search(r"https://(.+?)\.googlesource", git_repository).group(1) | 
|  | if not host: | 
|  | raise BuildkiteException("Couldn't get host name from %s" % git_repository) | 
|  | commit = os.getenv("BUILDKITE_COMMIT") | 
|  | line1 = f"The transformed code used in this pipeline can be found under https://{host}-review.googlesource.com/q/{commit}" | 
|  | line2 = f"\n\nFetch the source with `git fetch https://{host}.googlesource.com/bazel {commit}  && git checkout FETCH_HEAD`" | 
|  | commands = [ | 
|  | "buildkite-agent annotate --style=info '{}' --context 'gerrit'".format(line1), | 
|  | "buildkite-agent annotate --style=info '{}' --append --context 'gerrit'".format(line2), | 
|  | ] | 
|  | pipeline_steps.append( | 
|  | create_step( | 
|  | label=":pipeline: Print information about Gerrit Review Link", | 
|  | commands=commands, | 
|  | platform=DEFAULT_PLATFORM, | 
|  | ) | 
|  | ) | 
|  |  | 
|  |  | 
|  | def is_git_on_borg_repo(git_repository): | 
|  | return git_repository and "googlesource.com" in git_repository | 
|  |  | 
|  |  | 
|  | def hash_task_config(task_name, task_config): | 
|  | # Two task configs c1 and c2 have the same hash iff they lead to two functionally identical jobs | 
|  | # in the downstream pipeline. This function discards the "bazel" field (since it's being | 
|  | # overridden) and the "name" field (since it has no effect on the actual work). | 
|  | # Moreover, it adds an explicit "platform" field if that's missing. | 
|  | cpy = task_config.copy() | 
|  | cpy.pop("bazel", None) | 
|  | cpy.pop("name", None) | 
|  | if "platform" not in cpy: | 
|  | cpy["platform"] = task_name | 
|  |  | 
|  | m = hashlib.md5() | 
|  | # Technically we should sort cpy[key] if it's a list of entries | 
|  | # whose order does not matter (e.g. targets). | 
|  | # However, this seems to be overkill for the current use cases. | 
|  | for key in sorted(cpy): | 
|  | value = "%s:%s;" % (key, cpy[key]) | 
|  | m.update(value.encode("utf-8")) | 
|  |  | 
|  | return m.digest() | 
|  |  | 
|  |  | 
|  | def get_platform_for_task(task, task_config): | 
|  | # Most pipeline configurations have exactly one task per platform, which makes it | 
|  | # convenient to use the platform name as task ID. Consequently, we use the | 
|  | # task ID as platform if there is no explicit "platform" field. | 
|  | return task_config.get("platform", task) | 
|  |  | 
|  |  | 
|  | def create_config_validation_steps(git_commit): | 
|  | config_files = [ | 
|  | path | 
|  | for path in get_modified_files(git_commit) | 
|  | if path.startswith(".bazelci/") and os.path.splitext(path)[1] in CONFIG_FILE_EXTENSIONS | 
|  | ] | 
|  | return [ | 
|  | create_step( | 
|  | label=":cop: Validate {}".format(f), | 
|  | commands=[ | 
|  | fetch_bazelcipy_command(), | 
|  | "{} bazelci.py project_pipeline --file_config={}".format( | 
|  | PLATFORMS[DEFAULT_PLATFORM]["python"], f | 
|  | ), | 
|  | ], | 
|  | platform=DEFAULT_PLATFORM, | 
|  | ) | 
|  | for f in config_files | 
|  | ] | 
|  |  | 
|  |  | 
|  | def get_modified_files(git_commit): | 
|  | fetch_base_branch() | 
|  | merge_base_commit = execute_command_and_get_output( | 
|  | ["git", "merge-base", git_commit, "FETCH_HEAD"] | 
|  | ).strip() | 
|  | output = execute_command_and_get_output( | 
|  | [ | 
|  | "git", | 
|  | "diff-tree", | 
|  | "--no-commit-id", | 
|  | "--name-only", | 
|  | "-r", | 
|  | "{}..{}".format(merge_base_commit, git_commit), | 
|  | ] | 
|  | ) | 
|  | return output.split("\n") | 
|  |  | 
|  |  | 
|  | def print_pipeline_steps(pipeline_steps, handle_emergencies=True): | 
|  | if handle_emergencies: | 
|  | emergency_step = create_emergency_announcement_step_if_necessary() | 
|  | if emergency_step: | 
|  | pipeline_steps.insert(0, emergency_step) | 
|  |  | 
|  | print(yaml.dump({"steps": pipeline_steps})) | 
|  |  | 
|  |  | 
|  | def create_emergency_announcement_step_if_necessary(): | 
|  | style = "error" | 
|  | message, issue_url, last_good_bazel = None, None, None | 
|  | try: | 
|  | emergency_settings = load_remote_yaml_file(EMERGENCY_FILE_URL) | 
|  | message = emergency_settings.get("message") | 
|  | issue_url = emergency_settings.get("issue_url") | 
|  | last_good_bazel = emergency_settings.get("last_good_bazel") | 
|  | except urllib.error.HTTPError as ex: | 
|  | message = str(ex) | 
|  | style = "warning" | 
|  |  | 
|  | if not any([message, issue_url, last_good_bazel]): | 
|  | return | 
|  |  | 
|  | text = '<span class="h1">:rotating_light: Emergency :rotating_light:</span>\n' | 
|  | if message: | 
|  | text += "- {}\n".format(message) | 
|  | if issue_url: | 
|  | text += '- Please check this <a href="{}">issue</a> for more details.\n'.format(issue_url) | 
|  | if last_good_bazel: | 
|  | text += ( | 
|  | "- Default Bazel version is *{}*, " | 
|  | "unless the pipeline configuration specifies an explicit version." | 
|  | ).format(last_good_bazel) | 
|  |  | 
|  | return create_step( | 
|  | label=":rotating_light: Emergency :rotating_light:", | 
|  | commands=[ | 
|  | 'buildkite-agent annotate --append --style={} --context "omg" "{}"'.format(style, text) | 
|  | ], | 
|  | platform=DEFAULT_PLATFORM, | 
|  | ) | 
|  |  | 
|  |  | 
|  | def runner_step( | 
|  | platform, | 
|  | task, | 
|  | task_name=None, | 
|  | project_name=None, | 
|  | http_config=None, | 
|  | file_config=None, | 
|  | git_repository=None, | 
|  | git_commit=None, | 
|  | monitor_flaky_tests=False, | 
|  | use_but=False, | 
|  | shards=1, | 
|  | soft_fail=None, | 
|  | ): | 
|  | command = PLATFORMS[platform]["python"] + " bazelci.py runner --task=" + task | 
|  | if http_config: | 
|  | command += " --http_config=" + http_config | 
|  | if file_config: | 
|  | command += " --file_config=" + file_config | 
|  | if git_repository: | 
|  | command += " --git_repository=" + git_repository | 
|  | if git_commit: | 
|  | command += " --git_commit=" + git_commit | 
|  | if monitor_flaky_tests: | 
|  | command += " --monitor_flaky_tests" | 
|  | if use_but: | 
|  | command += " --use_but" | 
|  | label = create_label(platform, project_name, task_name=task_name) | 
|  | return create_step( | 
|  | label=label, | 
|  | commands=[fetch_bazelcipy_command(), command], | 
|  | platform=platform, | 
|  | shards=shards, | 
|  | soft_fail=soft_fail, | 
|  | ) | 
|  |  | 
|  |  | 
|  | def fetch_bazelcipy_command(): | 
|  | return "curl -sS {0}?{1} -o bazelci.py".format(SCRIPT_URL, int(time.time())) | 
|  |  | 
|  |  | 
|  | def fetch_aggregate_incompatible_flags_test_result_command(): | 
|  | return "curl -sS {0} -o aggregate_incompatible_flags_test_result.py".format( | 
|  | AGGREGATE_INCOMPATIBLE_TEST_RESULT_URL | 
|  | ) | 
|  |  | 
|  |  | 
|  | def upload_project_pipeline_step(project_name, git_repository, http_config, file_config): | 
|  | pipeline_command = ( | 
|  | '{0} bazelci.py project_pipeline --project_name="{1}" ' + "--git_repository={2}" | 
|  | ).format(PLATFORMS[DEFAULT_PLATFORM]["python"], project_name, git_repository) | 
|  | pipeline_command += " --use_but" | 
|  | if http_config: | 
|  | pipeline_command += " --http_config=" + http_config | 
|  | if file_config: | 
|  | pipeline_command += " --file_config=" + file_config | 
|  | pipeline_command += " | tee /dev/tty | buildkite-agent pipeline upload" | 
|  |  | 
|  | return create_step( | 
|  | label="Setup {0}".format(project_name), | 
|  | commands=[fetch_bazelcipy_command(), pipeline_command], | 
|  | platform=DEFAULT_PLATFORM, | 
|  | ) | 
|  |  | 
|  |  | 
|  | def create_label(platform, project_name, build_only=False, test_only=False, task_name=None): | 
|  | if build_only and test_only: | 
|  | raise BuildkiteException("build_only and test_only cannot be true at the same time") | 
|  | platform_display_name = PLATFORMS[platform]["emoji-name"] | 
|  |  | 
|  | if build_only: | 
|  | label = "Build " | 
|  | elif test_only: | 
|  | label = "Test " | 
|  | else: | 
|  | label = "" | 
|  |  | 
|  | platform_label = ( | 
|  | "{0} on {1}".format(task_name, platform_display_name) | 
|  | if task_name | 
|  | else platform_display_name | 
|  | ) | 
|  |  | 
|  | if project_name: | 
|  | # Update get_project_name_from_job in bazel_auto_sheriff.py if you change | 
|  | # the expected format of "Project Foo (Task bar on OS)" | 
|  | label += "{0} ({1})".format(project_name, platform_label) | 
|  | else: | 
|  | label += platform_label | 
|  |  | 
|  | return label | 
|  |  | 
|  |  | 
|  | def bazel_build_step( | 
|  | task, | 
|  | platform, | 
|  | project_name, | 
|  | http_config=None, | 
|  | file_config=None, | 
|  | build_only=False, | 
|  | test_only=False, | 
|  | ): | 
|  | pipeline_command = PLATFORMS[platform]["python"] + " bazelci.py runner" | 
|  | if build_only: | 
|  | pipeline_command += " --build_only --save_but" | 
|  | if test_only: | 
|  | pipeline_command += " --test_only" | 
|  | if http_config: | 
|  | pipeline_command += " --http_config=" + http_config | 
|  | if file_config: | 
|  | pipeline_command += " --file_config=" + file_config | 
|  | pipeline_command += " --task=" + task | 
|  |  | 
|  | step = create_step( | 
|  | label=create_label(platform, project_name, build_only, test_only), | 
|  | commands=[fetch_bazelcipy_command(), pipeline_command], | 
|  | platform=platform, | 
|  | ) | 
|  | # Always try to automatically retry the bazel build step, this will make | 
|  | # the publish bazel binaries pipeline more reliable. | 
|  | step["retry"] = { | 
|  | "automatic": [ | 
|  | {"exit_status": "*", "limit": 3}, | 
|  | ] | 
|  | } | 
|  | return step | 
|  |  | 
|  |  | 
|  | def filter_tasks_that_should_be_skipped(task_configs, pipeline_steps): | 
|  | skip_tasks = get_skip_tasks() | 
|  | if not skip_tasks: | 
|  | return task_configs | 
|  |  | 
|  | actually_skipped = [] | 
|  | skip_tasks = set(skip_tasks) | 
|  | for task in list(task_configs.keys()): | 
|  | if task in skip_tasks: | 
|  | actually_skipped.append(task) | 
|  | del task_configs[task] | 
|  | skip_tasks.remove(task) | 
|  |  | 
|  | if not task_configs: | 
|  | raise BuildkiteException( | 
|  | "Nothing to do since all tasks in the configuration should be skipped." | 
|  | ) | 
|  |  | 
|  | annotations = [] | 
|  | if actually_skipped: | 
|  | annotations.append( | 
|  | ("info", "Skipping the following task(s): {}".format(", ".join(actually_skipped))) | 
|  | ) | 
|  |  | 
|  | if skip_tasks: | 
|  | annotations.append( | 
|  | ( | 
|  | "warning", | 
|  | ( | 
|  | "The following tasks should have been skipped, " | 
|  | "but were not part of the configuration: {}" | 
|  | ).format(", ".join(skip_tasks)), | 
|  | ) | 
|  | ) | 
|  |  | 
|  | if annotations: | 
|  | print_skip_task_annotations(annotations, pipeline_steps) | 
|  |  | 
|  | return task_configs | 
|  |  | 
|  |  | 
|  | def get_skip_tasks(): | 
|  | value = os.getenv(SKIP_TASKS_ENV_VAR, "") | 
|  | return [v for v in value.split(",") if v] | 
|  |  | 
|  |  | 
|  | def print_skip_task_annotations(annotations, pipeline_steps): | 
|  | commands = [ | 
|  | "buildkite-agent annotate --style={} '{}'  --context 'ctx-{}'".format(s, t, hash(t)) | 
|  | for s, t in annotations | 
|  | ] | 
|  | pipeline_steps.append( | 
|  | create_step( | 
|  | label=":pipeline: Print information about skipped tasks", | 
|  | commands=commands, | 
|  | platform=DEFAULT_PLATFORM, | 
|  | ) | 
|  | ) | 
|  |  | 
|  |  | 
|  | def print_bazel_publish_binaries_pipeline(task_configs, http_config, file_config): | 
|  | if not task_configs: | 
|  | raise BuildkiteException("Bazel publish binaries pipeline configuration is empty.") | 
|  |  | 
|  | pipeline_steps = [] | 
|  | task_configs = filter_tasks_that_should_be_skipped(task_configs, pipeline_steps) | 
|  |  | 
|  | platforms = [get_platform_for_task(t, tc) for t, tc in task_configs.items()] | 
|  |  | 
|  | # These are the platforms that the bazel_publish_binaries.yml config is actually building. | 
|  | configured_platforms = set(filter(should_publish_binaries_for_platform, platforms)) | 
|  |  | 
|  | # These are the platforms that we want to build and publish according to this script. | 
|  | expected_platforms = set(filter(should_publish_binaries_for_platform, PLATFORMS)) | 
|  |  | 
|  | # We can skip this check if we're not on the main branch, because then we're probably | 
|  | # building a one-off custom debugging binary anyway. | 
|  | if current_branch_is_main_branch(): | 
|  | missing = expected_platforms.difference(configured_platforms) | 
|  | if missing: | 
|  | raise BuildkiteException( | 
|  | ( | 
|  | "Bazel publish binaries pipeline needs to build Bazel for every commit on all publish_binary-enabled platforms. " | 
|  | "Please add jobs for the missing platform(s) to the pipeline config: {}".format( | 
|  | ", ".join(missing) | 
|  | ) | 
|  | ) | 
|  | ) | 
|  |  | 
|  | # Build Bazel | 
|  | for task, task_config in task_configs.items(): | 
|  | pipeline_steps.append( | 
|  | bazel_build_step( | 
|  | task, | 
|  | get_platform_for_task(task, task_config), | 
|  | "Bazel", | 
|  | http_config, | 
|  | file_config, | 
|  | build_only=True, | 
|  | ) | 
|  | ) | 
|  |  | 
|  | pipeline_steps.append({"wait": None, "continue_on_failure": True}) | 
|  |  | 
|  | # If all builds succeed, publish the Bazel binaries to GCS. | 
|  | pipeline_steps.append( | 
|  | create_step( | 
|  | label="Publish Bazel Binaries", | 
|  | commands=[ | 
|  | fetch_bazelcipy_command(), | 
|  | PLATFORMS[DEFAULT_PLATFORM]["python"] + " bazelci.py publish_binaries", | 
|  | ], | 
|  | platform=DEFAULT_PLATFORM, | 
|  | ) | 
|  | ) | 
|  |  | 
|  | print_pipeline_steps(pipeline_steps) | 
|  |  | 
|  |  | 
|  | def should_publish_binaries_for_platform(platform): | 
|  | if platform not in PLATFORMS: | 
|  | raise BuildkiteException("Unknown platform '{}'".format(platform)) | 
|  |  | 
|  | return PLATFORMS[platform]["publish_binary"] | 
|  |  | 
|  |  | 
|  | def print_disabled_projects_info_box_step(): | 
|  | info_text = ["Downstream testing is disabled for the following projects :sadpanda:"] | 
|  | for project, config in DOWNSTREAM_PROJECTS.items(): | 
|  | disabled_reason = config.get("disabled_reason", None) | 
|  | if disabled_reason: | 
|  | info_text.append("* **%s**: %s" % (project, disabled_reason)) | 
|  |  | 
|  | if len(info_text) == 1: | 
|  | return None | 
|  | return create_step( | 
|  | label=":sadpanda:", | 
|  | commands=[ | 
|  | 'buildkite-agent annotate --append --style=info "\n' + "\n".join(info_text) + '\n"' | 
|  | ], | 
|  | platform=DEFAULT_PLATFORM, | 
|  | ) | 
|  |  | 
|  |  | 
|  | def print_incompatible_flags_info_box_step(incompatible_flags_map): | 
|  | info_text = ["Build and test with the following incompatible flags:"] | 
|  |  | 
|  | for flag in incompatible_flags_map: | 
|  | info_text.append("* **%s**: %s" % (flag, incompatible_flags_map[flag])) | 
|  |  | 
|  | if len(info_text) == 1: | 
|  | return None | 
|  | return create_step( | 
|  | label="Incompatible flags info", | 
|  | commands=[ | 
|  | 'buildkite-agent annotate --append --style=info "\n' + "\n".join(info_text) + '\n"' | 
|  | ], | 
|  | platform=DEFAULT_PLATFORM, | 
|  | ) | 
|  |  | 
|  |  | 
|  | def fetch_incompatible_flags(): | 
|  | """ | 
|  | Return a list of incompatible flags to be tested. The key is the flag name and the value is its Github URL. | 
|  | """ | 
|  | output = subprocess.check_output( | 
|  | [ | 
|  | # Query for open issues with "incompatible-change" and "migration-ready" label. | 
|  | "curl", | 
|  | "https://api.github.com/search/issues?per_page=100&q=repo:bazelbuild/bazel+label:incompatible-change+label:migration-ready+state:open", | 
|  | ] | 
|  | ).decode("utf-8") | 
|  | issue_info = json.loads(output) | 
|  |  | 
|  | FLAG_PATTERN = re.compile(r"^--[a-z][a-z0-9_]*$") | 
|  | incompatible_flags = {} | 
|  | for issue in issue_info["items"]: | 
|  | name = "--" + issue["title"].split(":")[0] | 
|  | url = issue["html_url"] | 
|  | if FLAG_PATTERN.match(name): | 
|  | incompatible_flags[name] = url | 
|  | else: | 
|  | eprint( | 
|  | f"{name} is not recognized as an incompatible flag, please modify the issue title " | 
|  | f'of {url} to "<incompatible flag name (without --)>:..."' | 
|  | ) | 
|  |  | 
|  | # If INCOMPATIBLE_FLAGS is set manually, we test those flags, try to keep the URL info if possible. | 
|  | if "INCOMPATIBLE_FLAGS" in os.environ: | 
|  | given_incompatible_flags = {} | 
|  | for flag in os.environ["INCOMPATIBLE_FLAGS"].split(","): | 
|  | given_incompatible_flags[flag] = incompatible_flags.get(flag, "") | 
|  | return given_incompatible_flags | 
|  |  | 
|  | return incompatible_flags | 
|  |  | 
|  |  | 
|  | def print_bazel_downstream_pipeline( | 
|  | task_configs, http_config, file_config, test_disabled_projects, notify | 
|  | ): | 
|  | pipeline_steps = [] | 
|  |  | 
|  | info_box_step = print_disabled_projects_info_box_step() | 
|  | if info_box_step is not None: | 
|  | pipeline_steps.append(info_box_step) | 
|  |  | 
|  | if not use_bazelisk_migrate(): | 
|  | if not task_configs: | 
|  | raise BuildkiteException("Bazel downstream pipeline configuration is empty.") | 
|  | for task, task_config in task_configs.items(): | 
|  | pipeline_steps.append( | 
|  | bazel_build_step( | 
|  | task, | 
|  | get_platform_for_task(task, task_config), | 
|  | "Bazel", | 
|  | http_config, | 
|  | file_config, | 
|  | build_only=True, | 
|  | ) | 
|  | ) | 
|  |  | 
|  | pipeline_steps.append("wait") | 
|  | else: | 
|  | incompatible_flags_map = fetch_incompatible_flags() | 
|  | if not incompatible_flags_map: | 
|  | step = create_step( | 
|  | label="No Incompatible flags info", | 
|  | commands=[ | 
|  | 'buildkite-agent annotate --style=error "No incompatible flag issue is found on github for current version of Bazel." --context "noinc"' | 
|  | ], | 
|  | platform=DEFAULT_PLATFORM, | 
|  | ) | 
|  | pipeline_steps.append(step) | 
|  | print_pipeline_steps(pipeline_steps) | 
|  | return | 
|  |  | 
|  | info_box_step = print_incompatible_flags_info_box_step(incompatible_flags_map) | 
|  | if info_box_step is not None: | 
|  | pipeline_steps.append(info_box_step) | 
|  |  | 
|  | for project, config in DOWNSTREAM_PROJECTS.items(): | 
|  | disabled_reason = config.get("disabled_reason", None) | 
|  | # If test_disabled_projects is true, we add configs for disabled projects. | 
|  | # If test_disabled_projects is false, we add configs for not disabled projects. | 
|  | if (test_disabled_projects and disabled_reason) or ( | 
|  | not test_disabled_projects and not disabled_reason | 
|  | ): | 
|  | pipeline_steps.append( | 
|  | upload_project_pipeline_step( | 
|  | project_name=project, | 
|  | git_repository=config["git_repository"], | 
|  | http_config=config.get("http_config", None), | 
|  | file_config=config.get("file_config", None), | 
|  | ) | 
|  | ) | 
|  |  | 
|  | if use_bazelisk_migrate(): | 
|  | current_build_number = os.environ.get("BUILDKITE_BUILD_NUMBER", None) | 
|  | if not current_build_number: | 
|  | raise BuildkiteException("Not running inside Buildkite") | 
|  |  | 
|  | pipeline_steps += get_steps_for_aggregating_migration_results(current_build_number, notify) | 
|  |  | 
|  | if ( | 
|  | not test_disabled_projects | 
|  | and not use_bazelisk_migrate() | 
|  | and current_branch_is_main_branch() | 
|  | ): | 
|  | # Only update the last green downstream commit in the regular Bazel@HEAD + Downstream pipeline. | 
|  | pipeline_steps.append("wait") | 
|  | pipeline_steps.append( | 
|  | create_step( | 
|  | label="Try Update Last Green Downstream Commit", | 
|  | commands=[ | 
|  | fetch_bazelcipy_command(), | 
|  | PLATFORMS[DEFAULT_PLATFORM]["python"] | 
|  | + " bazelci.py try_update_last_green_downstream_commit", | 
|  | ], | 
|  | platform=DEFAULT_PLATFORM, | 
|  | ) | 
|  | ) | 
|  |  | 
|  | print_pipeline_steps(pipeline_steps) | 
|  |  | 
|  |  | 
|  | def get_steps_for_aggregating_migration_results(current_build_number, notify): | 
|  | parts = [ | 
|  | PLATFORMS[DEFAULT_PLATFORM]["python"], | 
|  | "aggregate_incompatible_flags_test_result.py", | 
|  | "--build_number=%s" % current_build_number, | 
|  | ] | 
|  | if notify: | 
|  | parts.append("--notify") | 
|  | return [ | 
|  | {"wait": "~", "continue_on_failure": "true"}, | 
|  | create_step( | 
|  | label="Aggregate incompatible flags test result", | 
|  | commands=[ | 
|  | fetch_bazelcipy_command(), | 
|  | fetch_aggregate_incompatible_flags_test_result_command(), | 
|  | " ".join(parts), | 
|  | ], | 
|  | platform=DEFAULT_PLATFORM, | 
|  | ), | 
|  | ] | 
|  |  | 
|  |  | 
|  | def bazelci_builds_download_url(platform, git_commit): | 
|  | bucket_name = "bazel-testing-builds" if THIS_IS_TESTING else "bazel-builds" | 
|  | return "https://storage.googleapis.com/{}/artifacts/{}/{}/bazel".format( | 
|  | bucket_name, platform, git_commit | 
|  | ) | 
|  |  | 
|  |  | 
|  | def bazelci_builds_nojdk_download_url(platform, git_commit): | 
|  | bucket_name = "bazel-testing-builds" if THIS_IS_TESTING else "bazel-builds" | 
|  | return "https://storage.googleapis.com/{}/artifacts/{}/{}/bazel_nojdk".format( | 
|  | bucket_name, platform, git_commit | 
|  | ) | 
|  |  | 
|  |  | 
|  | def bazelci_builds_gs_url(platform, git_commit): | 
|  | bucket_name = "bazel-testing-builds" if THIS_IS_TESTING else "bazel-builds" | 
|  | return "gs://{}/artifacts/{}/{}/bazel".format(bucket_name, platform, git_commit) | 
|  |  | 
|  |  | 
|  | def bazelci_builds_nojdk_gs_url(platform, git_commit): | 
|  | bucket_name = "bazel-testing-builds" if THIS_IS_TESTING else "bazel-builds" | 
|  | return "gs://{}/artifacts/{}/{}/bazel_nojdk".format(bucket_name, platform, git_commit) | 
|  |  | 
|  |  | 
|  | def bazelci_latest_build_metadata_url(): | 
|  | bucket_name = "bazel-testing-builds" if THIS_IS_TESTING else "bazel-builds" | 
|  | return "gs://{}/metadata/latest.json".format(bucket_name) | 
|  |  | 
|  |  | 
|  | def bazelci_builds_metadata_url(git_commit): | 
|  | bucket_name = "bazel-testing-builds" if THIS_IS_TESTING else "bazel-builds" | 
|  | return "gs://{}/metadata/{}.json".format(bucket_name, git_commit) | 
|  |  | 
|  |  | 
|  | def bazelci_last_green_commit_url(git_repository, pipeline_slug): | 
|  | bucket_name = "bazel-testing-builds" if THIS_IS_TESTING else "bazel-untrusted-builds" | 
|  | return "gs://{}/last_green_commit/{}/{}".format( | 
|  | bucket_name, git_repository[len("https://") :], pipeline_slug | 
|  | ) | 
|  |  | 
|  |  | 
|  | def bazelci_last_green_downstream_commit_url(): | 
|  | bucket_name = "bazel-testing-builds" if THIS_IS_TESTING else "bazel-untrusted-builds" | 
|  | return "gs://{}/last_green_commit/downstream_pipeline".format(bucket_name) | 
|  |  | 
|  |  | 
|  | def get_last_green_commit_by_url(last_green_commit_url): | 
|  | try: | 
|  | return ( | 
|  | subprocess.check_output( | 
|  | [gsutil_command(), "cat", last_green_commit_url], env=os.environ | 
|  | ) | 
|  | .decode("utf-8") | 
|  | .strip() | 
|  | ) | 
|  | except subprocess.CalledProcessError: | 
|  | return None | 
|  |  | 
|  |  | 
|  | def get_last_green_commit(project_name): | 
|  | last_green_commit_url = bazelci_last_green_commit_url( | 
|  | DOWNSTREAM_PROJECTS[project_name]["git_repository"], | 
|  | DOWNSTREAM_PROJECTS[project_name]["pipeline_slug"], | 
|  | ) | 
|  | return get_last_green_commit_by_url(last_green_commit_url) | 
|  |  | 
|  |  | 
|  | def try_update_last_green_commit(): | 
|  | org_slug = os.getenv("BUILDKITE_ORGANIZATION_SLUG") | 
|  | pipeline_slug = os.getenv("BUILDKITE_PIPELINE_SLUG") | 
|  | build_number = os.getenv("BUILDKITE_BUILD_NUMBER") | 
|  | current_job_id = os.getenv("BUILDKITE_JOB_ID") | 
|  |  | 
|  | client = BuildkiteClient(org=org_slug, pipeline=pipeline_slug) | 
|  | build_info = client.get_build_info(build_number) | 
|  |  | 
|  | # Find any failing steps other than Buildifier and steps with soft_fail enabled then "try update last green". | 
|  | def has_failed(job): | 
|  | state = job.get("state") | 
|  | # Ignore steps that don't have a state (like "wait"). | 
|  | return ( | 
|  | state | 
|  | and state != "passed" | 
|  | and not job.get("soft_failed") | 
|  | and job["id"] != current_job_id | 
|  | and job["name"] not in (BUILDIFIER_STEP_NAME, SHARD_SUMMARY_STEP_NAME) | 
|  | ) | 
|  |  | 
|  | failing_jobs = [j["name"] for j in build_info["jobs"] if has_failed(j)] | 
|  | if failing_jobs: | 
|  | raise BuildkiteException( | 
|  | "Cannot update last green commit due to {} failing step(s): {}".format( | 
|  | len(failing_jobs), ", ".join(failing_jobs) | 
|  | ) | 
|  | ) | 
|  |  | 
|  | git_repository = os.getenv("BUILDKITE_REPO") | 
|  | last_green_commit_url = bazelci_last_green_commit_url(git_repository, pipeline_slug) | 
|  | update_last_green_commit_if_newer(last_green_commit_url) | 
|  |  | 
|  |  | 
|  | def update_last_green_commit_if_newer(last_green_commit_url): | 
|  | last_green_commit = get_last_green_commit_by_url(last_green_commit_url) | 
|  | current_commit = resolve_revision("HEAD") | 
|  | if last_green_commit: | 
|  | success = False | 
|  | try: | 
|  | execute_command(["git", "fetch", "-v", "origin", last_green_commit]) | 
|  | success = True | 
|  | except subprocess.CalledProcessError: | 
|  | # If there was an error fetching the commit it typically means | 
|  | # that the commit does not exist anymore - due to a force push. In | 
|  | # order to recover from that assume that the current commit is the | 
|  | # newest commit. | 
|  | result = [current_commit] | 
|  | finally: | 
|  | if success: | 
|  | result = ( | 
|  | subprocess.check_output( | 
|  | ["git", "rev-list", "%s..%s" % (last_green_commit, current_commit)] | 
|  | ) | 
|  | .decode("utf-8") | 
|  | .strip() | 
|  | ) | 
|  | else: | 
|  | result = None | 
|  |  | 
|  | # If current_commit is newer that last_green_commit, `git rev-list A..B` will output a bunch of | 
|  | # commits, otherwise the output should be empty. | 
|  | if not last_green_commit or result: | 
|  | execute_command( | 
|  | [ | 
|  | "echo %s | %s -h 'Cache-Control: no-store' cp - %s" | 
|  | % (current_commit, gsutil_command(), last_green_commit_url) | 
|  | ], | 
|  | shell=True, | 
|  | ) | 
|  | else: | 
|  | eprint( | 
|  | "Updating abandoned: last green commit (%s) is not older than current commit (%s)." | 
|  | % (last_green_commit, current_commit) | 
|  | ) | 
|  |  | 
|  |  | 
|  | def resolve_revision(rev): | 
|  | return subprocess.check_output(["git", "rev-parse", rev]).decode("utf-8").strip() | 
|  |  | 
|  |  | 
|  | def try_update_last_green_downstream_commit(): | 
|  | last_green_commit_url = bazelci_last_green_downstream_commit_url() | 
|  | update_last_green_commit_if_newer(last_green_commit_url) | 
|  |  | 
|  |  | 
|  | def latest_generation_and_build_number(): | 
|  | generation = None | 
|  | output = None | 
|  | for attempt in range(5): | 
|  | output = subprocess.check_output( | 
|  | [gsutil_command(), "stat", bazelci_latest_build_metadata_url()], env=os.environ | 
|  | ) | 
|  | match = re.search("Generation:[ ]*([0-9]+)", output.decode("utf-8")) | 
|  | if not match: | 
|  | raise BuildkiteException("Couldn't parse generation. gsutil output format changed?") | 
|  | generation = match.group(1) | 
|  |  | 
|  | match = re.search(r"Hash \(md5\):[ ]*([^\s]+)", output.decode("utf-8")) | 
|  | if not match: | 
|  | raise BuildkiteException("Couldn't parse md5 hash. gsutil output format changed?") | 
|  | expected_md5hash = base64.b64decode(match.group(1)) | 
|  |  | 
|  | output = subprocess.check_output( | 
|  | [gsutil_command(), "cat", bazelci_latest_build_metadata_url()], env=os.environ | 
|  | ) | 
|  | hasher = hashlib.md5() | 
|  | hasher.update(output) | 
|  | actual_md5hash = hasher.digest() | 
|  |  | 
|  | if expected_md5hash == actual_md5hash: | 
|  | break | 
|  | info = json.loads(output.decode("utf-8")) | 
|  | return generation, info["build_number"] | 
|  |  | 
|  |  | 
|  | def sha256_hexdigest(filename): | 
|  | sha256 = hashlib.sha256() | 
|  | with open(filename, "rb") as f: | 
|  | for block in iter(lambda: f.read(65536), b""): | 
|  | sha256.update(block) | 
|  | return sha256.hexdigest() | 
|  |  | 
|  |  | 
|  | def print_shard_summary(): | 
|  | tmpdir = tempfile.mkdtemp() | 
|  | try: | 
|  | print_collapsed_group("Fetching test artifacts...") | 
|  | all_test_artifacts = get_test_artifacts() | 
|  | print_collapsed_group("Downloading & parsing BEP files...") | 
|  | perf_stats = [] | 
|  | for base_task, current_test_artifacts in all_test_artifacts.items(): | 
|  | failures = [] | 
|  | for test_artifact in current_test_artifacts: | 
|  | local_bep_path = test_artifact.download_bep(tmpdir) | 
|  | if not local_bep_path: | 
|  | eprint( | 
|  | f"Skipping step {test_artifact.job_id} since " | 
|  | "{test_artifact.relative_bep_path} could not be downloaded..." | 
|  | ) | 
|  | continue | 
|  |  | 
|  | for test_execution in parse_bep(local_bep_path): | 
|  | for shard in test_execution.shards: | 
|  | perf_stats.append(shard.get_metrics(test_execution.label, test_artifact)) | 
|  |  | 
|  | if test_execution.overall_status == "PASSED": | 
|  | continue | 
|  | if len(test_execution.shards) == 1: | 
|  | continue | 
|  |  | 
|  | failures.append( | 
|  | test_execution.Format(test_artifact.job_id, test_artifact.is_windows) | 
|  | ) | 
|  |  | 
|  | if failures: | 
|  | message = "\n".join(failures) | 
|  | execute_command( | 
|  | [ | 
|  | "buildkite-agent", | 
|  | "annotate", | 
|  | "--style=error", | 
|  | f"**{base_task} Failures**\n\n{message}", | 
|  | "--context", | 
|  | f"{base_task}", | 
|  | ] | 
|  | ) | 
|  |  | 
|  | if perf_stats: | 
|  | slowest = sorted(perf_stats, key=lambda x: -x[0]) | 
|  | message = "\n".join( | 
|  | f"{rank}. {tpl[1]}" for rank, tpl in enumerate(slowest[:_SLOWEST_N_TARGETS], 1) | 
|  | ) | 
|  | execute_command( | 
|  | [ | 
|  | "buildkite-agent", | 
|  | "annotate", | 
|  | "--style=warning", | 
|  | f"**Slowest {_SLOWEST_N_TARGETS} targets**\n\n{message}", | 
|  | "--context", | 
|  | "slowest_targets", | 
|  | ] | 
|  | ) | 
|  | except Exception as ex: | 
|  | eprint(f"Failed to print shard summary: {ex}") | 
|  | finally: | 
|  | shutil.rmtree(tmpdir) | 
|  |  | 
|  |  | 
|  | def get_log_path_for_label(label, shard, total_shards, attempt, total_attempts, is_windows): | 
|  | parts = [label.lstrip("/").replace(":", "/")] | 
|  | if total_shards > 1: | 
|  | parts.append(f"shard_{shard}_of_{total_shards}") | 
|  | if total_attempts > 1: | 
|  | parts.append(f"test_attempts/attempt_{attempt}.log") | 
|  | else: | 
|  | parts.append("test.log") | 
|  |  | 
|  | path = "/".join(parts) | 
|  | return path.replace("/", ESCAPED_BACKSLASH) if is_windows else path | 
|  |  | 
|  |  | 
|  | def get_test_artifacts(): | 
|  | org_slug = os.getenv("BUILDKITE_ORGANIZATION_SLUG") | 
|  | pipeline_slug = os.getenv("BUILDKITE_PIPELINE_SLUG") | 
|  | build_number = os.getenv("BUILDKITE_BUILD_NUMBER") | 
|  |  | 
|  | client = BuildkiteClient(org=org_slug, pipeline=pipeline_slug) | 
|  | build_info = client.get_build_info(build_number) | 
|  |  | 
|  | paths = collections.defaultdict(list) | 
|  | for job in build_info["jobs"]: | 
|  | state = job.get("state") | 
|  | if not state:  # Wait steps etc. | 
|  | continue | 
|  |  | 
|  | job_name = job.get("name", "") | 
|  | # This is a bit hacky, but saves us one API request per job (to check for BUILDKITE_PARALLEL_JOB) | 
|  | match = _SHARD_RE.search(job_name) | 
|  | base_task = match.group(1) if match else job_name | 
|  |  | 
|  | relative_bep_path, relative_log_paths = get_test_file_paths(job["id"]) | 
|  | if not relative_bep_path: | 
|  | continue | 
|  |  | 
|  | ta = TestArtifacts( | 
|  | job_id=job["id"], | 
|  | job_name=job_name, | 
|  | job_url=job.get("web_url"), | 
|  | relative_bep_path=relative_bep_path, | 
|  | relative_log_paths=relative_log_paths, | 
|  | ) | 
|  | paths[base_task].append(ta) | 
|  |  | 
|  | return paths | 
|  |  | 
|  |  | 
|  | class TestArtifacts: | 
|  | def __init__(self, job_id, job_name, job_url, relative_bep_path, relative_log_paths) -> None: | 
|  | self.job_id = job_id | 
|  | self.job_name = job_name | 
|  | self.job_url = job_url | 
|  | self.relative_bep_path = relative_bep_path | 
|  | self.relative_log_paths = relative_log_paths | 
|  |  | 
|  | def download_bep(self, dest_dir: str) -> str: | 
|  | if not LOG_BUCKET: | 
|  | return None | 
|  |  | 
|  | job_dir = os.path.join(dest_dir, self.job_id) | 
|  | os.makedirs(job_dir) | 
|  |  | 
|  | # We cannot use `buildkite agent download *` since it cannot handle backslashes in Windows artifact paths. | 
|  | # Consequently, we just escape all backslashes and download files directly from GCS. | 
|  | url = "/".join( | 
|  | [LOG_BUCKET, self.job_id, self.relative_bep_path.replace("\\", ESCAPED_BACKSLASH)] | 
|  | ) | 
|  | try: | 
|  | return download_file(url, job_dir, _TEST_BEP_FILE) | 
|  | except: | 
|  | # TODO: handle exception | 
|  | return None | 
|  |  | 
|  | @property | 
|  | def is_windows(self) -> bool: | 
|  | # Dirty hack | 
|  | return "\\" in self.relative_bep_path | 
|  |  | 
|  |  | 
|  | def get_test_file_paths(job_id): | 
|  | bep_path = None | 
|  | log_paths = [] | 
|  |  | 
|  | output = execute_command_and_get_output( | 
|  | [ | 
|  | "buildkite-agent", | 
|  | "artifact", | 
|  | "search", | 
|  | "*", | 
|  | "--step", | 
|  | job_id, | 
|  | ], | 
|  | fail_if_nonzero=False, | 
|  | ).strip() | 
|  |  | 
|  | if not output or "no matches found" in output: | 
|  | return None, [] | 
|  |  | 
|  | for line in output.split("\n"): | 
|  | parts = line.split(" ") | 
|  | # Expected format: | 
|  | # JOB_ID FILE_PATH TIMESTAMP | 
|  | if len(parts) != 3: | 
|  | continue | 
|  |  | 
|  | path = parts[1] | 
|  | if path.endswith(_TEST_BEP_FILE): | 
|  | bep_path = path | 
|  | elif path.endswith(".log"): | 
|  | log_paths.append(path) | 
|  |  | 
|  | return bep_path, log_paths | 
|  |  | 
|  |  | 
|  | def format_millis(millis): | 
|  | def fmt(ms): | 
|  | return "{:.1f}s".format(ms / 1000) | 
|  |  | 
|  | if len(millis) == 1: | 
|  | return fmt(millis[0]) | 
|  |  | 
|  | total = sum(millis) | 
|  | return f"{fmt(total)} ({' + '.join(fmt(ms) for ms in millis)})" | 
|  |  | 
|  |  | 
|  | def format_test_status(status): | 
|  | cls = {"PASSED": "green", "FLAKY": "purple"}.get(status, "red") | 
|  | return f"<span class='{cls}'>{status}</span>" | 
|  |  | 
|  |  | 
|  | # TODO here and below: use @dataclasses.dataclass(frozen=True) once Python has been updated on Docker machines | 
|  | class TestAttempt: | 
|  | def __init__(self, number, status, millis, root_cause) -> None: | 
|  | self.number = number | 
|  | self.status = status | 
|  | self.millis = millis | 
|  | self.root_cause = root_cause | 
|  |  | 
|  |  | 
|  | class TestShard: | 
|  | def __init__(self, number, total_shards, attempts) -> None: | 
|  | self.number = number | 
|  | self.total_shards = total_shards | 
|  | self.attempts = attempts | 
|  |  | 
|  | def _get_detailed_overall_status(self): | 
|  | counter = collections.Counter([a.status for a in self.attempts]) | 
|  | passed = counter["PASSED"] | 
|  | no_attempts = len(self.attempts) | 
|  | if passed == no_attempts: | 
|  | return "PASSED", no_attempts, no_attempts | 
|  | elif passed and passed < no_attempts: | 
|  | return "FLAKY", no_attempts - passed, no_attempts | 
|  | elif counter["FAILED"]: | 
|  | return "FAILED", counter["FAILED"], no_attempts | 
|  |  | 
|  | [(status, count)] = counter.most_common(1) | 
|  | return status, count, no_attempts | 
|  |  | 
|  | def get_details(self): | 
|  | overall, bad_runs, total_runs = self._get_detailed_overall_status() | 
|  | qualifier = "" if not bad_runs else f"{bad_runs} out of " | 
|  | runs = f" in {qualifier}{total_runs} runs" if total_runs > 1 else "" | 
|  | time = f" in {format_millis(self.attempt_millis)}" if self.attempt_millis else "" | 
|  | cause = f" because of {self.root_cause}" if self.root_cause else "" | 
|  | return overall, f"{runs}{time}{cause}" | 
|  |  | 
|  | def get_metrics(self, label, test_artifact): | 
|  | total_time = sum(self.attempt_millis) | 
|  | shard_info = "" if self.total_shards == 1 else f" (shard {self.number}/{self.total_shards})" | 
|  | overall, stats = self.get_details() | 
|  | return ( | 
|  | total_time, | 
|  | f"{label}{shard_info} on {test_artifact.job_name}: {format_test_status(overall)} {stats} ([log]({test_artifact.job_url}))", | 
|  | ) | 
|  |  | 
|  | @property | 
|  | def overall_status(self): | 
|  | return self._get_detailed_overall_status()[0] | 
|  |  | 
|  | @property | 
|  | def attempt_millis(self): | 
|  | return [a.millis for a in self.attempts] | 
|  |  | 
|  | @property | 
|  | def root_cause(self): | 
|  | for a in self.attempts: | 
|  | if a.root_cause: | 
|  | return a.root_cause | 
|  |  | 
|  |  | 
|  | class TestExecution: | 
|  | def __init__(self, label, shards) -> None: | 
|  | self.label = label | 
|  | self.shards = shards | 
|  |  | 
|  | @property | 
|  | def overall_status(self): | 
|  | status_set = set(s.overall_status for s in self.shards) | 
|  | if len(status_set) > 1: | 
|  | for status in ( | 
|  | "FAILED", | 
|  | "TIMEOUT", | 
|  | "NO_STATUS", | 
|  | "INCOMPLETE", | 
|  | "REMOTE_FAILURE", | 
|  | "FAILED_TO_BUILD", | 
|  | "PASSED", | 
|  | ): | 
|  | if status in status_set: | 
|  | return status | 
|  |  | 
|  | return next(iter(status_set)) | 
|  |  | 
|  | @property | 
|  | def critical_path(self): | 
|  | max_millis = 0 | 
|  | path = None | 
|  |  | 
|  | for s in self.shards: | 
|  | duration_millis = sum(s.attempt_millis) | 
|  | if duration_millis > max_millis: | 
|  | max_millis = duration_millis | 
|  | path = s.attempt_millis | 
|  |  | 
|  | return format_millis(path) | 
|  |  | 
|  | def Format(self, job_id: str, is_windows: bool) -> str: | 
|  | def get_log_url_for_shard(s): | 
|  | local_log_path = get_log_path_for_label( | 
|  | self.label, | 
|  | s.number, | 
|  | len(self.shards), | 
|  | 1, | 
|  | len(s.attempts), | 
|  | is_windows, | 
|  | ) | 
|  | # TODO: check in relative_log_paths if log really exists? | 
|  | return os.path.join(LOG_BUCKET, job_id, local_log_path) | 
|  |  | 
|  | def format_shard(s): | 
|  | overall, statistics = s.get_details() | 
|  | # There are no per-target log files for FAILED_TO_BUILD tests, so we simply | 
|  | # link to the step's test log in this case. | 
|  | log = f"#{job_id}" if overall == "FAILED_TO_BUILD" else get_log_url_for_shard(s) | 
|  | return f"{format_test_status(overall)} {statistics}: [log]({log})" | 
|  |  | 
|  | failing_shards = [s for s in self.shards if s.overall_status != "PASSED"] | 
|  | if len(failing_shards) == 1: | 
|  | [shard] = failing_shards | 
|  | # TODO: show log links for failing attempts > 1? | 
|  | return f"- {self.label} {format_shard(shard)}" | 
|  |  | 
|  | shard_info = "\n".join( | 
|  | f"  - Shard {s.number}/{len(self.shards)}: {format_shard(s)}" for s in failing_shards | 
|  | ) | 
|  | return f"- {self.label}\n{shard_info}" | 
|  |  | 
|  |  | 
|  | def parse_bep(path): | 
|  | data = collections.defaultdict(dict) | 
|  | for test, shard, attempt, status, millis, root_cause in get_test_results_from_bep(path): | 
|  | ta = TestAttempt(number=attempt, status=status, millis=millis, root_cause=root_cause) | 
|  | if shard not in data[test]: | 
|  | data[test][shard] = [] | 
|  |  | 
|  | data[test][shard].append(ta) | 
|  |  | 
|  | tests = [] | 
|  | for test, attempts_per_shard in data.items(): | 
|  | shards = [ | 
|  | TestShard( | 
|  | number=shard, total_shards=len(data[test]), attempts=attempts_per_shard[shard] | 
|  | ) | 
|  | for shard in sorted(attempts_per_shard.keys()) | 
|  | ] | 
|  | tests.append(TestExecution(label=test, shards=shards)) | 
|  |  | 
|  | return tests | 
|  |  | 
|  |  | 
|  | def get_test_results_from_bep(path): | 
|  | with open(path, "rt") as f: | 
|  | for line in f: | 
|  | data = json.loads(line) | 
|  | entry = data.get("id", {}) | 
|  | test_result = entry.get("testResult") | 
|  | target_completed = entry.get("targetCompleted") | 
|  |  | 
|  | if test_result: | 
|  | if "testResult" not in data: | 
|  | continue | 
|  |  | 
|  | yield ( | 
|  | test_result["label"], | 
|  | test_result["shard"], | 
|  | test_result["attempt"], | 
|  | data["testResult"]["status"], | 
|  | int(data["testResult"]["testAttemptDurationMillis"]), | 
|  | None, | 
|  | ) | 
|  | elif target_completed: | 
|  | # There are no "testResult" events for targets that fail to build, | 
|  | # so we have to check for targetCompleted events that have a failureDetail | 
|  | # message. | 
|  | if data.get("completed", {}).get("failureDetail"): | 
|  | yield ( | 
|  | target_completed["label"], | 
|  | 1, | 
|  | 1, | 
|  | "FAILED_TO_BUILD", | 
|  | 0, | 
|  | get_root_cause(data), | 
|  | ) | 
|  |  | 
|  |  | 
|  | def get_root_cause(bep_event): | 
|  | for c in bep_event.get("children", ()): | 
|  | for v in c.values(): | 
|  | if "label" in v: | 
|  | return v["label"] | 
|  |  | 
|  |  | 
|  | def upload_bazel_binaries(): | 
|  | """ | 
|  | Uploads all Bazel binaries to a deterministic URL based on the current Git commit. | 
|  |  | 
|  | Returns maps of platform names to sha256 hashes of the corresponding bazel and bazel_nojdk binaries. | 
|  | """ | 
|  | bazel_hashes = {} | 
|  | bazel_nojdk_hashes = {} | 
|  | error = None | 
|  | for platform_name, platform in PLATFORMS.items(): | 
|  | if not should_publish_binaries_for_platform(platform_name): | 
|  | continue | 
|  | tmpdir = tempfile.mkdtemp() | 
|  | try: | 
|  | bazel_binary_path = download_bazel_binary(tmpdir, platform_name) | 
|  | # One platform that we build on can generate binaries for multiple platforms, e.g. | 
|  | # the centos7 platform generates binaries for the "centos7" platform, but also | 
|  | # for the generic "linux" platform. | 
|  | for target_platform_name in platform["publish_binary"]: | 
|  | execute_command( | 
|  | [ | 
|  | gsutil_command(), | 
|  | "cp", | 
|  | bazel_binary_path, | 
|  | bazelci_builds_gs_url(target_platform_name, os.environ["BUILDKITE_COMMIT"]), | 
|  | ] | 
|  | ) | 
|  | bazel_hashes[target_platform_name] = sha256_hexdigest(bazel_binary_path) | 
|  |  | 
|  | # Also publish bazel_nojdk binaries. | 
|  | bazel_nojdk_binary_path = download_bazel_nojdk_binary(tmpdir, platform_name) | 
|  | for target_platform_name in platform["publish_binary"]: | 
|  | execute_command( | 
|  | [ | 
|  | gsutil_command(), | 
|  | "cp", | 
|  | bazel_nojdk_binary_path, | 
|  | bazelci_builds_nojdk_gs_url( | 
|  | target_platform_name, os.environ["BUILDKITE_COMMIT"] | 
|  | ), | 
|  | ] | 
|  | ) | 
|  | bazel_nojdk_hashes[target_platform_name] = sha256_hexdigest(bazel_nojdk_binary_path) | 
|  | except subprocess.CalledProcessError as e: | 
|  | eprint( | 
|  | "Failured to download and publish Bazel binary for platform {}: {}".format( | 
|  | platform_name, e | 
|  | ) | 
|  | ) | 
|  | error = e | 
|  | finally: | 
|  | shutil.rmtree(tmpdir) | 
|  | # If we're not on the main branch, we're probably building a custom one-off binary and | 
|  | # ignore failures for individual platforms (it's possible that we didn't build binaries | 
|  | # for all platforms). | 
|  | if error and current_branch_is_main_branch(): | 
|  | raise error | 
|  | return bazel_hashes, bazel_nojdk_hashes | 
|  |  | 
|  |  | 
|  | def try_publish_binaries(bazel_hashes, bazel_nojdk_hashes, build_number, expected_generation): | 
|  | """ | 
|  | Uploads the info.json file that contains information about the latest Bazel commit that was | 
|  | successfully built on CI. | 
|  | """ | 
|  | now = datetime.datetime.now() | 
|  | git_commit = os.environ["BUILDKITE_COMMIT"] | 
|  | info = { | 
|  | "build_number": build_number, | 
|  | "build_time": now.strftime("%d-%m-%Y %H:%M"), | 
|  | "git_commit": git_commit, | 
|  | "platforms": {}, | 
|  | } | 
|  | for platform, sha256 in bazel_hashes.items(): | 
|  | info["platforms"][platform] = { | 
|  | "url": bazelci_builds_download_url(platform, git_commit), | 
|  | "sha256": sha256, | 
|  | "nojdk_url": bazelci_builds_nojdk_download_url(platform, git_commit), | 
|  | "nojdk_sha256": bazel_nojdk_hashes[platform], | 
|  | } | 
|  | tmpdir = tempfile.mkdtemp() | 
|  | try: | 
|  | info_file = os.path.join(tmpdir, "info.json") | 
|  | with open(info_file, mode="w", encoding="utf-8") as fp: | 
|  | json.dump(info, fp, indent=2, sort_keys=True) | 
|  |  | 
|  | try: | 
|  | execute_command( | 
|  | [ | 
|  | gsutil_command(), | 
|  | "-h", | 
|  | "x-goog-if-generation-match:" + expected_generation, | 
|  | "-h", | 
|  | "Content-Type:application/json", | 
|  | "cp", | 
|  | info_file, | 
|  | bazelci_latest_build_metadata_url(), | 
|  | ] | 
|  | ) | 
|  | except subprocess.CalledProcessError: | 
|  | raise BinaryUploadRaceException() | 
|  |  | 
|  | execute_command( | 
|  | [ | 
|  | gsutil_command(), | 
|  | "cp", | 
|  | bazelci_latest_build_metadata_url(), | 
|  | bazelci_builds_metadata_url(git_commit), | 
|  | ] | 
|  | ) | 
|  | finally: | 
|  | shutil.rmtree(tmpdir) | 
|  |  | 
|  |  | 
|  | def publish_binaries(): | 
|  | """ | 
|  | Publish Bazel binaries to GCS. | 
|  | """ | 
|  | current_build_number = os.environ.get("BUILDKITE_BUILD_NUMBER", None) | 
|  | if not current_build_number: | 
|  | raise BuildkiteException("Not running inside Buildkite") | 
|  | current_build_number = int(current_build_number) | 
|  |  | 
|  | # Upload the Bazel binaries for this commit. | 
|  | bazel_hashes, bazel_nojdk_hashes = upload_bazel_binaries() | 
|  |  | 
|  | # Try to update the info.json with data about our build. This will fail (expectedly) if we're | 
|  | # not the latest build. Only do this if we're building binaries from the main branch to avoid | 
|  | # accidentally publishing a custom debug build as the "latest" Bazel binary. | 
|  | if current_branch_is_main_branch(): | 
|  | for _ in range(5): | 
|  | latest_generation, latest_build_number = latest_generation_and_build_number() | 
|  |  | 
|  | if current_build_number <= latest_build_number: | 
|  | eprint( | 
|  | ( | 
|  | "Current build '{0}' is not newer than latest published '{1}'. " | 
|  | + "Skipping publishing of binaries." | 
|  | ).format(current_build_number, latest_build_number) | 
|  | ) | 
|  | break | 
|  |  | 
|  | try: | 
|  | try_publish_binaries( | 
|  | bazel_hashes, bazel_nojdk_hashes, current_build_number, latest_generation | 
|  | ) | 
|  | except BinaryUploadRaceException: | 
|  | # Retry. | 
|  | continue | 
|  |  | 
|  | eprint( | 
|  | "Successfully updated '{0}' to binaries from build {1}.".format( | 
|  | bazelci_latest_build_metadata_url(), current_build_number | 
|  | ) | 
|  | ) | 
|  | break | 
|  | else: | 
|  | raise BuildkiteException("Could not publish binaries, ran out of attempts.") | 
|  |  | 
|  |  | 
|  | # This is so that multiline python strings are represented as YAML | 
|  | # block strings. | 
|  | def str_presenter(dumper, data): | 
|  | if len(data.splitlines()) > 1:  # check for multiline string | 
|  | return dumper.represent_scalar("tag:yaml.org,2002:str", data, style="|") | 
|  | return dumper.represent_scalar("tag:yaml.org,2002:str", data) | 
|  |  | 
|  |  | 
|  | def main(argv=None): | 
|  | if argv is None: | 
|  | argv = sys.argv[1:] | 
|  |  | 
|  | yaml.add_representer(str, str_presenter) | 
|  |  | 
|  | parser = argparse.ArgumentParser(description="Bazel Continuous Integration Script") | 
|  | parser.add_argument("--script", type=str) | 
|  |  | 
|  | subparsers = parser.add_subparsers(dest="subparsers_name") | 
|  |  | 
|  | bazel_publish_binaries_pipeline = subparsers.add_parser("bazel_publish_binaries_pipeline") | 
|  | bazel_publish_binaries_pipeline.add_argument("--file_config", type=str) | 
|  | bazel_publish_binaries_pipeline.add_argument("--http_config", type=str) | 
|  | bazel_publish_binaries_pipeline.add_argument("--git_repository", type=str) | 
|  |  | 
|  | bazel_downstream_pipeline = subparsers.add_parser("bazel_downstream_pipeline") | 
|  | bazel_downstream_pipeline.add_argument("--file_config", type=str) | 
|  | bazel_downstream_pipeline.add_argument("--http_config", type=str) | 
|  | bazel_downstream_pipeline.add_argument("--git_repository", type=str) | 
|  | bazel_downstream_pipeline.add_argument( | 
|  | "--test_disabled_projects", type=bool, nargs="?", const=True | 
|  | ) | 
|  | bazel_downstream_pipeline.add_argument("--notify", type=bool, nargs="?", const=True) | 
|  |  | 
|  | project_pipeline = subparsers.add_parser("project_pipeline") | 
|  | project_pipeline.add_argument("--project_name", type=str) | 
|  | project_pipeline.add_argument("--file_config", type=str) | 
|  | project_pipeline.add_argument("--http_config", type=str) | 
|  | project_pipeline.add_argument("--git_repository", type=str) | 
|  | project_pipeline.add_argument("--monitor_flaky_tests", type=bool, nargs="?", const=True) | 
|  | project_pipeline.add_argument("--use_but", type=bool, nargs="?", const=True) | 
|  | project_pipeline.add_argument("--notify", type=bool, nargs="?", const=True) | 
|  | project_pipeline.add_argument("--print_shard_summary", type=bool, nargs="?", const=True) | 
|  |  | 
|  | runner = subparsers.add_parser("runner") | 
|  | runner.add_argument("--task", action="store", type=str, default="") | 
|  | runner.add_argument("--file_config", type=str) | 
|  | runner.add_argument("--http_config", type=str) | 
|  | runner.add_argument("--git_repository", type=str) | 
|  | runner.add_argument( | 
|  | "--git_commit", type=str, help="Reset the git repository to this commit after cloning it" | 
|  | ) | 
|  | runner.add_argument( | 
|  | "--repo_location", | 
|  | type=str, | 
|  | help="Use an existing repository instead of cloning from github", | 
|  | ) | 
|  | runner.add_argument( | 
|  | "--use_bazel_at_commit", type=str, help="Use Bazel binary built at a specific commit" | 
|  | ) | 
|  | runner.add_argument("--use_but", type=bool, nargs="?", const=True) | 
|  | runner.add_argument("--save_but", type=bool, nargs="?", const=True) | 
|  | runner.add_argument("--needs_clean", type=bool, nargs="?", const=True) | 
|  | runner.add_argument("--build_only", type=bool, nargs="?", const=True) | 
|  | runner.add_argument("--test_only", type=bool, nargs="?", const=True) | 
|  | runner.add_argument("--monitor_flaky_tests", type=bool, nargs="?", const=True) | 
|  |  | 
|  | subparsers.add_parser("publish_binaries") | 
|  | subparsers.add_parser("try_update_last_green_commit") | 
|  | subparsers.add_parser("try_update_last_green_downstream_commit") | 
|  | subparsers.add_parser("print_shard_summary") | 
|  |  | 
|  | args = parser.parse_args(argv) | 
|  |  | 
|  | if args.script: | 
|  | global SCRIPT_URL | 
|  | SCRIPT_URL = args.script | 
|  |  | 
|  | try: | 
|  | if args.subparsers_name == "bazel_publish_binaries_pipeline": | 
|  | configs = fetch_configs(args.http_config, args.file_config) | 
|  | print_bazel_publish_binaries_pipeline( | 
|  | task_configs=configs.get("tasks", None), | 
|  | http_config=args.http_config, | 
|  | file_config=args.file_config, | 
|  | ) | 
|  | elif args.subparsers_name == "bazel_downstream_pipeline": | 
|  | # If USE_BAZELISK_MIGRATE is true, we don't need to fetch task configs for Bazel | 
|  | # since we use Bazelisk to fetch Bazel binaries. | 
|  | configs = ( | 
|  | {} if use_bazelisk_migrate() else fetch_configs(args.http_config, args.file_config) | 
|  | ) | 
|  | print_bazel_downstream_pipeline( | 
|  | task_configs=configs.get("tasks", None), | 
|  | http_config=args.http_config, | 
|  | file_config=args.file_config, | 
|  | test_disabled_projects=args.test_disabled_projects, | 
|  | notify=args.notify, | 
|  | ) | 
|  | elif args.subparsers_name == "project_pipeline": | 
|  | # Fetch the repo in case we need to use file_config. | 
|  | if args.git_repository: | 
|  | git_commit = ( | 
|  | get_last_green_commit(args.project_name) if is_downstream_pipeline() else None | 
|  | ) | 
|  | clone_git_repository(args.git_repository, git_commit, suppress_stdout=True) | 
|  |  | 
|  | configs = fetch_configs(args.http_config, args.file_config) | 
|  | print_project_pipeline( | 
|  | configs=configs, | 
|  | project_name=args.project_name, | 
|  | http_config=args.http_config, | 
|  | file_config=args.file_config, | 
|  | git_repository=args.git_repository, | 
|  | monitor_flaky_tests=args.monitor_flaky_tests, | 
|  | use_but=args.use_but, | 
|  | notify=args.notify, | 
|  | print_shard_summary=args.print_shard_summary, | 
|  | ) | 
|  | elif args.subparsers_name == "runner": | 
|  | # Fetch the repo in case we need to use file_config. | 
|  | if args.repo_location: | 
|  | os.chdir(args.repo_location) | 
|  | elif args.git_repository: | 
|  | clone_git_repository(args.git_repository, args.git_commit) | 
|  |  | 
|  | configs = fetch_configs(args.http_config, args.file_config) | 
|  | tasks = configs.get("tasks", {}) | 
|  | task_config = tasks.get(args.task) | 
|  | if not task_config: | 
|  | raise BuildkiteException( | 
|  | "No such task '{}' in configuration. Available: {}".format( | 
|  | args.task, ", ".join(tasks) | 
|  | ) | 
|  | ) | 
|  |  | 
|  | os.environ["BAZELCI_TASK"] = args.task | 
|  |  | 
|  | platform = get_platform_for_task(args.task, task_config) | 
|  |  | 
|  | # The value of `BUILDKITE_MESSAGE` defaults to the commit message, which can be too large | 
|  | # on Windows, therefore we truncate the value to 1000 characters. | 
|  | # See https://github.com/bazelbuild/continuous-integration/issues/1218 | 
|  | if "BUILDKITE_MESSAGE" in os.environ: | 
|  | os.environ["BUILDKITE_MESSAGE"] = os.environ["BUILDKITE_MESSAGE"][:1000] | 
|  |  | 
|  | execute_commands( | 
|  | task_config=task_config, | 
|  | platform=platform, | 
|  | use_bazel_at_commit=args.use_bazel_at_commit, | 
|  | use_but=args.use_but, | 
|  | save_but=args.save_but, | 
|  | needs_clean=args.needs_clean, | 
|  | build_only=args.build_only, | 
|  | test_only=args.test_only, | 
|  | monitor_flaky_tests=args.monitor_flaky_tests, | 
|  | bazel_version=task_config.get("bazel") or configs.get("bazel"), | 
|  | ) | 
|  | elif args.subparsers_name == "publish_binaries": | 
|  | publish_binaries() | 
|  | elif args.subparsers_name == "try_update_last_green_commit": | 
|  | # Update the last green commit of a project pipeline | 
|  | try_update_last_green_commit() | 
|  | elif args.subparsers_name == "try_update_last_green_downstream_commit": | 
|  | # Update the last green commit of the downstream pipeline | 
|  | try_update_last_green_downstream_commit() | 
|  | elif args.subparsers_name == "print_shard_summary": | 
|  | print_shard_summary() | 
|  | else: | 
|  | parser.print_help() | 
|  | return 2 | 
|  | except BuildkiteException as e: | 
|  | eprint(str(e)) | 
|  | return 1 | 
|  | return 0 | 
|  |  | 
|  |  | 
|  | if __name__ == "__main__": | 
|  | sys.exit(main()) |