From 41720e98afba7582b89b5425642109acfe75130a Mon Sep 17 00:00:00 2001 From: David Dunleavy Date: Tue, 11 Feb 2025 14:03:27 -0800 Subject: [PATCH] Delete TensorFlow GPU Kokoro build now that GitHub Actions build runs continuously Also don't `bazel build --nobuild` for the new TensorFlow GHA GPU build. PiperOrigin-RevId: 725764079 --- build_tools/ci/build.py | 30 +----------------------------- build_tools/ci/golden_commands.txt | 10 ---------- 2 files changed, 1 insertion(+), 39 deletions(-) diff --git a/build_tools/ci/build.py b/build_tools/ci/build.py index b717fcea1f4b9..389003720662c 100755 --- a/build_tools/ci/build.py +++ b/build_tools/ci/build.py @@ -101,7 +101,6 @@ class BuildType(enum.Enum): JAX_X86_GPU_T4_SELF_HOSTED = enum.auto() TENSORFLOW_CPU_SELF_HOSTED = enum.auto() - TENSORFLOW_GPU = enum.auto() TENSORFLOW_X86_GPU_T4_SELF_HOSTED = enum.auto() @@ -214,7 +213,7 @@ def commands(self) -> List[List[str]]: # manually). if self.type_ not in ( BuildType.TENSORFLOW_CPU_SELF_HOSTED, - BuildType.TENSORFLOW_GPU, + BuildType.TENSORFLOW_X86_GPU_T4_SELF_HOSTED, BuildType.MACOS_CPU_X86, ): cmds.append( @@ -460,32 +459,6 @@ def nvidia_gpu_build_with_compute_capability( ), ) -_TENSORFLOW_GPU_BUILD = Build( - type_=BuildType.TENSORFLOW_GPU, - repo="tensorflow/tensorflow", - image_url=_ML_BUILD_IMAGE, - configs=( - "release_gpu_linux", - "rbe_linux_cuda", - "linux_cuda_pycpp_test_filters", - ), - target_patterns=( - "//tensorflow/compiler/...", - "-//tensorflow/compiler/tf2tensorrt/...", - "//tensorflow/python/...", - "-//tensorflow/python/distribute/...", - "-//tensorflow/python/compiler/tensorrt/...", - ), - build_tag_filters=("-no_oss", "+gpu"), - test_tag_filters=("-no_oss", "+gpu"), - options=dict( - verbose_failures=True, - test_output="errors", - override_repository="xla=/github/xla", - profile="profile.json.gz", - ), -) - _TENSORFLOW_GPU_SELF_HOSTED_BUILD = Build( type_=BuildType.TENSORFLOW_X86_GPU_T4_SELF_HOSTED, repo="tensorflow/tensorflow", @@ -517,7 +490,6 @@ def nvidia_gpu_build_with_compute_capability( "tensorflow/xla/linux/github_continuous/build_gpu": _GPU_BUILD, "tensorflow/xla/macos/github_continuous/cpu_py39_full": _MACOS_X86_BUILD, "tensorflow/xla/jax/gpu/build_gpu": _JAX_GPU_BUILD, - "tensorflow/xla/tensorflow/gpu/build_gpu": _TENSORFLOW_GPU_BUILD, "xla-linux-x86-cpu": _CPU_X86_SELF_HOSTED_BUILD, "xla-linux-arm64-cpu": _CPU_ARM64_SELF_HOSTED_BUILD, "xla-linux-x86-gpu-t4": _GPU_T4_SELF_HOSTED_BUILD, diff --git a/build_tools/ci/golden_commands.txt b/build_tools/ci/golden_commands.txt index 9491e34c80b52..70c2f324650ad 100644 --- a/build_tools/ci/golden_commands.txt +++ b/build_tools/ci/golden_commands.txt @@ -57,17 +57,7 @@ bazel analyze-profile profile.json.gz bazel test --build_tag_filters= --test_tag_filters= --config=release_cpu_linux --config=rbe_linux_cpu --config=linux_cpu_pycpp_test_filters --verbose_failures --test_output=errors --override_repository=xla=$GITHUB_WORKSPACE/openxla/xla --profile=profile.json.gz -- //tensorflow/compiler/... -//tensorflow/compiler/tf2tensorrt/... //tensorflow/python/... -//tensorflow/python/distribute/... -//tensorflow/python/compiler/tensorrt/... bazel analyze-profile profile.json.gz # END BuildType.TENSORFLOW_CPU_SELF_HOSTED -# BEGIN BuildType.TENSORFLOW_GPU -$KOKORO_ARTIFACTS_DIR/github/xla/.kokoro/generate_index_html.sh index.html -git clone --depth=1 https://github.com/tensorflow/tensorflow ./github/tensorflow -parallel --ungroup --retries 3 --delay 15 --nonall -- docker pull us-central1-docker.pkg.dev/tensorflow-sigs/tensorflow/ml-build:latest -docker run --detach --name=xla_ci --rm --interactive --tty --volume=./github:/github --workdir=/github/tensorflow us-central1-docker.pkg.dev/tensorflow-sigs/tensorflow/ml-build:latest bash -docker exec xla_ci bazel test --build_tag_filters=-no_oss,+gpu --test_tag_filters=-no_oss,+gpu --config=release_gpu_linux --config=rbe_linux_cuda --config=linux_cuda_pycpp_test_filters --verbose_failures --test_output=errors --override_repository=xla=/github/xla --profile=profile.json.gz -- //tensorflow/compiler/... -//tensorflow/compiler/tf2tensorrt/... //tensorflow/python/... -//tensorflow/python/distribute/... -//tensorflow/python/compiler/tensorrt/... -docker exec xla_ci bazel analyze-profile profile.json.gz -docker stop xla_ci -# END BuildType.TENSORFLOW_GPU # BEGIN BuildType.TENSORFLOW_X86_GPU_T4_SELF_HOSTED -parallel --ungroup --retries 3 --delay 15 --nonall -- bazel build --build_tag_filters=-no_oss,+gpu --test_tag_filters=-no_oss,+gpu --config=release_gpu_linux --config=rbe_linux_cuda --config=linux_cuda_pycpp_test_filters --verbose_failures --test_output=errors --override_repository=xla=$GITHUB_WORKSPACE/openxla/xla --profile=profile.json.gz --nobuild -- //tensorflow/compiler/... -//tensorflow/compiler/tf2tensorrt/... //tensorflow/python/... -//tensorflow/python/distribute/... -//tensorflow/python/compiler/tensorrt/... bazel test --build_tag_filters=-no_oss,+gpu --test_tag_filters=-no_oss,+gpu --config=release_gpu_linux --config=rbe_linux_cuda --config=linux_cuda_pycpp_test_filters --verbose_failures --test_output=errors --override_repository=xla=$GITHUB_WORKSPACE/openxla/xla --profile=profile.json.gz -- //tensorflow/compiler/... -//tensorflow/compiler/tf2tensorrt/... //tensorflow/python/... -//tensorflow/python/distribute/... -//tensorflow/python/compiler/tensorrt/... bazel analyze-profile profile.json.gz # END BuildType.TENSORFLOW_X86_GPU_T4_SELF_HOSTED