STT-tensorflow/tensorflow/tools/ci_build
A. Unique TensorFlower f41584353a Add the "--define=no_tensorflow_py_deps=true" flag for the windows cpu release
builds.

PiperOrigin-RevId: 317968971
Change-Id: I7d4db21474d85620928f3a5ffb1e4cfebaa2be9f
2020-06-23 17:33:10 -07:00
..
build_scripts Create an integration_testing directory for tests that are only run on nightly and release pip packages. 2020-03-26 15:24:30 -07:00
builds Create a nightly libtensorflow.so symlink script that is copied into the GCS directory for nightly builds. 2020-06-11 11:09:35 -07:00
ctpu Kokoro related change for switching nighlty-2.x to the nightly and removing older nightly which was pointing to nightly-1.15. 2020-03-04 13:27:39 -08:00
devtoolset Install multiple python versions simultaneously in our remote build docker. 2020-04-25 05:19:30 -07:00
gpu_build Updating that parallel_gpu_execute.sh script such that it works as expected for AMD GPus 2019-08-08 16:02:37 +00:00
install Update TF bazel version requirements. 2020-06-09 08:04:07 -07:00
linux Reverting == in Dockerfile 2020-06-18 07:27:13 -07:00
osx Filter out tpu wherever we filter out gpu. 2019-09-19 17:18:44 -07:00
per_release/scripts Experimental internal CI changes 2020-06-19 16:47:47 -07:00
pi Fix RPI CI build failure 2020-04-27 16:32:30 -07:00
presubmit Remove installation of Python 2 pylint - it has been disabled in CI a while ago. 2020-06-04 08:15:26 -07:00
protobuf
release Add the "--define=no_tensorflow_py_deps=true" flag for the windows cpu release 2020-06-23 17:33:10 -07:00
remote Remove read-only copy of MLIR core in TF and use version in LLVM 2019-12-30 07:00:20 -08:00
windows toolchains: migrate platforms to use exec_properties 2020-01-30 13:08:52 -08:00
xla/linux [ROCm] Updating Dockerfile.rocm / CI scripts to use ROCm 3.3 2020-04-20 14:51:13 +00:00
build_rbe.sh Add a simple script and config for building RBE images with Cloud Build. 2020-01-13 15:41:37 -08:00
ci_build.sh [ROCm] Update ROCm CI builds to use ROCm 2.8 2019-11-15 19:19:43 +00:00
ci_sanity.sh Have sanity build output a Bazel test summary XML file. 2020-05-20 17:30:39 -07:00
cloudbuild.yaml Add a simple script and config for building RBE images with Cloud Build. 2020-01-13 15:41:37 -08:00
code_link_check.sh Merge changes from github. 2017-07-19 15:12:25 -07:00
copy_binary.py Add tensorflow{,_core} to zip_these_files in copy_binary.py for py33 and py34 packages. 2019-06-12 10:31:13 -07:00
cuda-clang.patch Add Dockerfile to build TensorFlow with cuda-clang. 2019-10-23 10:38:35 -07:00
Dockerfile.android Switch to NDK API level 21 2019-11-15 17:13:47 -08:00
Dockerfile.cmake Fold Keras Applications into tf.keras, since it is no longer necessary to share 2019-12-12 17:09:51 -08:00
Dockerfile.cpu Upgrade all TF base images to ubuntu 16. (#15458) 2017-12-31 22:35:56 -08:00
Dockerfile.cpu.ppc64le Build OpenBLAS 0.3.0 on ppc64le for TF tests 2018-06-27 14:49:58 -05:00
Dockerfile.cuda-clang Add Dockerfile to build TensorFlow with cuda-clang. 2019-10-23 10:38:35 -07:00
Dockerfile.custom_op_gpu Add gpu custom op docker file. 2019-03-14 17:20:40 -07:00
Dockerfile.custom_op_ubuntu_16 Install python3 package by default 2020-06-05 11:54:29 -07:00
Dockerfile.custom_op_ubuntu_16_cuda10.0 Install python3 package by default 2020-06-05 11:54:29 -07:00
Dockerfile.custom_op_ubuntu_16_cuda10.1 Install python3 package by default 2020-06-05 11:54:29 -07:00
Dockerfile.debian.jessie.cpu Use archive.debian.org for jessie-backports as they got removed from main. 2019-07-10 13:17:05 -07:00
Dockerfile.gpu Automated rollback of commit 220e8e059a 2019-05-20 08:28:11 -07:00
Dockerfile.gpu.ppc64le Remove TF_NCCL_VERSION from Dockerfile.gpu.ppc64le 2018-10-18 17:56:01 -04:00
Dockerfile.hadoop Upgrade all TF base images to ubuntu 16. (#15458) 2017-12-31 22:35:56 -08:00
Dockerfile.micro Merge pull request #32287 from jenselofsson:micro_recognition 2020-04-23 12:50:03 -07:00
Dockerfile.pi Add a way to build TFLite PIP package with Flex support 2020-05-28 18:07:13 -07:00
Dockerfile.pi-python3 Add a way to build TFLite PIP package with Flex support 2020-05-28 18:07:13 -07:00
Dockerfile.pi-python37 Add a way to build TFLite PIP package with Flex support 2020-05-28 18:07:13 -07:00
Dockerfile.rbe.cpu Update RBE CPU Dockerfile with instructions on how to build/push. 2019-01-04 07:32:55 -08:00
Dockerfile.rbe.cuda9.0-cudnn7-ubuntu14.04 Automated rollback of commit 220e8e059a 2019-05-20 08:28:11 -07:00
Dockerfile.rbe.cuda10.0-cudnn7-ubuntu14.04 Automated rollback of commit 220e8e059a 2019-05-20 08:28:11 -07:00
Dockerfile.rbe.cuda10.0-cudnn7-ubuntu16.04-manylinux2010 Update the Dockerfiles for manylinux2010 toolchains with Python3.8 installations. 2020-02-20 15:33:06 -08:00
Dockerfile.rbe.cuda10.1-cudnn7-ubuntu16.04-manylinux2010 Update the OSS image to pickup latest tf estimator PIP package. 2020-05-12 16:03:19 -07:00
Dockerfile.rbe.cuda10.1-cudnn7-ubuntu16.04-manylinux2010-multipython Add bz2-devel so python will be compiled with bz2 support. 2020-05-12 11:36:14 -07:00
Dockerfile.rbe.cuda10.1-cudnn7-ubuntu18.04-manylinux2010-multipython Add RBE multipython Dockerfile for ubuntu 18.04. 2020-05-29 13:54:16 -07:00
Dockerfile.rbe.cuda11.0-cudnn8-ubuntu18.04-manylinux2010-multipython Add RBE Dockerfile for CUDA11+cuDNN8. 2020-06-22 04:02:13 -07:00
Dockerfile.rbe.gpu Upgrade ci docker images to cuda 10. 2019-01-04 14:36:41 -08:00
Dockerfile.rbe.rocm-ubuntu16.04 sync (ROCm fork --> TF repo) the dockerfile+scripts used for ROCm CI 2019-12-19 17:38:38 +00:00
Dockerfile.rbe.ubuntu16.04-manylinux2010 Update the OSS image to pickup latest tf estimator PIP package. 2020-05-12 16:03:19 -07:00
Dockerfile.rocm [ROCm] Updating Dockerfile.rocm / CI scripts to use ROCm 3.3 2020-04-20 14:51:13 +00:00
pep8 Add pep8 check option to ci_sanity.sh. 2016-07-07 13:32:47 -07:00
pylintrc Add C0301 line-too-long error to pylint sanity check. 2018-01-26 16:59:01 -08:00
README.md minor spelling tweaks 2020-03-24 12:34:02 +09:00
update_version.py Merge pull request #26113 from angersson:angerson-remove-old-dockerfiles 2019-04-30 10:19:44 -07:00

TensorFlow Builds

This directory contains all the files and setup instructions to run all the important builds and tests. You can run it yourself!

Run It Yourself

You have two options when running TensorFlow tests locally on your machine. First, using docker, you can run our Continuous Integration (CI) scripts on tensorflow devel images. The other option is to install all TensorFlow dependencies on your machine and run the scripts natively on your system.

Run TensorFlow CI Scripts using Docker

  1. Install Docker following the instructions on the docker website.

  2. Start a container with one of the devel images here: https://hub.docker.com/r/tensorflow/tensorflow/tags/.

  3. Based on your choice of the image, pick one of the scripts under https://github.com/tensorflow/tensorflow/tree/master/tensorflow/tools/ci_build/linux and run them from the TensorFlow repository root.

Run TensorFlow CI Scripts Natively on your Machine

  1. Follow the instructions at https://www.tensorflow.org/install/source, but stop when you get to the section "Configure the installation". You do not need to configure the installation to run the CI scripts.

  2. Pick the appropriate OS and python version you have installed, and run the script under tensorflow/tools/ci_build/.

TensorFlow Continuous Integration

To verify that new changes dont break TensorFlow, we run builds and tests on either Jenkins or a CI system internal to Google.

We can trigger builds and tests on updates to master or on each pull request. Contact one of the repository maintainers to trigger builds on your pull request.

View CI Results

The Pull Request will show if the change passed or failed the checks.

From the pull request, click Show all checks to see the list of builds and tests. Click on Details to see the results from Jenkins or the internal CI system.

Results from Jenkins are displayed in the Jenkins UI. For more information, see the Jenkins documentation.

Results from the internal CI system are displayed in the Build Status UI. In this UI, to see the logs for a failed build:

  • Click on the INVOCATION LOG tab to see the invocation log.

  • Click on the ARTIFACTS tab to see a list of all artifacts, including logs.

  • Individual test logs may be available. To see these logs, from the TARGETS tab, click on the failed target. Then, click on the TARGET LOG tab to see its test log.

    If youre looking at target that is sharded or a test that is flaky, then the build tool divided the target into multiple shards or ran the test multiple times. Each test log is specific to the shard, run, and attempt. To see a specific log:

    1. Click on the log icon that is on the right next to the shard, run, and attempt number.

    2. In the grid that appears on the right, click on the specific shard, run, and attempt to view its log. You can also type the desired shard, run, or attempt number in the field above its grid.

Third party TensorFlow CI

Mellanox TensorFlow CI

How to start CI
  • Submit special pull request (PR) comment to trigger CI: bot:mlx:test
  • Test session is run automatically.
  • Test results and artifacts (log files) are reported via PR comments
CI Steps

CI includes the following steps: * Build TensorFlow (GPU version) * Run TensorFlow tests: * TF CNN benchmarks (TensorFlow 1.13 and less) * TF models (TensorFlow 2.0): ResNet, synthetic data, NCCL, multi_worker_mirrored distributed strategy

Test Environment

CI is run in the Mellanox lab on a 2-node cluster with the following parameters:

  • Hardware * IB: 1x ConnectX-6 HCA (connected to Mellanox Quantum™ HDR switch) * GPU: 1x Nvidia Tesla K40m * Software * Ubuntu 16.04.6 * Internal stable MLNX_OFED, HPC-X™ and SHARP™ versions
Support (Mellanox)

With any questions/suggestions or in case of issues contact Artem Ryabov.