tensorflow / cloud
The TensorFlow Cloud repository provides APIs that will allow to easily go from debugging and training your Keras and TensorFlow code in a local environment to distributed training in the cloud.
README
TensorFlow Cloud
The TensorFlow Cloud repository provides APIs that will allow to easily go from
debugging, training, tuning your Keras and TensorFlow code in a local
environment to distributed training/tuning on Cloud.
Introduction
TensorFlow Cloud run API for GCP training/tuning
Installation
Requirements
-
Python >= 3.6
-
Google AI platform APIs enabled for
your GCP account. We use the AI platform for deploying docker images on GCP. -
Either a functioning version of
docker if you want to use a local
docker process for your build, or
create a cloud storage bucket
to use with Google Cloud build for
docker image build and publishing. -
(optional) nbconvert if you
are using a notebook file asentry_pointas shown in
usage guide #4.
For detailed end to end setup instructions, please see
Setup instructions.
Install latest release
pip install -U tensorflow-cloud
Install from source
git clone https://github.com/tensorflow/cloud.git
cd cloud
pip install src/python/.
High level overview
TensorFlow Cloud package provides the run API for training your models on GCP.
To start, let's walk through a simple workflow using this API.
-
Let's begin with a Keras model training code such as the following, saved as
mnist_example.py.import tensorflow as tf (x_train, y_train), (_, _) = tf.keras.datasets.mnist.load_data() x_train = x_train.reshape((60000, 28 * 28)) x_train = x_train.astype('float32') / 255 model = tf.keras.Sequential([ tf.keras.layers.Dense(512, activation='relu', input_shape=(28 * 28,)), tf.keras.layers.Dropout(0.2), tf.keras.layers.Dense(10, activation='softmax') ]) model.compile(loss='sparse_categorical_crossentropy', optimizer=tf.keras.optimizers.Adam(), metrics=['accuracy']) model.fit(x_train, y_train, epochs=10, batch_size=128) -
After you have tested this model on your local environment for a few epochs,
probably with a small dataset, you can train the model on Google Cloud by
writing the following simple scriptscale_mnist.py.import tensorflow_cloud as tfc tfc.run(entry_point='mnist_example.py')Running
scale_mnist.pywill automatically apply TensorFlow
one device strategy
and train your model at scale on Google Cloud Platform. Please see the
usage guide section for detailed instructions and additional
API parameters. -
You will see an output similar to the following on your console. This
information can be used to track the training job status.user@desktop$ python scale_mnist.py Job submitted successfully. Your job ID is: tf_cloud_train_519ec89c_a876_49a9_b578_4fe300f8865e Please access your job logs at the following URL: https://console.cloud.google.com/mlengine/jobs/tf_cloud_train_519ec89c_a876_49a9_b578_4fe300f8865e?project=prod-123
Setup instructions
End to end instructions to help set up your environment for Tensorflow Cloud.
You use one of the following notebooks to setup your project or follow the
instructions below.
<table align="left">
<td>
<a href="https://colab.research.google.com/github/tensorflow/cloud/blob/master/examples/google_cloud_project_setup_instructions.ipynb">
<img width="50" src="https://cloud.google.com/ml-engine/images/colab-logo-32px.png" alt="Colab logo">Run in Colab
</a>
</td>
<td>
<a href="https://github.com/tensorflow/cloud/blob/master/examples/google_cloud_project_setup_instructions.ipynb">
<img src="https://cloud.google.com/ml-engine/images/github-logo-32px.png" alt="GitHub logo">View on GitHub
</a>
</td>
<td>
<a href="https://www.kaggle.com/nitric/google-cloud-project-setup-instructions">
<img width="90" src="https://www.kaggle.com/static/images/site-logo.png" alt="Kaggle logo">Run in Kaggle
</a>
</td>
</table>
-
Create a new local directory
mkdir tensorflow_cloud cd tensorflow_cloud -
Make sure you have
python >= 3.6python -V -
Set up virtual environment
virtualenv tfcloud --python=python3 source tfcloud/bin/activate -
Set up your Google Cloud project
Verify that gcloud sdk is installed.
which gcloudSet default gcloud project
export PROJECT_ID=<your-project-id> gcloud config set project $PROJECT_ID -
Create a service account.
export SA_NAME=<your-sa-name> gcloud iam service-accounts create $SA_NAME gcloud projects add-iam-policy-binding $PROJECT_ID \ --member serviceAccount:$SA_NAME@$PROJECT_ID.iam.gserviceaccount.com \ --role 'roles/editor'Create a key for your service account.
gcloud iam service-accounts keys create ~/key.json --iam-account $SA_NAME@$PROJECT_ID.iam.gserviceaccount.comCreate the GOOGLE_APPLICATION_CREDENTIALS environment variable.
export GOOGLE_APPLICATION_CREDENTIALS=~/key.json -
Create a Cloud Storage bucket.
Using Google Cloud build is the
recommended method for building and publishing docker images, although we
optionally allow for local
docker daemon process
depending on your specific needs.BUCKET_NAME="your-bucket-name" REGION="us-central1" gcloud auth login gsutil mb -l $REGION gs://$BUCKET_NAME(optional for local docker setup)
shell sudo dockerd -
Authenticate access to Google Cloud registry.
gcloud auth configure-docker -
Install nbconvert if you plan
to use a notebook fileentry_pointas shown in
usage guide #4.pip install nbconvert -
Install latest release of tensorflow-cloud
pip install tensorflow-cloud
Usage guide
As described in the high level overview, the run API
allows you to train your models at scale on GCP. The
run
API can be used in four different ways. This is defined by where you are running
the API (Terminal vs IPython notebook), and your entry_point parameter.
entry_point is an optional Python script or notebook file path to the file
that contains your TensorFlow Keras training code. This is the most important
parameter in the API.
run(entry_point=None,
requirements_txt=None,
distribution_strategy='auto',
docker_config='auto',
chief_config='auto',
worker_config='auto',
worker_count=0,
entry_point_args=None,
stream_logs=False,
job_labels=None,
**kwargs)
-
Using a python file as
entry_point.If you have your
tf.kerasmodel in a python file (mnist_example.py),
then you can write the following simple script (scale_mnist.py) to scale
your model on GCP.import tensorflow_cloud as tfc tfc.run(entry_point='mnist_example.py')Please note that all the files in the same directory tree as
entry_point
will be packaged in the docker image created, along with theentry_point
file. It's recommended to create a new directory to house each cloud project
which includes necessary files and nothing else, to optimize image build
times. -
Using a notebook file as
entry_point.If you have your
tf.kerasmodel in a notebook file
(mnist_example.ipynb), then you can write the following simple script
(scale_mnist.py) to scale your model on GCP.import tensorflow_cloud as tfc tfc.run(entry_point='mnist_example.ipynb')Please note that all the files in the same directory tree as
entry_point
will be packaged in the docker image created, along with theentry_point
file. Like the python scriptentry_pointabove, we recommended creating a
new directory to house each cloud project which includes necessary files and
nothing else, to optimize image build times. -
Using
runwithin a python script that contains thetf.kerasmodel.You can use the
runAPI from within your python file that contains the
tf.kerasmodel (mnist_scale.py). In this use case,entry_pointshould
beNone. TherunAPI can be called anywhere and the entire file will be
executed remotely. The API can be called at the end to run the script
locally for debugging purposes (possibly with fewer epochs and other flags).import tensorflow_datasets as tfds import tensorflow as tf import tensorflow_cloud as tfc tfc.run( entry_point=None, distribution_strategy='auto', requirements_txt='requirements.txt', chief_config=tfc.MachineConfig( cpu_cores=8, memory=30, accelerator_type=tfc.AcceleratorType.NVIDIA_TESLA_T4, accelerator_count=2), worker_count=0) datasets, info = tfds.load(name='mnist', with_info=True, as_supervised=True) mnist_train, mnist_test = datasets['train'], datasets['test'] num_train_examples = info.splits['train'].num_examples num_test_examples = info.splits['test'].num_examples BUFFER_SIZE = 10000 BATCH_SIZE = 64 def scale(image, label): image = tf.cast(image, tf.float32) image /= 255 return image, label train_dataset = mnist_train.map(scale).cache() train_dataset = train_dataset.shuffle(BUFFER_SIZE).batch(BATCH_SIZE) model = tf.keras.Sequential([ tf.keras.layers.Conv2D(32, 3, activation='relu', input_shape=( 28, 28, 1)), tf.keras.layers.MaxPooling2D(), tf.keras.layers.Flatten(), tf.keras.layers.Dense(64, activation='relu'), tf.keras.layers.Dense(10, activation='softmax') ]) model.compile(loss='sparse_categorical_crossentropy', optimizer=tf.keras.optimizers.Adam(), metrics=['accuracy']) model.fit(train_dataset, epochs=12)Please note that all the files in the same directory tree as the python
script will be packaged in the docker image created, along with the python
file. It's recommended to create a new directory to house each cloud project
which includes necessary files and nothing else, to optimize image build
times. -
Using
runwithin a notebook script that contains thetf.kerasmodel.
In this use case,
entry_pointshould beNoneand
docker_config.image_build_bucketmust be specified, to ensure the build
can be stored and published.Cluster and distribution strategy configuration
By default,
runAPI takes care of wrapping your model code in a TensorFlow
distribution strategy based on the cluster configuration you have provided.No distribution
CPU chief config and no additional workers
tfc.run(entry_point='mnist_example.py', chief_config=tfc.COMMON_MACHINE_CONFIGS['CPU'])OneDeviceStrategy
1 GPU on chief (defaults to
AcceleratorType.NVIDIA_TESLA_T4) and no
additional workers.tfc.run(entry_point='mnist_example.py')MirroredStrategy
Chief config with multiple GPUS (
AcceleratorType.NVIDIA_TESLA_V100).tfc.run(entry_point='mnist_example.py', chief_config=tfc.COMMON_MACHINE_CONFIGS['V100_4X'])MultiWorkerMirroredStrategy
Chief config with 1 GPU and 2 workers each with 8 GPUs
(AcceleratorType.NVIDIA_TESLA_V100).tfc.run(entry_point='mnist_example.py', chief_config=tfc.COMMON_MACHINE_CONFIGS['V100_1X'], worker_count=2, worker_config=tfc.COMMON_MACHINE_CONFIGS['V100_8X'])TPUStrategy
Chief config with 1 CPU and 1 worker with TPU.
tfc.run(entry_point="mnist_example.py", chief_config=tfc.COMMON_MACHINE_CONFIGS["CPU"], worker_count=1, worker_config=tfc.COMMON_MACHINE_CONFIGS["TPU"])Please note that TPUStrategy with TensorFlow Cloud works only with TF
version 2.1 as this is the latest version supported by
AI Platform cloud TPUCustom distribution strategy
If you would like to take care of specifying distribution strategy in your
model code and do not wantrunAPI to create a strategy, then set
distribution_stategyasNone. This will be required for example when you
are usingstrategy.experimental_distribute_dataset.tfc.run(entry_point='mnist_example.py', distribution_strategy=None, worker_count=2)
What happens when you call run?
The API call will encompass the following:
- Making code entities such as a Keras script/notebook, cloud and
distribution ready. - Converting this distribution entity into a docker container with the
required dependencies. - Deploy this container at scale and train using TensorFlow distribution
strategies. - Stream logs and monitor them on hosted TensorBoard, manage checkpoint
storage.
By default, we will use local docker daemon for building and publishing docker
images to Google container registry. Images are published to
gcr.io/your-gcp-project-id. If you specify docker_config.image_build_bucket,
then we will use Google Cloud build to
build and publish docker images.
We use Google AI platform for deploying
docker images on GCP.
Please note that, when entry_point argument is specified, all the files in the
same directory tree as entry_point will be packaged in the docker image
created, along with the entry_point file.
Please see run API documentation for detailed information on the parameters
and how you can modify the above processes to suit your needs.
End to end examples
cd src/python/tensorflow_cloud/core
python tests/examples/call_run_on_script_with_keras_fit.py
- Using a python file as
entry_point(Keras fit API). - Using a python file as
entry_point(Keras custom training loop). - Using a python file as
entry_point(Keras save and load). - Using a notebook file as
entry_point. - Using
runwithin a python script that contains thetf.kerasmodel. - Using cloud build instead of local docker.
- Run AutoKeras with TensorFlow Cloud.
Running unit tests
pytest src/python/tensorflow_cloud/core/tests/unit/
Local vs remote training
Things to keep in mind when running your jobs remotely:
[Coming soon]
Debugging workflow
Here are some tips for fixing unexpected issues.
Operation disallowed within distribution strategy scope
Error like: Creating a generator within a strategy scope is disallowed,
because there is ambiguity on how to replicate a generator (e.g. should it be
copied so that each replica gets the same random numbers, or 'split' so that
each replica gets different random numbers).
Solution: Passing distribution_strategy='auto' to run API wraps all of
your script in a TF distribution strategy based on the cluster configuration
provided. You will see the above error or something similar to it, if for some
reason an operation is not allowed inside distribution strategy scope. To fix
the error, please pass None to the distribution_strategy param and create a
strategy instance as part of your training code as shown in
this
example.
Docker image build timeout
Error like: requests.exceptions.ConnectionError: ('Connection aborted.',
timeout('The write operation timed out'))
Solution: The directory being used as an entry point likely has too much
data for the image to successfully build, and there may be extraneous data
included in the build. Reformat your directory structure such that the folder
which contains the entry point only includes files necessary for the current
project.
Version not supported for TPU training
Error like: There was an error submitting the job.Field: tpu_tf_version
Error: The specified runtime version '2.3' is not supported for TPU training.
Please specify a different runtime version.
Solution: Please use TF version 2.1. See TPU Strategy in
Cluster and distribution strategy configuration section.
TF nightly build.
Warning like: Docker parent image '2.4.0.dev20200720' does not exist. Using
the latest TF nightly build.
Solution: If you do not provide docker_config.parent_image param, then by
default we use pre-built TF docker images as parent image. If you do not have TF
installed on the environment where run is called, then TF docker image for the
latest stable release will be used. Otherwise, the version of the docker image
will match the locally installed TF version. However, pre-built TF docker images
aren't available for TF nightlies except for the latest. So, if your local TF is
an older nightly version, we upgrade to the latest nightly automatically and
raise this warning.
Mixing distribution strategy objects.
Error like: RuntimeError: Mixing different tf.distribute.Strategy objects.
Solution: Please provide distribution_strategy=None when you already have
a distribution strategy defined in your model code. Specifying
distribution_strategy'='auto', will wrap your code in a TensorFlow
distribution strategy. This will cause the above error, if there is a strategy
object already used in your code.
Coming up
- Distributed Keras tuner support.
Contributing
We welcome community contributions, see CONTRIBUTING.md and,
for style help,
Writing TensorFlow documentation
guide.
License
Privacy Notice
This application reports technical and operational details of your usage of
Cloud Services in accordance with Google privacy policy, for more information
please refer to https://policies.google.com/privacy. If you wish to opt-out, you
may do so by running
tensorflow_cloud.utils.google_api_client.optout_metrics_reporting().
