batch-shipyard/recipes/TensorFlow-Distributed
Fred Park 7f2200a31d
Update recipes to refer to platform image docs
- Resolves #186
2018-04-18 12:35:26 -07:00
..
config Update recipes SSH username 2017-11-13 09:25:20 -08:00
docker Update TensorFlow recipes to 1.2.1 2017-08-02 11:00:53 -07:00
README.md Update recipes to refer to platform image docs 2018-04-18 12:35:26 -07:00

README.md

TensorFlow-Distributed

This recipe shows how to run TensorFlow in distributed mode across multiple CPUs or GPUs (either single node or multinode) using N-series Azure VM instances in an Azure Batch compute pool.

Configuration

Please see refer to this set of sample configuration files for this recipe.

Pool Configuration

The pool configuration should enable the following properties if on multiple GPUs:

  • vm_size must be a GPU enabled VM size if using GPUs. Because TensorFlow is a GPU-accelerated compute application, you should choose a GPU compute accelerated VM instance size. If not using GPUs, any other appropriate CPU-based VM size can be selected.
  • vm_configuration is the VM configuration. Please select an appropriate platform_image with GPU as supported by Batch Shipyard. If not using GPUs, you can select any appropriate platform image.

If on multiple CPUs:

  • max_tasks_per_node must be set to 1 or omitted

Other pool properties such as publisher, offer, sku, vm_size and vm_count should be set to your desired values for multiple CPU configuration.

Global Configuration

The global configuration should set the following properties:

  • docker_images array must have a reference to a valid TensorFlow Docker image that can work with multi-instance tasks. The alfpark/tensorflow images have been prepared by using Google's TensorFlow Dockerfile as a base and extending the image to work with Batch Shipyard.

Jobs Configuration

The jobs configuration should set the following properties within the tasks array which should have a task definition containing:

  • docker_image should be the name of the Docker image for this container invocation, e.g., alfpark/tensorflow:1.2.1-gpu or alfpark/tensorflow:1.2.1-cpu
  • command should contain the command to pass to the Docker run invocation. To run the example MNIST replica example, the command would look like: "/bin/bash -c \"/shipyard/launcher.sh /shipyard/mnist_replica.py\"". The launcher will automatically detect the number of GPUs and pass the correct number to the TensorFlow script. Please see the launcher.sh for the launcher source.
  • gpu must be set to true if run on GPUs. This enables invoking the nvidia-docker wrapper. This property should be omitted or set to false if run on CPUs.
  • multi_instance property must be defined
    • num_instances should be set to pool_specification_vm_count_dedicated, pool_vm_count_low_priority, pool_current_dedicated, or pool_current_low_priority
    • coordination_command should be unset or null. For pools with native container support, this command should be supplied if a non-standard sshd is required.
    • resource_files should be unset or the array can be empty

Dockerfile and supplementary files

The Dockerfile for the Docker images can be found here.

You must agree to the following license prior to use: