a2145cc30b
Adding Microsoft SECURITY.MD |
||
---|---|---|
docs | ||
examples | ||
profiling | ||
tools | ||
varuna | ||
vmss_scripts | ||
.DS_Store | ||
.gitignore | ||
README.md | ||
SECURITY.md | ||
apex.patch | ||
requirements.txt | ||
setup.py | ||
varuna.pdf |
README.md
Varuna
Varuna is a tool for efficient training of large DNN models on commodity GPUs and networking. It implements a combination of pipeline parallelism and data parallelism in PyTorch, and enables training on a changing set of resources smoothly.
This repository is an implementation of the paper:
"Varuna: Scalable, Low-cost Training of Massive Deep Learning Models", to appear in EuroSys'22.
Setup & Installation
Varuna requires python 3, PyTorch (1.5+) and apex.
The patch apex.patch
in this directory needs to be applied to apex before building it. Varuna's code and this patch has been tested for this commit of apex.
git clone https://github.com/NVIDIA/apex
cp apex.patch /path/to/apex/
cd /path/to/apex
git apply apex.patch
pip install -v --disable-pip-version-check --no-cache-dir --global-option="--cpp_ext" --global-option="--cuda_ext" ./
To install, clone this repository, cd into it and run
python setup.py install
Running
Varuna trains large DNN models by parallelising them into sequential pipeline stages and data parallel replicas across a set of GPUs. These methods are called pipeline parallelism and data parallelism respectively.
To enable parallel training with Varuna, there are several steps the user must follow. Detailed docs are in the docs/
folder as webpages (html/index.html
) or pdf (varuna.pdf
).
Examples of models working with varuna can also be found in examples/
. Please see this folder to run examples with BERT and Megatron-LM.
Some of the steps for are briefly described below.
CutPoint demarcation
Varuna slices a DNN model into sequential pipeline stages. For this, the model should be annotated with varuna CutPoint
instances between different operations/ parts of model computation. These are nn.Module instances that are potential slice points in the mode. For each CutPoint, Varuna can either ignore it or activate it as a partition boundary. (see CutPoints)
from varuna import CutPoint
class SampleModel(nn.Module):
def __init__(...):
....
self.cutpoints = [CutPoint() for i in range(num_cutpoints)]
....
def forward(input...):
input = self.some_operation(input)
input = self.cutpoints[0](input) # marked as a potential stage boundary
input = self.some_other_operation(input)
....
for i in range(sub_modules):
x = sub_module_i(input, ...)
x = self.cutpoints[i+1](x) # each cutpoint instance should be used only once in a model
....
Operations separated by CutPoints should preferably have no shared modules/parameters. For weight sharing between different parts of the module, you should register separate nn.Parameter instances (even for the same tensor) and pass the pair of parameter names as shared_weights to Varuna.
Wrapping the model in Varuna
The nn.Module for your DNN instance should be wrapped in a Varuna
instance before training and before optimizer creation. (see Varuna) Wrapping in Varuna
returns a model partitioned according to the given stage_to_rank_map (which is passed by the varuna launcher) and moved to the GPU. After this initialization, each rank in the job has only the parts of the model required by it. Varuna internally handles fp16 mixed precision training and shared parameters (such as the initial and last embedding weights in BERT/GPT-2).
Optimizer creation should be after this since it requires model parameters as input. The optimizer needs to be registered with Varuna using a setter.
model = MyModel() # full model on CPU
# provide dummy input function to varuna for initialization. Inputs mst be in dictionary form.
def get_batch_fn(size, device=None):
batch = dataset[:size]
if device is not None:
batch = [t.to(device) for t in batch]
inputs, mask = batch
return {'inputs': inputs, 'mask': mask, 'extra_norm': True }
shared_weights = [("language_model.embedding.word_embeddings.weight","lm_head_weight")] # parameter sharing between stages
model = Varuna( model, args.stage_to_rank_map, dry_run_input, global_batch_size,
args.chunk_size, args.fp16, local_rank=args.local_rank,
device=args.local_rank, shared_weights=shared_weights)
# now model is a subset of the original model, moved to the GPU on each process
optimizer = get_optimizer(model)
model.set_optimizer(optimizer)
Training loop.
The Varuna training loop does not require a separate forward & backward step, the script may just call the step
function. The input to this function should be of the per-process batch size (batch_size / data_parallel_workers), and should be a dictionary with arg names and values. The step function makes micro-batches out of this input batch, completes the fwd/bwd pipeline schedule and reduces the gradients/overflow over the whole job, returning the loss and overflow boolean.
inputs = dict({
"input_ids": tokens,
"position_ids": position_ids,
"attention_mask": attention_mask,
"loss_mask": loss_mask,
"labels": labels
})
loss, overflow = model.step(inputs)
loss = torch.Tensor([loss])
if not overflow:
optimizer.step()
Launcher and Arguments
To launch a distributed training process using Varuna, use the run_varuna.py script as follows:
python -m varuna.run_varuna --machine_list <file_with_ips> --gpus_per_node <num_gpus_per_node> --batch-size <total_effective_batch_size> --nstages <number_of_pipeline_stages> --chunk_size <micro_batch_size_for_pipeline> --code_dir <working_dir_for_training> user_training_script.py <...user args...>
See Launching varuna.
This expects all machines in the machine_list to be set up with necessary code/libraries in code_dir and have gpus_per_node GPUs working. The job is launched with all workers running the user_training_script and args.
This launcher passes a few arguments to the user training script for Varuna. These should be passed during Varuna
initialisation in the python script:
- rank: process rank in overall distributed job
- local_rank: process rank in the local node
- stage_to_rank_map: varuna config info about stage placement
- chunk_size: micro batch size for Varuna pipeline
- batch-size: per process batch size
Changing resources: job morphing
Varuna enables training on a changing set of nodes/gpus. This is through monitoring the machine_list text file of IPs with the set of available nodes at any time. Training jobs are launched from a long-living manager. On detecting a change, Varuna checkpoints, stops and relaunches the job from the manager. To allow for this on-demand checkpoint/stop, varuna relies on user signals (SIGUSR1 in unix). The user therefore needs to add a simple handler for this signal to their training script. See Morphing.
if __name__ == "__main__":
def handler(signum,_):
save_checkpoint(iteration, model, optimizer, lr_scheduler)
exit()
signal.signal(signal.SIGUSR1, handler)
Profiling, config selection
Varuna supports auto-configuration of data-parallel and pipeline-parallel dimensions that saves the user from running and comparing different configs for better performance. To enable this, the user needs to run one-time profiling of the model and network conditions using the Profiler
class in Varuna.
See Varuna Profiling.
This is instantiated similar to Varuna and runs as a distributed process:
model = BuildModel()
profiler = Profiler(model, args.device, fp16=args.fp16)
def get_batch(size):
# function to get sample batches of given size for profiling
return batch
profiler.initialize(get_batch)
microbatches_to_profile = list(range(1,max_micro_BS))
profile = profiler.profile_all(get_batch, microbatches_to_profile, out_folder=args.save)
Each process profiles the compute of different cutpoints and at the same time measures communication with other processes. This builds and saves a profile of the model in the specified location, from where it can be accessed by the AutoConfig
class. AutoConfig
calculates the different configs for a given number of GPUs and simulates them using information from the pre-built profile to compare and return the best performing setting in a few seconds. This calculation is triggered by run_varuna when no nstages and chunk_size arguments are given and a profile location is passed.