# PyTorch Notes | |
This document lists nuances of pytorch relevant to our work. | |
## Distributed Launcher | |
### pt <= 1.8.1 | |
The good old `torch.distributed.launch` works here: | |
``` | |
export LAUNCHER="python -u -m torch.distributed.launch \ | |
--nproc_per_node $GPUS_PER_NODE \ | |
--nnodes $NNODES \ | |
--master_addr $MASTER_ADDR \ | |
--master_port $MASTER_PORT \ | |
" | |
``` | |
### pt >= 1.9 | |
pytorch switched to elastic in 1.9. `torch.distributed.launch` is supposed to be backward compatible, but it's not. Under multi-node it results in `RuntimeError: Address already in use` error. | |
Therefore for pt 1.9 and higher you must use the following launcher syntax: | |
``` | |
export LAUNCHER="python -u -m torch.distributed.run \ | |
--nproc_per_node $GPUS_PER_NODE \ | |
--nnodes $NNODES \ | |
--rdzv_endpoint $MASTER_ADDR:$MASTER_PORT \ | |
--rdzv_backend c10d \ | |
--max_restarts 0 \ | |
" | |
``` | |
For full details see: https://pytorch.org/docs/1.9.0/elastic/quickstart.html | |
Note: If you're using the `deepspeed` launcher (which we can't use in the slurm environment), it should continue working as before with either pytorch version. | |