Skip to content

Commit

Permalink
minor fix
Browse files Browse the repository at this point in the history
Signed-off-by: YunLiu <[email protected]>
  • Loading branch information
KumoLiu committed Apr 15, 2024
1 parent 27ac56a commit eab7fce
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions acceleration/distributed_training/distributed_training.md
Original file line number Diff line number Diff line change
Expand Up @@ -12,11 +12,11 @@ root_dir="/path/to/your/directory" # Change this to your desired directory

## Multi-GPU Training

Users can set your `NUM_GPUS_PER_NODE`, `NUM_NODES`, `INDEX_CURRENT_NODE`, as well as `DIR_OF_TESTDATA` for the directory of the test dataset.
Users can set your `NUM_GPUS_PER_NODE`, `NUM_NODES`, `INDEX_CURRENT_NODE`, as well as `DIR_OF_DATA` for the directory of the test dataset.
Then users can execute the following command to start multi-GPU model training:

```
torchrun --nproc_per_node=NUM_GPUS_PER_NODE --nnodes=NUM_NODES brats_training_ddp.py -d DIR_OF_TESTDATA
torchrun --nproc_per_node=NUM_GPUS_PER_NODE --nnodes=NUM_NODES brats_training_ddp.py -d DIR_OF_DATA
```

## Multi-Node Training
Expand Down

0 comments on commit eab7fce

Please sign in to comment.