Skip to content

feat(embed): launch finetune with torchrun for multi-GPU support#166

Draft
shan-nvidia wants to merge 1 commit intomainfrom
steve/embed_finetune_multi_gpu
Draft

feat(embed): launch finetune with torchrun for multi-GPU support#166
shan-nvidia wants to merge 1 commit intomainfrom
steve/embed_finetune_multi_gpu

Conversation

@shan-nvidia
Copy link
Copy Markdown
Contributor

Use torch.distributed.run with --nproc_per_node=gpu so training automatically uses all available GPUs (works correctly with 1 GPU too).

Use torch.distributed.run with --nproc_per_node=gpu so training
automatically uses all available GPUs (works correctly with 1 GPU too).

Mirrors the rerank recipe change in 756e4f2.

Signed-off-by: Steve Han <sthan@nvidia.com>
Made-with: Cursor
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant