Skip to content

Conversation

@apbose
Copy link
Collaborator

@apbose apbose commented Nov 14, 2025

This PR

  1. Adds rank based logging for the distributed examples
  2. Corrects the fallback to pytorch case for NCCL converters
  3. This with Changes to TRT-LLM download tool for multigpu distributed case  #3830 provides utilities for running distributed tensor parallel examples using torch.distributed

@meta-cla meta-cla bot added the cla signed label Nov 14, 2025
@github-actions github-actions bot added component: tests Issues re: Tests component: conversion Issues re: Conversion stage component: api [Python] Issues re: Python API component: dynamo Issues relating to the `torch.compile` or `torch._dynamo.export` paths labels Nov 14, 2025
@apbose apbose changed the title Adding rank based logging for torch distributed examples. Also correc… Adding rank based logging for torch distributed examples Nov 14, 2025
@github-actions github-actions bot requested a review from narendasan November 14, 2025 00:05
@apbose apbose marked this pull request as draft November 14, 2025 00:05
@apbose apbose changed the title Adding rank based logging for torch distributed examples Adding rank based logging for torch distributed examples[WIP] Nov 14, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

cla signed component: api [Python] Issues re: Python API component: conversion Issues re: Conversion stage component: dynamo Issues relating to the `torch.compile` or `torch._dynamo.export` paths component: tests Issues re: Tests

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants