Unverified Commit 9c088829 authored by Lianmin Zheng's avatar Lianmin Zheng Committed by GitHub
Browse files

Revert "Use device_id in dist init to reduce NCCL communicator warmup & creation overhead" (#5786)

parent 005aad32
......@@ -1055,11 +1055,6 @@ def init_distributed_environment(
world_size=world_size,
rank=rank,
timeout=timeout,
device_id=torch.device(
f"cuda:{torch.cuda.current_device()}"
if hasattr(torch, "cuda") and torch.cuda.is_available()
else None
), # Allow NCCL to eagerly init communicator
)
# set the local rank
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment