Source code for torch.distributed
import torch
def is_available():
return hasattr(torch._C, "_c10d_init")
if is_available() and not torch._C._c10d_init():
raise RuntimeError("Failed to initialize PyTorch distributed support")
if is_available():
from .distributed_c10d import * # noqa: F401
# Variables prefixed with underscore are not auto imported
# See the comment in `distributed_c10d.py` above `_backend` on why we expose
# this.
from .distributed_c10d import _backend # noqa: F401