vllm.distributed.device_communicators.cpu_communicator
CpuCommunicator
¶
Bases: DeviceCommunicatorBase
Source code in vllm/distributed/device_communicators/cpu_communicator.py
__init__
¶
__init__(
cpu_group: ProcessGroup,
device: Optional[device] = None,
device_group: Optional[ProcessGroup] = None,
unique_name: str = "",
)
Source code in vllm/distributed/device_communicators/cpu_communicator.py
all_gather
¶
Source code in vllm/distributed/device_communicators/cpu_communicator.py
all_reduce
¶
gather
¶
NOTE: We assume that the input tensor is on the same device across
all the ranks.
NOTE: dst
is the local rank of the destination rank.
Source code in vllm/distributed/device_communicators/cpu_communicator.py
_CPUSHMDistributed
¶
Source code in vllm/distributed/device_communicators/cpu_communicator.py
group_name
instance-attribute
¶
__init__
¶
__init__(communicator: CpuCommunicator)
Source code in vllm/distributed/device_communicators/cpu_communicator.py
_init_cpu_shm
¶
_init_cpu_shm() -> int
Source code in vllm/distributed/device_communicators/cpu_communicator.py
all_gather_into_tensor
¶
all_reduce
¶
gather
¶
gather(
input: Tensor,
gather_list: Optional[list[Tensor]],
dst: int = -1,
group: Optional[ProcessGroup] = None,
) -> None