vllm.distributed.communication_op
broadcast_tensor_dict
¶
broadcast_tensor_dict(
tensor_dict: Optional[
dict[Any, Union[Tensor, Any]]
] = None,
src: int = 0,
)
Source code in vllm/distributed/communication_op.py
tensor_model_parallel_all_gather
¶
All-gather the input tensor across model parallel group.
tensor_model_parallel_all_reduce
¶
tensor_model_parallel_gather
¶
Gather the input tensor across model parallel group.
tensor_model_parallel_reduce_scatter
¶
Reduce-Scatter the input tensor across model parallel group.