vllm.distributed.device_communicators.all2all ¶
AgRsAll2AllManager ¶
Bases: All2AllManagerBase
An implementation of all2all communication based on all-gather (dispatch) and reduce-scatter (combine).
Source code in vllm/distributed/device_communicators/all2all.py
__init__ ¶
combine ¶
Reduce-scatter hidden_states across all dp ranks.
Source code in vllm/distributed/device_communicators/all2all.py
destroy ¶
dispatch ¶
dispatch(
hidden_states: Tensor,
topk_weights: Tensor,
topk_ids: Tensor,
is_sequence_parallel: bool = False,
extra_tensors: list[Tensor] | None = None,
) -> (
tuple[Tensor, Tensor, Tensor]
| tuple[Tensor, Tensor, Tensor, list[Tensor]]
)
Gather hidden_states and router_logits from all dp ranks.
Source code in vllm/distributed/device_communicators/all2all.py
DeepEPAll2AllManagerBase ¶
Bases: All2AllManagerBase
All2All communication based on DeepEP High-Throughput kernels.
Source code in vllm/distributed/device_communicators/all2all.py
__init__ ¶
Source code in vllm/distributed/device_communicators/all2all.py
combine ¶
destroy ¶
DeepEPHTAll2AllManager ¶
Bases: DeepEPAll2AllManagerBase
All2All communication based on DeepEP High-Throughput kernels.
Source code in vllm/distributed/device_communicators/all2all.py
__init__ ¶
_make_all2all_kwargs ¶
Source code in vllm/distributed/device_communicators/all2all.py
get_handle ¶
Source code in vllm/distributed/device_communicators/all2all.py
set_num_sms ¶
set_num_sms(num_sms: int)
Source code in vllm/distributed/device_communicators/all2all.py
DeepEPLLAll2AllManager ¶
Bases: DeepEPAll2AllManagerBase
All2All communication based on DeepEP Low-Latency kernels.
Source code in vllm/distributed/device_communicators/all2all.py
__init__ ¶
_make_all2all_kwargs ¶
_make_all2all_kwargs(
max_num_tokens_per_dp_rank: int,
token_hidden_size: int,
num_ep_ranks: int,
num_global_experts: int,
num_local_experts: int,
) -> dict[Any, Any]
the maximum number of tokens a DP rank
can dispatch all the ranks must hold the same value.
token_hidden_size: the hidden dimension of each token. num_ep_ranks: the number of EP group ranks. num_global_experts: Number of experts in the model. num_local_experts: Number of experts in an EP rank.
Source code in vllm/distributed/device_communicators/all2all.py
get_handle ¶
The kwargs for DeepEPLLAll2AllManager is dictated by _make_all2all_kwargs.
Source code in vllm/distributed/device_communicators/all2all.py
FlashInferAllToAllManager ¶
Bases: All2AllManagerBase
All2All communication based on flashinfer kernels.
Source code in vllm/distributed/device_communicators/all2all.py
430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 | |
__init__ ¶
Source code in vllm/distributed/device_communicators/all2all.py
cleanup ¶
Clean up workspace
Source code in vllm/distributed/device_communicators/all2all.py
ensure_alltoall_workspace_initialized ¶
Ensure workspace is initialized
Source code in vllm/distributed/device_communicators/all2all.py
get_handle ¶
initialize ¶
Initialize workspace
Source code in vllm/distributed/device_communicators/all2all.py
NaiveAll2AllManager ¶
Bases: All2AllManagerBase
A naive implementation of all2all communication. It uses all-reduce under the hood, which is not efficient at all. The main purpose is for testing and debugging.
Source code in vllm/distributed/device_communicators/all2all.py
__init__ ¶
combine ¶
Source code in vllm/distributed/device_communicators/all2all.py
destroy ¶
dispatch ¶
dispatch(
hidden_states: Tensor,
topk_weights: Tensor,
topk_ids: Tensor,
is_sequence_parallel: bool = False,
extra_tensors: list[Tensor] | None = None,
) -> (
tuple[Tensor, Tensor, Tensor]
| tuple[Tensor, Tensor, Tensor, list[Tensor]]
)
Source code in vllm/distributed/device_communicators/all2all.py
naive_multicast ¶
naive_multicast(
x: Tensor,
cu_tokens_across_sp_cpu: Tensor,
is_sequence_parallel: bool,
) -> Tensor
Source code in vllm/distributed/device_communicators/all2all.py
PPLXAll2AllManager ¶
Bases: All2AllManagerBase
All2All communication based on PPLX kernels.
Source code in vllm/distributed/device_communicators/all2all.py
190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 | |
__init__ ¶
Source code in vllm/distributed/device_communicators/all2all.py
combine ¶
destroy ¶
Source code in vllm/distributed/device_communicators/all2all.py
dispatch ¶
dispatch(
hidden_states: Tensor,
topk_weights: Tensor,
topk_ids: Tensor,
is_sequence_parallel: bool = False,
extra_tensors: list[Tensor] | None = None,
) -> tuple[Tensor, Tensor, Tensor]