vllm.v1.attention.backends
Modules:
Name | Description |
---|---|
cpu_attn |
|
flash_attn |
Attention layer with FlashAttention. |
flashinfer |
Attention layer with FlashInfer. |
flex_attention |
Attention layer with FlashAttention. |
mamba_attn |
|
mla |
|
pallas |
|
rocm_aiter_fa |
Attention layer with AiterFlashAttention. |
triton_attn |
Attention layer with PagedAttention and Triton prefix prefill. |
utils |
|