[doc] Update document for flashinfer mla (#3907)

This commit is contained in:
Baizhou Zhang
2025-02-26 20:40:45 -08:00
committed by GitHub
parent 8b681d7724
commit 71ed01833d
2 changed files with 2 additions and 1 deletions

View File

@@ -133,6 +133,7 @@ Please consult the documentation below to learn more about the parameters you ma
* `attention_backend`: The backend for attention computation and KV cache management.
* `sampling_backend`: The backend for sampling.
* `enable_flashinfer_mla`: The backend for flashinfer MLA wrapper. It can optimize the throughput of deepseek models.
## Constrained Decoding