[Feature] add multi-rank support for Lora (#4492)

Co-authored-by: rudy152 <czh1137892874@gmail.com>
This commit is contained in:
chaobo jia
2025-03-29 00:38:44 +08:00
committed by GitHub
parent 6dea5c96bf
commit ef9a378a20
16 changed files with 292 additions and 97 deletions

View File

@@ -31,8 +31,8 @@ class LoRAModelCase:
base: str
adaptors: List[LoRAAdaptor]
tp_size: int = 1
prefill_tolerance: float = 5e-2
decode_tolerance: float = 5e-2
prefill_tolerance: float = 1e-1
decode_tolerance: float = 1e-1
rouge_l_tolerance: float = 1.0
max_loras_per_batch: int = 1
skip_long_prompt: bool = False