[Feature] add multi-rank support for Lora (#4492)
Co-authored-by: rudy152 <czh1137892874@gmail.com>
This commit is contained in:
@@ -31,8 +31,8 @@ class LoRAModelCase:
|
||||
base: str
|
||||
adaptors: List[LoRAAdaptor]
|
||||
tp_size: int = 1
|
||||
prefill_tolerance: float = 5e-2
|
||||
decode_tolerance: float = 5e-2
|
||||
prefill_tolerance: float = 1e-1
|
||||
decode_tolerance: float = 1e-1
|
||||
rouge_l_tolerance: float = 1.0
|
||||
max_loras_per_batch: int = 1
|
||||
skip_long_prompt: bool = False
|
||||
|
||||
Reference in New Issue
Block a user