Enable native ModelOpt quantization support (1/3) (#7149)
Signed-off-by: Zhiyu Cheng <zhiyuc@nvidia.com>
This commit is contained in:
@@ -125,6 +125,7 @@ suites = {
|
||||
TestFile("test_vlm_input_format.py", 300),
|
||||
TestFile("test_vision_openai_server_a.py", 724),
|
||||
TestFile("test_vision_openai_server_b.py", 446),
|
||||
TestFile("test_modelopt_loader.py", 30),
|
||||
],
|
||||
"per-commit-2-gpu": [
|
||||
TestFile("lora/test_lora_tp.py", 116),
|
||||
|
||||
Reference in New Issue
Block a user