Enable native ModelOpt quantization support (1/3) (#7149)
Signed-off-by: Zhiyu Cheng <zhiyuc@nvidia.com>
This commit is contained in:
@@ -880,7 +880,7 @@ class ModelRunner:
|
||||
load_config = LoadConfig(load_format=load_format)
|
||||
|
||||
# Only support DefaultModelLoader for now
|
||||
loader = get_model_loader(load_config)
|
||||
loader = get_model_loader(load_config, self.model_config)
|
||||
if not isinstance(loader, DefaultModelLoader):
|
||||
message = f"Failed to get model loader: {loader}."
|
||||
return False, message
|
||||
|
||||
Reference in New Issue
Block a user