Small fixes for torchao quant (#2476)

This commit is contained in:
Jerry Zhang
2024-12-16 14:08:12 -08:00
committed by GitHub
parent 7154b4b1df
commit 82699474fd
2 changed files with 6 additions and 5 deletions

View File

@@ -157,6 +157,10 @@ class ModelRunner:
self.sampler = Sampler()
self.load_model()
apply_torchao_config_to_model(
self.model, global_server_args_dict["torchao_config"]
)
# Apply torch TP if the model supports it
supports_torch_tp = getattr(self.model, "supports_torch_tp", False)
if self.tp_size > 1 and supports_torch_tp:
@@ -165,10 +169,6 @@ class ModelRunner:
else:
self.torch_tp_applied = False
apply_torchao_config_to_model(
self.model, global_server_args_dict["torchao_config"]
)
# Init memory pool and attention backends
if server_args.lora_paths is not None:
self.init_lora_manager()