更新 transformers_server.py
All checks were successful
Docker Build and Push / docker (push) Successful in 2m24s
All checks were successful
Docker Build and Push / docker (push) Successful in 2m24s
This commit is contained in:
@@ -79,7 +79,7 @@ def load_model():
|
|||||||
processor_class = resolve_transformers_class(config["processer_class"])
|
processor_class = resolve_transformers_class(config["processer_class"])
|
||||||
model_class = resolve_transformers_class(config["model_class"])
|
model_class = resolve_transformers_class(config["model_class"])
|
||||||
torch_dtype = resolve_torch_dtype(config["torch_dtype"])
|
torch_dtype = resolve_torch_dtype(config["torch_dtype"])
|
||||||
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
device = torch.device("cuda")
|
||||||
|
|
||||||
logger.info(
|
logger.info(
|
||||||
f"model config: model_class={config['model_class']}, "
|
f"model config: model_class={config['model_class']}, "
|
||||||
|
|||||||
Reference in New Issue
Block a user