初始化项目,由ModelHub XC社区提供模型
Model: JdoubleU/careconnect-llama3.2-3b Source: Original Platform
This commit is contained in:
80
handler.py
Normal file
80
handler.py
Normal file
@@ -0,0 +1,80 @@
|
||||
from typing import Any, Dict, List
|
||||
|
||||
import torch
|
||||
from transformers import AutoModelForCausalLM, AutoTokenizer
|
||||
|
||||
|
||||
class EndpointHandler:
|
||||
def __init__(self, path: str = ""):
|
||||
self.tokenizer = AutoTokenizer.from_pretrained(path)
|
||||
|
||||
if self.tokenizer.pad_token is None:
|
||||
self.tokenizer.pad_token = self.tokenizer.eos_token
|
||||
|
||||
self.model = AutoModelForCausalLM.from_pretrained(
|
||||
path,
|
||||
device_map="auto",
|
||||
torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32,
|
||||
)
|
||||
self.model.eval()
|
||||
|
||||
def __call__(self, data: Dict[str, Any]) -> Dict[str, Any]:
|
||||
inputs = data.get("inputs", "")
|
||||
parameters = data.get("parameters", {}) or {}
|
||||
|
||||
if isinstance(inputs, list):
|
||||
prompt = "\n".join(str(x) for x in inputs)
|
||||
else:
|
||||
prompt = str(inputs)
|
||||
|
||||
max_new_tokens = int(parameters.get("max_new_tokens", 256))
|
||||
temperature = float(parameters.get("temperature", 0.7))
|
||||
top_p = float(parameters.get("top_p", 0.9))
|
||||
do_sample = bool(parameters.get("do_sample", True))
|
||||
|
||||
messages = parameters.get("messages")
|
||||
if messages and isinstance(messages, list):
|
||||
if hasattr(self.tokenizer, "apply_chat_template"):
|
||||
formatted_prompt = self.tokenizer.apply_chat_template(
|
||||
messages,
|
||||
tokenize=False,
|
||||
add_generation_prompt=True,
|
||||
)
|
||||
else:
|
||||
formatted_prompt = "\n".join(
|
||||
f"{m.get('role', 'user')}: {m.get('content', '')}" for m in messages
|
||||
)
|
||||
else:
|
||||
formatted_prompt = prompt
|
||||
|
||||
model_inputs = self.tokenizer(
|
||||
formatted_prompt,
|
||||
return_tensors="pt",
|
||||
padding=True,
|
||||
truncation=True,
|
||||
)
|
||||
|
||||
model_inputs = {k: v.to(self.model.device) for k, v in model_inputs.items()}
|
||||
|
||||
with torch.no_grad():
|
||||
outputs = self.model.generate(
|
||||
**model_inputs,
|
||||
max_new_tokens=max_new_tokens,
|
||||
temperature=temperature,
|
||||
top_p=top_p,
|
||||
do_sample=do_sample,
|
||||
pad_token_id=self.tokenizer.pad_token_id,
|
||||
eos_token_id=self.tokenizer.eos_token_id,
|
||||
)
|
||||
|
||||
generated_text = self.tokenizer.decode(
|
||||
outputs[0],
|
||||
skip_special_tokens=True,
|
||||
)
|
||||
|
||||
if generated_text.startswith(formatted_prompt):
|
||||
generated_text = generated_text[len(formatted_prompt):].strip()
|
||||
|
||||
return {
|
||||
"generated_text": generated_text
|
||||
}
|
||||
Reference in New Issue
Block a user