doc: emphasize and notify the usage of chat_template (#3589)
Co-authored-by: Chayenne <zhaochen20@outlook.com>
This commit is contained in:
@@ -24,7 +24,8 @@
|
||||
"\n",
|
||||
"Launch the server in your terminal and wait for it to initialize.\n",
|
||||
"\n",
|
||||
"Remember to add `--chat-template llama_3_vision` to specify the vision chat template, otherwise the server only supports text.\n",
|
||||
"**Remember to add `--chat-template llama_3_vision` to specify the vision chat template, otherwise the server only supports text, and performance degradation may occur.**\n",
|
||||
"\n",
|
||||
"We need to specify `--chat-template` for vision language models because the chat template provided in Hugging Face tokenizer only supports text."
|
||||
]
|
||||
},
|
||||
|
||||
@@ -56,6 +56,9 @@ Please consult the documentation below to learn more about the parameters you ma
|
||||
* `json_model_override_args`: Override model config with the provided JSON.
|
||||
* `delete_ckpt_after_loading`: Delete the model checkpoint after loading the model.
|
||||
|
||||
> [!IMPORTANT]
|
||||
> **Make sure the correct `chat_template` is passed, or performance degradation may occur.**
|
||||
|
||||
## Serving: HTTP & API
|
||||
|
||||
### HTTP Server configuration
|
||||
|
||||
Reference in New Issue
Block a user