From 7c99103f4c2a5e92e34bbfc5f3779fe9a447c8ee Mon Sep 17 00:00:00 2001 From: Michael Yao Date: Wed, 23 Apr 2025 23:21:55 +0800 Subject: [PATCH] [Doc] Fix two 404 links caused by sglang typo (#5667) Signed-off-by: windsonsea --- docs/supported_models/support_new_models.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/docs/supported_models/support_new_models.md b/docs/supported_models/support_new_models.md index d2eda6e10..ae8b664bd 100644 --- a/docs/supported_models/support_new_models.md +++ b/docs/supported_models/support_new_models.md @@ -4,7 +4,7 @@ This document explains how to add support for new language models and vision‐l ## How to Support a new Language Model -To support a new model in SGLang, you only need to add a single file under the [SGLang Models Directory](https://github.com/sgl-project/sglang/tree/main/python/sgllang/srt/models). You can learn from existing model implementations and create a new file for your model. For most models, you should be able to find a similar model to start with (e.g., starting from Llama). Also refer how to [port a Model from vLLM to SGLang](#port-a-model-from-vllm-to-sglang) +To support a new model in SGLang, you only need to add a single file under the [SGLang Models Directory](https://github.com/sgl-project/sglang/tree/main/python/sglang/srt/models). You can learn from existing model implementations and create a new file for your model. For most models, you should be able to find a similar model to start with (e.g., starting from Llama). Also refer how to [port a Model from vLLM to SGLang](#port-a-model-from-vllm-to-sglang) ## How to Support a new Vision-Language model @@ -58,7 +58,7 @@ The [vLLM Models Directory](https://github.com/vllm-project/vllm/tree/main/vllm/ To port a model from vLLM to SGLang: - Compare these two files for guidance: - - [SGLang Llama Implementation](https://github.com/sgl-project/sglang/blob/main/python/sgllang/srt/models/llama.py) + - [SGLang Llama Implementation](https://github.com/sgl-project/sglang/blob/main/python/sglang/srt/models/llama.py) - [vLLM Llama Implementation](https://github.com/vllm-project/vllm/blob/main/vllm/model_executor/models/llama.py) - The major differences include: - **Replace vLLM’s `Attention` with `RadixAttention`** (ensure you pass `layer_id` to `RadixAttention`).