Fix docs (#1890)
This commit is contained in:
@@ -26,7 +26,7 @@
|
||||
"\n",
|
||||
"```bash\n",
|
||||
"python3 -m sglang.launch_server --model-path meta-llama/Llama-3.2-11B-Vision-Instruct \\\n",
|
||||
" --port 30010 --chat-template llama_3_vision\n",
|
||||
" --port 30000 --chat-template llama_3_vision\n",
|
||||
"```\n",
|
||||
"in your terminal and wait for the server to be ready.\n",
|
||||
"\n",
|
||||
@@ -50,11 +50,11 @@
|
||||
"embedding_process = execute_shell_command(\n",
|
||||
" \"\"\"\n",
|
||||
"python3 -m sglang.launch_server --model-path meta-llama/Llama-3.2-11B-Vision-Instruct \\\n",
|
||||
" --port=30010 --chat-template=llama_3_vision\n",
|
||||
" --port=30000 --chat-template=llama_3_vision\n",
|
||||
"\"\"\"\n",
|
||||
")\n",
|
||||
"\n",
|
||||
"wait_for_server(\"http://localhost:30010\")"
|
||||
"wait_for_server(\"http://localhost:30000\")"
|
||||
]
|
||||
},
|
||||
{
|
||||
@@ -75,7 +75,7 @@
|
||||
"import subprocess\n",
|
||||
"\n",
|
||||
"curl_command = \"\"\"\n",
|
||||
"curl -s http://localhost:30010/v1/chat/completions \\\n",
|
||||
"curl -s http://localhost:30000/v1/chat/completions \\\n",
|
||||
" -d '{\n",
|
||||
" \"model\": \"meta-llama/Llama-3.2-11B-Vision-Instruct\",\n",
|
||||
" \"messages\": [\n",
|
||||
@@ -118,7 +118,7 @@
|
||||
"source": [
|
||||
"import requests\n",
|
||||
"\n",
|
||||
"url = \"http://localhost:30010/v1/chat/completions\"\n",
|
||||
"url = \"http://localhost:30000/v1/chat/completions\"\n",
|
||||
"\n",
|
||||
"data = {\n",
|
||||
" \"model\": \"meta-llama/Llama-3.2-11B-Vision-Instruct\",\n",
|
||||
@@ -161,7 +161,7 @@
|
||||
"source": [
|
||||
"from openai import OpenAI\n",
|
||||
"\n",
|
||||
"client = OpenAI(base_url=\"http://localhost:30010/v1\", api_key=\"None\")\n",
|
||||
"client = OpenAI(base_url=\"http://localhost:30000/v1\", api_key=\"None\")\n",
|
||||
"\n",
|
||||
"response = client.chat.completions.create(\n",
|
||||
" model=\"meta-llama/Llama-3.2-11B-Vision-Instruct\",\n",
|
||||
@@ -205,7 +205,7 @@
|
||||
"source": [
|
||||
"from openai import OpenAI\n",
|
||||
"\n",
|
||||
"client = OpenAI(base_url=\"http://localhost:30010/v1\", api_key=\"None\")\n",
|
||||
"client = OpenAI(base_url=\"http://localhost:30000/v1\", api_key=\"None\")\n",
|
||||
"\n",
|
||||
"response = client.chat.completions.create(\n",
|
||||
" model=\"meta-llama/Llama-3.2-11B-Vision-Instruct\",\n",
|
||||
|
||||
Reference in New Issue
Block a user