diff --git a/README.md b/README.md index 44c8c05..36ef5dd 100644 --- a/README.md +++ b/README.md @@ -45,9 +45,7 @@ docker run -it --rm \ curl -X POST http://localhost:10086/generate \ -H "Content-Type: application/json" \ -d '{ - "model": "qwen3-8b", "messages": [{"role": "user", "content": "你好"}], - "stream": true }' ``` diff --git a/app.py b/app.py index dc6a6a5..1794210 100644 --- a/app.py +++ b/app.py @@ -14,7 +14,7 @@ import torch from modelscope import AutoModel, AutoTokenizer # -------- Configuration -------- -MODEL_DIR = os.environ.get("DEESEEK_MODEL_DIR", "/mnt/models") +MODEL_DIR = os.environ.get("DEESEEK_MODEL_DIR", "/model") MODEL_PREFERRED_DTYPE = os.environ.get("DEESEEK_DTYPE", "bfloat16") # or float16/float32 # -------- FastAPI app --------