upload fp16 weights

This commit is contained in:
ai-modelscope
2025-02-27 01:09:11 +08:00
parent a76d8364ca
commit 9672140f8b

View File

@@ -40,14 +40,14 @@ Check out our [llama.cpp documentation](https://qwen.readthedocs.io/en/latest/ru
We advise you to clone [`llama.cpp`](https://github.com/ggerganov/llama.cpp) and install it following the official guide. We follow the latest version of llama.cpp. We advise you to clone [`llama.cpp`](https://github.com/ggerganov/llama.cpp) and install it following the official guide. We follow the latest version of llama.cpp.
In the following demonstration, we assume that you are running commands under the repository `llama.cpp`. In the following demonstration, we assume that you are running commands under the repository `llama.cpp`.
Since cloning the entire repo may be inefficient, you can manually download the GGUF file that you need or use `modelscope`: Since cloning the entire repo may be inefficient, you can manually download the GGUF file that you need or use `huggingface-cli`:
1. Install 1. Install
```shell ```shell
pip install -U modelscope pip install -U huggingface_hub
``` ```
2. Download: 2. Download:
```shell ```shell
modelscope download --model=qwen/Qwen2.5-1.5B-Instruct-GGUF --local_dir . qwen2.5-1.5b-instruct-q5_k_m.gguf huggingface-cli download Qwen/Qwen2.5-1.5B-Instruct-GGUF qwen2.5-1.5b-instruct-q5_k_m.gguf --local-dir . --local-dir-use-symlinks False
``` ```
For users, to achieve chatbot-like experience, it is recommended to commence in the conversation mode: For users, to achieve chatbot-like experience, it is recommended to commence in the conversation mode: