From 2f74c508c97cf4a677afc02e3b00a0d4d4453358 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Mon, 20 Apr 2026 08:49:43 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: nuupy/HY-MT1.5-1.8B-Q4_K_M-GGUF Source: Original Platform --- .gitattributes | 38 +++++++++++++++++ HY-MT1.5-1.8B-Q8_0.gguf | 3 ++ README.md | 89 +++++++++++++++++++++++++++++++++++++++ hy-mt1.5-1.8b-f16.gguf | 3 ++ hy-mt1.5-1.8b-q4_k_m.gguf | 3 ++ 5 files changed, 136 insertions(+) create mode 100644 .gitattributes create mode 100644 HY-MT1.5-1.8B-Q8_0.gguf create mode 100644 README.md create mode 100644 hy-mt1.5-1.8b-f16.gguf create mode 100644 hy-mt1.5-1.8b-q4_k_m.gguf diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..08f52c9 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,38 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +hy-mt1.5-1.8b-q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text +hy-mt1.5-1.8b-f16.gguf filter=lfs diff=lfs merge=lfs -text +HY-MT1.5-1.8B-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/HY-MT1.5-1.8B-Q8_0.gguf b/HY-MT1.5-1.8B-Q8_0.gguf new file mode 100644 index 0000000..88944f9 --- /dev/null +++ b/HY-MT1.5-1.8B-Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6789b06d0902f2f5312c0e1703d56ccbddfcfb6c653d22519b7c720f7db9a98e +size 1908528288 diff --git a/README.md b/README.md new file mode 100644 index 0000000..a51ede4 --- /dev/null +++ b/README.md @@ -0,0 +1,89 @@ +--- +library_name: transformers +tags: +- translation +- llama-cpp +- gguf-my-repo +language: +- zh +- en +- fr +- pt +- es +- ja +- tr +- ru +- ar +- ko +- th +- it +- de +- vi +- ms +- id +- tl +- hi +- pl +- cs +- nl +- km +- my +- fa +- gu +- ur +- te +- mr +- he +- bn +- ta +- uk +- bo +- kk +- mn +- ug +base_model: tencent/HY-MT1.5-1.8B +--- + +# nuupy/HY-MT1.5-1.8B-Q4_K_M-GGUF +This model was converted to GGUF format from [`tencent/HY-MT1.5-1.8B`](https://huggingface.co/tencent/HY-MT1.5-1.8B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space. +Refer to the [original model card](https://huggingface.co/tencent/HY-MT1.5-1.8B) for more details on the model. + +## Use with llama.cpp +Install llama.cpp through brew (works on Mac and Linux) + +```bash +brew install llama.cpp + +``` +Invoke the llama.cpp server or the CLI. + +### CLI: +```bash +llama-cli --hf-repo nuupy/HY-MT1.5-1.8B-Q4_K_M-GGUF --hf-file hy-mt1.5-1.8b-q4_k_m.gguf -p "The meaning to life and the universe is" +``` + +### Server: +```bash +llama-server --hf-repo nuupy/HY-MT1.5-1.8B-Q4_K_M-GGUF --hf-file hy-mt1.5-1.8b-q4_k_m.gguf -c 2048 +``` + +Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well. + +Step 1: Clone llama.cpp from GitHub. +``` +git clone https://github.com/ggerganov/llama.cpp +``` + +Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux). +``` +cd llama.cpp && LLAMA_CURL=1 make +``` + +Step 3: Run inference through the main binary. +``` +./llama-cli --hf-repo nuupy/HY-MT1.5-1.8B-Q4_K_M-GGUF --hf-file hy-mt1.5-1.8b-q4_k_m.gguf -p "The meaning to life and the universe is" +``` +or +``` +./llama-server --hf-repo nuupy/HY-MT1.5-1.8B-Q4_K_M-GGUF --hf-file hy-mt1.5-1.8b-q4_k_m.gguf -c 2048 +``` diff --git a/hy-mt1.5-1.8b-f16.gguf b/hy-mt1.5-1.8b-f16.gguf new file mode 100644 index 0000000..b78cd59 --- /dev/null +++ b/hy-mt1.5-1.8b-f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:951ee1329b851dbab0cfd00aab9aa733279fb3325a5750d0de346f2c3451013e +size 3587534144 diff --git a/hy-mt1.5-1.8b-q4_k_m.gguf b/hy-mt1.5-1.8b-q4_k_m.gguf new file mode 100644 index 0000000..d614112 --- /dev/null +++ b/hy-mt1.5-1.8b-q4_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b9474cac7e0c640eb1949b9c6007550d11cfca504bc6d5393300134811e9628 +size 1133080224