From 25c5ea5170a49b5153fbfb850eb88be98e2eea14 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Tue, 28 Apr 2026 11:12:05 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: artificialguybr/Gemma2-2B-OpenHermes2.5-gguf Source: Original Platform --- .gitattributes | 47 +++++++++++++++++++++++++++++++++++++++++++++++ README.md | 46 ++++++++++++++++++++++++++++++++++++++++++++++ model-Q2_K.gguf | 3 +++ model-Q3_K_L.gguf | 3 +++ model-Q3_K_M.gguf | 3 +++ model-Q3_K_S.gguf | 3 +++ model-Q4_0.gguf | 3 +++ model-Q4_K_M.gguf | 3 +++ model-Q4_K_S.gguf | 3 +++ model-Q5_0.gguf | 3 +++ model-Q5_K_M.gguf | 3 +++ model-Q5_K_S.gguf | 3 +++ model-Q6_K.gguf | 3 +++ model-Q8_0.gguf | 3 +++ 14 files changed, 129 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 model-Q2_K.gguf create mode 100644 model-Q3_K_L.gguf create mode 100644 model-Q3_K_M.gguf create mode 100644 model-Q3_K_S.gguf create mode 100644 model-Q4_0.gguf create mode 100644 model-Q4_K_M.gguf create mode 100644 model-Q4_K_S.gguf create mode 100644 model-Q5_0.gguf create mode 100644 model-Q5_K_M.gguf create mode 100644 model-Q5_K_S.gguf create mode 100644 model-Q6_K.gguf create mode 100644 model-Q8_0.gguf diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..ee83470 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,47 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +model-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +model-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +model-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +model-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +model-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +model-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +model-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +model-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text +model-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +model-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +model-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +model-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..91a6773 --- /dev/null +++ b/README.md @@ -0,0 +1,46 @@ +--- +base_model: artificialguybr/Gemma2-2B-OpenHermes2.5 +datasets: +- teknium/OpenHermes-2.5 +language: +- en +library_name: transformers +license: apache-2.0 +quantized_by: artificialguybr +tags: +- gemma +- instruct +- finetune +- chatml +- gpt4 +- synthetic data +- distillation +--- +## About + + + +--- +### 🌐 Website +You can find more of my models, projects, and information on my official website: +- **[artificialguy.com](https://artificialguy.com/)** + + +### 🚀 Prompt Hub +Need high-quality prompts for image models and LLMs? Explore **[findgoodprompt.com](https://findgoodprompt.com)**. +### 💖 Support My Work +If you find this model useful, please consider supporting my work. It helps me cover server costs and dedicate more time to new open-source projects. +- **Patreon:** [Support on Patreon](https://www.patreon.com/user?u=81570187) +- **Ko-fi:** [Buy me a Ko-fi](https://ko-fi.com/artificialguybr) +- **Buy Me a Coffee:** [Buy me a Coffee](https://buymeacoffee.com/jvkape) + + + + +Quantization for: https://huggingface.co/artificialguybr/Gemma2-2B-OpenHermes2.5 + +## How to use + +If you are unsure how to use GGUF files, look at the [TheBloke +READMEs](https://huggingface.co/TheBloke/CodeLlama-70B-Python-GGUF) for +more details, including on how to concatenate multi-part files. diff --git a/model-Q2_K.gguf b/model-Q2_K.gguf new file mode 100644 index 0000000..1d32299 --- /dev/null +++ b/model-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68a9f0e478954cd4869d77436a804d71a2e21b522afbe5865dba51ec34d51f65 +size 1229829024 diff --git a/model-Q3_K_L.gguf b/model-Q3_K_L.gguf new file mode 100644 index 0000000..9b37a91 --- /dev/null +++ b/model-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1255289309c29a410daf3173d9f19dd476b918534e5cdbcc426fc6cdbf999bf2 +size 1550435232 diff --git a/model-Q3_K_M.gguf b/model-Q3_K_M.gguf new file mode 100644 index 0000000..d62668f --- /dev/null +++ b/model-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d483fe5195ad83a712bf5478f9f2c3a45c52f2a1d7e722d137b2c8d8bd46dc3e +size 1461666720 diff --git a/model-Q3_K_S.gguf b/model-Q3_K_S.gguf new file mode 100644 index 0000000..dbabf1c --- /dev/null +++ b/model-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35d300baeeb197749b209bbd51086ae8b14530a0000cdfc31c6af2aba953676f +size 1360659360 diff --git a/model-Q4_0.gguf b/model-Q4_0.gguf new file mode 100644 index 0000000..07a6bd3 --- /dev/null +++ b/model-Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59bf6afe78535039af31a77816fb3d43457c4ddcd40b3bcb5dc611db8e925455 +size 1629508512 diff --git a/model-Q4_K_M.gguf b/model-Q4_K_M.gguf new file mode 100644 index 0000000..586223b --- /dev/null +++ b/model-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:26aa21d90cce3a797f694ee8caddeebb04af3655280b699301fc61cd239e0869 +size 1708581792 diff --git a/model-Q4_K_S.gguf b/model-Q4_K_S.gguf new file mode 100644 index 0000000..7b7a4ea --- /dev/null +++ b/model-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2943fc9ccba665a9099b94597a9e99b8274b3be641b35869bf2b1dbbeaedea1e +size 1638650784 diff --git a/model-Q5_0.gguf b/model-Q5_0.gguf new file mode 100644 index 0000000..8404a26 --- /dev/null +++ b/model-Q5_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2b034a59a9d69b5177303bd68919f469a39a19d11f788a28673b08149651243f +size 1882543008 diff --git a/model-Q5_K_M.gguf b/model-Q5_K_M.gguf new file mode 100644 index 0000000..200320a --- /dev/null +++ b/model-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a207b63c2f3642240fb8b4a8621a71e4980eb15b6ec2736d6e77879c1bf0cc23 +size 1923277728 diff --git a/model-Q5_K_S.gguf b/model-Q5_K_S.gguf new file mode 100644 index 0000000..e531650 --- /dev/null +++ b/model-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6cfc5ae2bd10d7e0eed1fe338737e38d5c9dbc7cfe7450c314be0ea556a27e52 +size 1882543008 diff --git a/model-Q6_K.gguf b/model-Q6_K.gguf new file mode 100644 index 0000000..19cce57 --- /dev/null +++ b/model-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c187b3cc42646a9ba0059c0492f19b330baa0d1d46b22de51ae124f155051ae +size 2151392160 diff --git a/model-Q8_0.gguf b/model-Q8_0.gguf new file mode 100644 index 0000000..6d6ef06 --- /dev/null +++ b/model-Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d4c2f972d84eaeba28b8aa18b462ef4be4f46e0cfa65ed836b4183aabe091ccf +size 2784494496