From 5dd84e9c9b00e152f59ed24385cff751674a6037 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Tue, 21 Apr 2026 19:00:33 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: mradermacher/Apollo2-1.5B-GGUF Source: Original Platform --- .gitattributes | 47 +++++++++++++++++ Apollo2-1.5B.IQ4_XS.gguf | 3 ++ Apollo2-1.5B.Q2_K.gguf | 3 ++ Apollo2-1.5B.Q3_K_L.gguf | 3 ++ Apollo2-1.5B.Q3_K_M.gguf | 3 ++ Apollo2-1.5B.Q3_K_S.gguf | 3 ++ Apollo2-1.5B.Q4_K_M.gguf | 3 ++ Apollo2-1.5B.Q4_K_S.gguf | 3 ++ Apollo2-1.5B.Q5_K_M.gguf | 3 ++ Apollo2-1.5B.Q5_K_S.gguf | 3 ++ Apollo2-1.5B.Q6_K.gguf | 3 ++ Apollo2-1.5B.Q8_0.gguf | 3 ++ Apollo2-1.5B.f16.gguf | 3 ++ README.md | 111 +++++++++++++++++++++++++++++++++++++++ 14 files changed, 194 insertions(+) create mode 100644 .gitattributes create mode 100644 Apollo2-1.5B.IQ4_XS.gguf create mode 100644 Apollo2-1.5B.Q2_K.gguf create mode 100644 Apollo2-1.5B.Q3_K_L.gguf create mode 100644 Apollo2-1.5B.Q3_K_M.gguf create mode 100644 Apollo2-1.5B.Q3_K_S.gguf create mode 100644 Apollo2-1.5B.Q4_K_M.gguf create mode 100644 Apollo2-1.5B.Q4_K_S.gguf create mode 100644 Apollo2-1.5B.Q5_K_M.gguf create mode 100644 Apollo2-1.5B.Q5_K_S.gguf create mode 100644 Apollo2-1.5B.Q6_K.gguf create mode 100644 Apollo2-1.5B.Q8_0.gguf create mode 100644 Apollo2-1.5B.f16.gguf create mode 100644 README.md diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..ad57488 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,47 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +Apollo2-1.5B.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +Apollo2-1.5B.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +Apollo2-1.5B.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +Apollo2-1.5B.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Apollo2-1.5B.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Apollo2-1.5B.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Apollo2-1.5B.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Apollo2-1.5B.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Apollo2-1.5B.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Apollo2-1.5B.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Apollo2-1.5B.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +Apollo2-1.5B.f16.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/Apollo2-1.5B.IQ4_XS.gguf b/Apollo2-1.5B.IQ4_XS.gguf new file mode 100644 index 0000000..76e1070 --- /dev/null +++ b/Apollo2-1.5B.IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71e612f93d461fc65af3d20e006a80deefbadae88ddfa45969505e83af0a5be3 +size 1026161440 diff --git a/Apollo2-1.5B.Q2_K.gguf b/Apollo2-1.5B.Q2_K.gguf new file mode 100644 index 0000000..2645913 --- /dev/null +++ b/Apollo2-1.5B.Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b2d934545c60986f1015d9f7b7811165c3332e70543ed3525522e305ca377a52 +size 752879392 diff --git a/Apollo2-1.5B.Q3_K_L.gguf b/Apollo2-1.5B.Q3_K_L.gguf new file mode 100644 index 0000000..cf2a3b4 --- /dev/null +++ b/Apollo2-1.5B.Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e1369f653cfd5168936abf5568edf8128a51bca02aab3e4e6096b3561328a7c0 +size 980439328 diff --git a/Apollo2-1.5B.Q3_K_M.gguf b/Apollo2-1.5B.Q3_K_M.gguf new file mode 100644 index 0000000..754c5c0 --- /dev/null +++ b/Apollo2-1.5B.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f09bea14155e7723e8ff242827bfbf1eb10ad5c4afda8ee338260dae2b29c32a +size 924455200 diff --git a/Apollo2-1.5B.Q3_K_S.gguf b/Apollo2-1.5B.Q3_K_S.gguf new file mode 100644 index 0000000..1d46222 --- /dev/null +++ b/Apollo2-1.5B.Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:188a2ae6fb255ee1100c0c34169c149f35fbb75c1aaa0c73481c52bde093d832 +size 861221152 diff --git a/Apollo2-1.5B.Q4_K_M.gguf b/Apollo2-1.5B.Q4_K_M.gguf new file mode 100644 index 0000000..b8318ff --- /dev/null +++ b/Apollo2-1.5B.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:210a4ac022269da312ca4f965ea4b33bfcbe8b1133faedc35ee4c5dd349e9b39 +size 1117319968 diff --git a/Apollo2-1.5B.Q4_K_S.gguf b/Apollo2-1.5B.Q4_K_S.gguf new file mode 100644 index 0000000..9de7240 --- /dev/null +++ b/Apollo2-1.5B.Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3dd8bd8c5e729896a7344a822d6a2ebd06cc510d2ea8c8af4006408fa39e07de +size 1071584032 diff --git a/Apollo2-1.5B.Q5_K_M.gguf b/Apollo2-1.5B.Q5_K_M.gguf new file mode 100644 index 0000000..4ff305b --- /dev/null +++ b/Apollo2-1.5B.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a811cfdcd412d0f189b1dab49aae1497f50c1c9bd3683b4b6a8c343b6051137 +size 1285493536 diff --git a/Apollo2-1.5B.Q5_K_S.gguf b/Apollo2-1.5B.Q5_K_S.gguf new file mode 100644 index 0000000..28d238d --- /dev/null +++ b/Apollo2-1.5B.Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a07bc38cdd8ef4cb86e7a2f51e99e437673ec5630af73b89c09828e58f5314d +size 1259172640 diff --git a/Apollo2-1.5B.Q6_K.gguf b/Apollo2-1.5B.Q6_K.gguf new file mode 100644 index 0000000..3d75f33 --- /dev/null +++ b/Apollo2-1.5B.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:da09514f808e5512f26b93a8484c75ee9f18823c2554cf6ec5ae2cd131f9c5e2 +size 1464177952 diff --git a/Apollo2-1.5B.Q8_0.gguf b/Apollo2-1.5B.Q8_0.gguf new file mode 100644 index 0000000..5609883 --- /dev/null +++ b/Apollo2-1.5B.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3a2243a1a79bffd27bee3ab4de10ae7132b1579201ce545e5ccee9b560e29227 +size 1894531360 diff --git a/Apollo2-1.5B.f16.gguf b/Apollo2-1.5B.f16.gguf new file mode 100644 index 0000000..3b54f16 --- /dev/null +++ b/Apollo2-1.5B.f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:976b3c55b6166a2831855dcfa6220861a10ff2715af93394050c3895caddd0c1 +size 3560415520 diff --git a/README.md b/README.md new file mode 100644 index 0000000..6bd7c65 --- /dev/null +++ b/README.md @@ -0,0 +1,111 @@ +--- +base_model: FreedomIntelligence/Apollo2-1.5B +datasets: +- FreedomIntelligence/ApolloMoEDataset +language: +- ar +- en +- zh +- ko +- ja +- mn +- th +- vi +- lo +- mg +- de +- pt +- es +- fr +- ru +- it +- hr +- gl +- cs +- co +- la +- uk +- bs +- bg +- eo +- sq +- da +- sa +- gn +- sr +- sk +- gd +- lb +- hi +- ku +- mt +- he +- ln +- bm +- sw +- ig +- rw +- ha +library_name: transformers +license: apache-2.0 +quantized_by: mradermacher +tags: +- biology +- medical +--- +## About + + + + + + +static quants of https://huggingface.co/FreedomIntelligence/Apollo2-1.5B + + +weighted/imatrix quants are available at https://huggingface.co/mradermacher/Apollo2-1.5B-i1-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/Apollo2-1.5B-GGUF/resolve/main/Apollo2-1.5B.Q2_K.gguf) | Q2_K | 0.9 | | +| [GGUF](https://huggingface.co/mradermacher/Apollo2-1.5B-GGUF/resolve/main/Apollo2-1.5B.Q3_K_S.gguf) | Q3_K_S | 1.0 | | +| [GGUF](https://huggingface.co/mradermacher/Apollo2-1.5B-GGUF/resolve/main/Apollo2-1.5B.Q3_K_M.gguf) | Q3_K_M | 1.0 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/Apollo2-1.5B-GGUF/resolve/main/Apollo2-1.5B.Q3_K_L.gguf) | Q3_K_L | 1.1 | | +| [GGUF](https://huggingface.co/mradermacher/Apollo2-1.5B-GGUF/resolve/main/Apollo2-1.5B.IQ4_XS.gguf) | IQ4_XS | 1.1 | | +| [GGUF](https://huggingface.co/mradermacher/Apollo2-1.5B-GGUF/resolve/main/Apollo2-1.5B.Q4_K_S.gguf) | Q4_K_S | 1.2 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Apollo2-1.5B-GGUF/resolve/main/Apollo2-1.5B.Q4_K_M.gguf) | Q4_K_M | 1.2 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Apollo2-1.5B-GGUF/resolve/main/Apollo2-1.5B.Q5_K_S.gguf) | Q5_K_S | 1.4 | | +| [GGUF](https://huggingface.co/mradermacher/Apollo2-1.5B-GGUF/resolve/main/Apollo2-1.5B.Q5_K_M.gguf) | Q5_K_M | 1.4 | | +| [GGUF](https://huggingface.co/mradermacher/Apollo2-1.5B-GGUF/resolve/main/Apollo2-1.5B.Q6_K.gguf) | Q6_K | 1.6 | very good quality | +| [GGUF](https://huggingface.co/mradermacher/Apollo2-1.5B-GGUF/resolve/main/Apollo2-1.5B.Q8_0.gguf) | Q8_0 | 2.0 | fast, best quality | +| [GGUF](https://huggingface.co/mradermacher/Apollo2-1.5B-GGUF/resolve/main/Apollo2-1.5B.f16.gguf) | f16 | 3.7 | 16 bpw, overkill | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. + +