commit 4ce2ace9d136fdf340bb3d76feaa663a4869a2ee Author: ModelHub XC Date: Wed May 6 21:25:37 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: mradermacher/Beagle14-7B-i1-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..9e9921d --- /dev/null +++ b/.gitattributes @@ -0,0 +1,58 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +imatrix.dat filter=lfs diff=lfs merge=lfs -text +Beagle14-7B.i1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +Beagle14-7B.i1-IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text +Beagle14-7B.i1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Beagle14-7B.i1-IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text +Beagle14-7B.i1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Beagle14-7B.i1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Beagle14-7B.i1-IQ2_M.gguf filter=lfs diff=lfs merge=lfs -text +Beagle14-7B.i1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Beagle14-7B.i1-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +Beagle14-7B.i1-Q2_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Beagle14-7B.i1-IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text +Beagle14-7B.i1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Beagle14-7B.i1-IQ2_XXS.gguf filter=lfs diff=lfs merge=lfs -text +Beagle14-7B.i1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +Beagle14-7B.i1-IQ2_XS.gguf filter=lfs diff=lfs merge=lfs -text +Beagle14-7B.i1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Beagle14-7B.i1-IQ2_S.gguf filter=lfs diff=lfs merge=lfs -text +Beagle14-7B.i1-IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text +Beagle14-7B.i1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Beagle14-7B.i1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +Beagle14-7B.i1-IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text +Beagle14-7B.i1-IQ3_S.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/Beagle14-7B.i1-IQ1_M.gguf b/Beagle14-7B.i1-IQ1_M.gguf new file mode 100644 index 0000000..e13d549 --- /dev/null +++ b/Beagle14-7B.i1-IQ1_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c6dc3e1d31f644a4a3fa8ee81d79db96aacebcc7b90c33bee6b79efaa5ba1f4 +size 1754447232 diff --git a/Beagle14-7B.i1-IQ1_S.gguf b/Beagle14-7B.i1-IQ1_S.gguf new file mode 100644 index 0000000..f438c58 --- /dev/null +++ b/Beagle14-7B.i1-IQ1_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:72d121a3205b8520ff91a3b2cbf7a52e6b7a5ecefc5ab9bcf2f4860707b3869c +size 1612103040 diff --git a/Beagle14-7B.i1-IQ2_M.gguf b/Beagle14-7B.i1-IQ2_M.gguf new file mode 100644 index 0000000..efa68a2 --- /dev/null +++ b/Beagle14-7B.i1-IQ2_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:29f825240743f906a347d49a37545fc581a4fddfd9496a6a97311cb29092169f +size 2500713856 diff --git a/Beagle14-7B.i1-IQ2_S.gguf b/Beagle14-7B.i1-IQ2_S.gguf new file mode 100644 index 0000000..4974967 --- /dev/null +++ b/Beagle14-7B.i1-IQ2_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f89cadae8fcf109c25fe0a92c31e1602e7f345e9cd11744df0e4f197ae183fe1 +size 2310921600 diff --git a/Beagle14-7B.i1-IQ2_XS.gguf b/Beagle14-7B.i1-IQ2_XS.gguf new file mode 100644 index 0000000..ab323b2 --- /dev/null +++ b/Beagle14-7B.i1-IQ2_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d96aefa28e77a53a62b3a6a2ea5a8ab2782e3cb05d13e04c78364f4ac7d8ec53 +size 2198257024 diff --git a/Beagle14-7B.i1-IQ2_XXS.gguf b/Beagle14-7B.i1-IQ2_XXS.gguf new file mode 100644 index 0000000..92b7d55 --- /dev/null +++ b/Beagle14-7B.i1-IQ2_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f6b46ce010189631be00f40420cf6da70e9c1a312ef0192bcc133a4b0142069b +size 1991687552 diff --git a/Beagle14-7B.i1-IQ3_M.gguf b/Beagle14-7B.i1-IQ3_M.gguf new file mode 100644 index 0000000..bc03630 --- /dev/null +++ b/Beagle14-7B.i1-IQ3_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:92c771a89a0b972eaaf68a52ed21d1b05ee76bdd94a1531baff735556f47ae36 +size 3284893056 diff --git a/Beagle14-7B.i1-IQ3_S.gguf b/Beagle14-7B.i1-IQ3_S.gguf new file mode 100644 index 0000000..1414c3e --- /dev/null +++ b/Beagle14-7B.i1-IQ3_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9228aca6d180f81a00aedaaa95d5301121f2df6876c7262dd95ee1bce108614b +size 3182394752 diff --git a/Beagle14-7B.i1-IQ3_XS.gguf b/Beagle14-7B.i1-IQ3_XS.gguf new file mode 100644 index 0000000..8078447 --- /dev/null +++ b/Beagle14-7B.i1-IQ3_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:18920f03bef45e6ca5bd2189ce835a6f1e26935355a20c86733a5e9eae7b18e0 +size 3018816896 diff --git a/Beagle14-7B.i1-IQ3_XXS.gguf b/Beagle14-7B.i1-IQ3_XXS.gguf new file mode 100644 index 0000000..20a0298 --- /dev/null +++ b/Beagle14-7B.i1-IQ3_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aef9f44c84dd1ba3289532ca947d753d6af4faaa54da0fcede096165e67a9ccd +size 2827345280 diff --git a/Beagle14-7B.i1-IQ4_XS.gguf b/Beagle14-7B.i1-IQ4_XS.gguf new file mode 100644 index 0000000..f1bba0b --- /dev/null +++ b/Beagle14-7B.i1-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e70ff477039d6d53221be9a10fb9282c4803acb5ac30d99da080ff587c86e65e +size 3907689856 diff --git a/Beagle14-7B.i1-Q2_K.gguf b/Beagle14-7B.i1-Q2_K.gguf new file mode 100644 index 0000000..bce8317 --- /dev/null +++ b/Beagle14-7B.i1-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3f83fd0a0908a29384cb1fe0c675ce02463f283ac5a009b000151c74d19c41de +size 2719243648 diff --git a/Beagle14-7B.i1-Q2_K_S.gguf b/Beagle14-7B.i1-Q2_K_S.gguf new file mode 100644 index 0000000..ebe68c5 --- /dev/null +++ b/Beagle14-7B.i1-Q2_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b531f8d2b903f14186110bc0c023d25120ffc29d4e456ccc190682c4af7b8616 +size 2528927104 diff --git a/Beagle14-7B.i1-Q3_K_L.gguf b/Beagle14-7B.i1-Q3_K_L.gguf new file mode 100644 index 0000000..ed94512 --- /dev/null +++ b/Beagle14-7B.i1-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a4258ba9147906e2257ae04f5f085cfd212395a7df5c98f01522881f99689f07 +size 3822026112 diff --git a/Beagle14-7B.i1-Q3_K_M.gguf b/Beagle14-7B.i1-Q3_K_M.gguf new file mode 100644 index 0000000..8c92d11 --- /dev/null +++ b/Beagle14-7B.i1-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:048461de91ad2cdcfd471ab63a1e79e3d730926a52a9f9a76a70f360ccf6e349 +size 3518987648 diff --git a/Beagle14-7B.i1-Q3_K_S.gguf b/Beagle14-7B.i1-Q3_K_S.gguf new file mode 100644 index 0000000..d80f6da --- /dev/null +++ b/Beagle14-7B.i1-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d649df914bf2b4d95f2fd4e760d5b53d919db1bb32358629cf6106574841c2b +size 3164568960 diff --git a/Beagle14-7B.i1-Q4_0.gguf b/Beagle14-7B.i1-Q4_0.gguf new file mode 100644 index 0000000..237f98f --- /dev/null +++ b/Beagle14-7B.i1-Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:487cd71282daede8d16327dedf34ae578893a513a5fbe0df4f0238cdd4786d52 +size 4123598208 diff --git a/Beagle14-7B.i1-Q4_K_M.gguf b/Beagle14-7B.i1-Q4_K_M.gguf new file mode 100644 index 0000000..8b9b02f --- /dev/null +++ b/Beagle14-7B.i1-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b06a2ce106a3940347323586db59b6b851df086985178c83018d475282267313 +size 4368440704 diff --git a/Beagle14-7B.i1-Q4_K_S.gguf b/Beagle14-7B.i1-Q4_K_S.gguf new file mode 100644 index 0000000..1ee6cc8 --- /dev/null +++ b/Beagle14-7B.i1-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:58b3638caaf2f7c85cb5a49142cc02d24c5f9ba5ff3786b173633f8289dd2d5e +size 4140375424 diff --git a/Beagle14-7B.i1-Q5_K_M.gguf b/Beagle14-7B.i1-Q5_K_M.gguf new file mode 100644 index 0000000..6527b9b --- /dev/null +++ b/Beagle14-7B.i1-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:885d3baa2119116d5210f400257533f0b90700ef5146b75bc237f8673c7a8443 +size 5131410816 diff --git a/Beagle14-7B.i1-Q5_K_S.gguf b/Beagle14-7B.i1-Q5_K_S.gguf new file mode 100644 index 0000000..634b471 --- /dev/null +++ b/Beagle14-7B.i1-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:89cb952200bbdbec1dd45d451844ecd1c09bee6f1e0648a1142a6a9c8f351d9a +size 4997717376 diff --git a/Beagle14-7B.i1-Q6_K.gguf b/Beagle14-7B.i1-Q6_K.gguf new file mode 100644 index 0000000..855c9d7 --- /dev/null +++ b/Beagle14-7B.i1-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b8b88e2dd3416fd04a79ea825b17ab1b3ef4576e7bb8b25e1d695496469e5d16 +size 5942066560 diff --git a/README.md b/README.md new file mode 100644 index 0000000..869b2e0 --- /dev/null +++ b/README.md @@ -0,0 +1,80 @@ +--- +base_model: mlabonne/Beagle14-7B +language: +- en +library_name: transformers +license: cc-by-nc-4.0 +quantized_by: mradermacher +tags: +- merge +- mergekit +- lazymergekit +- fblgit/UNA-TheBeagle-7b-v1 +- argilla/distilabeled-Marcoro14-7B-slerp +--- +## About + + + + + + +weighted/imatrix quants of https://huggingface.co/mlabonne/Beagle14-7B + + +static quants are available at https://huggingface.co/mradermacher/Beagle14-7B-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/Beagle14-7B-i1-GGUF/resolve/main/Beagle14-7B.i1-IQ1_S.gguf) | i1-IQ1_S | 1.7 | for the desperate | +| [GGUF](https://huggingface.co/mradermacher/Beagle14-7B-i1-GGUF/resolve/main/Beagle14-7B.i1-IQ1_M.gguf) | i1-IQ1_M | 1.9 | mostly desperate | +| [GGUF](https://huggingface.co/mradermacher/Beagle14-7B-i1-GGUF/resolve/main/Beagle14-7B.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 2.1 | | +| [GGUF](https://huggingface.co/mradermacher/Beagle14-7B-i1-GGUF/resolve/main/Beagle14-7B.i1-IQ2_XS.gguf) | i1-IQ2_XS | 2.3 | | +| [GGUF](https://huggingface.co/mradermacher/Beagle14-7B-i1-GGUF/resolve/main/Beagle14-7B.i1-IQ2_S.gguf) | i1-IQ2_S | 2.4 | | +| [GGUF](https://huggingface.co/mradermacher/Beagle14-7B-i1-GGUF/resolve/main/Beagle14-7B.i1-IQ2_M.gguf) | i1-IQ2_M | 2.6 | | +| [GGUF](https://huggingface.co/mradermacher/Beagle14-7B-i1-GGUF/resolve/main/Beagle14-7B.i1-Q2_K_S.gguf) | i1-Q2_K_S | 2.6 | very low quality | +| [GGUF](https://huggingface.co/mradermacher/Beagle14-7B-i1-GGUF/resolve/main/Beagle14-7B.i1-Q2_K.gguf) | i1-Q2_K | 2.8 | IQ3_XXS probably better | +| [GGUF](https://huggingface.co/mradermacher/Beagle14-7B-i1-GGUF/resolve/main/Beagle14-7B.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 2.9 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/Beagle14-7B-i1-GGUF/resolve/main/Beagle14-7B.i1-IQ3_XS.gguf) | i1-IQ3_XS | 3.1 | | +| [GGUF](https://huggingface.co/mradermacher/Beagle14-7B-i1-GGUF/resolve/main/Beagle14-7B.i1-Q3_K_S.gguf) | i1-Q3_K_S | 3.3 | IQ3_XS probably better | +| [GGUF](https://huggingface.co/mradermacher/Beagle14-7B-i1-GGUF/resolve/main/Beagle14-7B.i1-IQ3_S.gguf) | i1-IQ3_S | 3.3 | beats Q3_K* | +| [GGUF](https://huggingface.co/mradermacher/Beagle14-7B-i1-GGUF/resolve/main/Beagle14-7B.i1-IQ3_M.gguf) | i1-IQ3_M | 3.4 | | +| [GGUF](https://huggingface.co/mradermacher/Beagle14-7B-i1-GGUF/resolve/main/Beagle14-7B.i1-Q3_K_M.gguf) | i1-Q3_K_M | 3.6 | IQ3_S probably better | +| [GGUF](https://huggingface.co/mradermacher/Beagle14-7B-i1-GGUF/resolve/main/Beagle14-7B.i1-Q3_K_L.gguf) | i1-Q3_K_L | 3.9 | IQ3_M probably better | +| [GGUF](https://huggingface.co/mradermacher/Beagle14-7B-i1-GGUF/resolve/main/Beagle14-7B.i1-IQ4_XS.gguf) | i1-IQ4_XS | 4.0 | | +| [GGUF](https://huggingface.co/mradermacher/Beagle14-7B-i1-GGUF/resolve/main/Beagle14-7B.i1-Q4_0.gguf) | i1-Q4_0 | 4.2 | fast, low quality | +| [GGUF](https://huggingface.co/mradermacher/Beagle14-7B-i1-GGUF/resolve/main/Beagle14-7B.i1-Q4_K_S.gguf) | i1-Q4_K_S | 4.2 | optimal size/speed/quality | +| [GGUF](https://huggingface.co/mradermacher/Beagle14-7B-i1-GGUF/resolve/main/Beagle14-7B.i1-Q4_K_M.gguf) | i1-Q4_K_M | 4.5 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Beagle14-7B-i1-GGUF/resolve/main/Beagle14-7B.i1-Q5_K_S.gguf) | i1-Q5_K_S | 5.1 | | +| [GGUF](https://huggingface.co/mradermacher/Beagle14-7B-i1-GGUF/resolve/main/Beagle14-7B.i1-Q5_K_M.gguf) | i1-Q5_K_M | 5.2 | | +| [GGUF](https://huggingface.co/mradermacher/Beagle14-7B-i1-GGUF/resolve/main/Beagle14-7B.i1-Q6_K.gguf) | i1-Q6_K | 6.0 | practically like static Q6_K | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to. + + diff --git a/imatrix.dat b/imatrix.dat new file mode 100644 index 0000000..27c2a27 --- /dev/null +++ b/imatrix.dat @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:650a3aed709026a1cad9d34281e1552bb75e153985b57bc07b62e376f816f07f +size 4988157