From b1b030124987d11006bc9ca41b5db4586748e062 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Sat, 9 May 2026 13:28:39 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: mradermacher/DS-Qwen-7b-GG-CalibratedConfRL-GGUF Source: Original Platform --- .gitattributes | 47 ++++++++++++++++ DS-Qwen-7b-GG-CalibratedConfRL.IQ4_XS.gguf | 3 + DS-Qwen-7b-GG-CalibratedConfRL.Q2_K.gguf | 3 + DS-Qwen-7b-GG-CalibratedConfRL.Q3_K_L.gguf | 3 + DS-Qwen-7b-GG-CalibratedConfRL.Q3_K_M.gguf | 3 + DS-Qwen-7b-GG-CalibratedConfRL.Q3_K_S.gguf | 3 + DS-Qwen-7b-GG-CalibratedConfRL.Q4_K_M.gguf | 3 + DS-Qwen-7b-GG-CalibratedConfRL.Q4_K_S.gguf | 3 + DS-Qwen-7b-GG-CalibratedConfRL.Q5_K_M.gguf | 3 + DS-Qwen-7b-GG-CalibratedConfRL.Q5_K_S.gguf | 3 + DS-Qwen-7b-GG-CalibratedConfRL.Q6_K.gguf | 3 + DS-Qwen-7b-GG-CalibratedConfRL.Q8_0.gguf | 3 + DS-Qwen-7b-GG-CalibratedConfRL.f16.gguf | 3 + README.md | 64 ++++++++++++++++++++++ 14 files changed, 147 insertions(+) create mode 100644 .gitattributes create mode 100644 DS-Qwen-7b-GG-CalibratedConfRL.IQ4_XS.gguf create mode 100644 DS-Qwen-7b-GG-CalibratedConfRL.Q2_K.gguf create mode 100644 DS-Qwen-7b-GG-CalibratedConfRL.Q3_K_L.gguf create mode 100644 DS-Qwen-7b-GG-CalibratedConfRL.Q3_K_M.gguf create mode 100644 DS-Qwen-7b-GG-CalibratedConfRL.Q3_K_S.gguf create mode 100644 DS-Qwen-7b-GG-CalibratedConfRL.Q4_K_M.gguf create mode 100644 DS-Qwen-7b-GG-CalibratedConfRL.Q4_K_S.gguf create mode 100644 DS-Qwen-7b-GG-CalibratedConfRL.Q5_K_M.gguf create mode 100644 DS-Qwen-7b-GG-CalibratedConfRL.Q5_K_S.gguf create mode 100644 DS-Qwen-7b-GG-CalibratedConfRL.Q6_K.gguf create mode 100644 DS-Qwen-7b-GG-CalibratedConfRL.Q8_0.gguf create mode 100644 DS-Qwen-7b-GG-CalibratedConfRL.f16.gguf create mode 100644 README.md diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..993ff0e --- /dev/null +++ b/.gitattributes @@ -0,0 +1,47 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +DS-Qwen-7b-GG-CalibratedConfRL.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +DS-Qwen-7b-GG-CalibratedConfRL.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +DS-Qwen-7b-GG-CalibratedConfRL.f16.gguf filter=lfs diff=lfs merge=lfs -text +DS-Qwen-7b-GG-CalibratedConfRL.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +DS-Qwen-7b-GG-CalibratedConfRL.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +DS-Qwen-7b-GG-CalibratedConfRL.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +DS-Qwen-7b-GG-CalibratedConfRL.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +DS-Qwen-7b-GG-CalibratedConfRL.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +DS-Qwen-7b-GG-CalibratedConfRL.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +DS-Qwen-7b-GG-CalibratedConfRL.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +DS-Qwen-7b-GG-CalibratedConfRL.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +DS-Qwen-7b-GG-CalibratedConfRL.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/DS-Qwen-7b-GG-CalibratedConfRL.IQ4_XS.gguf b/DS-Qwen-7b-GG-CalibratedConfRL.IQ4_XS.gguf new file mode 100644 index 0000000..c7a189f --- /dev/null +++ b/DS-Qwen-7b-GG-CalibratedConfRL.IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:826706fa6e22a188df66f758fbe2318e60bc963b1a1ced6a08f5536d4638382c +size 4250298560 diff --git a/DS-Qwen-7b-GG-CalibratedConfRL.Q2_K.gguf b/DS-Qwen-7b-GG-CalibratedConfRL.Q2_K.gguf new file mode 100644 index 0000000..4422e9d --- /dev/null +++ b/DS-Qwen-7b-GG-CalibratedConfRL.Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66dd8d1ffdf3f47ddccc5d563697560364352b603df5bb25fd4c296a3d9467be +size 3015940288 diff --git a/DS-Qwen-7b-GG-CalibratedConfRL.Q3_K_L.gguf b/DS-Qwen-7b-GG-CalibratedConfRL.Q3_K_L.gguf new file mode 100644 index 0000000..02ac161 --- /dev/null +++ b/DS-Qwen-7b-GG-CalibratedConfRL.Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:27fcea63ad6e65309fd8a46241e27ad7aab5c27fa0c5e55310f4ef3d3cc91192 +size 4088459456 diff --git a/DS-Qwen-7b-GG-CalibratedConfRL.Q3_K_M.gguf b/DS-Qwen-7b-GG-CalibratedConfRL.Q3_K_M.gguf new file mode 100644 index 0000000..1afca23 --- /dev/null +++ b/DS-Qwen-7b-GG-CalibratedConfRL.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b554827193304aa0137580ca97f9cc8723ea1a4524235a6492ec82d70b1bdb75 +size 3808391360 diff --git a/DS-Qwen-7b-GG-CalibratedConfRL.Q3_K_S.gguf b/DS-Qwen-7b-GG-CalibratedConfRL.Q3_K_S.gguf new file mode 100644 index 0000000..baa0981 --- /dev/null +++ b/DS-Qwen-7b-GG-CalibratedConfRL.Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d9c25a4c103efd2d1432441ac076e28ce0e4b48a053586dea7301575e35b7da9 +size 3492368576 diff --git a/DS-Qwen-7b-GG-CalibratedConfRL.Q4_K_M.gguf b/DS-Qwen-7b-GG-CalibratedConfRL.Q4_K_M.gguf new file mode 100644 index 0000000..d8d583e --- /dev/null +++ b/DS-Qwen-7b-GG-CalibratedConfRL.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a8f5f57f79d5ac858d0f368e6f44fb1dc0d9e4806a3ca56df4df339b02a13ba +size 4683073728 diff --git a/DS-Qwen-7b-GG-CalibratedConfRL.Q4_K_S.gguf b/DS-Qwen-7b-GG-CalibratedConfRL.Q4_K_S.gguf new file mode 100644 index 0000000..7f0434e --- /dev/null +++ b/DS-Qwen-7b-GG-CalibratedConfRL.Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c6fc392c48fc6c936bbf93ec2da70e97efbb984ef3155733aeb4d07a4b82d55f +size 4457769152 diff --git a/DS-Qwen-7b-GG-CalibratedConfRL.Q5_K_M.gguf b/DS-Qwen-7b-GG-CalibratedConfRL.Q5_K_M.gguf new file mode 100644 index 0000000..7c37ba2 --- /dev/null +++ b/DS-Qwen-7b-GG-CalibratedConfRL.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f5ed3fb0b03cfd0f5297b018e8c19fb4c8711d815749ae980e1372e638e5352b +size 5444831424 diff --git a/DS-Qwen-7b-GG-CalibratedConfRL.Q5_K_S.gguf b/DS-Qwen-7b-GG-CalibratedConfRL.Q5_K_S.gguf new file mode 100644 index 0000000..94306cc --- /dev/null +++ b/DS-Qwen-7b-GG-CalibratedConfRL.Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:88e49b93c1de250bc6e595748fb3108617c6557ab81bc2ffd48cffe0c5d82202 +size 5315176640 diff --git a/DS-Qwen-7b-GG-CalibratedConfRL.Q6_K.gguf b/DS-Qwen-7b-GG-CalibratedConfRL.Q6_K.gguf new file mode 100644 index 0000000..c676d86 --- /dev/null +++ b/DS-Qwen-7b-GG-CalibratedConfRL.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a8c55e190b54fd1593dcafb87d7c72ddcf15fc57367f6bab4c6b56d66b0bd0c +size 6254198976 diff --git a/DS-Qwen-7b-GG-CalibratedConfRL.Q8_0.gguf b/DS-Qwen-7b-GG-CalibratedConfRL.Q8_0.gguf new file mode 100644 index 0000000..7e5c14f --- /dev/null +++ b/DS-Qwen-7b-GG-CalibratedConfRL.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec9d3c1643354883b51551780c37536a004c7d8cb3904c00ae3544e441bf128c +size 8098525376 diff --git a/DS-Qwen-7b-GG-CalibratedConfRL.f16.gguf b/DS-Qwen-7b-GG-CalibratedConfRL.f16.gguf new file mode 100644 index 0000000..2e94b34 --- /dev/null +++ b/DS-Qwen-7b-GG-CalibratedConfRL.f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b82e30374a57aba3eef729d46c4d8d417af2ea72a38eb4eecf01ee3724de3795 +size 15237853376 diff --git a/README.md b/README.md new file mode 100644 index 0000000..129066c --- /dev/null +++ b/README.md @@ -0,0 +1,64 @@ +--- +base_model: AmirhoseinGH/DS-Qwen-7b-GG-CalibratedConfRL +language: +- en +library_name: transformers +license: apache-2.0 +quantized_by: mradermacher +--- +## About + + + + + + +static quants of https://huggingface.co/AmirhoseinGH/DS-Qwen-7b-GG-CalibratedConfRL + + +weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion. +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/DS-Qwen-7b-GG-CalibratedConfRL-GGUF/resolve/main/DS-Qwen-7b-GG-CalibratedConfRL.Q2_K.gguf) | Q2_K | 3.1 | | +| [GGUF](https://huggingface.co/mradermacher/DS-Qwen-7b-GG-CalibratedConfRL-GGUF/resolve/main/DS-Qwen-7b-GG-CalibratedConfRL.Q3_K_S.gguf) | Q3_K_S | 3.6 | | +| [GGUF](https://huggingface.co/mradermacher/DS-Qwen-7b-GG-CalibratedConfRL-GGUF/resolve/main/DS-Qwen-7b-GG-CalibratedConfRL.Q3_K_M.gguf) | Q3_K_M | 3.9 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/DS-Qwen-7b-GG-CalibratedConfRL-GGUF/resolve/main/DS-Qwen-7b-GG-CalibratedConfRL.Q3_K_L.gguf) | Q3_K_L | 4.2 | | +| [GGUF](https://huggingface.co/mradermacher/DS-Qwen-7b-GG-CalibratedConfRL-GGUF/resolve/main/DS-Qwen-7b-GG-CalibratedConfRL.IQ4_XS.gguf) | IQ4_XS | 4.4 | | +| [GGUF](https://huggingface.co/mradermacher/DS-Qwen-7b-GG-CalibratedConfRL-GGUF/resolve/main/DS-Qwen-7b-GG-CalibratedConfRL.Q4_K_S.gguf) | Q4_K_S | 4.6 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/DS-Qwen-7b-GG-CalibratedConfRL-GGUF/resolve/main/DS-Qwen-7b-GG-CalibratedConfRL.Q4_K_M.gguf) | Q4_K_M | 4.8 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/DS-Qwen-7b-GG-CalibratedConfRL-GGUF/resolve/main/DS-Qwen-7b-GG-CalibratedConfRL.Q5_K_S.gguf) | Q5_K_S | 5.4 | | +| [GGUF](https://huggingface.co/mradermacher/DS-Qwen-7b-GG-CalibratedConfRL-GGUF/resolve/main/DS-Qwen-7b-GG-CalibratedConfRL.Q5_K_M.gguf) | Q5_K_M | 5.5 | | +| [GGUF](https://huggingface.co/mradermacher/DS-Qwen-7b-GG-CalibratedConfRL-GGUF/resolve/main/DS-Qwen-7b-GG-CalibratedConfRL.Q6_K.gguf) | Q6_K | 6.4 | very good quality | +| [GGUF](https://huggingface.co/mradermacher/DS-Qwen-7b-GG-CalibratedConfRL-GGUF/resolve/main/DS-Qwen-7b-GG-CalibratedConfRL.Q8_0.gguf) | Q8_0 | 8.2 | fast, best quality | +| [GGUF](https://huggingface.co/mradermacher/DS-Qwen-7b-GG-CalibratedConfRL-GGUF/resolve/main/DS-Qwen-7b-GG-CalibratedConfRL.f16.gguf) | f16 | 15.3 | 16 bpw, overkill | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. + +