commit c9c0a9ddc575a34c35d3164591998534f7ad860d Author: ModelHub XC Date: Fri May 8 22:28:04 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: mradermacher/RC-Qwen2VL-2b-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..293204a --- /dev/null +++ b/.gitattributes @@ -0,0 +1,49 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +RC-Qwen2VL-2b.mmproj-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +RC-Qwen2VL-2b.mmproj-f16.gguf filter=lfs diff=lfs merge=lfs -text +RC-Qwen2VL-2b.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +RC-Qwen2VL-2b.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +RC-Qwen2VL-2b.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +RC-Qwen2VL-2b.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +RC-Qwen2VL-2b.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +RC-Qwen2VL-2b.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +RC-Qwen2VL-2b.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +RC-Qwen2VL-2b.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +RC-Qwen2VL-2b.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +RC-Qwen2VL-2b.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +RC-Qwen2VL-2b.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +RC-Qwen2VL-2b.f16.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/RC-Qwen2VL-2b.IQ4_XS.gguf b/RC-Qwen2VL-2b.IQ4_XS.gguf new file mode 100644 index 0000000..6efad14 --- /dev/null +++ b/RC-Qwen2VL-2b.IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:47f360de7dd2decccd027e6a28cef6aad9de746135a014f0e43a4ca9257d8240 +size 902181856 diff --git a/RC-Qwen2VL-2b.Q2_K.gguf b/RC-Qwen2VL-2b.Q2_K.gguf new file mode 100644 index 0000000..a23c770 --- /dev/null +++ b/RC-Qwen2VL-2b.Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:41a9c62f368a9544ca5ac051237966cc094cc976203036bd7acbe1aaf3436490 +size 676303840 diff --git a/RC-Qwen2VL-2b.Q3_K_L.gguf b/RC-Qwen2VL-2b.Q3_K_L.gguf new file mode 100644 index 0000000..0d2626e --- /dev/null +++ b/RC-Qwen2VL-2b.Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cc06deac7d1e3874fbf4537f4d591a89733404919e01b115a4dffbf4c51a8505 +size 880161760 diff --git a/RC-Qwen2VL-2b.Q3_K_M.gguf b/RC-Qwen2VL-2b.Q3_K_M.gguf new file mode 100644 index 0000000..e887a46 --- /dev/null +++ b/RC-Qwen2VL-2b.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:947137c05bd1e3e77aaaa593d8282ef12a533fbce1a255aa71f2a342feb2d3d3 +size 824177632 diff --git a/RC-Qwen2VL-2b.Q3_K_S.gguf b/RC-Qwen2VL-2b.Q3_K_S.gguf new file mode 100644 index 0000000..332930f --- /dev/null +++ b/RC-Qwen2VL-2b.Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9419cf6e16e1dbcf8a75354923eabdf395bb87ac8cd9c234f9ecf7dbb4e2c755 +size 760943584 diff --git a/RC-Qwen2VL-2b.Q4_K_M.gguf b/RC-Qwen2VL-2b.Q4_K_M.gguf new file mode 100644 index 0000000..a7a384e --- /dev/null +++ b/RC-Qwen2VL-2b.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:50a08a466187f8b6fcc55b361d9eb3eb4f185b37ac3b43e9e1ce473202f8ca9f +size 986047456 diff --git a/RC-Qwen2VL-2b.Q4_K_S.gguf b/RC-Qwen2VL-2b.Q4_K_S.gguf new file mode 100644 index 0000000..5028db8 --- /dev/null +++ b/RC-Qwen2VL-2b.Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a05d2ae00a790480822221e364c26dacab30dd25034238965ae019a60f869b84 +size 940311520 diff --git a/RC-Qwen2VL-2b.Q5_K_M.gguf b/RC-Qwen2VL-2b.Q5_K_M.gguf new file mode 100644 index 0000000..445de44 --- /dev/null +++ b/RC-Qwen2VL-2b.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:433ebbd0c4ef2d1de525018544b48cfb9abfc7be8acc33e1ecd0d4286307d155 +size 1125049312 diff --git a/RC-Qwen2VL-2b.Q5_K_S.gguf b/RC-Qwen2VL-2b.Q5_K_S.gguf new file mode 100644 index 0000000..57a0534 --- /dev/null +++ b/RC-Qwen2VL-2b.Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd4444c891c104abe2ad8253b76400340e60a2d37c8f385e3831e9e12e6b98d3 +size 1098728416 diff --git a/RC-Qwen2VL-2b.Q6_K.gguf b/RC-Qwen2VL-2b.Q6_K.gguf new file mode 100644 index 0000000..36ed1da --- /dev/null +++ b/RC-Qwen2VL-2b.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d62c4ab92101e2174e400fe4d27549bda2e1204e13e926b77659a9413224467 +size 1272738784 diff --git a/RC-Qwen2VL-2b.Q8_0.gguf b/RC-Qwen2VL-2b.Q8_0.gguf new file mode 100644 index 0000000..100c61a --- /dev/null +++ b/RC-Qwen2VL-2b.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b896c7322bc15966590beab8918a69b5c5f23f57bf3820483de78845e6f70973 +size 1646572000 diff --git a/RC-Qwen2VL-2b.f16.gguf b/RC-Qwen2VL-2b.f16.gguf new file mode 100644 index 0000000..182eacd --- /dev/null +++ b/RC-Qwen2VL-2b.f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8cf5b3214a5ade20e9e5756d2de8768206b90ffcdc8e4915cf1e97797b2e5ee5 +size 3093668320 diff --git a/RC-Qwen2VL-2b.mmproj-Q8_0.gguf b/RC-Qwen2VL-2b.mmproj-Q8_0.gguf new file mode 100644 index 0000000..50a0337 --- /dev/null +++ b/RC-Qwen2VL-2b.mmproj-Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2f260de81f030c6795ed34a6f454d5bf7fd7b23daf6ccd53cf33e65b90f1785e +size 712893984 diff --git a/RC-Qwen2VL-2b.mmproj-f16.gguf b/RC-Qwen2VL-2b.mmproj-f16.gguf new file mode 100644 index 0000000..0682bd4 --- /dev/null +++ b/RC-Qwen2VL-2b.mmproj-f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c71698aa935913f1a69bf2607208edf876c6f993f297f12a021e0fc42a5f2cb +size 1331656224 diff --git a/README.md b/README.md new file mode 100644 index 0000000..68dfbe5 --- /dev/null +++ b/README.md @@ -0,0 +1,78 @@ +--- +base_model: weihongliang/RC-Qwen2VL-2b +language: +- en +library_name: transformers +license: apache-2.0 +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +tags: +- multimodal +- llm +- personalized_multimodal_understanding +--- +## About + + + + + + + + + +static quants of https://huggingface.co/weihongliang/RC-Qwen2VL-2b + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#RC-Qwen2VL-2b-GGUF).*** + +weighted/imatrix quants are available at https://huggingface.co/mradermacher/RC-Qwen2VL-2b-i1-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/RC-Qwen2VL-2b-GGUF/resolve/main/RC-Qwen2VL-2b.Q2_K.gguf) | Q2_K | 0.8 | | +| [GGUF](https://huggingface.co/mradermacher/RC-Qwen2VL-2b-GGUF/resolve/main/RC-Qwen2VL-2b.mmproj-Q8_0.gguf) | mmproj-Q8_0 | 0.8 | multi-modal supplement | +| [GGUF](https://huggingface.co/mradermacher/RC-Qwen2VL-2b-GGUF/resolve/main/RC-Qwen2VL-2b.Q3_K_S.gguf) | Q3_K_S | 0.9 | | +| [GGUF](https://huggingface.co/mradermacher/RC-Qwen2VL-2b-GGUF/resolve/main/RC-Qwen2VL-2b.Q3_K_M.gguf) | Q3_K_M | 0.9 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/RC-Qwen2VL-2b-GGUF/resolve/main/RC-Qwen2VL-2b.Q3_K_L.gguf) | Q3_K_L | 1.0 | | +| [GGUF](https://huggingface.co/mradermacher/RC-Qwen2VL-2b-GGUF/resolve/main/RC-Qwen2VL-2b.IQ4_XS.gguf) | IQ4_XS | 1.0 | | +| [GGUF](https://huggingface.co/mradermacher/RC-Qwen2VL-2b-GGUF/resolve/main/RC-Qwen2VL-2b.Q4_K_S.gguf) | Q4_K_S | 1.0 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/RC-Qwen2VL-2b-GGUF/resolve/main/RC-Qwen2VL-2b.Q4_K_M.gguf) | Q4_K_M | 1.1 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/RC-Qwen2VL-2b-GGUF/resolve/main/RC-Qwen2VL-2b.Q5_K_S.gguf) | Q5_K_S | 1.2 | | +| [GGUF](https://huggingface.co/mradermacher/RC-Qwen2VL-2b-GGUF/resolve/main/RC-Qwen2VL-2b.Q5_K_M.gguf) | Q5_K_M | 1.2 | | +| [GGUF](https://huggingface.co/mradermacher/RC-Qwen2VL-2b-GGUF/resolve/main/RC-Qwen2VL-2b.Q6_K.gguf) | Q6_K | 1.4 | very good quality | +| [GGUF](https://huggingface.co/mradermacher/RC-Qwen2VL-2b-GGUF/resolve/main/RC-Qwen2VL-2b.mmproj-f16.gguf) | mmproj-f16 | 1.4 | multi-modal supplement | +| [GGUF](https://huggingface.co/mradermacher/RC-Qwen2VL-2b-GGUF/resolve/main/RC-Qwen2VL-2b.Q8_0.gguf) | Q8_0 | 1.7 | fast, best quality | +| [GGUF](https://huggingface.co/mradermacher/RC-Qwen2VL-2b-GGUF/resolve/main/RC-Qwen2VL-2b.f16.gguf) | f16 | 3.2 | 16 bpw, overkill | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. + +