commit f098740925edcf182c7d590246a3be200739fd2c Author: ModelHub XC Date: Thu May 7 03:37:02 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: mradermacher/Qwen-SEA-LION-v4-4B-VL-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..06bfa13 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,49 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +Qwen-SEA-LION-v4-4B-VL.mmproj-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +Qwen-SEA-LION-v4-4B-VL.mmproj-f16.gguf filter=lfs diff=lfs merge=lfs -text +Qwen-SEA-LION-v4-4B-VL.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +Qwen-SEA-LION-v4-4B-VL.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +Qwen-SEA-LION-v4-4B-VL.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +Qwen-SEA-LION-v4-4B-VL.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Qwen-SEA-LION-v4-4B-VL.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Qwen-SEA-LION-v4-4B-VL.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Qwen-SEA-LION-v4-4B-VL.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Qwen-SEA-LION-v4-4B-VL.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Qwen-SEA-LION-v4-4B-VL.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Qwen-SEA-LION-v4-4B-VL.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Qwen-SEA-LION-v4-4B-VL.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +Qwen-SEA-LION-v4-4B-VL.f16.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/Qwen-SEA-LION-v4-4B-VL.IQ4_XS.gguf b/Qwen-SEA-LION-v4-4B-VL.IQ4_XS.gguf new file mode 100644 index 0000000..18ea6a4 --- /dev/null +++ b/Qwen-SEA-LION-v4-4B-VL.IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8d057b9dd4d073d14770a6f4b9b27094e27e9858364850c57b4695f417d184c7 +size 2286318112 diff --git a/Qwen-SEA-LION-v4-4B-VL.Q2_K.gguf b/Qwen-SEA-LION-v4-4B-VL.Q2_K.gguf new file mode 100644 index 0000000..5b23100 --- /dev/null +++ b/Qwen-SEA-LION-v4-4B-VL.Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:20db7cf362604ace846712a849abe3a206ae38293fe75ccc7b948be8405f8b8f +size 1669501472 diff --git a/Qwen-SEA-LION-v4-4B-VL.Q3_K_L.gguf b/Qwen-SEA-LION-v4-4B-VL.Q3_K_L.gguf new file mode 100644 index 0000000..6b67799 --- /dev/null +++ b/Qwen-SEA-LION-v4-4B-VL.Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d895a46e772aa2664608a28c5194cf30faed1ca7254d0ca57dfd39d66d980ce4 +size 2239787552 diff --git a/Qwen-SEA-LION-v4-4B-VL.Q3_K_M.gguf b/Qwen-SEA-LION-v4-4B-VL.Q3_K_M.gguf new file mode 100644 index 0000000..c6ac396 --- /dev/null +++ b/Qwen-SEA-LION-v4-4B-VL.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:255986928bc9e1fef9f42615454566e7a952958510ee21d6c7186887392dc9ca +size 2075619872 diff --git a/Qwen-SEA-LION-v4-4B-VL.Q3_K_S.gguf b/Qwen-SEA-LION-v4-4B-VL.Q3_K_S.gguf new file mode 100644 index 0000000..caf676e --- /dev/null +++ b/Qwen-SEA-LION-v4-4B-VL.Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:76c3586fb8aa9343e74cf1d4913fe4ddd460a820a8cad56cd40ec487cdfc8f22 +size 1886999072 diff --git a/Qwen-SEA-LION-v4-4B-VL.Q4_K_M.gguf b/Qwen-SEA-LION-v4-4B-VL.Q4_K_M.gguf new file mode 100644 index 0000000..ccff8af --- /dev/null +++ b/Qwen-SEA-LION-v4-4B-VL.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:58e326a32a9de47295c9f4501d4fc3bd82b293c3220b0f46103f70ba41e10eb2 +size 2497282592 diff --git a/Qwen-SEA-LION-v4-4B-VL.Q4_K_S.gguf b/Qwen-SEA-LION-v4-4B-VL.Q4_K_S.gguf new file mode 100644 index 0000000..9cac027 --- /dev/null +++ b/Qwen-SEA-LION-v4-4B-VL.Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f54e48519f03a9ea1af8b2beb23d828eefd09c599225522a5ccd1e25600a03cb +size 2383311392 diff --git a/Qwen-SEA-LION-v4-4B-VL.Q5_K_M.gguf b/Qwen-SEA-LION-v4-4B-VL.Q5_K_M.gguf new file mode 100644 index 0000000..87a4cfe --- /dev/null +++ b/Qwen-SEA-LION-v4-4B-VL.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:db6dd05be8bca881815244e5f71d9b904ebc87f37571435fe6b1481bcb16e53e +size 2889515552 diff --git a/Qwen-SEA-LION-v4-4B-VL.Q5_K_S.gguf b/Qwen-SEA-LION-v4-4B-VL.Q5_K_S.gguf new file mode 100644 index 0000000..0faf7b4 --- /dev/null +++ b/Qwen-SEA-LION-v4-4B-VL.Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:78cc30a8c3f7d4ac0db2a01c338f708bda1cb8b8847be8df0a2132b9a5b856f7 +size 2823713312 diff --git a/Qwen-SEA-LION-v4-4B-VL.Q6_K.gguf b/Qwen-SEA-LION-v4-4B-VL.Q6_K.gguf new file mode 100644 index 0000000..ea453ec --- /dev/null +++ b/Qwen-SEA-LION-v4-4B-VL.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3ff9481192b160766bcb29cc2e64e66e1e89f8db629ea960862e72a285fd744c +size 3306263072 diff --git a/Qwen-SEA-LION-v4-4B-VL.Q8_0.gguf b/Qwen-SEA-LION-v4-4B-VL.Q8_0.gguf new file mode 100644 index 0000000..1b5947b --- /dev/null +++ b/Qwen-SEA-LION-v4-4B-VL.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:afb7b6419fee58067e4ab23e85ae3ac4a085120d9f8a10b69656c0c89b4885ed +size 4280407072 diff --git a/Qwen-SEA-LION-v4-4B-VL.f16.gguf b/Qwen-SEA-LION-v4-4B-VL.f16.gguf new file mode 100644 index 0000000..4369838 --- /dev/null +++ b/Qwen-SEA-LION-v4-4B-VL.f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0719d5a2ce9b73c5455d13682f2168dc68db39b14a324358932777fd7434c305 +size 8051287072 diff --git a/Qwen-SEA-LION-v4-4B-VL.mmproj-Q8_0.gguf b/Qwen-SEA-LION-v4-4B-VL.mmproj-Q8_0.gguf new file mode 100644 index 0000000..8a05365 --- /dev/null +++ b/Qwen-SEA-LION-v4-4B-VL.mmproj-Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:95e7f825c022e4b3860e8a07f94a751027240b354305ce75409f46d3e6ba90dc +size 453974752 diff --git a/Qwen-SEA-LION-v4-4B-VL.mmproj-f16.gguf b/Qwen-SEA-LION-v4-4B-VL.mmproj-f16.gguf new file mode 100644 index 0000000..bb66361 --- /dev/null +++ b/Qwen-SEA-LION-v4-4B-VL.mmproj-f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1c8908268febbce506dd1094953cafa972928a9afaaf9243bf446703b0b746e9 +size 836180704 diff --git a/README.md b/README.md new file mode 100644 index 0000000..af2a646 --- /dev/null +++ b/README.md @@ -0,0 +1,80 @@ +--- +base_model: aisingapore/Qwen-SEA-LION-v4-4B-VL +language: +- en +- vi +- id +- th +- my +- ta +- tl +- ms +library_name: transformers +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +--- +## About + + + + + + + + + +static quants of https://huggingface.co/aisingapore/Qwen-SEA-LION-v4-4B-VL + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#Qwen-SEA-LION-v4-4B-VL-GGUF).*** + +weighted/imatrix quants are available at https://huggingface.co/mradermacher/Qwen-SEA-LION-v4-4B-VL-i1-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/Qwen-SEA-LION-v4-4B-VL-GGUF/resolve/main/Qwen-SEA-LION-v4-4B-VL.mmproj-Q8_0.gguf) | mmproj-Q8_0 | 0.6 | multi-modal supplement | +| [GGUF](https://huggingface.co/mradermacher/Qwen-SEA-LION-v4-4B-VL-GGUF/resolve/main/Qwen-SEA-LION-v4-4B-VL.mmproj-f16.gguf) | mmproj-f16 | 0.9 | multi-modal supplement | +| [GGUF](https://huggingface.co/mradermacher/Qwen-SEA-LION-v4-4B-VL-GGUF/resolve/main/Qwen-SEA-LION-v4-4B-VL.Q2_K.gguf) | Q2_K | 1.8 | | +| [GGUF](https://huggingface.co/mradermacher/Qwen-SEA-LION-v4-4B-VL-GGUF/resolve/main/Qwen-SEA-LION-v4-4B-VL.Q3_K_S.gguf) | Q3_K_S | 2.0 | | +| [GGUF](https://huggingface.co/mradermacher/Qwen-SEA-LION-v4-4B-VL-GGUF/resolve/main/Qwen-SEA-LION-v4-4B-VL.Q3_K_M.gguf) | Q3_K_M | 2.2 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/Qwen-SEA-LION-v4-4B-VL-GGUF/resolve/main/Qwen-SEA-LION-v4-4B-VL.Q3_K_L.gguf) | Q3_K_L | 2.3 | | +| [GGUF](https://huggingface.co/mradermacher/Qwen-SEA-LION-v4-4B-VL-GGUF/resolve/main/Qwen-SEA-LION-v4-4B-VL.IQ4_XS.gguf) | IQ4_XS | 2.4 | | +| [GGUF](https://huggingface.co/mradermacher/Qwen-SEA-LION-v4-4B-VL-GGUF/resolve/main/Qwen-SEA-LION-v4-4B-VL.Q4_K_S.gguf) | Q4_K_S | 2.5 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Qwen-SEA-LION-v4-4B-VL-GGUF/resolve/main/Qwen-SEA-LION-v4-4B-VL.Q4_K_M.gguf) | Q4_K_M | 2.6 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Qwen-SEA-LION-v4-4B-VL-GGUF/resolve/main/Qwen-SEA-LION-v4-4B-VL.Q5_K_S.gguf) | Q5_K_S | 2.9 | | +| [GGUF](https://huggingface.co/mradermacher/Qwen-SEA-LION-v4-4B-VL-GGUF/resolve/main/Qwen-SEA-LION-v4-4B-VL.Q5_K_M.gguf) | Q5_K_M | 3.0 | | +| [GGUF](https://huggingface.co/mradermacher/Qwen-SEA-LION-v4-4B-VL-GGUF/resolve/main/Qwen-SEA-LION-v4-4B-VL.Q6_K.gguf) | Q6_K | 3.4 | very good quality | +| [GGUF](https://huggingface.co/mradermacher/Qwen-SEA-LION-v4-4B-VL-GGUF/resolve/main/Qwen-SEA-LION-v4-4B-VL.Q8_0.gguf) | Q8_0 | 4.4 | fast, best quality | +| [GGUF](https://huggingface.co/mradermacher/Qwen-SEA-LION-v4-4B-VL-GGUF/resolve/main/Qwen-SEA-LION-v4-4B-VL.f16.gguf) | f16 | 8.2 | 16 bpw, overkill | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. + +