commit f80e66fc48b2714cd379cd09395e873280daeb2d Author: ModelHub XC Date: Thu May 7 01:08:40 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: mradermacher/Viel-v2-i1-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..8ad66a6 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,60 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +imatrix.dat filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-IQ4_NL.gguf filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-IQ2_M.gguf filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-Q2_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-IQ2_XXS.gguf filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-IQ2_XS.gguf filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-IQ2_S.gguf filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-Q4_1.gguf filter=lfs diff=lfs merge=lfs -text +Viel-v2.i1-IQ3_S.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..6986ba1 --- /dev/null +++ b/README.md @@ -0,0 +1,78 @@ +--- +base_model: ArsParadox/Viel-v2 +datasets: +- ArsParadox/Viel_Dataset_Lite +language: +- en +library_name: transformers +license: mit +quantized_by: mradermacher +--- +## About + + + + + + +weighted/imatrix quants of https://huggingface.co/ArsParadox/Viel-v2 + + +static quants are available at https://huggingface.co/mradermacher/Viel-v2-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-IQ1_S.gguf) | i1-IQ1_S | 2.1 | for the desperate | +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-IQ1_M.gguf) | i1-IQ1_M | 2.3 | mostly desperate | +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 2.5 | | +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-IQ2_XS.gguf) | i1-IQ2_XS | 2.7 | | +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-IQ2_S.gguf) | i1-IQ2_S | 2.9 | | +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-IQ2_M.gguf) | i1-IQ2_M | 3.0 | | +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-Q2_K_S.gguf) | i1-Q2_K_S | 3.1 | very low quality | +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-Q2_K.gguf) | i1-Q2_K | 3.3 | IQ3_XXS probably better | +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 3.4 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-IQ3_XS.gguf) | i1-IQ3_XS | 3.6 | | +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-Q3_K_S.gguf) | i1-Q3_K_S | 3.8 | IQ3_XS probably better | +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-IQ3_S.gguf) | i1-IQ3_S | 3.8 | beats Q3_K* | +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-IQ3_M.gguf) | i1-IQ3_M | 3.9 | | +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-Q3_K_M.gguf) | i1-Q3_K_M | 4.1 | IQ3_S probably better | +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-Q3_K_L.gguf) | i1-Q3_K_L | 4.4 | IQ3_M probably better | +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-IQ4_XS.gguf) | i1-IQ4_XS | 4.5 | | +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-Q4_0.gguf) | i1-Q4_0 | 4.8 | fast, low quality | +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-IQ4_NL.gguf) | i1-IQ4_NL | 4.8 | prefer IQ4_XS | +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-Q4_K_S.gguf) | i1-Q4_K_S | 4.8 | optimal size/speed/quality | +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-Q4_K_M.gguf) | i1-Q4_K_M | 5.0 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-Q4_1.gguf) | i1-Q4_1 | 5.2 | | +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-Q5_K_S.gguf) | i1-Q5_K_S | 5.7 | | +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-Q5_K_M.gguf) | i1-Q5_K_M | 5.8 | | +| [GGUF](https://huggingface.co/mradermacher/Viel-v2-i1-GGUF/resolve/main/Viel-v2.i1-Q6_K.gguf) | i1-Q6_K | 6.7 | practically like static Q6_K | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to. + + diff --git a/Viel-v2.i1-IQ1_M.gguf b/Viel-v2.i1-IQ1_M.gguf new file mode 100644 index 0000000..c1e0bea --- /dev/null +++ b/Viel-v2.i1-IQ1_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b2766dd048f8a423c7f306dafc9c293ad3f2ad79901db60ffb38023608f047db +size 2161972704 diff --git a/Viel-v2.i1-IQ1_S.gguf b/Viel-v2.i1-IQ1_S.gguf new file mode 100644 index 0000000..f0a593a --- /dev/null +++ b/Viel-v2.i1-IQ1_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:01fb4f322306056fcd73c54a12610ec87abdf7edf21763740154cc528c5b59d9 +size 2019628512 diff --git a/Viel-v2.i1-IQ2_M.gguf b/Viel-v2.i1-IQ2_M.gguf new file mode 100644 index 0000000..e004e10 --- /dev/null +++ b/Viel-v2.i1-IQ2_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:115e342dc94f4ae9527011ef1ef3cd26508996c91651f6808278b4c8d76ff778 +size 2948281824 diff --git a/Viel-v2.i1-IQ2_S.gguf b/Viel-v2.i1-IQ2_S.gguf new file mode 100644 index 0000000..332f517 --- /dev/null +++ b/Viel-v2.i1-IQ2_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:43b7f717f882c71e29a7c7a05952519a96a1bf5729358f3cabfb51cb55c20ffd +size 2758489568 diff --git a/Viel-v2.i1-IQ2_XS.gguf b/Viel-v2.i1-IQ2_XS.gguf new file mode 100644 index 0000000..f9ff158 --- /dev/null +++ b/Viel-v2.i1-IQ2_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8fc974ebbc6c38e39f38bb9bf94c0b66734e1d13450c43fab3eb3f5c0c61d30c +size 2605782496 diff --git a/Viel-v2.i1-IQ2_XXS.gguf b/Viel-v2.i1-IQ2_XXS.gguf new file mode 100644 index 0000000..6367ee4 --- /dev/null +++ b/Viel-v2.i1-IQ2_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b5024ebccdeee683d7b54b4dd6757789f26a9483d58929ed949e4ad2368ce1b4 +size 2399213024 diff --git a/Viel-v2.i1-IQ3_M.gguf b/Viel-v2.i1-IQ3_M.gguf new file mode 100644 index 0000000..9b7977f --- /dev/null +++ b/Viel-v2.i1-IQ3_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc68c1bb2d9a22ab52c1a03ab5de2739ddc7cc623a00bb720c8159f4b456aa51 +size 3784824288 diff --git a/Viel-v2.i1-IQ3_S.gguf b/Viel-v2.i1-IQ3_S.gguf new file mode 100644 index 0000000..3c5cbc1 --- /dev/null +++ b/Viel-v2.i1-IQ3_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2cc81b1b745a0ae3be55441e41964c520e8f0e84f033ccb9cb994a656bce48ca +size 3682325984 diff --git a/Viel-v2.i1-IQ3_XS.gguf b/Viel-v2.i1-IQ3_XS.gguf new file mode 100644 index 0000000..2288c83 --- /dev/null +++ b/Viel-v2.i1-IQ3_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1c4b44b40e994c751b6d252e63f93eedbb5fafb62657dbaded80678cd93d3e80 +size 3518748128 diff --git a/Viel-v2.i1-IQ3_XXS.gguf b/Viel-v2.i1-IQ3_XXS.gguf new file mode 100644 index 0000000..8ddb579 --- /dev/null +++ b/Viel-v2.i1-IQ3_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0640a053b8f43bea10fc86cfc5ecd0b5bc01b3954ac7dbe641f05cdab6e83dc4 +size 3274913248 diff --git a/Viel-v2.i1-IQ4_NL.gguf b/Viel-v2.i1-IQ4_NL.gguf new file mode 100644 index 0000000..f50e267 --- /dev/null +++ b/Viel-v2.i1-IQ4_NL.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c6f03d6ede476e6b28f9bc36a274b09c539831745e8213c936e9b2e4acc9555 +size 4677989856 diff --git a/Viel-v2.i1-IQ4_XS.gguf b/Viel-v2.i1-IQ4_XS.gguf new file mode 100644 index 0000000..c4c8dd7 --- /dev/null +++ b/Viel-v2.i1-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:41b66d8dcc14ad3a63832eaa0f364d016e4c8f5720e7f998fbe3b6d91216c0bc +size 4447663584 diff --git a/Viel-v2.i1-Q2_K.gguf b/Viel-v2.i1-Q2_K.gguf new file mode 100644 index 0000000..c9dfbe6 --- /dev/null +++ b/Viel-v2.i1-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6fd22025a259d9d7e8590e4e3d209e75a903d8b61f13c3654c08e87dc03e38ab +size 3179132384 diff --git a/Viel-v2.i1-Q2_K_S.gguf b/Viel-v2.i1-Q2_K_S.gguf new file mode 100644 index 0000000..948c5d3 --- /dev/null +++ b/Viel-v2.i1-Q2_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fdc7b46131691df695bca630566d23c49f6c887ea5bd8a777eb74a6d20c4daaa +size 2988815840 diff --git a/Viel-v2.i1-Q3_K_L.gguf b/Viel-v2.i1-Q3_K_L.gguf new file mode 100644 index 0000000..574dea2 --- /dev/null +++ b/Viel-v2.i1-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7962783af3d7118fd258e7bb1649c3c51610a38de4621e9f9f9e758d56f440f2 +size 4321957344 diff --git a/Viel-v2.i1-Q3_K_M.gguf b/Viel-v2.i1-Q3_K_M.gguf new file mode 100644 index 0000000..2a1c2e4 --- /dev/null +++ b/Viel-v2.i1-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:236663ce5d0056a1883aeb6afe8c6a0aaeaf5b48c5d45cef18c29f66719c9ea6 +size 4018918880 diff --git a/Viel-v2.i1-Q3_K_S.gguf b/Viel-v2.i1-Q3_K_S.gguf new file mode 100644 index 0000000..98cba2b --- /dev/null +++ b/Viel-v2.i1-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aad79c15abd12033a4542e6e441869c7e1b4f3cee8969d4a0e0ef2d9345c0d51 +size 3664500192 diff --git a/Viel-v2.i1-Q4_0.gguf b/Viel-v2.i1-Q4_0.gguf new file mode 100644 index 0000000..5f42f1d --- /dev/null +++ b/Viel-v2.i1-Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:214a1ee39ab20ba262d4e28603d53b5849f3b459d0522e1b5d58f79e4f57176b +size 4675892704 diff --git a/Viel-v2.i1-Q4_1.gguf b/Viel-v2.i1-Q4_1.gguf new file mode 100644 index 0000000..5991f29 --- /dev/null +++ b/Viel-v2.i1-Q4_1.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9781d6f8de3ff8d945c96c03ddd9f086d1bdeffb62b717b3ba0d0ca11516b3ec +size 5130253792 diff --git a/Viel-v2.i1-Q4_K_M.gguf b/Viel-v2.i1-Q4_K_M.gguf new file mode 100644 index 0000000..84f5594 --- /dev/null +++ b/Viel-v2.i1-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68bbcd5adf1cc740f3a28e3ef518bf8bad1c258b863911b040057e49b7bba4ad +size 4920735200 diff --git a/Viel-v2.i1-Q4_K_S.gguf b/Viel-v2.i1-Q4_K_S.gguf new file mode 100644 index 0000000..824c3fe --- /dev/null +++ b/Viel-v2.i1-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fea6cd88ccc84f862e31115b1d2b4a0fb22c17dcee5ac42e3403a9b3af451fd3 +size 4692669920 diff --git a/Viel-v2.i1-Q5_K_M.gguf b/Viel-v2.i1-Q5_K_M.gguf new file mode 100644 index 0000000..5701220 --- /dev/null +++ b/Viel-v2.i1-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cfe62405c85c0f11e31a13810042ecf07134a1f277d6a342d8b5620aa3b77bf1 +size 5732988384 diff --git a/Viel-v2.i1-Q5_K_S.gguf b/Viel-v2.i1-Q5_K_S.gguf new file mode 100644 index 0000000..7ac4d0b --- /dev/null +++ b/Viel-v2.i1-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9bf388b7a3e9b5c53bc7723b14708462f36151ab0a3b32cc559f45655fa332e3 +size 5599294944 diff --git a/Viel-v2.i1-Q6_K.gguf b/Viel-v2.i1-Q6_K.gguf new file mode 100644 index 0000000..f26c39b --- /dev/null +++ b/Viel-v2.i1-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9ff96f6baaaee7689c6ad28fe31a0c3e6751376dad6034a86beaa724ebdbfd96 +size 6596007392 diff --git a/imatrix.dat b/imatrix.dat new file mode 100644 index 0000000..c724a85 --- /dev/null +++ b/imatrix.dat @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a9c511342a4fcc32cd99a96d888d65a4d662096845e1929ed1d92d2e3e04b6b7 +size 4988157