commit 92c3ad542b586c8d2893fa8c4195bb9b61547334 Author: ModelHub XC Date: Sun Apr 26 22:30:04 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: mradermacher/Gemma-3-4B-IT-TL-SynthDolly-1A-E3-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..42b089d --- /dev/null +++ b/.gitattributes @@ -0,0 +1,49 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +Gemma-3-4B-IT-TL-SynthDolly-1A-E3.mmproj-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +Gemma-3-4B-IT-TL-SynthDolly-1A-E3.mmproj-f16.gguf filter=lfs diff=lfs merge=lfs -text +Gemma-3-4B-IT-TL-SynthDolly-1A-E3.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +Gemma-3-4B-IT-TL-SynthDolly-1A-E3.f16.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.IQ4_XS.gguf b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.IQ4_XS.gguf new file mode 100644 index 0000000..94acfd9 --- /dev/null +++ b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:630f8524290e2e991052ac2023b9103fbfcdb25734e7a6564ec0e20c1e45513f +size 2279626528 diff --git a/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q2_K.gguf b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q2_K.gguf new file mode 100644 index 0000000..3853700 --- /dev/null +++ b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd53d8db393f6fa9ed62494ab79f4725e518165758e86a3343933e1d321f4d5d +size 1729165088 diff --git a/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q3_K_L.gguf b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q3_K_L.gguf new file mode 100644 index 0000000..556d337 --- /dev/null +++ b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:da3689b63550473d6f82ca57ffea2f8e1562dae1904f4ffaa1e78a2c3f483289 +size 2236086048 diff --git a/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q3_K_M.gguf b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q3_K_M.gguf new file mode 100644 index 0000000..105248c --- /dev/null +++ b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:167b6c1be03c57ed30af5a115a178cfe5087a43627a4e956d5fe8aebcaf0792c +size 2098460448 diff --git a/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q3_K_S.gguf b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q3_K_S.gguf new file mode 100644 index 0000000..faf4cb9 --- /dev/null +++ b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cc57360f9ef2675a6015e182af78f5faa8e81d7d9f2ce615bf1ca9b279783215 +size 1937364768 diff --git a/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q4_K_M.gguf b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q4_K_M.gguf new file mode 100644 index 0000000..036816b --- /dev/null +++ b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4dfa1fcd8cd6147424bb2b7c970e5848fa101338229c85520d461ffda9f37c61 +size 2489894688 diff --git a/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q4_K_S.gguf b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q4_K_S.gguf new file mode 100644 index 0000000..90b304f --- /dev/null +++ b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:31668b1b7a26a4b3f32cd5e05c6fd93f513ab0925159de2ba4cce9ff06b79d11 +size 2377930528 diff --git a/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q5_K_M.gguf b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q5_K_M.gguf new file mode 100644 index 0000000..cfdfe85 --- /dev/null +++ b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b329f51890cf5d712bbfd0d583e280eca0039a92f510a566861d967185fecc37 +size 2829698848 diff --git a/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q5_K_S.gguf b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q5_K_S.gguf new file mode 100644 index 0000000..d8b1c74 --- /dev/null +++ b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c956c091de60886c52a7e5fd9ddd23da223325d2972aca0c5f1e6c2d7af546dd +size 2764592928 diff --git a/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q6_K.gguf b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q6_K.gguf new file mode 100644 index 0000000..9433563 --- /dev/null +++ b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6127c6d55e774794a32402b0c9fa576a90749b1badf98e695c01325393d1cce6 +size 3190740768 diff --git a/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q8_0.gguf b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q8_0.gguf new file mode 100644 index 0000000..c03abca --- /dev/null +++ b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cc4209641b83e4a5fb75c336c22b5e1b53ceb02f0787737023ad6b939e2a9b67 +size 4130402848 diff --git a/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.f16.gguf b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.f16.gguf new file mode 100644 index 0000000..a22bdc8 --- /dev/null +++ b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:89988871a278644ef8dbc888cb60dabd032e214a1a9b297c52390f2ac81abe9a +size 7767804448 diff --git a/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.mmproj-Q8_0.gguf b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.mmproj-Q8_0.gguf new file mode 100644 index 0000000..cc7a434 --- /dev/null +++ b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.mmproj-Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b4e945c780acc1462cbc3093867bd352a8d40d23b54c3aa88c677fad3271d82 +size 591377888 diff --git a/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.mmproj-f16.gguf b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.mmproj-f16.gguf new file mode 100644 index 0000000..978ae94 --- /dev/null +++ b/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.mmproj-f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d66bef47cc36eb2da427aa9be8fd920345798b413522fabf055c2b57c550a17 +size 851251808 diff --git a/README.md b/README.md new file mode 100644 index 0000000..b82ba2e --- /dev/null +++ b/README.md @@ -0,0 +1,79 @@ +--- +base_model: kairawal/Gemma-3-4B-IT-TL-SynthDolly-1A-E3 +language: +- en +library_name: transformers +license: apache-2.0 +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +tags: +- text-generation-inference +- transformers +- unsloth +- gemma3 +--- +## About + + + + + + + + + +static quants of https://huggingface.co/kairawal/Gemma-3-4B-IT-TL-SynthDolly-1A-E3 + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#Gemma-3-4B-IT-TL-SynthDolly-1A-E3-GGUF).*** + +weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion. +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/Gemma-3-4B-IT-TL-SynthDolly-1A-E3-GGUF/resolve/main/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.mmproj-Q8_0.gguf) | mmproj-Q8_0 | 0.7 | multi-modal supplement | +| [GGUF](https://huggingface.co/mradermacher/Gemma-3-4B-IT-TL-SynthDolly-1A-E3-GGUF/resolve/main/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.mmproj-f16.gguf) | mmproj-f16 | 1.0 | multi-modal supplement | +| [GGUF](https://huggingface.co/mradermacher/Gemma-3-4B-IT-TL-SynthDolly-1A-E3-GGUF/resolve/main/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q2_K.gguf) | Q2_K | 1.8 | | +| [GGUF](https://huggingface.co/mradermacher/Gemma-3-4B-IT-TL-SynthDolly-1A-E3-GGUF/resolve/main/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q3_K_S.gguf) | Q3_K_S | 2.0 | | +| [GGUF](https://huggingface.co/mradermacher/Gemma-3-4B-IT-TL-SynthDolly-1A-E3-GGUF/resolve/main/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q3_K_M.gguf) | Q3_K_M | 2.2 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/Gemma-3-4B-IT-TL-SynthDolly-1A-E3-GGUF/resolve/main/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q3_K_L.gguf) | Q3_K_L | 2.3 | | +| [GGUF](https://huggingface.co/mradermacher/Gemma-3-4B-IT-TL-SynthDolly-1A-E3-GGUF/resolve/main/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.IQ4_XS.gguf) | IQ4_XS | 2.4 | | +| [GGUF](https://huggingface.co/mradermacher/Gemma-3-4B-IT-TL-SynthDolly-1A-E3-GGUF/resolve/main/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q4_K_S.gguf) | Q4_K_S | 2.5 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Gemma-3-4B-IT-TL-SynthDolly-1A-E3-GGUF/resolve/main/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q4_K_M.gguf) | Q4_K_M | 2.6 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/Gemma-3-4B-IT-TL-SynthDolly-1A-E3-GGUF/resolve/main/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q5_K_S.gguf) | Q5_K_S | 2.9 | | +| [GGUF](https://huggingface.co/mradermacher/Gemma-3-4B-IT-TL-SynthDolly-1A-E3-GGUF/resolve/main/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q5_K_M.gguf) | Q5_K_M | 2.9 | | +| [GGUF](https://huggingface.co/mradermacher/Gemma-3-4B-IT-TL-SynthDolly-1A-E3-GGUF/resolve/main/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q6_K.gguf) | Q6_K | 3.3 | very good quality | +| [GGUF](https://huggingface.co/mradermacher/Gemma-3-4B-IT-TL-SynthDolly-1A-E3-GGUF/resolve/main/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.Q8_0.gguf) | Q8_0 | 4.2 | fast, best quality | +| [GGUF](https://huggingface.co/mradermacher/Gemma-3-4B-IT-TL-SynthDolly-1A-E3-GGUF/resolve/main/Gemma-3-4B-IT-TL-SynthDolly-1A-E3.f16.gguf) | f16 | 7.9 | 16 bpw, overkill | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. + +