commit c69ff4cd702ec5a186fcacd8fc1c173bc8426a72 Author: ModelHub XC Date: Sat May 9 02:00:59 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: mradermacher/BarcenasMexico-270m-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..88cf700 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,47 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +BarcenasMexico-270m.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +BarcenasMexico-270m.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +BarcenasMexico-270m.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +BarcenasMexico-270m.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +BarcenasMexico-270m.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +BarcenasMexico-270m.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +BarcenasMexico-270m.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +BarcenasMexico-270m.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +BarcenasMexico-270m.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +BarcenasMexico-270m.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +BarcenasMexico-270m.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +BarcenasMexico-270m.f16.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/BarcenasMexico-270m.IQ4_XS.gguf b/BarcenasMexico-270m.IQ4_XS.gguf new file mode 100644 index 0000000..0ab5d49 --- /dev/null +++ b/BarcenasMexico-270m.IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:abb50612fc3373577b8902eacc5f0a90086451fa54ca26b16fb177c6adffafa1 +size 241268032 diff --git a/BarcenasMexico-270m.Q2_K.gguf b/BarcenasMexico-270m.Q2_K.gguf new file mode 100644 index 0000000..f2d4cf2 --- /dev/null +++ b/BarcenasMexico-270m.Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:30463311dbbf57fcfaf6ec4b4cd74d848748459c671557fe2d9037f30ac49ca6 +size 237079872 diff --git a/BarcenasMexico-270m.Q3_K_L.gguf b/BarcenasMexico-270m.Q3_K_L.gguf new file mode 100644 index 0000000..a4a7b81 --- /dev/null +++ b/BarcenasMexico-270m.Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:67937f74a4a75535b954771ad0240516b145b9bc090bd7250fc4d89fdf972a58 +size 246388032 diff --git a/BarcenasMexico-270m.Q3_K_M.gguf b/BarcenasMexico-270m.Q3_K_M.gguf new file mode 100644 index 0000000..d10b309 --- /dev/null +++ b/BarcenasMexico-270m.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e9ba5404346f258f298352b5f8ea8eff823b582fde1975f7107d199d4e82d056 +size 241964352 diff --git a/BarcenasMexico-270m.Q3_K_S.gguf b/BarcenasMexico-270m.Q3_K_S.gguf new file mode 100644 index 0000000..26c842e --- /dev/null +++ b/BarcenasMexico-270m.Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6fb4b6a12d322264c1f5ac4ac89041d6830e3af607f819b58e55c26c7076cc14 +size 236711232 diff --git a/BarcenasMexico-270m.Q4_K_M.gguf b/BarcenasMexico-270m.Q4_K_M.gguf new file mode 100644 index 0000000..d621c1c --- /dev/null +++ b/BarcenasMexico-270m.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d110b9915de4f42aa74e036fada1dc794e69e2a16895f9a4fa4db0d16bd49fde +size 253115712 diff --git a/BarcenasMexico-270m.Q4_K_S.gguf b/BarcenasMexico-270m.Q4_K_S.gguf new file mode 100644 index 0000000..30b6b89 --- /dev/null +++ b/BarcenasMexico-270m.Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b5aa86b227d1f60b1c1bf46b01f15c1c61519202e03f0006991d504fb40fe69b +size 249890112 diff --git a/BarcenasMexico-270m.Q5_K_M.gguf b/BarcenasMexico-270m.Q5_K_M.gguf new file mode 100644 index 0000000..0648c5c --- /dev/null +++ b/BarcenasMexico-270m.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:47fffc849246313c358679077999b03ed4d1a2c7d5ca20fd741fbc855274a6c9 +size 260027712 diff --git a/BarcenasMexico-270m.Q5_K_S.gguf b/BarcenasMexico-270m.Q5_K_S.gguf new file mode 100644 index 0000000..0e7c679 --- /dev/null +++ b/BarcenasMexico-270m.Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b21e25f0f39e50ab298dc533382f0a88b6b61d2c3ccd97fb58a2b7a3cf47f773 +size 258000192 diff --git a/BarcenasMexico-270m.Q6_K.gguf b/BarcenasMexico-270m.Q6_K.gguf new file mode 100644 index 0000000..f475040 --- /dev/null +++ b/BarcenasMexico-270m.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0f67f07f41d5b971ff17a7f8f3cfa3527ac497daee7588ca5f0c62c24201967 +size 282975552 diff --git a/BarcenasMexico-270m.Q8_0.gguf b/BarcenasMexico-270m.Q8_0.gguf new file mode 100644 index 0000000..3fb8dcc --- /dev/null +++ b/BarcenasMexico-270m.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9748c5a09518d48cd3098ae073ee54c5da269dd4459a488698f60095fb21fbba +size 291546432 diff --git a/BarcenasMexico-270m.f16.gguf b/BarcenasMexico-270m.f16.gguf new file mode 100644 index 0000000..540c5f5 --- /dev/null +++ b/BarcenasMexico-270m.f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab56bb40e78b8c142333d8ebe4b70fd4d54e9546057e1c4c2c7875096111c86b +size 542836032 diff --git a/README.md b/README.md new file mode 100644 index 0000000..5de73a8 --- /dev/null +++ b/README.md @@ -0,0 +1,74 @@ +--- +base_model: Danielbrdz/BarcenasMexico-270m +datasets: +- Danielbrdz/BarcenasMexico +language: +- es +library_name: transformers +license: gemma +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +--- +## About + + + + + + + + + +static quants of https://huggingface.co/Danielbrdz/BarcenasMexico-270m + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#BarcenasMexico-270m-GGUF).*** + +weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion. +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/BarcenasMexico-270m-GGUF/resolve/main/BarcenasMexico-270m.Q3_K_S.gguf) | Q3_K_S | 0.3 | | +| [GGUF](https://huggingface.co/mradermacher/BarcenasMexico-270m-GGUF/resolve/main/BarcenasMexico-270m.Q2_K.gguf) | Q2_K | 0.3 | | +| [GGUF](https://huggingface.co/mradermacher/BarcenasMexico-270m-GGUF/resolve/main/BarcenasMexico-270m.IQ4_XS.gguf) | IQ4_XS | 0.3 | | +| [GGUF](https://huggingface.co/mradermacher/BarcenasMexico-270m-GGUF/resolve/main/BarcenasMexico-270m.Q3_K_M.gguf) | Q3_K_M | 0.3 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/BarcenasMexico-270m-GGUF/resolve/main/BarcenasMexico-270m.Q3_K_L.gguf) | Q3_K_L | 0.3 | | +| [GGUF](https://huggingface.co/mradermacher/BarcenasMexico-270m-GGUF/resolve/main/BarcenasMexico-270m.Q4_K_S.gguf) | Q4_K_S | 0.3 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/BarcenasMexico-270m-GGUF/resolve/main/BarcenasMexico-270m.Q4_K_M.gguf) | Q4_K_M | 0.4 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/BarcenasMexico-270m-GGUF/resolve/main/BarcenasMexico-270m.Q5_K_S.gguf) | Q5_K_S | 0.4 | | +| [GGUF](https://huggingface.co/mradermacher/BarcenasMexico-270m-GGUF/resolve/main/BarcenasMexico-270m.Q5_K_M.gguf) | Q5_K_M | 0.4 | | +| [GGUF](https://huggingface.co/mradermacher/BarcenasMexico-270m-GGUF/resolve/main/BarcenasMexico-270m.Q6_K.gguf) | Q6_K | 0.4 | very good quality | +| [GGUF](https://huggingface.co/mradermacher/BarcenasMexico-270m-GGUF/resolve/main/BarcenasMexico-270m.Q8_0.gguf) | Q8_0 | 0.4 | fast, best quality | +| [GGUF](https://huggingface.co/mradermacher/BarcenasMexico-270m-GGUF/resolve/main/BarcenasMexico-270m.f16.gguf) | f16 | 0.6 | 16 bpw, overkill | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. + +