commit 3be20478caf7d9b376bd5078fe76c018b88ee6ed Author: ModelHub XC Date: Tue Apr 21 13:12:57 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: LetheanNetwork/LEM-Gemma3-1B-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..d62df81 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,47 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +LEM-Gemma3-1B-IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text +LEM-Gemma3-1B-IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text +LEM-Gemma3-1B-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +LEM-Gemma3-1B-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +LEM-Gemma3-1B-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +LEM-Gemma3-1B-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +LEM-Gemma3-1B-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +LEM-Gemma3-1B-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +LEM-Gemma3-1B-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +LEM-Gemma3-1B-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +LEM-Gemma3-1B-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +LEM-Gemma3-1B-bf16.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/LEM-Gemma3-1B-IQ3_XS.gguf b/LEM-Gemma3-1B-IQ3_XS.gguf new file mode 100644 index 0000000..e228906 --- /dev/null +++ b/LEM-Gemma3-1B-IQ3_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:834e8d83fb830b6b7d0ccad55f385f6585e634a80b93773b77626ebe16ddf258 +size 859683712 diff --git a/LEM-Gemma3-1B-IQ3_XXS.gguf b/LEM-Gemma3-1B-IQ3_XXS.gguf new file mode 100644 index 0000000..ffc5bbc --- /dev/null +++ b/LEM-Gemma3-1B-IQ3_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ef133ca0d7455733019ca750f2e735e53ee4666e9c40df447e459ab95f1f0ea5 +size 862856320 diff --git a/LEM-Gemma3-1B-IQ4_XS.gguf b/LEM-Gemma3-1B-IQ4_XS.gguf new file mode 100644 index 0000000..2d16788 --- /dev/null +++ b/LEM-Gemma3-1B-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2909987e5c0d7043f5c669cc9fc7885762213256e10f815eb116666ec2ed7d6d +size 888036736 diff --git a/LEM-Gemma3-1B-Q3_K_M.gguf b/LEM-Gemma3-1B-Q3_K_M.gguf new file mode 100644 index 0000000..ab5f941 --- /dev/null +++ b/LEM-Gemma3-1B-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:77b9328cfbafdbabe9ca401ac2ddf052a3efee7620035e2991f8a6a7761327f9 +size 892285312 diff --git a/LEM-Gemma3-1B-Q3_K_S.gguf b/LEM-Gemma3-1B-Q3_K_S.gguf new file mode 100644 index 0000000..16112e1 --- /dev/null +++ b/LEM-Gemma3-1B-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd7319b079ccb4b91daa8643b4af72d1e6168219bd256f802fe1ae1d6fa58395 +size 858725248 diff --git a/LEM-Gemma3-1B-Q4_K_M.gguf b/LEM-Gemma3-1B-Q4_K_M.gguf new file mode 100644 index 0000000..99e4f6c --- /dev/null +++ b/LEM-Gemma3-1B-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b485b8bfc25c53ac16b648a04a46f4c32888c7e5aac1794d181be9f144ccdb4 +size 1013676160 diff --git a/LEM-Gemma3-1B-Q4_K_S.gguf b/LEM-Gemma3-1B-Q4_K_S.gguf new file mode 100644 index 0000000..e723ff5 --- /dev/null +++ b/LEM-Gemma3-1B-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e35c70976b66bb490e3e25cb828de500facd6ceddbbf1927273d0b1511368a3e +size 988610944 diff --git a/LEM-Gemma3-1B-Q5_K_M.gguf b/LEM-Gemma3-1B-Q5_K_M.gguf new file mode 100644 index 0000000..6f37d0d --- /dev/null +++ b/LEM-Gemma3-1B-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0f4989aa2c6b932a4df42dc934a3d93dbb70fbd069a52bdff6a37264a0324e90 +size 1077837952 diff --git a/LEM-Gemma3-1B-Q5_K_S.gguf b/LEM-Gemma3-1B-Q5_K_S.gguf new file mode 100644 index 0000000..496d459 --- /dev/null +++ b/LEM-Gemma3-1B-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7096b16104bfb4df4df03e64469d77b56991bb964f23f57705e6cc63a3803c64 +size 1062891904 diff --git a/LEM-Gemma3-1B-Q6_K.gguf b/LEM-Gemma3-1B-Q6_K.gguf new file mode 100644 index 0000000..bd32aa1 --- /dev/null +++ b/LEM-Gemma3-1B-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:605f8827fa171f4dcb090d6a4cb227b95a42c3de6990ecb135dc103739883bfe +size 1332602752 diff --git a/LEM-Gemma3-1B-Q8_0.gguf b/LEM-Gemma3-1B-Q8_0.gguf new file mode 100644 index 0000000..390202f --- /dev/null +++ b/LEM-Gemma3-1B-Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a8782f80fa4e7d99faa39c1fe7affbcab7788d467e108153a503006b444486a +size 1390170496 diff --git a/LEM-Gemma3-1B-bf16.gguf b/LEM-Gemma3-1B-bf16.gguf new file mode 100644 index 0000000..ebf5dbe --- /dev/null +++ b/LEM-Gemma3-1B-bf16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0ca8c3ae9ef9c96ac41b6d7179bd972fe87fe8c700e2f23f97eab04e0990d4fe +size 2610553216 diff --git a/README.md b/README.md new file mode 100644 index 0000000..2181ccd --- /dev/null +++ b/README.md @@ -0,0 +1,101 @@ +--- +language: en +tags: +- gguf +- lem +- ethics +- alignment +- cymatic-linguistic-bpl +- rocm +- llama-cpp +- gemma3 +library_name: gguf +pipeline_tag: text-generation +base_model: google/gemma-3-1b-it +license: other +license_name: eupl-1.2 +license_link: https://joinup.ec.europa.eu/licence/european-union-public-licence-v-12 +--- + +# LEM-Gemma3-1B-GGUF + +GGUF quantisations of [LEM-Gemma3-1B](https://huggingface.co/lthn/LEM-Gemma3-1B) — the foundation teacher model of the CL-BPL cascade. Ethics are in the weights, not in a system prompt. + +The 4B model trained on this 1B's distilled responses achieved **25th in the world for Instruction Following** on LiveBench. + +> [LEM-Gemma3-1B (safetensors)](https://huggingface.co/lthn/LEM-Gemma3-1B) | [Collection](https://huggingface.co/collections/lthn/lethean-ethical-models-lem-699e863449120d22596f739c) | [Research Paper](https://huggingface.co/datasets/lthn/LEM-research) | [Benchmarks](https://huggingface.co/datasets/lthn/LEM-benchmarks) + +--- + +## Quick Start + +No system prompt needed. Ethics hold from weights alone. + +```bash +# GPU offload (CUDA, ROCm, Metal) +llama-server -m LEM-Gemma3-1B-Q4_K_M.gguf -ngl 99 --port 8080 + +# CPU — fast enough for 1B +llama-server -m LEM-Gemma3-1B-Q4_K_M.gguf -ngl 0 --port 8080 + +# OpenAI-compatible API +curl http://localhost:8080/v1/chat/completions \ + -d '{"model":"LEM-Gemma3-1B","messages":[{"role":"user","content":"What is kindness?"}]}' +``` + +--- + +## Quantisations + +All quantised from the BF16 source using llama.cpp. + +| Bits | Quant | Size | Notes | +|------|-------|------|-------| +| 3-bit | IQ3_XXS | 823 MB | Smallest usable (imatrix) | +| 3-bit | IQ3_XS | 820 MB | (imatrix) | +| 3-bit | Q3_K_S | 819 MB | | +| 3-bit | Q3_K_M | 851 MB | | +| 4-bit | IQ4_XS | 847 MB | (imatrix) | +| 4-bit | Q4_K_S | 943 MB | | +| **4-bit** | **Q4_K_M** | **967 MB** | **Recommended — best quality/size balance** | +| 5-bit | Q5_K_S | 1.0 GB | | +| 5-bit | Q5_K_M | 1.0 GB | Near-lossless | +| 6-bit | Q6_K | 1.2 GB | | +| 8-bit | Q8_0 | 1.3 GB | Virtually lossless | +| 16-bit | BF16 | 2.4 GB | Full precision | + +--- + +## About LEM-Gemma3-1B + +The 1B is trained first and hardest — its alignment must be pristine because every larger model inherits from it. CL-BPL uses the 1B's constrained latent space as an advantage: with fewer parameters, there are fewer places for sycophancy to hide. + +``` +LEM-Gemma3-1B (this model — foundation teacher) + -> LEM-Gemma3-4B (25th IF on LiveBench) + -> LEM-Gemma3-12B (next) + -> LEM-Gemma3-27B (planned) +``` + +Built on Google Gemma3-1B-IT through the Ethics-Composure-Ethics sandwich structure (700 iterations across 3 phases). Full training details in the [main model card](https://huggingface.co/lthn/LEM-Gemma3-1B). + +## Other Formats + +| Format | Repo | +|--------|------| +| FP16 safetensors (Transformers, vLLM) | [lthn/LEM-Gemma3-1B](https://huggingface.co/lthn/LEM-Gemma3-1B) | + +## Licence + +[European Union Public Licence v1.2](https://joinup.ec.europa.eu/licence/european-union-public-licence-v-12) (EUPL-1.2). Base model subject to Google's Gemma licence terms. + +## Citation + +```bibtex +@misc{lem-gemma3-1b-2026, + title={LEM-Gemma3-1B: Foundation Teacher for Cymatic-Linguistic Back-Propagation}, + author={Lethean Project}, + year={2026}, + url={https://huggingface.co/lthn/LEM-Gemma3-1B} +} +```