commit 7c31c15e97bb7931e30cfa2d1227b2d7d2eb2e4e Author: ModelHub XC Date: Wed May 6 14:05:46 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: mradermacher/EverythingLM-13b-V2-16k-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..e80f2d3 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,46 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +EverythingLM-13b-V2-16k.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +EverythingLM-13b-V2-16k.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +EverythingLM-13b-V2-16k.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +EverythingLM-13b-V2-16k.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +EverythingLM-13b-V2-16k.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +EverythingLM-13b-V2-16k.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +EverythingLM-13b-V2-16k.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +EverythingLM-13b-V2-16k.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +EverythingLM-13b-V2-16k.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +EverythingLM-13b-V2-16k.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +EverythingLM-13b-V2-16k.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/EverythingLM-13b-V2-16k.IQ4_XS.gguf b/EverythingLM-13b-V2-16k.IQ4_XS.gguf new file mode 100644 index 0000000..4f5915d --- /dev/null +++ b/EverythingLM-13b-V2-16k.IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9f90e3470b5bcc32c975e6379bb2b75803beeca7128004aca827dae354ef6b2a +size 7019519200 diff --git a/EverythingLM-13b-V2-16k.Q2_K.gguf b/EverythingLM-13b-V2-16k.Q2_K.gguf new file mode 100644 index 0000000..2ebae1f --- /dev/null +++ b/EverythingLM-13b-V2-16k.Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4d054c20817de8ddb2d1dd5560d4b8674a3c1252abdd16d153bdae04dca43499 +size 4854271200 diff --git a/EverythingLM-13b-V2-16k.Q3_K_L.gguf b/EverythingLM-13b-V2-16k.Q3_K_L.gguf new file mode 100644 index 0000000..09c8ef7 --- /dev/null +++ b/EverythingLM-13b-V2-16k.Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:85054f2d3dbd100d4132ccd0fcf2bd44dd8d8eb11196f62057698b845843917f +size 6929560800 diff --git a/EverythingLM-13b-V2-16k.Q3_K_M.gguf b/EverythingLM-13b-V2-16k.Q3_K_M.gguf new file mode 100644 index 0000000..0eea994 --- /dev/null +++ b/EverythingLM-13b-V2-16k.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab6cd0c7e6905a4ce37b64fb17a49d10873c89a66cae103e41bfe61a89c5f7f4 +size 6337770720 diff --git a/EverythingLM-13b-V2-16k.Q3_K_S.gguf b/EverythingLM-13b-V2-16k.Q3_K_S.gguf new file mode 100644 index 0000000..b29e85b --- /dev/null +++ b/EverythingLM-13b-V2-16k.Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b030c0d8b7f923ec1662a09a099d08f0296fbb19925a6d2d15efee011bbc60ac +size 5658981600 diff --git a/EverythingLM-13b-V2-16k.Q4_K_M.gguf b/EverythingLM-13b-V2-16k.Q4_K_M.gguf new file mode 100644 index 0000000..42b3097 --- /dev/null +++ b/EverythingLM-13b-V2-16k.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9d6d5e417c125031fed699bc8fbb34688d3ddab0c1839325ea32ee449dd322dc +size 7865957600 diff --git a/EverythingLM-13b-V2-16k.Q4_K_S.gguf b/EverythingLM-13b-V2-16k.Q4_K_S.gguf new file mode 100644 index 0000000..d1c13c9 --- /dev/null +++ b/EverythingLM-13b-V2-16k.Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:560addf95f01ba56ef6b21f4159a0b4d7e8045fd5eb4de311066f7ff0efc1ed1 +size 7423180000 diff --git a/EverythingLM-13b-V2-16k.Q5_K_M.gguf b/EverythingLM-13b-V2-16k.Q5_K_M.gguf new file mode 100644 index 0000000..4422532 --- /dev/null +++ b/EverythingLM-13b-V2-16k.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:de9da82452470f4826b71d4b60fcc83e61045c5fb52500b8b8a303dfe48202f2 +size 9229925600 diff --git a/EverythingLM-13b-V2-16k.Q5_K_S.gguf b/EverythingLM-13b-V2-16k.Q5_K_S.gguf new file mode 100644 index 0000000..e10e68e --- /dev/null +++ b/EverythingLM-13b-V2-16k.Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2798623527a76ef942e65f202a436ca8468728ed4a2f2127f9e9d0cb1e650bc2 +size 8972287200 diff --git a/EverythingLM-13b-V2-16k.Q6_K.gguf b/EverythingLM-13b-V2-16k.Q6_K.gguf new file mode 100644 index 0000000..98abea7 --- /dev/null +++ b/EverythingLM-13b-V2-16k.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a520be52ffce79137d2ae7c88a8f9bce046773c2ef6f74ab12326d8093d75f4f +size 10679141600 diff --git a/EverythingLM-13b-V2-16k.Q8_0.gguf b/EverythingLM-13b-V2-16k.Q8_0.gguf new file mode 100644 index 0000000..c250926 --- /dev/null +++ b/EverythingLM-13b-V2-16k.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4beb0c7737160f2efeadf397e6469b728d6c8b7ecb3ddecaafc7037d83e5c372 +size 13831320800 diff --git a/README.md b/README.md new file mode 100644 index 0000000..67f2f16 --- /dev/null +++ b/README.md @@ -0,0 +1,70 @@ +--- +base_model: totally-not-an-llm/EverythingLM-13b-V2-16k +datasets: +- totally-not-an-llm/EverythingLM-data-V2 +language: +- en +library_name: transformers +license: llama2 +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +--- +## About + + + + + + +static quants of https://huggingface.co/totally-not-an-llm/EverythingLM-13b-V2-16k + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#EverythingLM-13b-V2-16k-GGUF).*** + +weighted/imatrix quants are available at https://huggingface.co/mradermacher/EverythingLM-13b-V2-16k-i1-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/EverythingLM-13b-V2-16k-GGUF/resolve/main/EverythingLM-13b-V2-16k.Q2_K.gguf) | Q2_K | 5.0 | | +| [GGUF](https://huggingface.co/mradermacher/EverythingLM-13b-V2-16k-GGUF/resolve/main/EverythingLM-13b-V2-16k.Q3_K_S.gguf) | Q3_K_S | 5.8 | | +| [GGUF](https://huggingface.co/mradermacher/EverythingLM-13b-V2-16k-GGUF/resolve/main/EverythingLM-13b-V2-16k.Q3_K_M.gguf) | Q3_K_M | 6.4 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/EverythingLM-13b-V2-16k-GGUF/resolve/main/EverythingLM-13b-V2-16k.Q3_K_L.gguf) | Q3_K_L | 7.0 | | +| [GGUF](https://huggingface.co/mradermacher/EverythingLM-13b-V2-16k-GGUF/resolve/main/EverythingLM-13b-V2-16k.IQ4_XS.gguf) | IQ4_XS | 7.1 | | +| [GGUF](https://huggingface.co/mradermacher/EverythingLM-13b-V2-16k-GGUF/resolve/main/EverythingLM-13b-V2-16k.Q4_K_S.gguf) | Q4_K_S | 7.5 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/EverythingLM-13b-V2-16k-GGUF/resolve/main/EverythingLM-13b-V2-16k.Q4_K_M.gguf) | Q4_K_M | 8.0 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/EverythingLM-13b-V2-16k-GGUF/resolve/main/EverythingLM-13b-V2-16k.Q5_K_S.gguf) | Q5_K_S | 9.1 | | +| [GGUF](https://huggingface.co/mradermacher/EverythingLM-13b-V2-16k-GGUF/resolve/main/EverythingLM-13b-V2-16k.Q5_K_M.gguf) | Q5_K_M | 9.3 | | +| [GGUF](https://huggingface.co/mradermacher/EverythingLM-13b-V2-16k-GGUF/resolve/main/EverythingLM-13b-V2-16k.Q6_K.gguf) | Q6_K | 10.8 | very good quality | +| [GGUF](https://huggingface.co/mradermacher/EverythingLM-13b-V2-16k-GGUF/resolve/main/EverythingLM-13b-V2-16k.Q8_0.gguf) | Q8_0 | 13.9 | fast, best quality | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. + +