commit ed8fc483e9fc43f6be0261141e647ec77a1a39ec Author: ModelHub XC Date: Fri May 8 19:29:03 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: mradermacher/icarus-1-8b-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..9af2836 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,47 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.f16.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..fab76e0 --- /dev/null +++ b/README.md @@ -0,0 +1,77 @@ +--- +base_model: APRKDEV/icarus-1-8b +language: +- en +library_name: transformers +license: mit +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +tags: +- neonaut-studio +- icarus-project +- deep-reasoning +- proprietary-architecture +--- +## About + + + + + + + + + +static quants of https://huggingface.co/APRKDEV/icarus-1-8b + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#icarus-1-8b-GGUF).*** + +weighted/imatrix quants are available at https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-GGUF/resolve/main/icarus-1-8b.Q2_K.gguf) | Q2_K | 3.3 | | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-GGUF/resolve/main/icarus-1-8b.Q3_K_S.gguf) | Q3_K_S | 3.8 | | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-GGUF/resolve/main/icarus-1-8b.Q3_K_M.gguf) | Q3_K_M | 4.1 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-GGUF/resolve/main/icarus-1-8b.Q3_K_L.gguf) | Q3_K_L | 4.4 | | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-GGUF/resolve/main/icarus-1-8b.IQ4_XS.gguf) | IQ4_XS | 4.6 | | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-GGUF/resolve/main/icarus-1-8b.Q4_K_S.gguf) | Q4_K_S | 4.8 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-GGUF/resolve/main/icarus-1-8b.Q4_K_M.gguf) | Q4_K_M | 5.0 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-GGUF/resolve/main/icarus-1-8b.Q5_K_S.gguf) | Q5_K_S | 5.7 | | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-GGUF/resolve/main/icarus-1-8b.Q5_K_M.gguf) | Q5_K_M | 5.8 | | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-GGUF/resolve/main/icarus-1-8b.Q6_K.gguf) | Q6_K | 6.7 | very good quality | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-GGUF/resolve/main/icarus-1-8b.Q8_0.gguf) | Q8_0 | 8.6 | fast, best quality | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-GGUF/resolve/main/icarus-1-8b.f16.gguf) | f16 | 16.2 | 16 bpw, overkill | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. + + diff --git a/icarus-1-8b.IQ4_XS.gguf b/icarus-1-8b.IQ4_XS.gguf new file mode 100644 index 0000000..b084d3f --- /dev/null +++ b/icarus-1-8b.IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b617efef832a203ec31dd7094b13415d5affc77a76f453efd53a217b2bb9d64 +size 4484363520 diff --git a/icarus-1-8b.Q2_K.gguf b/icarus-1-8b.Q2_K.gguf new file mode 100644 index 0000000..4bf0aeb --- /dev/null +++ b/icarus-1-8b.Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d54ba949d08f85d1c6ea4d00337f42e7284864709c4f19fae7474dcf633db36f +size 3179132160 diff --git a/icarus-1-8b.Q3_K_L.gguf b/icarus-1-8b.Q3_K_L.gguf new file mode 100644 index 0000000..e94ef74 --- /dev/null +++ b/icarus-1-8b.Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d5c4a626a457f79ff25001f062bfc3041ed60263226ffc3c084b48888ba266be +size 4321957120 diff --git a/icarus-1-8b.Q3_K_M.gguf b/icarus-1-8b.Q3_K_M.gguf new file mode 100644 index 0000000..b7c6676 --- /dev/null +++ b/icarus-1-8b.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d216cedbe95a7f3ff268edd5a53ce1cf27090231e5f68768886fa13282191caa +size 4018918656 diff --git a/icarus-1-8b.Q3_K_S.gguf b/icarus-1-8b.Q3_K_S.gguf new file mode 100644 index 0000000..d43512e --- /dev/null +++ b/icarus-1-8b.Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7c2b1406402b3a0e7760122219e803d63a79636c34716d1bcf0a3cf0637dbad0 +size 3664499968 diff --git a/icarus-1-8b.Q4_K_M.gguf b/icarus-1-8b.Q4_K_M.gguf new file mode 100644 index 0000000..129255c --- /dev/null +++ b/icarus-1-8b.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3aeff6f13cc0f8a88de647fd82f216532eec2b5c100a58e8fe55620a6975cd8d +size 4920734976 diff --git a/icarus-1-8b.Q4_K_S.gguf b/icarus-1-8b.Q4_K_S.gguf new file mode 100644 index 0000000..077b719 --- /dev/null +++ b/icarus-1-8b.Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b7a4ff89080d41a7122306f7aa059f0d088a17e4470d5b0b22d8acdb1f133e4 +size 4692669696 diff --git a/icarus-1-8b.Q5_K_M.gguf b/icarus-1-8b.Q5_K_M.gguf new file mode 100644 index 0000000..46c3574 --- /dev/null +++ b/icarus-1-8b.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5851954a96f4999197d36a3131f058850686bc71633a1115ffead7a634bcfdf4 +size 5732988160 diff --git a/icarus-1-8b.Q5_K_S.gguf b/icarus-1-8b.Q5_K_S.gguf new file mode 100644 index 0000000..21fab2a --- /dev/null +++ b/icarus-1-8b.Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:29107cb7fb138e309c2874849e6a69559faedcf6d2eb206285b4b8d6066f0eff +size 5599294720 diff --git a/icarus-1-8b.Q6_K.gguf b/icarus-1-8b.Q6_K.gguf new file mode 100644 index 0000000..9d44c10 --- /dev/null +++ b/icarus-1-8b.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2291149d2688cd357ed2360cd94fc0a9817fbac6834a3c3c6629a681644b8212 +size 6596007168 diff --git a/icarus-1-8b.Q8_0.gguf b/icarus-1-8b.Q8_0.gguf new file mode 100644 index 0000000..a499027 --- /dev/null +++ b/icarus-1-8b.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bdbd923e4f68c5d87ddfd81d06b66c18b4fdec7d1e60a48361f7d79570f3efa0 +size 8540771584 diff --git a/icarus-1-8b.f16.gguf b/icarus-1-8b.f16.gguf new file mode 100644 index 0000000..dadabf2 --- /dev/null +++ b/icarus-1-8b.f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:520c6c11f1b29014a67984a8d0c155d0334d6025db6c345ba24f340a3c73522f +size 16068891904