commit eeb1b9fb309dd6023912361dbc98fd33edbbbb0d Author: ModelHub XC Date: Thu May 7 17:06:03 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: mradermacher/icarus-1-8b-i1-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..8bc027d --- /dev/null +++ b/.gitattributes @@ -0,0 +1,60 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.imatrix.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-IQ4_NL.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-IQ2_M.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-Q2_K_S.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-IQ2_XXS.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-IQ2_XS.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-IQ2_S.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-Q4_1.gguf filter=lfs diff=lfs merge=lfs -text +icarus-1-8b.i1-IQ3_S.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..bfdd0f1 --- /dev/null +++ b/README.md @@ -0,0 +1,90 @@ +--- +base_model: APRKDEV/icarus-1-8b +language: +- en +library_name: transformers +license: mit +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +tags: +- neonaut-studio +- icarus-project +- deep-reasoning +- proprietary-architecture +--- +## About + + + + + + + + + +weighted/imatrix quants of https://huggingface.co/APRKDEV/icarus-1-8b + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#icarus-1-8b-i1-GGUF).*** + +static quants are available at https://huggingface.co/mradermacher/icarus-1-8b-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.imatrix.gguf) | imatrix | 0.1 | imatrix file (for creating your own quants) | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-IQ1_S.gguf) | i1-IQ1_S | 2.1 | for the desperate | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-IQ1_M.gguf) | i1-IQ1_M | 2.3 | mostly desperate | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 2.5 | | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-IQ2_XS.gguf) | i1-IQ2_XS | 2.7 | | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-IQ2_S.gguf) | i1-IQ2_S | 2.9 | | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-IQ2_M.gguf) | i1-IQ2_M | 3.0 | | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-Q2_K_S.gguf) | i1-Q2_K_S | 3.1 | very low quality | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-Q2_K.gguf) | i1-Q2_K | 3.3 | IQ3_XXS probably better | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 3.4 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-IQ3_XS.gguf) | i1-IQ3_XS | 3.6 | | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-Q3_K_S.gguf) | i1-Q3_K_S | 3.8 | IQ3_XS probably better | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-IQ3_S.gguf) | i1-IQ3_S | 3.8 | beats Q3_K* | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-IQ3_M.gguf) | i1-IQ3_M | 3.9 | | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-Q3_K_M.gguf) | i1-Q3_K_M | 4.1 | IQ3_S probably better | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-Q3_K_L.gguf) | i1-Q3_K_L | 4.4 | IQ3_M probably better | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-IQ4_XS.gguf) | i1-IQ4_XS | 4.5 | | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-Q4_0.gguf) | i1-Q4_0 | 4.8 | fast, low quality | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-IQ4_NL.gguf) | i1-IQ4_NL | 4.8 | prefer IQ4_XS | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-Q4_K_S.gguf) | i1-Q4_K_S | 4.8 | optimal size/speed/quality | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-Q4_K_M.gguf) | i1-Q4_K_M | 5.0 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-Q4_1.gguf) | i1-Q4_1 | 5.2 | | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-Q5_K_S.gguf) | i1-Q5_K_S | 5.7 | | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-Q5_K_M.gguf) | i1-Q5_K_M | 5.8 | | +| [GGUF](https://huggingface.co/mradermacher/icarus-1-8b-i1-GGUF/resolve/main/icarus-1-8b.i1-Q6_K.gguf) | i1-Q6_K | 6.7 | practically like static Q6_K | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to. + + diff --git a/icarus-1-8b.i1-IQ1_M.gguf b/icarus-1-8b.i1-IQ1_M.gguf new file mode 100644 index 0000000..ee75fc4 --- /dev/null +++ b/icarus-1-8b.i1-IQ1_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:97d8e07481b94e24a2352bd6da5c4075ebc7699840bc44c4322e601214e30dc5 +size 2161972736 diff --git a/icarus-1-8b.i1-IQ1_S.gguf b/icarus-1-8b.i1-IQ1_S.gguf new file mode 100644 index 0000000..519f551 --- /dev/null +++ b/icarus-1-8b.i1-IQ1_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a857754ab04b8984edd9baf361e60ba90fd23596aece75aed78764b89abfb60b +size 2019628544 diff --git a/icarus-1-8b.i1-IQ2_M.gguf b/icarus-1-8b.i1-IQ2_M.gguf new file mode 100644 index 0000000..ed576bc --- /dev/null +++ b/icarus-1-8b.i1-IQ2_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ac9432cd8445ef8d89f2854328df2b2adaf91d657f468c02d0904078b823832c +size 2948281856 diff --git a/icarus-1-8b.i1-IQ2_S.gguf b/icarus-1-8b.i1-IQ2_S.gguf new file mode 100644 index 0000000..950d85b --- /dev/null +++ b/icarus-1-8b.i1-IQ2_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1b52f012588b479dfe2154144a86c5588aed5c373d4fcab02564ef50ea3ec63c +size 2758489600 diff --git a/icarus-1-8b.i1-IQ2_XS.gguf b/icarus-1-8b.i1-IQ2_XS.gguf new file mode 100644 index 0000000..19a62d7 --- /dev/null +++ b/icarus-1-8b.i1-IQ2_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d93ca10c88b4b73e2bb93fd63e4db37fbf785f29ac2b37bb91481f9841abf1e0 +size 2605782528 diff --git a/icarus-1-8b.i1-IQ2_XXS.gguf b/icarus-1-8b.i1-IQ2_XXS.gguf new file mode 100644 index 0000000..e4891b9 --- /dev/null +++ b/icarus-1-8b.i1-IQ2_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:79c0e08835144cc2be2eedc9e5b0c1344029aeee11d5e97c007e09730b282025 +size 2399213056 diff --git a/icarus-1-8b.i1-IQ3_M.gguf b/icarus-1-8b.i1-IQ3_M.gguf new file mode 100644 index 0000000..22adbf7 --- /dev/null +++ b/icarus-1-8b.i1-IQ3_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a31f031490db26455104197ccc640e1c0fd451e74d3e9bf47b448a6bf61000e +size 3784824320 diff --git a/icarus-1-8b.i1-IQ3_S.gguf b/icarus-1-8b.i1-IQ3_S.gguf new file mode 100644 index 0000000..3aafb02 --- /dev/null +++ b/icarus-1-8b.i1-IQ3_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:909ec4df4087acebcf6c09dc628f6dca25009cc6baab445f10d663639312e246 +size 3682326016 diff --git a/icarus-1-8b.i1-IQ3_XS.gguf b/icarus-1-8b.i1-IQ3_XS.gguf new file mode 100644 index 0000000..671f932 --- /dev/null +++ b/icarus-1-8b.i1-IQ3_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd8b43cbba2a54d075c278adb2449c3ef1eb5a9faa4cd408673a5859c7dcdac5 +size 3518748160 diff --git a/icarus-1-8b.i1-IQ3_XXS.gguf b/icarus-1-8b.i1-IQ3_XXS.gguf new file mode 100644 index 0000000..4fb50cd --- /dev/null +++ b/icarus-1-8b.i1-IQ3_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5e48d39bf88a8d4d215f370a5f830d04a167769ab059423aae559a735d7365d5 +size 3274913280 diff --git a/icarus-1-8b.i1-IQ4_NL.gguf b/icarus-1-8b.i1-IQ4_NL.gguf new file mode 100644 index 0000000..d022aee --- /dev/null +++ b/icarus-1-8b.i1-IQ4_NL.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:26e3b383e741a0dd8db6bf016eec6e433778503ef90964c0ee27e1918fa561f7 +size 4677989888 diff --git a/icarus-1-8b.i1-IQ4_XS.gguf b/icarus-1-8b.i1-IQ4_XS.gguf new file mode 100644 index 0000000..736db97 --- /dev/null +++ b/icarus-1-8b.i1-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9fcf88c44440e7c9b547ae3cb5cc4ed08cc45b0734aaba1ddc7d547893dc55db +size 4447663616 diff --git a/icarus-1-8b.i1-Q2_K.gguf b/icarus-1-8b.i1-Q2_K.gguf new file mode 100644 index 0000000..159da1f --- /dev/null +++ b/icarus-1-8b.i1-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1984a5749fb7cbfe4998452b47583b8a87fa35c56ab65d43880ab4d5d9e645c7 +size 3179132416 diff --git a/icarus-1-8b.i1-Q2_K_S.gguf b/icarus-1-8b.i1-Q2_K_S.gguf new file mode 100644 index 0000000..ee4770f --- /dev/null +++ b/icarus-1-8b.i1-Q2_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b5836f00ce5d437dd8a7044f7c9dae2230712129b64e5018b261c22fc7e1499c +size 2988815872 diff --git a/icarus-1-8b.i1-Q3_K_L.gguf b/icarus-1-8b.i1-Q3_K_L.gguf new file mode 100644 index 0000000..b433b0d --- /dev/null +++ b/icarus-1-8b.i1-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:92114872907b55966106574f5a3cf8a0d60ffaf80b93fa835f9a19fed7593ea6 +size 4321957376 diff --git a/icarus-1-8b.i1-Q3_K_M.gguf b/icarus-1-8b.i1-Q3_K_M.gguf new file mode 100644 index 0000000..f4ecea1 --- /dev/null +++ b/icarus-1-8b.i1-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e431385f2978e9bd26aab54a095c55e9df81a3334afbc0338ca629c16ed71cad +size 4018918912 diff --git a/icarus-1-8b.i1-Q3_K_S.gguf b/icarus-1-8b.i1-Q3_K_S.gguf new file mode 100644 index 0000000..0878f0d --- /dev/null +++ b/icarus-1-8b.i1-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:21c32bfc4d1d492e085061dfbf93b9ad5f1e08c8b65612760350774d87e20f5b +size 3664500224 diff --git a/icarus-1-8b.i1-Q4_0.gguf b/icarus-1-8b.i1-Q4_0.gguf new file mode 100644 index 0000000..47a2f24 --- /dev/null +++ b/icarus-1-8b.i1-Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9ad0d3bf3149f00f2fc183df9564379803eae49208d38a39301525e2bbea2c71 +size 4675892736 diff --git a/icarus-1-8b.i1-Q4_1.gguf b/icarus-1-8b.i1-Q4_1.gguf new file mode 100644 index 0000000..731edec --- /dev/null +++ b/icarus-1-8b.i1-Q4_1.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9e1cb476ac4eed70bcc62486b2bafdf30619f3c00c2cbe0014a58f89b369162b +size 5130253824 diff --git a/icarus-1-8b.i1-Q4_K_M.gguf b/icarus-1-8b.i1-Q4_K_M.gguf new file mode 100644 index 0000000..491a8a8 --- /dev/null +++ b/icarus-1-8b.i1-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b65e4f9e3250880bd97e0b82fd2d18b283e5d34e2b912f1cffec5ea49dadb7a9 +size 4920735232 diff --git a/icarus-1-8b.i1-Q4_K_S.gguf b/icarus-1-8b.i1-Q4_K_S.gguf new file mode 100644 index 0000000..11cfd82 --- /dev/null +++ b/icarus-1-8b.i1-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b744484936034dbfd00283546a07c40229745d7d86f82c0373238209f04a08a +size 4692669952 diff --git a/icarus-1-8b.i1-Q5_K_M.gguf b/icarus-1-8b.i1-Q5_K_M.gguf new file mode 100644 index 0000000..244adeb --- /dev/null +++ b/icarus-1-8b.i1-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8851274a58cad8ef466737a63c2aca7d18d65fcb96610c9b677ac2b62236a7da +size 5732988416 diff --git a/icarus-1-8b.i1-Q5_K_S.gguf b/icarus-1-8b.i1-Q5_K_S.gguf new file mode 100644 index 0000000..ad46d39 --- /dev/null +++ b/icarus-1-8b.i1-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a3eb11ff7cd5450189633ada1bd084c60a492752fcfb523bca9d441e7fe5d42e +size 5599294976 diff --git a/icarus-1-8b.i1-Q6_K.gguf b/icarus-1-8b.i1-Q6_K.gguf new file mode 100644 index 0000000..3c34e5a --- /dev/null +++ b/icarus-1-8b.i1-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:659fbb7b23795a82835478860671915a329b1d6cad79e8ce5e210923358249b4 +size 6596007424 diff --git a/icarus-1-8b.imatrix.gguf b/icarus-1-8b.imatrix.gguf new file mode 100644 index 0000000..73da63f --- /dev/null +++ b/icarus-1-8b.imatrix.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f349807781121257436e4c57e6217a153cecb0cb02dfd91e47a6e5d867a4011e +size 5015200