commit f4c4c0747d674837427ae028aaa1c5dc19abf929 Author: ModelHub XC Date: Sun Apr 12 08:43:59 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: mradermacher/aria-1b-chat-i1-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..0a69e1c --- /dev/null +++ b/.gitattributes @@ -0,0 +1,60 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.imatrix.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-IQ2_M.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-IQ2_S.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-IQ2_XS.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-IQ2_XXS.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-IQ3_S.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-IQ4_NL.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-Q2_K_S.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-Q4_1.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +aria-1b-chat.i1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..5e40ea3 --- /dev/null +++ b/README.md @@ -0,0 +1,92 @@ +--- +base_model: dkumar15/aria-1b-chat +language: +- en +library_name: transformers +license: apache-2.0 +mradermacher: + readme_rev: 1 +quantized_by: mradermacher +tags: +- llama +- causal-lm +- from-scratch +- dpo +- chat +- text-generation +--- +## About + + + + + + + + + +weighted/imatrix quants of https://huggingface.co/dkumar15/aria-1b-chat + + + +***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#aria-1b-chat-i1-GGUF).*** + +static quants are available at https://huggingface.co/mradermacher/aria-1b-chat-GGUF +## Usage + +If you are unsure how to use GGUF files, refer to one of [TheBloke's +READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for +more details, including on how to concatenate multi-part files. + +## Provided Quants + +(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) + +| Link | Type | Size/GB | Notes | +|:-----|:-----|--------:|:------| +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.imatrix.gguf) | imatrix | 0.1 | imatrix file (for creating your own quants) | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-IQ1_S.gguf) | i1-IQ1_S | 0.5 | for the desperate | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-IQ1_M.gguf) | i1-IQ1_M | 0.5 | mostly desperate | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 0.5 | | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-IQ2_XS.gguf) | i1-IQ2_XS | 0.5 | | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-IQ2_S.gguf) | i1-IQ2_S | 0.5 | | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-IQ2_M.gguf) | i1-IQ2_M | 0.6 | | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-Q2_K_S.gguf) | i1-Q2_K_S | 0.6 | very low quality | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-Q2_K.gguf) | i1-Q2_K | 0.6 | IQ3_XXS probably better | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 0.6 | lower quality | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-IQ3_XS.gguf) | i1-IQ3_XS | 0.6 | | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-Q3_K_S.gguf) | i1-Q3_K_S | 0.6 | IQ3_XS probably better | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-IQ3_S.gguf) | i1-IQ3_S | 0.6 | beats Q3_K* | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-IQ3_M.gguf) | i1-IQ3_M | 0.7 | | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-Q3_K_M.gguf) | i1-Q3_K_M | 0.7 | IQ3_S probably better | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-Q3_K_L.gguf) | i1-Q3_K_L | 0.7 | IQ3_M probably better | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-IQ4_XS.gguf) | i1-IQ4_XS | 0.7 | | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-Q4_0.gguf) | i1-Q4_0 | 0.7 | fast, low quality | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-IQ4_NL.gguf) | i1-IQ4_NL | 0.7 | prefer IQ4_XS | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-Q4_K_S.gguf) | i1-Q4_K_S | 0.8 | optimal size/speed/quality | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-Q4_1.gguf) | i1-Q4_1 | 0.8 | | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-Q4_K_M.gguf) | i1-Q4_K_M | 0.8 | fast, recommended | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-Q5_K_S.gguf) | i1-Q5_K_S | 0.9 | | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-Q5_K_M.gguf) | i1-Q5_K_M | 0.9 | | +| [GGUF](https://huggingface.co/mradermacher/aria-1b-chat-i1-GGUF/resolve/main/aria-1b-chat.i1-Q6_K.gguf) | i1-Q6_K | 1.1 | practically like static Q6_K | + +Here is a handy graph by ikawrakow comparing some lower-quality quant +types (lower is better): + +![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) + +And here are Artefact2's thoughts on the matter: +https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 + +## FAQ / Model Request + +See https://huggingface.co/mradermacher/model_requests for some answers to +questions you might have and/or if you want some other model quantized. + +## Thanks + +I thank my company, [nethype GmbH](https://www.nethype.de/), for letting +me use its servers and providing upgrades to my workstation to enable +this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to. + + diff --git a/aria-1b-chat.i1-IQ1_M.gguf b/aria-1b-chat.i1-IQ1_M.gguf new file mode 100644 index 0000000..54e131a --- /dev/null +++ b/aria-1b-chat.i1-IQ1_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:024cbb7c8f46c8f5f19fc34f3ca87b1b16dd2f8b109638ff0b958d9988c6ab8e +size 377651552 diff --git a/aria-1b-chat.i1-IQ1_S.gguf b/aria-1b-chat.i1-IQ1_S.gguf new file mode 100644 index 0000000..a5d2446 --- /dev/null +++ b/aria-1b-chat.i1-IQ1_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:686537cda49001124d1bc3536f39b3a76ef18a6fbe24dc8aac87ed30b8fc4993 +size 363323744 diff --git a/aria-1b-chat.i1-IQ2_M.gguf b/aria-1b-chat.i1-IQ2_M.gguf new file mode 100644 index 0000000..4400039 --- /dev/null +++ b/aria-1b-chat.i1-IQ2_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bdb248c58044b07f10f95f031b15b979cb6de2afb222a9a0eb8a9b42a81a8ba2 +size 462255072 diff --git a/aria-1b-chat.i1-IQ2_S.gguf b/aria-1b-chat.i1-IQ2_S.gguf new file mode 100644 index 0000000..6c870fb --- /dev/null +++ b/aria-1b-chat.i1-IQ2_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:13390443da897284bbd5c0ab4e541339f87ed143532beb60bb14a094790cce57 +size 443151328 diff --git a/aria-1b-chat.i1-IQ2_XS.gguf b/aria-1b-chat.i1-IQ2_XS.gguf new file mode 100644 index 0000000..2de7cac --- /dev/null +++ b/aria-1b-chat.i1-IQ2_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a2e431caaedeee4d3ec80c40052f4f4c0b55a52fee926ccdee8b3c44cca283d +size 423518560 diff --git a/aria-1b-chat.i1-IQ2_XXS.gguf b/aria-1b-chat.i1-IQ2_XXS.gguf new file mode 100644 index 0000000..86df961 --- /dev/null +++ b/aria-1b-chat.i1-IQ2_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d17854d1e98d357e11ffdf2ce9488ce98c4175e517721e46867e1871065373ae +size 401531232 diff --git a/aria-1b-chat.i1-IQ3_M.gguf b/aria-1b-chat.i1-IQ3_M.gguf new file mode 100644 index 0000000..177707a --- /dev/null +++ b/aria-1b-chat.i1-IQ3_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:02f57ed6ebcde78809447460870bda3525bc7e4aafaffef279dfcd900143a874 +size 552896288 diff --git a/aria-1b-chat.i1-IQ3_S.gguf b/aria-1b-chat.i1-IQ3_S.gguf new file mode 100644 index 0000000..f959ce1 --- /dev/null +++ b/aria-1b-chat.i1-IQ3_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ceb359842fd64c54b4f94eb74089ca1cef7211e151e7fb371f88c0835f79cafb +size 537823008 diff --git a/aria-1b-chat.i1-IQ3_XS.gguf b/aria-1b-chat.i1-IQ3_XS.gguf new file mode 100644 index 0000000..d27fcbb --- /dev/null +++ b/aria-1b-chat.i1-IQ3_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6657347cac53f6368b0acebcbcc687f1abafa6a3408e2aba237b854b6039c433 +size 514451232 diff --git a/aria-1b-chat.i1-IQ3_XXS.gguf b/aria-1b-chat.i1-IQ3_XXS.gguf new file mode 100644 index 0000000..b8be5c1 --- /dev/null +++ b/aria-1b-chat.i1-IQ3_XXS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c8d7ea20e5721e997b839a63fa9f4a9f51aaa269d72a9237c5ea6d32083ea736 +size 493253600 diff --git a/aria-1b-chat.i1-IQ4_NL.gguf b/aria-1b-chat.i1-IQ4_NL.gguf new file mode 100644 index 0000000..ad23e52 --- /dev/null +++ b/aria-1b-chat.i1-IQ4_NL.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:188a2f4ca8f099366cdaed1d4ab8dfed20e36067398e7d1ea644783580aedf46 +size 642857536 diff --git a/aria-1b-chat.i1-IQ4_XS.gguf b/aria-1b-chat.i1-IQ4_XS.gguf new file mode 100644 index 0000000..7bbb784 --- /dev/null +++ b/aria-1b-chat.i1-IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9d7dedd396a1fc325203f8e63832c3d22d7075e6bd15c5a61ea2405f068a0713 +size 618822016 diff --git a/aria-1b-chat.i1-Q2_K.gguf b/aria-1b-chat.i1-Q2_K.gguf new file mode 100644 index 0000000..f499bc1 --- /dev/null +++ b/aria-1b-chat.i1-Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf1c5467d146604c7686f4c46064605d8eb614b92aca6e5203e393fb61d7dc2b +size 469079200 diff --git a/aria-1b-chat.i1-Q2_K_S.gguf b/aria-1b-chat.i1-Q2_K_S.gguf new file mode 100644 index 0000000..35e6c16 --- /dev/null +++ b/aria-1b-chat.i1-Q2_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e0193096f9e9570e001b4c505f56378942fe66861d85af1204f8b962c31d4369 +size 462525600 diff --git a/aria-1b-chat.i1-Q3_K_L.gguf b/aria-1b-chat.i1-Q3_K_L.gguf new file mode 100644 index 0000000..5ebd924 --- /dev/null +++ b/aria-1b-chat.i1-Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8c31cbad49876627dc8d0b9eebd66fdc214dd114662fb16a3871125cb9ec00f2 +size 610993952 diff --git a/aria-1b-chat.i1-Q3_K_M.gguf b/aria-1b-chat.i1-Q3_K_M.gguf new file mode 100644 index 0000000..3c05dc3 --- /dev/null +++ b/aria-1b-chat.i1-Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:96bcb01fb359fa9a998acf072e2f65446ca4d55186997df53168aaa75d5ab99e +size 582043424 diff --git a/aria-1b-chat.i1-Q3_K_S.gguf b/aria-1b-chat.i1-Q3_K_S.gguf new file mode 100644 index 0000000..b3156a9 --- /dev/null +++ b/aria-1b-chat.i1-Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7c37228fa19c3f37d61a193cea65361eb179dbc8ada6201082521a419137a82e +size 534759200 diff --git a/aria-1b-chat.i1-Q4_0.gguf b/aria-1b-chat.i1-Q4_0.gguf new file mode 100644 index 0000000..6abe566 --- /dev/null +++ b/aria-1b-chat.i1-Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5e194cde0a2114d1375f795b09103a47d983c64a97247ed62292b147a3169e52 +size 641382976 diff --git a/aria-1b-chat.i1-Q4_1.gguf b/aria-1b-chat.i1-Q4_1.gguf new file mode 100644 index 0000000..a647c42 --- /dev/null +++ b/aria-1b-chat.i1-Q4_1.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0215060b57f62377cf333380ddaa52f3a77ba0e86028248f953b5c2cea25b60 +size 704986048 diff --git a/aria-1b-chat.i1-Q4_K_M.gguf b/aria-1b-chat.i1-Q4_K_M.gguf new file mode 100644 index 0000000..884fbe7 --- /dev/null +++ b/aria-1b-chat.i1-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9e271ee887831598c3ac6fdc35cf51e04fd70496f1745d0aa010ea15459a8269 +size 715946560 diff --git a/aria-1b-chat.i1-Q4_K_S.gguf b/aria-1b-chat.i1-Q4_K_S.gguf new file mode 100644 index 0000000..c8c88eb --- /dev/null +++ b/aria-1b-chat.i1-Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:76457c13c70ec48d072a266c5b0d28104b6f99353e6fead80a25e182fe0d38ee +size 672905792 diff --git a/aria-1b-chat.i1-Q5_K_M.gguf b/aria-1b-chat.i1-Q5_K_M.gguf new file mode 100644 index 0000000..1d7de23 --- /dev/null +++ b/aria-1b-chat.i1-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b37a84cbcc012e4946b7e644de1eff232b89c5a7682bbed42a60a419cc46f08d +size 822115648 diff --git a/aria-1b-chat.i1-Q5_K_S.gguf b/aria-1b-chat.i1-Q5_K_S.gguf new file mode 100644 index 0000000..23f74dc --- /dev/null +++ b/aria-1b-chat.i1-Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9e036ea62d05fa64f7a906cd8fa59e9d8efe0efa6e5ae2dc95e47304bd5c4031 +size 785497408 diff --git a/aria-1b-chat.i1-Q6_K.gguf b/aria-1b-chat.i1-Q6_K.gguf new file mode 100644 index 0000000..ef9d618 --- /dev/null +++ b/aria-1b-chat.i1-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:473c7fca103d97a58b602edd4ffc409c2a1f246535eb22bb577dfc3dcc29f806 +size 968208512 diff --git a/aria-1b-chat.imatrix.gguf b/aria-1b-chat.imatrix.gguf new file mode 100644 index 0000000..f2c81b5 --- /dev/null +++ b/aria-1b-chat.imatrix.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a774f05ba6f89399931b5eaa56a037cb6e5971300da228d16c81b6babb1763f +size 1589440