From 3ea9f3409c134ad771927960fd7d237b14e2b4aa Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Wed, 29 Apr 2026 20:28:47 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: datatab/Yugo45A-GPT-Quantized-GGUF Source: Original Platform --- .gitattributes | 49 ++++++++++++++++++++++++++ README.md | 17 +++++++++ Yugo45A-GPT-Quantized-GGUF.Q3_K_M.gguf | 3 ++ Yugo45A-GPT-Quantized-GGUF.Q4_K_M.gguf | 3 ++ Yugo45A-GPT-Quantized-GGUF.Q4_K_S.gguf | 3 ++ Yugo45A-GPT-Quantized-GGUF.Q5_0.gguf | 3 ++ Yugo45A-GPT-Quantized-GGUF.Q5_K_M.gguf | 3 ++ Yugo45A-GPT-Quantized-GGUF.Q6_K.gguf | 3 ++ Yugo45A-GPT-Quantized-GGUF.Q8_0.gguf | 3 ++ config.json | 3 ++ 10 files changed, 90 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 Yugo45A-GPT-Quantized-GGUF.Q3_K_M.gguf create mode 100644 Yugo45A-GPT-Quantized-GGUF.Q4_K_M.gguf create mode 100644 Yugo45A-GPT-Quantized-GGUF.Q4_K_S.gguf create mode 100644 Yugo45A-GPT-Quantized-GGUF.Q5_0.gguf create mode 100644 Yugo45A-GPT-Quantized-GGUF.Q5_K_M.gguf create mode 100644 Yugo45A-GPT-Quantized-GGUF.Q6_K.gguf create mode 100644 Yugo45A-GPT-Quantized-GGUF.Q8_0.gguf create mode 100644 config.json diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..b4e1371 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,49 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +Yugo45A-GPT-Quantized-GGUF-unsloth.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Yugo45A-GPT-Quantized-GGUF.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Yugo45A-GPT-Quantized-GGUF-unsloth.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Yugo45A-GPT-Quantized-GGUF.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Yugo45A-GPT-Quantized-GGUF-unsloth.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Yugo45A-GPT-Quantized-GGUF.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Yugo45A-GPT-Quantized-GGUF-unsloth.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Yugo45A-GPT-Quantized-GGUF.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Yugo45A-GPT-Quantized-GGUF-unsloth.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +Yugo45A-GPT-Quantized-GGUF.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +Yugo45A-GPT-Quantized-GGUF-unsloth.Q5_0.gguf filter=lfs diff=lfs merge=lfs -text +Yugo45A-GPT-Quantized-GGUF.Q5_0.gguf filter=lfs diff=lfs merge=lfs -text +Yugo45A-GPT-Quantized-GGUF-unsloth.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +Yugo45A-GPT-Quantized-GGUF.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..9098c71 --- /dev/null +++ b/README.md @@ -0,0 +1,17 @@ +--- +language: +- sr +license: apache-2.0 +tags: +- text-generation-inference +- transformers +- mistral +- gguf +base_model: datatab/Yugo45A-GPT +--- + +# Uploaded model + +- **Developed by:** datatab +- **License:** apache-2.0 +- **Finetuned from model :** datatab/AlphaMonarch-7B \ No newline at end of file diff --git a/Yugo45A-GPT-Quantized-GGUF.Q3_K_M.gguf b/Yugo45A-GPT-Quantized-GGUF.Q3_K_M.gguf new file mode 100644 index 0000000..606cdb4 --- /dev/null +++ b/Yugo45A-GPT-Quantized-GGUF.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65cd973a39e89c1e29249ca59e442b94387a8e1eaede53b6596fc08483389f9f +size 3518986208 diff --git a/Yugo45A-GPT-Quantized-GGUF.Q4_K_M.gguf b/Yugo45A-GPT-Quantized-GGUF.Q4_K_M.gguf new file mode 100644 index 0000000..7bd930b --- /dev/null +++ b/Yugo45A-GPT-Quantized-GGUF.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:585500855be6f32273f4ad1e9db5d60681dd33bb703176fc8d7009b5c6d67049 +size 4368439264 diff --git a/Yugo45A-GPT-Quantized-GGUF.Q4_K_S.gguf b/Yugo45A-GPT-Quantized-GGUF.Q4_K_S.gguf new file mode 100644 index 0000000..5ddfa5d --- /dev/null +++ b/Yugo45A-GPT-Quantized-GGUF.Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:398e089667681d5226987866bfa5d10950c7db6bd2308a520551b24c1d27117d +size 4140373984 diff --git a/Yugo45A-GPT-Quantized-GGUF.Q5_0.gguf b/Yugo45A-GPT-Quantized-GGUF.Q5_0.gguf new file mode 100644 index 0000000..0d01fd0 --- /dev/null +++ b/Yugo45A-GPT-Quantized-GGUF.Q5_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bbffaf95e40125b97871bad96aae5640b7a477edd421d6d9960e272664a08033 +size 4997715936 diff --git a/Yugo45A-GPT-Quantized-GGUF.Q5_K_M.gguf b/Yugo45A-GPT-Quantized-GGUF.Q5_K_M.gguf new file mode 100644 index 0000000..ced9795 --- /dev/null +++ b/Yugo45A-GPT-Quantized-GGUF.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8bf9e8aaba9445432980b8e8e36fef062c6920963051c097e281d7648a5dbeb5 +size 5131409376 diff --git a/Yugo45A-GPT-Quantized-GGUF.Q6_K.gguf b/Yugo45A-GPT-Quantized-GGUF.Q6_K.gguf new file mode 100644 index 0000000..9ac11d6 --- /dev/null +++ b/Yugo45A-GPT-Quantized-GGUF.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3237ca4c64e8631ae77b4a4e97056fe1a63cc098f95aecc4bda91c0bfad3d5fd +size 5942065120 diff --git a/Yugo45A-GPT-Quantized-GGUF.Q8_0.gguf b/Yugo45A-GPT-Quantized-GGUF.Q8_0.gguf new file mode 100644 index 0000000..103e376 --- /dev/null +++ b/Yugo45A-GPT-Quantized-GGUF.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:11b567e2eae34e5158c0d534f3b81b2ce8bbcf3df144bcee0315c8dba37723df +size 7695857600 diff --git a/config.json b/config.json new file mode 100644 index 0000000..9f0f76f --- /dev/null +++ b/config.json @@ -0,0 +1,3 @@ +{ + "model_type": "mistral" +} \ No newline at end of file