From c54bf95e92e5a6d32c2d19f3ab53298ac4c47f5d Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Mon, 20 Apr 2026 19:05:23 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: ZeroWw/Phi-3.5-mini-instruct_Uncensored-GGUF Source: Original Platform --- .gitattributes | 41 ++++++++++++++++++++++ Phi-3.5-mini-instruct_Uncensored.f16.gguf | 3 ++ Phi-3.5-mini-instruct_Uncensored.q5_k.gguf | 3 ++ Phi-3.5-mini-instruct_Uncensored.q6_k.gguf | 3 ++ Phi-3.5-mini-instruct_Uncensored.q8_0.gguf | 3 ++ Phi-3.5-mini-instruct_Uncensored.q8_p.gguf | 3 ++ Phi-3.5-mini-instruct_Uncensored.q8q4.gguf | 3 ++ README.md | 17 +++++++++ 8 files changed, 76 insertions(+) create mode 100644 .gitattributes create mode 100644 Phi-3.5-mini-instruct_Uncensored.f16.gguf create mode 100644 Phi-3.5-mini-instruct_Uncensored.q5_k.gguf create mode 100644 Phi-3.5-mini-instruct_Uncensored.q6_k.gguf create mode 100644 Phi-3.5-mini-instruct_Uncensored.q8_0.gguf create mode 100644 Phi-3.5-mini-instruct_Uncensored.q8_p.gguf create mode 100644 Phi-3.5-mini-instruct_Uncensored.q8q4.gguf create mode 100644 README.md diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..627a464 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,41 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +Phi-3.5-mini-instruct_Uncensored.f16.gguf filter=lfs diff=lfs merge=lfs -text +Phi-3.5-mini-instruct_Uncensored.q5_k.gguf filter=lfs diff=lfs merge=lfs -text +Phi-3.5-mini-instruct_Uncensored.q6_k.gguf filter=lfs diff=lfs merge=lfs -text +Phi-3.5-mini-instruct_Uncensored.q8_0.gguf filter=lfs diff=lfs merge=lfs -text +Phi-3.5-mini-instruct_Uncensored.q8_p.gguf filter=lfs diff=lfs merge=lfs -text +Phi-3.5-mini-instruct_Uncensored.q8q4.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/Phi-3.5-mini-instruct_Uncensored.f16.gguf b/Phi-3.5-mini-instruct_Uncensored.f16.gguf new file mode 100644 index 0000000..f3c804c --- /dev/null +++ b/Phi-3.5-mini-instruct_Uncensored.f16.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4ec12ac0c10da8d24e3d753a9a7b3837c92fd24b430cea1e476a95c1cf37771d +size 7643297152 diff --git a/Phi-3.5-mini-instruct_Uncensored.q5_k.gguf b/Phi-3.5-mini-instruct_Uncensored.q5_k.gguf new file mode 100644 index 0000000..93fbb72 --- /dev/null +++ b/Phi-3.5-mini-instruct_Uncensored.q5_k.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d55fff7d10d15025968939b84bde0d561eb08639e28de8c17f1bf06bc5ee9c7f +size 3060757888 diff --git a/Phi-3.5-mini-instruct_Uncensored.q6_k.gguf b/Phi-3.5-mini-instruct_Uncensored.q6_k.gguf new file mode 100644 index 0000000..63cc2a9 --- /dev/null +++ b/Phi-3.5-mini-instruct_Uncensored.q6_k.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66506bedf97a318eb20f0cda986c1b37e868aa6a4f3e29e83b0ce82d8b9cc795 +size 3368252800 diff --git a/Phi-3.5-mini-instruct_Uncensored.q8_0.gguf b/Phi-3.5-mini-instruct_Uncensored.q8_0.gguf new file mode 100644 index 0000000..8d4aa01 --- /dev/null +++ b/Phi-3.5-mini-instruct_Uncensored.q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f50f70a74b1d80d2286e3dc771034cebcad8a62bb4b4711a5b237d1cee5562c +size 4245910912 diff --git a/Phi-3.5-mini-instruct_Uncensored.q8_p.gguf b/Phi-3.5-mini-instruct_Uncensored.q8_p.gguf new file mode 100644 index 0000000..b70cdbc --- /dev/null +++ b/Phi-3.5-mini-instruct_Uncensored.q8_p.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bfc8e17ecd9283532104c8e21e807db904737d7241ad0b6b40358be9e3954904 +size 4061222272 diff --git a/Phi-3.5-mini-instruct_Uncensored.q8q4.gguf b/Phi-3.5-mini-instruct_Uncensored.q8q4.gguf new file mode 100644 index 0000000..91a612c --- /dev/null +++ b/Phi-3.5-mini-instruct_Uncensored.q8q4.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1811e995dc53797087bfadbcd2ffcdd57dcfc0cc012ebe086ca530cf0b7a5c2d +size 2466338176 diff --git a/README.md b/README.md new file mode 100644 index 0000000..3ab75c1 --- /dev/null +++ b/README.md @@ -0,0 +1,17 @@ + +--- +license: mit +language: +- en +pipeline_tag: text-generation +--- + +My own (ZeroWw) quantizations. +output and embed tensors quantized to f16. +all other tensors quantized to q5_k or q6_k. + +Result: +both f16.q6 and f16.q5 are smaller than q8_0 standard quantization +and they perform as well as the pure f16. + +Updated on: Thu Aug 22, 13:56:30