commit ff88be34de623ab5b122626fc537ce660ee683ca Author: ModelHub XC Date: Tue Apr 21 17:29:48 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: roleplaiapp/DeepSeek-R1-Distill-Qwen-14B-Q8_0-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..e4e9cc5 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,36 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +deepseek-r1-distill-qwen-14b-q8_0.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..fdccd0f --- /dev/null +++ b/README.md @@ -0,0 +1,43 @@ +--- +base_model: deepseek-ai/DeepSeek-R1-Distill-Qwen-14B +library_name: transformers +pipeline_tag: text-generation +tags: +- llama-cpp +- DeepSeek-R1-Distill-Qwen-14B +- gguf +- Q8_0 +- 14b +- qwen +- DeepSeek-R1 +- llama-cpp +- deepseek-ai +- code +- math +- chat +- roleplay +- text-generation +- safetensors +- nlp +- code +--- + +# roleplaiapp/DeepSeek-R1-Distill-Qwen-14B-Q8_0-GGUF + +**Repo:** `roleplaiapp/DeepSeek-R1-Distill-Qwen-14B-Q8_0-GGUF` +**Original Model:** `DeepSeek-R1-Distill-Qwen-14B` +**Organization:** `deepseek-ai` +**Quantized File:** `deepseek-r1-distill-qwen-14b-q8_0.gguf` +**Quantization:** `GGUF` +**Quantization Method:** `Q8_0` +**Use Imatrix:** `False` +**Split Model:** `False` + +## Overview +This is an GGUF Q8_0 quantized version of [DeepSeek-R1-Distill-Qwen-14B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-14B). + +## Quantization By +I often have idle A100 GPUs while building/testing and training the RP app, so I put them to use quantizing models. +I hope the community finds these quantizations useful. + +Andrew Webby @ [RolePlai](https://roleplai.app/) diff --git a/deepseek-r1-distill-qwen-14b-q8_0.gguf b/deepseek-r1-distill-qwen-14b-q8_0.gguf new file mode 100644 index 0000000..9016438 --- /dev/null +++ b/deepseek-r1-distill-qwen-14b-q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b08ed57a4824de6ab414095cd20b53e8dfd375dd953f940ffbe7340e5c9adb5f +size 15701597312