From b1f2680ce252760707f681c96f4be39cdce6ceed Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Mon, 11 May 2026 13:29:28 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: SinpxAI/Neural-Chat-7B-v3.2-GGUF Source: Original Platform --- .gitattributes | 57 +++++++++++++++++++++++++++++++++ README.md | 26 +++++++++++++++ configuration.json | 1 + neural-chat-7b-v3.2.Q2_K.gguf | 3 ++ neural-chat-7b-v3.2.Q3_K_L.gguf | 3 ++ neural-chat-7b-v3.2.Q3_K_M.gguf | 3 ++ neural-chat-7b-v3.2.Q3_K_S.gguf | 3 ++ neural-chat-7b-v3.2.Q4_0.gguf | 3 ++ neural-chat-7b-v3.2.Q4_K_M.gguf | 3 ++ neural-chat-7b-v3.2.Q4_K_S.gguf | 3 ++ neural-chat-7b-v3.2.Q5_0.gguf | 3 ++ neural-chat-7b-v3.2.Q5_K_M.gguf | 3 ++ neural-chat-7b-v3.2.Q5_K_S.gguf | 3 ++ neural-chat-7b-v3.2.Q6_K.gguf | 3 ++ neural-chat-7b-v3.2.Q8_0.gguf | 3 ++ 15 files changed, 120 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 configuration.json create mode 100644 neural-chat-7b-v3.2.Q2_K.gguf create mode 100644 neural-chat-7b-v3.2.Q3_K_L.gguf create mode 100644 neural-chat-7b-v3.2.Q3_K_M.gguf create mode 100644 neural-chat-7b-v3.2.Q3_K_S.gguf create mode 100644 neural-chat-7b-v3.2.Q4_0.gguf create mode 100644 neural-chat-7b-v3.2.Q4_K_M.gguf create mode 100644 neural-chat-7b-v3.2.Q4_K_S.gguf create mode 100644 neural-chat-7b-v3.2.Q5_0.gguf create mode 100644 neural-chat-7b-v3.2.Q5_K_M.gguf create mode 100644 neural-chat-7b-v3.2.Q5_K_S.gguf create mode 100644 neural-chat-7b-v3.2.Q6_K.gguf create mode 100644 neural-chat-7b-v3.2.Q8_0.gguf diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..fd4c4b6 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,57 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bin.* filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zstandard filter=lfs diff=lfs merge=lfs -text +*.tfevents* filter=lfs diff=lfs merge=lfs -text +*.db* filter=lfs diff=lfs merge=lfs -text +*.ark* filter=lfs diff=lfs merge=lfs -text +**/*ckpt*data* filter=lfs diff=lfs merge=lfs -text +**/*ckpt*.meta filter=lfs diff=lfs merge=lfs -text +**/*ckpt*.index filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.gguf filter=lfs diff=lfs merge=lfs -text + +*.gguf filter=lfs diff=lfs merge=lfs -text + +*.gguf filter=lfs diff=lfs merge=lfs -text + +*.gguf filter=lfs diff=lfs merge=lfs -text + +*.gguf filter=lfs diff=lfs merge=lfs -text + +*.gguf filter=lfs diff=lfs merge=lfs -text + +*.gguf filter=lfs diff=lfs merge=lfs -text + +*.gguf filter=lfs diff=lfs merge=lfs -text + +*.gguf filter=lfs diff=lfs merge=lfs -text + +*.gguf filter=lfs diff=lfs merge=lfs -text + +*.gguf filter=lfs diff=lfs merge=lfs -text + +*.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..fb22f7b --- /dev/null +++ b/README.md @@ -0,0 +1,26 @@ +--- +#以下为文本生成的 "tasks"示例,您可以从此网页中了解更多相关信息:https://modelscope.cn/docs/%E4%BB%BB%E5%8A%A1%E7%9A%84%E4%BB%8B%E7%BB%8D +#tasks: +#- text-generation +license: Apache License 2.0 +--- +### 当前模型的贡献者未提供更加详细的模型介绍。模型文件和权重,可浏览“模型文件”页面获取。 +#### 您可以通过如下git clone命令,或者ModelScope SDK来下载模型 + +SDK下载 +```bash +#安装ModelScope +pip install modelscope +``` +```python +#SDK模型下载 +from modelscope import snapshot_download +model_dir = snapshot_download('SinpxAI/Neural-Chat-7B-v3.2-GGUF') +``` +Git下载 +``` +#Git模型下载 +git clone https://www.modelscope.cn/SinpxAI/Neural-Chat-7B-v3.2-GGUF.git +``` + +

如果您是本模型的贡献者,我们邀请您根据模型贡献文档,及时完善模型卡片内容。

\ No newline at end of file diff --git a/configuration.json b/configuration.json new file mode 100644 index 0000000..3822542 --- /dev/null +++ b/configuration.json @@ -0,0 +1 @@ +{"framework": "\u5176\u4ed6", "task": "text-generation"} \ No newline at end of file diff --git a/neural-chat-7b-v3.2.Q2_K.gguf b/neural-chat-7b-v3.2.Q2_K.gguf new file mode 100644 index 0000000..e8b3b85 --- /dev/null +++ b/neural-chat-7b-v3.2.Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3d91b80f515dd853baf3bd41dbfcf50785e892fbe4bd385e1d4e363d52f9b15f +size 2719241856 diff --git a/neural-chat-7b-v3.2.Q3_K_L.gguf b/neural-chat-7b-v3.2.Q3_K_L.gguf new file mode 100644 index 0000000..9c211b0 --- /dev/null +++ b/neural-chat-7b-v3.2.Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9961dcd302fe7f1ded83ccff683cd5e1ea1277c00ff08cbdaed1bef4864719aa +size 3822024320 diff --git a/neural-chat-7b-v3.2.Q3_K_M.gguf b/neural-chat-7b-v3.2.Q3_K_M.gguf new file mode 100644 index 0000000..8e4308c --- /dev/null +++ b/neural-chat-7b-v3.2.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c934919205aa322a7c16cea69c460858fd75c2b1df3832a136784171fcba2ef1 +size 3518985856 diff --git a/neural-chat-7b-v3.2.Q3_K_S.gguf b/neural-chat-7b-v3.2.Q3_K_S.gguf new file mode 100644 index 0000000..aecddd2 --- /dev/null +++ b/neural-chat-7b-v3.2.Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:27ca8e1da3b73301c169c853ceb0a487d4f9fe5d9b4f0457eadde5f1c5775e87 +size 3164567168 diff --git a/neural-chat-7b-v3.2.Q4_0.gguf b/neural-chat-7b-v3.2.Q4_0.gguf new file mode 100644 index 0000000..a34e699 --- /dev/null +++ b/neural-chat-7b-v3.2.Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dcc22dc8bc752ff74b652a7ce694b6b9ab5eb8cbe7ddb744f038aac852e54d7c +size 4108916352 diff --git a/neural-chat-7b-v3.2.Q4_K_M.gguf b/neural-chat-7b-v3.2.Q4_K_M.gguf new file mode 100644 index 0000000..c471240 --- /dev/null +++ b/neural-chat-7b-v3.2.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8011edfb8a27bbcf04173f45013a0950871f0aee619a2f8d7187e4e64be9ab6b +size 4368438912 diff --git a/neural-chat-7b-v3.2.Q4_K_S.gguf b/neural-chat-7b-v3.2.Q4_K_S.gguf new file mode 100644 index 0000000..2faeb17 --- /dev/null +++ b/neural-chat-7b-v3.2.Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:511de474218ce0fbf0a67f237d5288801f38866d9ebe6b5664a6820bbc731353 +size 4140373632 diff --git a/neural-chat-7b-v3.2.Q5_0.gguf b/neural-chat-7b-v3.2.Q5_0.gguf new file mode 100644 index 0000000..85556be --- /dev/null +++ b/neural-chat-7b-v3.2.Q5_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0c5868a2ee84334e30a417e4d914565a767729c76ec9a39199fd134ed5c50a2e +size 4997715584 diff --git a/neural-chat-7b-v3.2.Q5_K_M.gguf b/neural-chat-7b-v3.2.Q5_K_M.gguf new file mode 100644 index 0000000..a31a139 --- /dev/null +++ b/neural-chat-7b-v3.2.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3815996caa1f968cffef593f1a078add16e7e8e64c3a08a3aff3c5612aed4057 +size 5131409024 diff --git a/neural-chat-7b-v3.2.Q5_K_S.gguf b/neural-chat-7b-v3.2.Q5_K_S.gguf new file mode 100644 index 0000000..c699363 --- /dev/null +++ b/neural-chat-7b-v3.2.Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:56bb01684da11c1f2857ed00e88a4c0013db093653a6b83bbc8039fa2291e1ee +size 4997715584 diff --git a/neural-chat-7b-v3.2.Q6_K.gguf b/neural-chat-7b-v3.2.Q6_K.gguf new file mode 100644 index 0000000..ffd8189 --- /dev/null +++ b/neural-chat-7b-v3.2.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:76e7d11d6389d176ec69dcb945781e70ed46aa68eb5f979be9301e6601345685 +size 5942064768 diff --git a/neural-chat-7b-v3.2.Q8_0.gguf b/neural-chat-7b-v3.2.Q8_0.gguf new file mode 100644 index 0000000..9117296 --- /dev/null +++ b/neural-chat-7b-v3.2.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:854c56b09ae90dfb20ff362f85cd180ec6e4e31b43e377581f0ea6af312fc7ad +size 7695857280