Files
llama-3-chinese-8b-instruct…/README.md
ChineseAlpacaGroup 8895fd6daf Update README.md
2024-05-29 00:55:34 +00:00

109 lines
3.2 KiB
Markdown
Raw Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

---
frameworks:
- other
license: Apache License 2.0
model-type:
- llama
language:
- zh
- en
tools:
- llamacpp
#model-type:
##如 gpt、phi、llama、chatglm、baichuan 等
#- gpt
#domain:
##如 nlp、cv、audio、multi-modal
#- nlp
#language:
##语言代码列表 https://help.aliyun.com/document_detail/215387.html?spm=a2c4g.11186623.0.0.9f8d7467kni6Aa
#- cn
#metrics:
##如 CIDEr、Blue、ROUGE 等
#- CIDEr
#tags:
##各种自定义,包括 pretrained、fine-tuned、instruction-tuned、RL-tuned 等训练方法和其他
#- pretrained
#tools:
##如 vllm、fastchat、llamacpp、AdaSeq 等
#- vllm
---
---
license: apache-2.0
language:
- zh
- en
---
# Llama-3-Chinese-8B-Instruct-v3-GGUF
<p align="center">
<a href="https://github.com/ymcui/Chinese-LLaMA-Alpaca-3"><img src="https://ymcui.com/images/chinese-llama-alpaca-3-banner.png" width="600"/></a>
</p>
这个仓库包含了**Llama-3-Chinese-8B-Instruct-v3-GGUF**兼容llama.cpp/ollama等是[Llama-3-Chinese-8B-Instruct-v3](https://modelscope.cn/models/ChineseAlpacaGroup/llama-3-chinese-8b-instruct-v3)模型的量化版本。
**注意:这是一个指令模型,可以直接适用于对话、问答等任务。**
更多细节性能、使用方法等请参考GitHub项目页面https://github.com/ymcui/Chinese-LLaMA-Alpaca-3
## 量化性能
评测指标PPL**越低越好**
| Quant | Size | PPL |
| :---: | -------: | ------------------: |
| Q2_K | 2.96 GB | 10.0534 +/- 0.13135 |
| Q3_K | 3.74 GB | 6.3295 +/- 0.07816 |
| Q4_0 | 4.34 GB | 6.3200 +/- 0.07893 |
| Q4_K | 4.58 GB | 6.0042 +/- 0.07431 |
| Q5_0 | 5.21 GB | 6.0437 +/- 0.07526 |
| Q5_K | 5.34 GB | 5.9484 +/- 0.07399 |
| Q6_K | 6.14 GB | 5.9469 +/- 0.07404 |
| Q8_0 | 7.95 GB | 5.8933 +/- 0.07305 |
| F16 | 14.97 GB | 5.8902 +/- 0.07303 |
## 其他
- 完整模型https://modelscope.cn/models/ChineseAlpacaGroup/llama-3-chinese-8b-instruct-v3
- 关于本模型的提问,请通过 https://github.com/ymcui/Chinese-LLaMA-Alpaca-3 提交issue
----
This repository contains **Llama-3-Chinese-8B-Instruct-v3-GGUF** (llama.cpp/ollama/tgw, etc. compatible), which is the quantized version of [Llama-3-Chinese-8B-Instruct-v3](https://huggingface.co/hfl/llama-3-chinese-8b-instruct-v3).
**Note: this is an instruction (chat) model, which can be used for conversation, QA, etc.**
Further details (performance, usage, etc.) should refer to GitHub project page: https://github.com/ymcui/Chinese-LLaMA-Alpaca-3
## Performance
Metric: PPL, lower is better
| Quant | Size | PPL |
| :---: | -------: | ------------------: |
| Q2_K | 2.96 GB | 10.0534 +/- 0.13135 |
| Q3_K | 3.74 GB | 6.3295 +/- 0.07816 |
| Q4_0 | 4.34 GB | 6.3200 +/- 0.07893 |
| Q4_K | 4.58 GB | 6.0042 +/- 0.07431 |
| Q5_0 | 5.21 GB | 6.0437 +/- 0.07526 |
| Q5_K | 5.34 GB | 5.9484 +/- 0.07399 |
| Q6_K | 6.14 GB | 5.9469 +/- 0.07404 |
| Q8_0 | 7.95 GB | 5.8933 +/- 0.07305 |
| F16 | 14.97 GB | 5.8902 +/- 0.07303 |
## Others
- For full model, please see: https://huggingface.co/hfl/llama-3-chinese-8b-instruct-v3
- If you have questions/issues regarding this model, please submit an issue through https://github.com/ymcui/Chinese-LLaMA-Alpaca-3