diff --git a/.gitattributes b/.gitattributes index 53d7257..1036323 100644 --- a/.gitattributes +++ b/.gitattributes @@ -1,47 +1,47 @@ *.7z filter=lfs diff=lfs merge=lfs -text *.arrow filter=lfs diff=lfs merge=lfs -text *.bin filter=lfs diff=lfs merge=lfs -text -*.bin.* filter=lfs diff=lfs merge=lfs -text *.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text *.ftz filter=lfs diff=lfs merge=lfs -text *.gz filter=lfs diff=lfs merge=lfs -text *.h5 filter=lfs diff=lfs merge=lfs -text *.joblib filter=lfs diff=lfs merge=lfs -text *.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text *.model filter=lfs diff=lfs merge=lfs -text *.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text *.onnx filter=lfs diff=lfs merge=lfs -text *.ot filter=lfs diff=lfs merge=lfs -text *.parquet filter=lfs diff=lfs merge=lfs -text *.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text *.pt filter=lfs diff=lfs merge=lfs -text *.pth filter=lfs diff=lfs merge=lfs -text *.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text saved_model/**/* filter=lfs diff=lfs merge=lfs -text *.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text *.tflite filter=lfs diff=lfs merge=lfs -text *.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text *.xz filter=lfs diff=lfs merge=lfs -text *.zip filter=lfs diff=lfs merge=lfs -text -*.zstandard filter=lfs diff=lfs merge=lfs -text -*.tfevents* filter=lfs diff=lfs merge=lfs -text -*.db* filter=lfs diff=lfs merge=lfs -text -*.ark* filter=lfs diff=lfs merge=lfs -text -**/*ckpt*data* filter=lfs diff=lfs merge=lfs -text -**/*ckpt*.meta filter=lfs diff=lfs merge=lfs -text -**/*ckpt*.index filter=lfs diff=lfs merge=lfs -text -*.safetensors filter=lfs diff=lfs merge=lfs -text -*.ckpt filter=lfs diff=lfs merge=lfs -text -*.gguf* filter=lfs diff=lfs merge=lfs -text -*.ggml filter=lfs diff=lfs merge=lfs -text -*.llamafile* filter=lfs diff=lfs merge=lfs -text -*.pt2 filter=lfs diff=lfs merge=lfs -text -*.mlmodel filter=lfs diff=lfs merge=lfs -text -*.npy filter=lfs diff=lfs merge=lfs -text -*.npz filter=lfs diff=lfs merge=lfs -text -*.pickle filter=lfs diff=lfs merge=lfs -text -*.pkl filter=lfs diff=lfs merge=lfs -text -*.tar filter=lfs diff=lfs merge=lfs -text -*.wasm filter=lfs diff=lfs merge=lfs -text *.zst filter=lfs diff=lfs merge=lfs -text -*tfevents* filter=lfs diff=lfs merge=lfs -text \ No newline at end of file +*tfevents* filter=lfs diff=lfs merge=lfs -text +docsgpt-7b-mistral.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +docsgpt-7b-mistral.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +docsgpt-7b-mistral.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +docsgpt-7b-mistral.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +docsgpt-7b-mistral.Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +docsgpt-7b-mistral.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +docsgpt-7b-mistral.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +docsgpt-7b-mistral.Q5_0.gguf filter=lfs diff=lfs merge=lfs -text +docsgpt-7b-mistral.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +docsgpt-7b-mistral.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +docsgpt-7b-mistral.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +docsgpt-7b-mistral.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md index cd2bce0..122952a 100644 --- a/README.md +++ b/README.md @@ -1,47 +1,423 @@ --- -license: Apache License 2.0 +base_model: Arc53/docsgpt-7b-mistral +inference: false +license: apache-2.0 +model_creator: Arc53 +model_name: DocsGPT 7B Mistral +model_type: mistral +prompt_template: '### Instruction -#model-type: -##如 gpt、phi、llama、chatglm、baichuan 等 -#- gpt + {prompt} -#domain: -##如 nlp、cv、audio、multi-modal -#- nlp + ### Context -#language: -##语言代码列表 https://help.aliyun.com/document_detail/215387.html?spm=a2c4g.11186623.0.0.9f8d7467kni6Aa -#- cn + {{context}} -#metrics: -##如 CIDEr、Blue、ROUGE 等 -#- CIDEr + ### Answer -#tags: -##各种自定义,包括 pretrained、fine-tuned、instruction-tuned、RL-tuned 等训练方法和其他 -#- pretrained - -#tools: -##如 vllm、fastchat、llamacpp、AdaSeq 等 -#- vllm + ' +quantized_by: TheBloke +tags: +- rag +- closed-qa +- context +- mistral --- -### 当前模型的贡献者未提供更加详细的模型介绍。模型文件和权重,可浏览“模型文件”页面获取。 -#### 您可以通过如下git clone命令,或者ModelScope SDK来下载模型 + + + + +
+TheBloke's LLM work is generously supported by a grant from andreessen horowitz (a16z)
如果您是本模型的贡献者,我们邀请您根据模型贡献文档,及时完善模型卡片内容。
\ No newline at end of file +## How to use with LangChain + +Here are guides on using llama-cpp-python and ctransformers with LangChain: + +* [LangChain + llama-cpp-python](https://python.langchain.com/docs/integrations/llms/llamacpp) +* [LangChain + ctransformers](https://python.langchain.com/docs/integrations/providers/ctransformers) + + + + + +## Discord + +For further support, and discussions on these models and AI in general, join us at: + +[TheBloke AI's Discord server](https://discord.gg/theblokeai) + +## Thanks, and how to contribute + +Thanks to the [chirper.ai](https://chirper.ai) team! + +Thanks to Clay from [gpus.llm-utils.org](llm-utils)! + +I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training. + +If you're able and willing to contribute it will be most gratefully received and will help me to keep providing more models, and to start work on new AI projects. + +Donaters will get priority support on any and all AI/LLM/model questions and requests, access to a private Discord room, plus other benefits. + +* Patreon: https://patreon.com/TheBlokeAI +* Ko-Fi: https://ko-fi.com/TheBlokeAI + +**Special thanks to**: Aemon Algiz. + +**Patreon special mentions**: Michael Levine, 阿明, Trailburnt, Nikolai Manek, John Detwiler, Randy H, Will Dee, Sebastain Graf, NimbleBox.ai, Eugene Pentland, Emad Mostaque, Ai Maven, Jim Angel, Jeff Scroggin, Michael Davis, Manuel Alberto Morcote, Stephen Murray, Robert, Justin Joy, Luke @flexchar, Brandon Frisco, Elijah Stavena, S_X, Dan Guido, Undi ., Komninos Chatzipapas, Shadi, theTransient, Lone Striker, Raven Klaugh, jjj, Cap'n Zoog, Michel-Marie MAUDET (LINAGORA), Matthew Berman, David, Fen Risland, Omer Bin Jawed, Luke Pendergrass, Kalila, OG, Erik Bjäreholt, Rooh Singh, Joseph William Delisle, Dan Lewis, TL, John Villwock, AzureBlack, Brad, Pedro Madruga, Caitlyn Gatomon, K, jinyuan sun, Mano Prime, Alex, Jeffrey Morgan, Alicia Loh, Illia Dulskyi, Chadd, transmissions 11, fincy, Rainer Wilmers, ReadyPlayerEmma, knownsqashed, Mandus, biorpg, Deo Leter, Brandon Phillips, SuperWojo, Sean Connelly, Iucharbius, Jack West, Harry Royden McLaughlin, Nicholas, terasurfer, Vitor Caleffi, Duane Dunston, Johann-Peter Hartmann, David Ziegler, Olakabola, Ken Nordquist, Trenton Dambrowitz, Tom X Nguyen, Vadim, Ajan Kanaga, Leonard Tan, Clay Pascal, Alexandros Triantafyllidis, JM33133, Xule, vamX, ya boyyy, subjectnull, Talal Aujan, Alps Aficionado, wassieverse, Ari Malik, James Bentley, Woland, Spencer Kim, Michael Dempsey, Fred von Graf, Elle, zynix, William Richards, Stanislav Ovsiannikov, Edmond Seymore, Jonathan Leane, Martin Kemka, usrbinkat, Enrico Ros + + +Thank you to all my generous patrons and donaters! + +And thank you again to a16z for their generous grant. + + + + +# Original model card: Arc53's DocsGPT 7B Mistral + + + +DocsGPT is optimized for Documentation (RAG optimised): Specifically fine-tuned for providing answers that are based on context, making it particularly useful for developers and technical support teams. +We used the Lora fine tuning process. +This model is fine tuned on top of zephyr-7b-beta + + +It's an apache-2.0 license so you can use it for commercial purposes too. + + +Benchmarks: + +Bacon: +The BACON test is an internal assessment designed to evaluate the capabilities of neural networks in handling questions with substantial content. It focuses on testing the model's understanding of context-driven queries, as well as its tendency for hallucination and attention span. The questions in both parts are carefully crafted, drawing from diverse sources such as scientific papers, complex code problems, and instructional prompts, providing a comprehensive test of the model's ability to process and generate information in various domains. +| Model | Score | +|------------------------------|-------| +| gpt-4 | 8.74 | +| DocsGPT-7b-Mistral | 8.64 | +| gpt-3.5-turbo | 8.42 | +| zephyr-7b-beta | 8.37 | +| neural-chat-7b-v3-1 | 7.88 | +| Mistral-7B-Instruct-v0.1 | 7.44 | +| openinstruct-mistral-7b | 5.86 | +| llama-2-13b | 2.29 | + + + + + + + + + + +MTbench with llm judge: + + + +########## First turn ########## +| Model | Turn | Score | +|-----------------------|------|----------| +| gpt-4 | 1 | 8.956250 | +| gpt-3.5-turbo | 1 | 8.075000 | +| DocsGPT-7b-Mistral | 1 | 7.593750 | +| zephyr-7b-beta | 1 | 7.412500 | +| vicuna-13b-v1.3 | 1 | 6.812500 | +| alpaca-13b | 1 | 4.975000 | +| deepseek-coder-6.7b | 1 | 4.506329 | + +########## Second turn ########## +| Model | Turn | Score | +|-----------------------|------|----------| +| gpt-4 | 2 | 9.025000 | +| gpt-3.5-turbo | 2 | 7.812500 | +| DocsGPT-7b-Mistral | 2 | 6.740000 | +| zephyr-7b-beta | 2 | 6.650000 | +| vicuna-13b-v1.3 | 2 | 5.962500 | +| deepseek-coder-6.7b | 2 | 5.025641 | +| alpaca-13b | 2 | 4.087500 | + +########## Average ########## +| Model | Score | +|-----------------------|----------| +| gpt-4 | 8.990625 | +| gpt-3.5-turbo | 7.943750 | +| DocsGPT-7b-Mistral | 7.166875 | +| zephyr-7b-beta | 7.031250 | +| vicuna-13b-v1.3 | 6.387500 | +| deepseek-coder-6.7b | 4.764331 | +| alpaca-13b | 4.531250 | + + + + +To prepare your prompts make sure you keep this format: + +``` +### Instruction +(where the question goes) +### Context +(your document retrieval + system instructions) +### Answer +``` + + diff --git a/config.json b/config.json new file mode 100644 index 0000000..9f0f76f --- /dev/null +++ b/config.json @@ -0,0 +1,3 @@ +{ + "model_type": "mistral" +} \ No newline at end of file diff --git a/configuration.json b/configuration.json new file mode 100644 index 0000000..159097f --- /dev/null +++ b/configuration.json @@ -0,0 +1 @@ +{"framework": "pytorch", "task": "others", "allow_remote": true} \ No newline at end of file diff --git a/docsgpt-7b-mistral.Q2_K.gguf b/docsgpt-7b-mistral.Q2_K.gguf new file mode 100644 index 0000000..328c6b1 --- /dev/null +++ b/docsgpt-7b-mistral.Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:88b6e1bb167226d058384e2ad5bc4730256307173086d1937c3df3bdd3850a7b +size 3084034624 diff --git a/docsgpt-7b-mistral.Q3_K_L.gguf b/docsgpt-7b-mistral.Q3_K_L.gguf new file mode 100644 index 0000000..d5216f5 --- /dev/null +++ b/docsgpt-7b-mistral.Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2920082dee3cc4cc3f4ea11075aaabf3390e4cfed1f6a37994d40898036e8940 +size 3822961216 diff --git a/docsgpt-7b-mistral.Q3_K_M.gguf b/docsgpt-7b-mistral.Q3_K_M.gguf new file mode 100644 index 0000000..da3984e --- /dev/null +++ b/docsgpt-7b-mistral.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fce1b9025464415239df394e44c23a5d064486e07b76288c4af9e2328d3a5b1e +size 3519922752 diff --git a/docsgpt-7b-mistral.Q3_K_S.gguf b/docsgpt-7b-mistral.Q3_K_S.gguf new file mode 100644 index 0000000..e70b116 --- /dev/null +++ b/docsgpt-7b-mistral.Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e7aacf68f2e37e47fade55772f13aae6e1fd798fdb74436336441bcd271bbe70 +size 3165504064 diff --git a/docsgpt-7b-mistral.Q4_0.gguf b/docsgpt-7b-mistral.Q4_0.gguf new file mode 100644 index 0000000..8eafc3d --- /dev/null +++ b/docsgpt-7b-mistral.Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f021872457105152d3511b6aebba614cdccf2c786a882dee9806ac454aa63613 +size 4109853248 diff --git a/docsgpt-7b-mistral.Q4_K_M.gguf b/docsgpt-7b-mistral.Q4_K_M.gguf new file mode 100644 index 0000000..94c0d70 --- /dev/null +++ b/docsgpt-7b-mistral.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9aa516f6a25e73dd978314e96cbc1b174632e899fa890a800ac143111ead0796 +size 4369375808 diff --git a/docsgpt-7b-mistral.Q4_K_S.gguf b/docsgpt-7b-mistral.Q4_K_S.gguf new file mode 100644 index 0000000..91897e6 --- /dev/null +++ b/docsgpt-7b-mistral.Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e7f693b608b5af5176b4020c2a5f388b1c191775174e70b8cb5fb064196a5ea0 +size 4141310528 diff --git a/docsgpt-7b-mistral.Q5_0.gguf b/docsgpt-7b-mistral.Q5_0.gguf new file mode 100644 index 0000000..9193458 --- /dev/null +++ b/docsgpt-7b-mistral.Q5_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:720d8d271f216c6883922b8c3d1141dcd08b23948d37fa2dce1ae67af466c99a +size 4998652480 diff --git a/docsgpt-7b-mistral.Q5_K_M.gguf b/docsgpt-7b-mistral.Q5_K_M.gguf new file mode 100644 index 0000000..f12d4e4 --- /dev/null +++ b/docsgpt-7b-mistral.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dda2a1a4babfe227f723a73a935e1d910c9bdc5be29f52612fa63718981686ca +size 5132345920 diff --git a/docsgpt-7b-mistral.Q5_K_S.gguf b/docsgpt-7b-mistral.Q5_K_S.gguf new file mode 100644 index 0000000..bd7dd0c --- /dev/null +++ b/docsgpt-7b-mistral.Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:88ab4a012fc4da67affa01bb3ac601068d3b3748a17508e7432a53766b314b4c +size 4998652480 diff --git a/docsgpt-7b-mistral.Q6_K.gguf b/docsgpt-7b-mistral.Q6_K.gguf new file mode 100644 index 0000000..162fc43 --- /dev/null +++ b/docsgpt-7b-mistral.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:004e66b90c8fed30af628deb4a6009a612978a0746e0c8bfbca2a68a96994ccf +size 5943001664 diff --git a/docsgpt-7b-mistral.Q8_0.gguf b/docsgpt-7b-mistral.Q8_0.gguf new file mode 100644 index 0000000..7253ee4 --- /dev/null +++ b/docsgpt-7b-mistral.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8fd1ccdd97cae07913e4846c121fb316443a284ba084ca1c97b4df38f0a7e187 +size 7696794176