From 6016cda671d73a4d141099021397a807074c65a1 Mon Sep 17 00:00:00 2001 From: Aleksey Korshuk Date: Mon, 10 Apr 2023 21:34:38 +0000 Subject: [PATCH] Create README.md (#5) - Create README.md (3fba2749f1d88ca243cff234abb7b2b653254f3d) Co-authored-by: Tom Jobbins --- README.md | 46 ++++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 46 insertions(+) create mode 100644 README.md diff --git a/README.md b/README.md new file mode 100644 index 0000000..d067a06 --- /dev/null +++ b/README.md @@ -0,0 +1,46 @@ +--- +license: other +--- +# Vicuna 7B without "ethics" filtering + +This repository contains an alternative version of the [Vicuna 7B model](https://huggingface.co/lmsys/vicuna-7b-delta-v0). + +This model was natively fine-tuned using ShareGPT data, but without the "ethics" filtering used for the original Vicuna. + +[A GPTQ quantised 4-bit version is available here](https://huggingface.co/TheBloke/vicuna-AlekseyKorshuk-7B-GPTQ-4bit-128g). + +# Original Vicuna Model Card + +## Model details + +**Model type:** +Vicuna is an open-source chatbot trained by fine-tuning LLaMA on user-shared conversations collected from ShareGPT. +It is an auto-regressive language model, based on the transformer architecture. + +**Model date:** +Vicuna was trained between March 2023 and April 2023. + +**Organizations developing the model:** +The Vicuna team with members from UC Berkeley, CMU, Stanford, and UC San Diego. + +**Paper or resources for more information:** +https://vicuna.lmsys.org/ + +**License:** +Apache License 2.0 + +**Where to send questions or comments about the model:** +https://github.com/lm-sys/FastChat/issues + +## Intended use +**Primary intended uses:** +The primary use of Vicuna is research on large language models and chatbots. + +**Primary intended users:** +The primary intended users of the model are researchers and hobbyists in natural language processing, machine learning, and artificial intelligence. + +## Training dataset +70K conversations collected from ShareGPT.com. + +## Evaluation dataset +A preliminary evaluation of the model quality is conducted by creating a set of 80 diverse questions and utilizing GPT-4 to judge the model outputs. See https://vicuna.lmsys.org/ for more details. \ No newline at end of file