license, datasets, language, pipeline_tag
license datasets language pipeline_tag
apache-2.0
anon8231489123/ShareGPT_Vicuna_unfiltered
en
text-generation

Model description

This is a Vicuna-like model with only 160M parameters, which is fine-tuned from LLaMA-160m on ShareGPT data.

The training setup follows the Vicuna suite.

The model is mainly developed as a base Small Speculative Model in MCSD paper. As a comparison, it can be better aligned to the Vicuna models than LLaMA-160m with little loss of alignment to the LLaMA models.

Draft Model Target Model Alignment
LLaMA-68/160M LLaMA-13/33B 😃
LLaMA-68/160M Vicuna-13/33B 😟
Vicuna-68/160M LLaMA-13/33B 😃
Vicuna-68/160M Vicuna-13/33B 😃
Description
Model synced from source: double7/vicuna-160m
Readme 28 KiB