language, license, library_name, tags, model-index
| language |
license |
library_name |
tags |
model-index |
|
|
apache-2.0 |
transformers |
| roleplay |
| text-generation-inference |
| merge |
| not-for-all-audiences |
|
| name |
results |
| BigMaid-20B-v1.0 |
| task |
dataset |
metrics |
source |
| type |
name |
| text-generation |
Text Generation |
|
| name |
type |
config |
split |
args |
| AI2 Reasoning Challenge (25-Shot) |
ai2_arc |
ARC-Challenge |
test |
|
|
| type |
value |
name |
| acc_norm |
61.35 |
normalized accuracy |
|
|
|
|
| task |
dataset |
metrics |
source |
| type |
name |
| text-generation |
Text Generation |
|
| name |
type |
split |
args |
| HellaSwag (10-Shot) |
hellaswag |
validation |
|
|
| type |
value |
name |
| acc_norm |
85.26 |
normalized accuracy |
|
|
|
|
| task |
dataset |
metrics |
source |
| type |
name |
| text-generation |
Text Generation |
|
| name |
type |
config |
split |
args |
| MMLU (5-Shot) |
cais/mmlu |
all |
test |
|
|
| type |
value |
name |
| acc |
57.15 |
accuracy |
|
|
|
|
| task |
dataset |
metrics |
source |
| type |
name |
| text-generation |
Text Generation |
|
| name |
type |
config |
split |
args |
| TruthfulQA (0-shot) |
truthful_qa |
multiple_choice |
validation |
|
|
|
|
|
| task |
dataset |
metrics |
source |
| type |
name |
| text-generation |
Text Generation |
|
| name |
type |
config |
split |
args |
| Winogrande (5-shot) |
winogrande |
winogrande_xl |
validation |
|
|
| type |
value |
name |
| acc |
75.3 |
accuracy |
|
|
|
|
| task |
dataset |
metrics |
source |
| type |
name |
| text-generation |
Text Generation |
|
| name |
type |
config |
split |
args |
| GSM8k (5-shot) |
gsm8k |
main |
test |
|
|
| type |
value |
name |
| acc |
2.05 |
accuracy |
|
|
|
|
|
|
|
BigMaid-20B-v1.0

Model Details
Warning: This model can produce NSFW content!
Results
- Bigger version of original, uncensored like oryginal.
- Retains all good qualities of original with additional affinity for abstract and lighthearted humor
All comments are greatly appreciated, download, test and if you appreciate my work, consider buying me my fuel:

Detailed results can be found here
| Metric |
Value |
| Avg. |
56.07 |
| AI2 Reasoning Challenge (25-Shot) |
61.35 |
| HellaSwag (10-Shot) |
85.26 |
| MMLU (5-Shot) |
57.15 |
| TruthfulQA (0-shot) |
55.29 |
| Winogrande (5-shot) |
75.30 |
| GSM8k (5-shot) |
2.05 |