初始化项目,由ModelHub XC社区提供模型
Model: Gryphe/MergeMonster-13b-20231124 Source: Original Platform
This commit is contained in:
36
README.md
Normal file
36
README.md
Normal file
@@ -0,0 +1,36 @@
|
||||
---
|
||||
license: other
|
||||
language:
|
||||
- en
|
||||
---
|
||||
I received a request to upload the FP16 version of my first [MergeMonster](https://github.com/Gryphe/MergeMonster/tree/main) 13b merge, which had the sole focus of reducing overused ChatGPT roleplaying phrases. Many ministrations were murdered throughout this process.
|
||||
|
||||
GGUF quantizations can be found in my main [MergeMonster](https://huggingface.co/Gryphe/MergeMonster) repo.
|
||||
|
||||
## Merge Composition
|
||||
|
||||
Unsurprisingly, not much remained of MythoMax. I'm (obviously!) very fond of it, but it's not perfect.
|
||||
|
||||
Reminder: MergeMonster works with hundreds of pieces scattered throughout the final merge. If a model takes a large part of the composition this does not mean it is merged in a manner that might be logical to humans.
|
||||
```
|
||||
Base model: MythoMax-L2-13b
|
||||
|
||||
------ FINAL MERGE COMPOSITION ------
|
||||
LLaMA2-13B-Tiefighter: 0.27
|
||||
Emerhyst-13B: 0.26
|
||||
Orca-2-13b: 0.24
|
||||
Stheno-1.8-L2-13B: 0.09
|
||||
MythoMax-L2-13b: 0.09
|
||||
Athena-v4: 0.04
|
||||
```
|
||||
|
||||
## Prompt Format
|
||||
|
||||
As always, I suggest sticking to the Alpaca format.
|
||||
|
||||
```
|
||||
### Instruction:
|
||||
Your instruction or question here.
|
||||
|
||||
### Response:
|
||||
```
|
||||
Reference in New Issue
Block a user