初始化项目,由ModelHub XC社区提供模型

Model: nonetrix/pippafeet-11B-0.1
Source: Original Platform
This commit is contained in:
ModelHub XC
2026-04-13 03:10:58 +08:00
commit 45d9ee6411
12 changed files with 91366 additions and 0 deletions

35
.gitattributes vendored Normal file
View File

@@ -0,0 +1,35 @@
*.7z filter=lfs diff=lfs merge=lfs -text
*.arrow filter=lfs diff=lfs merge=lfs -text
*.bin filter=lfs diff=lfs merge=lfs -text
*.bz2 filter=lfs diff=lfs merge=lfs -text
*.ckpt filter=lfs diff=lfs merge=lfs -text
*.ftz filter=lfs diff=lfs merge=lfs -text
*.gz filter=lfs diff=lfs merge=lfs -text
*.h5 filter=lfs diff=lfs merge=lfs -text
*.joblib filter=lfs diff=lfs merge=lfs -text
*.lfs.* filter=lfs diff=lfs merge=lfs -text
*.mlmodel filter=lfs diff=lfs merge=lfs -text
*.model filter=lfs diff=lfs merge=lfs -text
*.msgpack filter=lfs diff=lfs merge=lfs -text
*.npy filter=lfs diff=lfs merge=lfs -text
*.npz filter=lfs diff=lfs merge=lfs -text
*.onnx filter=lfs diff=lfs merge=lfs -text
*.ot filter=lfs diff=lfs merge=lfs -text
*.parquet filter=lfs diff=lfs merge=lfs -text
*.pb filter=lfs diff=lfs merge=lfs -text
*.pickle filter=lfs diff=lfs merge=lfs -text
*.pkl filter=lfs diff=lfs merge=lfs -text
*.pt filter=lfs diff=lfs merge=lfs -text
*.pth filter=lfs diff=lfs merge=lfs -text
*.rar filter=lfs diff=lfs merge=lfs -text
*.safetensors filter=lfs diff=lfs merge=lfs -text
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
*.tar.* filter=lfs diff=lfs merge=lfs -text
*.tar filter=lfs diff=lfs merge=lfs -text
*.tflite filter=lfs diff=lfs merge=lfs -text
*.tgz filter=lfs diff=lfs merge=lfs -text
*.wasm filter=lfs diff=lfs merge=lfs -text
*.xz filter=lfs diff=lfs merge=lfs -text
*.zip filter=lfs diff=lfs merge=lfs -text
*.zst filter=lfs diff=lfs merge=lfs -text
*tfevents* filter=lfs diff=lfs merge=lfs -text

73
README.md Normal file
View File

@@ -0,0 +1,73 @@
---
base_model:
- Endevor/EndlessRP-v3-7B
- sanjiwatsuki/Loyal-Toppy-Bruins-Maid-7B-DARE
- SanjiWatsuki/Kunoichi-DPO-v2-7B
- undi95/Toppy-M-7B
- yam-peleg/Experiment30-7B
library_name: transformers
tags:
- mergekit
- merge
- not-for-all-audiences
license: apache-2.0
---
# Pippafeet-11B-0.1
This model is a mix of some of the "best 7B roleplaying LLMs". I selected a few models based on "creativity" from a random benchmark, and a final roleplaying LLM based on "IQ," and finally another LLM merged twice that "excels at general tasks" according to a separate benchmark for it's size. My goal was to combine the "most creative" smaller roleplaying LLMs, merge them, and enhance the intelligence by incorporating "decent general model" twice, along with a "smarter" roleplaying LLM. I don't really trust benchmarks much, but I thought it would at least give it some alignment perhaps, even if it is overfitted to a dataset to score well, I thought since it's a merge so it might negate overfitting somewhat, seems to have worked to some extent luckily.
In my limited testing, this model performs really well, giving decent replies most of the time.... That is if you ignore the fatal flaws, which are inherent to how this model was created unfortunately. Since it's made by directly stacking the weights of other models, it likes to constantly create new words and stutter and generally act stange, however if you ignore this and fill in the blanks yourself the model is quite decent. I plan to try to remove this weirdness if possible by using a LoRA but I am not sure if I will be able to, no promisses. If you have the compute to fine tune this model, I emplore you to because I think it is a promissing base.
Edit: Fine tune is pretty much impossible because ROCm is hot garbage and I should have never bought a AMD GPU, if someone has a functional GPU please fine tune it for me. Might be able to do it on CPU somehow but likely not FP16 and slow as fuck and in GGUF
Artwork source, please contact for me to remove it if wished: https://twitter.com/Kumaartsu/status/1756793643384402070
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6342619a9948f573f37a4a60/MUD4N762ncyUw2dPfzVJ_.png)
Note: this model is in no way affiliated with Phase Connect, Pipkin Pippa, or the artists artwork.
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
## Merge Details
### Merge Method
This model was merged using the [linear](https://arxiv.org/abs/2203.05482) merge method.
### Models Merged
The following models were included in the merge:
* [Endevor/EndlessRP-v3-7B](https://huggingface.co/Endevor/EndlessRP-v3-7B)
* [sanjiwatsuki/Loyal-Toppy-Bruins-Maid-7B-DARE](https://huggingface.co/sanjiwatsuki/Loyal-Toppy-Bruins-Maid-7B-DARE)
* [SanjiWatsuki/Kunoichi-DPO-v2-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-DPO-v2-7B)
* [undi95/Toppy-M-7B](https://huggingface.co/undi95/Toppy-M-7B)
* [yam-peleg/Experiment30-7B](https://huggingface.co/yam-peleg/Experiment30-7B)
### Configuration
The following YAML configuration was used to produce this model:
```yaml
merge_method: linear
parameters:
weight: 1.0
slices:
- sources:
- model: yam-peleg/Experiment30-7B
layer_range: [0, 16]
- sources:
- model: Endevor/EndlessRP-v3-7B
layer_range: [8, 24]
- sources:
- model: SanjiWatsuki/Kunoichi-DPO-v2-7B
layer_range: [17, 24]
- sources:
- model: undi95/Toppy-M-7B
layer_range: [20, 28]
- sources:
- model: sanjiwatsuki/Loyal-Toppy-Bruins-Maid-7B-DARE
layer_range: [28, 30]
- sources:
- model: yam-peleg/Experiment30-7B
layer_range: [29, 32]
dtype: float16
tokenizer_source: model:yam-peleg/Experiment30-7B
```

26
config.json Normal file
View File

@@ -0,0 +1,26 @@
{
"_name_or_path": "Endevor/EndlessRP-v3-7B",
"architectures": [
"MistralForCausalLM"
],
"attention_dropout": 0.0,
"bos_token_id": 1,
"eos_token_id": 2,
"hidden_act": "silu",
"hidden_size": 4096,
"initializer_range": 0.02,
"intermediate_size": 14336,
"max_position_embeddings": 32768,
"model_type": "mistral",
"num_attention_heads": 32,
"num_hidden_layers": 52,
"num_key_value_heads": 8,
"rms_norm_eps": 1e-05,
"rope_theta": 10000.0,
"sliding_window": 4096,
"tie_word_embeddings": false,
"torch_dtype": "float16",
"transformers_version": "4.39.1",
"use_cache": true,
"vocab_size": 32000
}

24
mergekit_config.yml Normal file
View File

@@ -0,0 +1,24 @@
merge_method: linear
parameters:
weight: 1.0
slices:
- sources:
- model: yam-peleg/Experiment30-7B
layer_range: [0, 16]
- sources:
- model: Endevor/EndlessRP-v3-7B
layer_range: [8, 24]
- sources:
- model: SanjiWatsuki/Kunoichi-DPO-v2-7B
layer_range: [17, 24]
- sources:
- model: undi95/Toppy-M-7B
layer_range: [20, 28]
- sources:
- model: sanjiwatsuki/Loyal-Toppy-Bruins-Maid-7B-DARE
layer_range: [28, 30]
- sources:
- model: yam-peleg/Experiment30-7B
layer_range: [29, 32]
dtype: float16
tokenizer_source: model:yam-peleg/Experiment30-7B

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:856fcdc130cb8082b3117122bc099d86e206453ee5864e82db9dd12ef8d5b4b3
size 9993305232

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:792276deef1c183ca5a3e7fcb6f1bd0143bdee0936bcb56a3dc7a7a950bb8320
size 9898949064

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:4a0a6155fa057bf779fab9e70a3f3477502afc58291005dbf6f5a65fc899de85
size 3315744288

File diff suppressed because one or more lines are too long

30
special_tokens_map.json Normal file
View File

@@ -0,0 +1,30 @@
{
"bos_token": {
"content": "<s>",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false
},
"eos_token": {
"content": "</s>",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false
},
"pad_token": {
"content": "<unk>",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false
},
"unk_token": {
"content": "<unk>",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false
}
}

91122
tokenizer.json Normal file

File diff suppressed because it is too large Load Diff

BIN
tokenizer.model (Stored with Git LFS) Normal file

Binary file not shown.

43
tokenizer_config.json Normal file
View File

@@ -0,0 +1,43 @@
{
"add_bos_token": true,
"add_eos_token": false,
"added_tokens_decoder": {
"0": {
"content": "<unk>",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"1": {
"content": "<s>",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"2": {
"content": "</s>",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
}
},
"additional_special_tokens": [],
"bos_token": "<s>",
"clean_up_tokenization_spaces": false,
"eos_token": "</s>",
"legacy": true,
"model_max_length": 32768,
"pad_token": "<unk>",
"padding_side": "right",
"sp_model_kwargs": {},
"spaces_between_special_tokens": false,
"tokenizer_class": "LlamaTokenizer",
"unk_token": "<unk>",
"use_default_system_prompt": false
}